{ "best_metric": null, "best_model_checkpoint": null, "epoch": 15.0, "eval_steps": 500, "global_step": 6135, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0024449877750611247, "grad_norm": 0.056970333624040005, "learning_rate": 4.9999996722213865e-06, "loss": 0.6502, "step": 1 }, { "epoch": 0.004889975550122249, "grad_norm": 0.0642923240948625, "learning_rate": 4.99999868888563e-06, "loss": 0.7052, "step": 2 }, { "epoch": 0.007334963325183374, "grad_norm": 0.07295774500273575, "learning_rate": 4.999997049992989e-06, "loss": 1.163, "step": 3 }, { "epoch": 0.009779951100244499, "grad_norm": 0.08454639435282074, "learning_rate": 4.999994755543892e-06, "loss": 0.7364, "step": 4 }, { "epoch": 0.012224938875305624, "grad_norm": 0.06591401969976246, "learning_rate": 4.999991805538943e-06, "loss": 0.8502, "step": 5 }, { "epoch": 0.014669926650366748, "grad_norm": 0.056687605494465804, "learning_rate": 4.999988199978913e-06, "loss": 0.5687, "step": 6 }, { "epoch": 0.017114914425427872, "grad_norm": 0.07161055534855067, "learning_rate": 4.999983938864751e-06, "loss": 0.6998, "step": 7 }, { "epoch": 0.019559902200488997, "grad_norm": 0.058300557445260116, "learning_rate": 4.999979022197571e-06, "loss": 0.6303, "step": 8 }, { "epoch": 0.022004889975550123, "grad_norm": 0.05437433502479284, "learning_rate": 4.999973449978664e-06, "loss": 0.5753, "step": 9 }, { "epoch": 0.02444987775061125, "grad_norm": 0.06723667478202587, "learning_rate": 4.999967222209489e-06, "loss": 0.9733, "step": 10 }, { "epoch": 0.02689486552567237, "grad_norm": 0.08315652473020697, "learning_rate": 4.999960338891682e-06, "loss": 1.1614, "step": 11 }, { "epoch": 0.029339853300733496, "grad_norm": 0.07785854113341294, "learning_rate": 4.999952800027046e-06, "loss": 0.7975, "step": 12 }, { "epoch": 0.03178484107579462, "grad_norm": 0.07307268247146252, "learning_rate": 4.999944605617559e-06, "loss": 1.0792, "step": 13 }, { "epoch": 0.034229828850855744, "grad_norm": 0.06504667944919901, "learning_rate": 4.999935755665369e-06, "loss": 0.6103, "step": 14 }, { "epoch": 0.03667481662591687, "grad_norm": 0.06531060627142818, "learning_rate": 4.999926250172798e-06, "loss": 0.6697, "step": 15 }, { "epoch": 0.039119804400977995, "grad_norm": 0.057727976859811216, "learning_rate": 4.999916089142336e-06, "loss": 0.6186, "step": 16 }, { "epoch": 0.04156479217603912, "grad_norm": 0.07073003216152718, "learning_rate": 4.99990527257665e-06, "loss": 1.0184, "step": 17 }, { "epoch": 0.044009779951100246, "grad_norm": 0.06607794689704095, "learning_rate": 4.999893800478573e-06, "loss": 0.6661, "step": 18 }, { "epoch": 0.04645476772616137, "grad_norm": 0.09361726958951792, "learning_rate": 4.999881672851118e-06, "loss": 0.9106, "step": 19 }, { "epoch": 0.0488997555012225, "grad_norm": 0.07718233312983057, "learning_rate": 4.9998688896974625e-06, "loss": 0.7479, "step": 20 }, { "epoch": 0.05134474327628362, "grad_norm": 0.06509269497362077, "learning_rate": 4.999855451020958e-06, "loss": 0.6082, "step": 21 }, { "epoch": 0.05378973105134474, "grad_norm": 0.07152985737396737, "learning_rate": 4.999841356825129e-06, "loss": 0.6822, "step": 22 }, { "epoch": 0.05623471882640587, "grad_norm": 0.07769426391009598, "learning_rate": 4.9998266071136714e-06, "loss": 0.7239, "step": 23 }, { "epoch": 0.05867970660146699, "grad_norm": 0.08276748551801542, "learning_rate": 4.999811201890453e-06, "loss": 0.6978, "step": 24 }, { "epoch": 0.061124694376528114, "grad_norm": 0.09628639792438537, "learning_rate": 4.999795141159514e-06, "loss": 0.8996, "step": 25 }, { "epoch": 0.06356968215158924, "grad_norm": 0.06258737253229976, "learning_rate": 4.999778424925065e-06, "loss": 0.5211, "step": 26 }, { "epoch": 0.06601466992665037, "grad_norm": 0.08811261681853647, "learning_rate": 4.99976105319149e-06, "loss": 0.7886, "step": 27 }, { "epoch": 0.06845965770171149, "grad_norm": 0.09319511948769583, "learning_rate": 4.999743025963343e-06, "loss": 0.9864, "step": 28 }, { "epoch": 0.07090464547677261, "grad_norm": 0.082190275231665, "learning_rate": 4.999724343245352e-06, "loss": 0.8088, "step": 29 }, { "epoch": 0.07334963325183375, "grad_norm": 0.10764497651069097, "learning_rate": 4.999705005042417e-06, "loss": 0.6992, "step": 30 }, { "epoch": 0.07579462102689487, "grad_norm": 0.09511750481525305, "learning_rate": 4.9996850113596075e-06, "loss": 0.8331, "step": 31 }, { "epoch": 0.07823960880195599, "grad_norm": 0.09242231072962526, "learning_rate": 4.999664362202167e-06, "loss": 0.708, "step": 32 }, { "epoch": 0.08068459657701711, "grad_norm": 0.08341700201579721, "learning_rate": 4.99964305757551e-06, "loss": 1.0805, "step": 33 }, { "epoch": 0.08312958435207823, "grad_norm": 0.06155034900954256, "learning_rate": 4.999621097485223e-06, "loss": 0.4807, "step": 34 }, { "epoch": 0.08557457212713937, "grad_norm": 0.08420997143915426, "learning_rate": 4.999598481937064e-06, "loss": 0.7115, "step": 35 }, { "epoch": 0.08801955990220049, "grad_norm": 0.08645529689872837, "learning_rate": 4.999575210936965e-06, "loss": 0.6044, "step": 36 }, { "epoch": 0.09046454767726161, "grad_norm": 0.07531317485402605, "learning_rate": 4.999551284491027e-06, "loss": 0.8446, "step": 37 }, { "epoch": 0.09290953545232274, "grad_norm": 0.09254647443966509, "learning_rate": 4.9995267026055225e-06, "loss": 1.0609, "step": 38 }, { "epoch": 0.09535452322738386, "grad_norm": 0.06902120774533693, "learning_rate": 4.9995014652869e-06, "loss": 0.5849, "step": 39 }, { "epoch": 0.097799511002445, "grad_norm": 0.11450312213133851, "learning_rate": 4.999475572541776e-06, "loss": 1.2489, "step": 40 }, { "epoch": 0.10024449877750612, "grad_norm": 0.12143513035395985, "learning_rate": 4.9994490243769406e-06, "loss": 0.8074, "step": 41 }, { "epoch": 0.10268948655256724, "grad_norm": 0.10095751319167787, "learning_rate": 4.999421820799355e-06, "loss": 0.8203, "step": 42 }, { "epoch": 0.10513447432762836, "grad_norm": 0.08911137880429622, "learning_rate": 4.999393961816154e-06, "loss": 0.7197, "step": 43 }, { "epoch": 0.10757946210268948, "grad_norm": 0.129400332443115, "learning_rate": 4.99936544743464e-06, "loss": 1.2678, "step": 44 }, { "epoch": 0.1100244498777506, "grad_norm": 0.10031345707812848, "learning_rate": 4.9993362776622924e-06, "loss": 1.5299, "step": 45 }, { "epoch": 0.11246943765281174, "grad_norm": 0.07753346019013206, "learning_rate": 4.999306452506759e-06, "loss": 0.6727, "step": 46 }, { "epoch": 0.11491442542787286, "grad_norm": 0.08482223258078687, "learning_rate": 4.999275971975862e-06, "loss": 0.4733, "step": 47 }, { "epoch": 0.11735941320293398, "grad_norm": 0.12129946290133806, "learning_rate": 4.999244836077593e-06, "loss": 0.8762, "step": 48 }, { "epoch": 0.1198044009779951, "grad_norm": 0.10928653456458395, "learning_rate": 4.9992130448201174e-06, "loss": 0.5832, "step": 49 }, { "epoch": 0.12224938875305623, "grad_norm": 0.10988768445643009, "learning_rate": 4.999180598211771e-06, "loss": 0.7233, "step": 50 }, { "epoch": 0.12469437652811736, "grad_norm": 0.1053130155345633, "learning_rate": 4.999147496261062e-06, "loss": 0.8271, "step": 51 }, { "epoch": 0.1271393643031785, "grad_norm": 0.10759620901058406, "learning_rate": 4.999113738976669e-06, "loss": 0.6885, "step": 52 }, { "epoch": 0.1295843520782396, "grad_norm": 0.11577751614456157, "learning_rate": 4.999079326367448e-06, "loss": 0.701, "step": 53 }, { "epoch": 0.13202933985330073, "grad_norm": 0.09535428814755058, "learning_rate": 4.999044258442419e-06, "loss": 0.6498, "step": 54 }, { "epoch": 0.13447432762836187, "grad_norm": 0.16209145805899675, "learning_rate": 4.999008535210778e-06, "loss": 0.691, "step": 55 }, { "epoch": 0.13691931540342298, "grad_norm": 0.12612499277133693, "learning_rate": 4.998972156681895e-06, "loss": 0.7276, "step": 56 }, { "epoch": 0.1393643031784841, "grad_norm": 0.10406117297804278, "learning_rate": 4.9989351228653075e-06, "loss": 0.7196, "step": 57 }, { "epoch": 0.14180929095354522, "grad_norm": 0.11912931166857814, "learning_rate": 4.998897433770726e-06, "loss": 0.9538, "step": 58 }, { "epoch": 0.14425427872860636, "grad_norm": 0.09672628819333685, "learning_rate": 4.998859089408034e-06, "loss": 1.1227, "step": 59 }, { "epoch": 0.1466992665036675, "grad_norm": 0.10215295318189853, "learning_rate": 4.998820089787287e-06, "loss": 0.6743, "step": 60 }, { "epoch": 0.1491442542787286, "grad_norm": 0.11280790975393021, "learning_rate": 4.998780434918712e-06, "loss": 0.5333, "step": 61 }, { "epoch": 0.15158924205378974, "grad_norm": 0.1774356389549198, "learning_rate": 4.998740124812704e-06, "loss": 0.9371, "step": 62 }, { "epoch": 0.15403422982885084, "grad_norm": 0.10426167157172311, "learning_rate": 4.9986991594798375e-06, "loss": 0.5998, "step": 63 }, { "epoch": 0.15647921760391198, "grad_norm": 0.11819600790108543, "learning_rate": 4.998657538930852e-06, "loss": 0.8305, "step": 64 }, { "epoch": 0.15892420537897312, "grad_norm": 0.10182772481798326, "learning_rate": 4.998615263176663e-06, "loss": 0.6427, "step": 65 }, { "epoch": 0.16136919315403422, "grad_norm": 0.10620039619674311, "learning_rate": 4.998572332228355e-06, "loss": 0.5889, "step": 66 }, { "epoch": 0.16381418092909536, "grad_norm": 0.12021868850600481, "learning_rate": 4.9985287460971856e-06, "loss": 0.6706, "step": 67 }, { "epoch": 0.16625916870415647, "grad_norm": 0.10893869913926743, "learning_rate": 4.998484504794584e-06, "loss": 0.6363, "step": 68 }, { "epoch": 0.1687041564792176, "grad_norm": 0.13194236620987754, "learning_rate": 4.998439608332152e-06, "loss": 1.3301, "step": 69 }, { "epoch": 0.17114914425427874, "grad_norm": 0.17068657483797256, "learning_rate": 4.998394056721661e-06, "loss": 1.0071, "step": 70 }, { "epoch": 0.17359413202933985, "grad_norm": 0.10829113413634824, "learning_rate": 4.998347849975057e-06, "loss": 0.7784, "step": 71 }, { "epoch": 0.17603911980440098, "grad_norm": 0.1291929918135214, "learning_rate": 4.998300988104457e-06, "loss": 0.7124, "step": 72 }, { "epoch": 0.1784841075794621, "grad_norm": 0.1235367727071986, "learning_rate": 4.998253471122148e-06, "loss": 0.6257, "step": 73 }, { "epoch": 0.18092909535452323, "grad_norm": 0.12967411869926201, "learning_rate": 4.9982052990405895e-06, "loss": 0.6845, "step": 74 }, { "epoch": 0.18337408312958436, "grad_norm": 0.11283010140159752, "learning_rate": 4.998156471872415e-06, "loss": 0.7685, "step": 75 }, { "epoch": 0.18581907090464547, "grad_norm": 0.14859626261811418, "learning_rate": 4.998106989630428e-06, "loss": 0.9241, "step": 76 }, { "epoch": 0.1882640586797066, "grad_norm": 0.10453236434422224, "learning_rate": 4.998056852327602e-06, "loss": 0.8729, "step": 77 }, { "epoch": 0.19070904645476772, "grad_norm": 0.16878089089242582, "learning_rate": 4.998006059977087e-06, "loss": 0.6309, "step": 78 }, { "epoch": 0.19315403422982885, "grad_norm": 0.11545542520278311, "learning_rate": 4.997954612592198e-06, "loss": 0.7567, "step": 79 }, { "epoch": 0.19559902200489, "grad_norm": 0.1375865837735131, "learning_rate": 4.9979025101864294e-06, "loss": 1.0424, "step": 80 }, { "epoch": 0.1980440097799511, "grad_norm": 0.12810477596360154, "learning_rate": 4.997849752773443e-06, "loss": 0.7293, "step": 81 }, { "epoch": 0.20048899755501223, "grad_norm": 0.13802070118221518, "learning_rate": 4.997796340367071e-06, "loss": 0.9855, "step": 82 }, { "epoch": 0.20293398533007334, "grad_norm": 0.12223213324491261, "learning_rate": 4.99774227298132e-06, "loss": 0.6349, "step": 83 }, { "epoch": 0.20537897310513448, "grad_norm": 0.12684251294283894, "learning_rate": 4.99768755063037e-06, "loss": 0.8088, "step": 84 }, { "epoch": 0.2078239608801956, "grad_norm": 0.13824509671408847, "learning_rate": 4.997632173328568e-06, "loss": 0.6177, "step": 85 }, { "epoch": 0.21026894865525672, "grad_norm": 0.15409914738654823, "learning_rate": 4.997576141090436e-06, "loss": 0.9274, "step": 86 }, { "epoch": 0.21271393643031786, "grad_norm": 0.1465455864569219, "learning_rate": 4.9975194539306675e-06, "loss": 1.0025, "step": 87 }, { "epoch": 0.21515892420537897, "grad_norm": 0.08760359472319448, "learning_rate": 4.9974621118641255e-06, "loss": 0.6117, "step": 88 }, { "epoch": 0.2176039119804401, "grad_norm": 0.10955988313080253, "learning_rate": 4.997404114905847e-06, "loss": 0.6335, "step": 89 }, { "epoch": 0.2200488997555012, "grad_norm": 0.13032374035452762, "learning_rate": 4.997345463071041e-06, "loss": 0.9161, "step": 90 }, { "epoch": 0.22249388753056235, "grad_norm": 0.10333150887648376, "learning_rate": 4.997286156375088e-06, "loss": 0.5768, "step": 91 }, { "epoch": 0.22493887530562348, "grad_norm": 0.11095398870073611, "learning_rate": 4.997226194833537e-06, "loss": 0.6353, "step": 92 }, { "epoch": 0.2273838630806846, "grad_norm": 0.11422289268164039, "learning_rate": 4.9971655784621146e-06, "loss": 1.0907, "step": 93 }, { "epoch": 0.22982885085574573, "grad_norm": 0.10393048195804086, "learning_rate": 4.9971043072767126e-06, "loss": 0.7662, "step": 94 }, { "epoch": 0.23227383863080683, "grad_norm": 0.17083560410434256, "learning_rate": 4.9970423812934e-06, "loss": 1.3592, "step": 95 }, { "epoch": 0.23471882640586797, "grad_norm": 0.12888083529675434, "learning_rate": 4.996979800528413e-06, "loss": 0.6691, "step": 96 }, { "epoch": 0.2371638141809291, "grad_norm": 0.09896919452550294, "learning_rate": 4.996916564998165e-06, "loss": 0.646, "step": 97 }, { "epoch": 0.2396088019559902, "grad_norm": 0.13022303525623774, "learning_rate": 4.996852674719234e-06, "loss": 0.962, "step": 98 }, { "epoch": 0.24205378973105135, "grad_norm": 0.13590235028802428, "learning_rate": 4.996788129708375e-06, "loss": 1.3377, "step": 99 }, { "epoch": 0.24449877750611246, "grad_norm": 0.10680948537478709, "learning_rate": 4.996722929982515e-06, "loss": 0.6796, "step": 100 }, { "epoch": 0.2469437652811736, "grad_norm": 0.13975673207873412, "learning_rate": 4.996657075558749e-06, "loss": 0.7666, "step": 101 }, { "epoch": 0.24938875305623473, "grad_norm": 0.11324922241216796, "learning_rate": 4.9965905664543455e-06, "loss": 0.8291, "step": 102 }, { "epoch": 0.25183374083129584, "grad_norm": 0.11304832716241121, "learning_rate": 4.996523402686745e-06, "loss": 0.9367, "step": 103 }, { "epoch": 0.254278728606357, "grad_norm": 0.12834753980055527, "learning_rate": 4.9964555842735605e-06, "loss": 0.6685, "step": 104 }, { "epoch": 0.2567237163814181, "grad_norm": 0.08988623728847779, "learning_rate": 4.996387111232574e-06, "loss": 0.6556, "step": 105 }, { "epoch": 0.2591687041564792, "grad_norm": 0.12673117114876908, "learning_rate": 4.99631798358174e-06, "loss": 1.2048, "step": 106 }, { "epoch": 0.2616136919315403, "grad_norm": 0.12374158370008943, "learning_rate": 4.996248201339188e-06, "loss": 0.6128, "step": 107 }, { "epoch": 0.26405867970660146, "grad_norm": 0.09576497724521663, "learning_rate": 4.996177764523215e-06, "loss": 0.9087, "step": 108 }, { "epoch": 0.2665036674816626, "grad_norm": 0.1299861945815397, "learning_rate": 4.996106673152291e-06, "loss": 0.5431, "step": 109 }, { "epoch": 0.26894865525672373, "grad_norm": 0.11488409073340845, "learning_rate": 4.996034927245057e-06, "loss": 0.832, "step": 110 }, { "epoch": 0.2713936430317848, "grad_norm": 0.16739301953270402, "learning_rate": 4.9959625268203284e-06, "loss": 0.7255, "step": 111 }, { "epoch": 0.27383863080684595, "grad_norm": 0.0911522743248785, "learning_rate": 4.9958894718970895e-06, "loss": 0.5702, "step": 112 }, { "epoch": 0.2762836185819071, "grad_norm": 0.13020888607160633, "learning_rate": 4.995815762494496e-06, "loss": 0.7724, "step": 113 }, { "epoch": 0.2787286063569682, "grad_norm": 0.07606783563632581, "learning_rate": 4.995741398631878e-06, "loss": 0.5307, "step": 114 }, { "epoch": 0.28117359413202936, "grad_norm": 0.09308595601431244, "learning_rate": 4.995666380328733e-06, "loss": 0.9479, "step": 115 }, { "epoch": 0.28361858190709044, "grad_norm": 0.07887678996394007, "learning_rate": 4.995590707604735e-06, "loss": 0.6134, "step": 116 }, { "epoch": 0.2860635696821516, "grad_norm": 0.07836620825794727, "learning_rate": 4.9955143804797255e-06, "loss": 0.6522, "step": 117 }, { "epoch": 0.2885085574572127, "grad_norm": 0.10758103214466581, "learning_rate": 4.995437398973719e-06, "loss": 0.7114, "step": 118 }, { "epoch": 0.29095354523227385, "grad_norm": 0.1028771022908388, "learning_rate": 4.995359763106904e-06, "loss": 0.6825, "step": 119 }, { "epoch": 0.293398533007335, "grad_norm": 0.08658710914184022, "learning_rate": 4.995281472899636e-06, "loss": 0.5607, "step": 120 }, { "epoch": 0.29584352078239606, "grad_norm": 0.10877855073612133, "learning_rate": 4.995202528372446e-06, "loss": 0.8574, "step": 121 }, { "epoch": 0.2982885085574572, "grad_norm": 0.08475753927946016, "learning_rate": 4.995122929546034e-06, "loss": 0.5852, "step": 122 }, { "epoch": 0.30073349633251834, "grad_norm": 0.07951060549243796, "learning_rate": 4.995042676441273e-06, "loss": 0.6117, "step": 123 }, { "epoch": 0.30317848410757947, "grad_norm": 0.09217102200073693, "learning_rate": 4.994961769079207e-06, "loss": 0.5907, "step": 124 }, { "epoch": 0.3056234718826406, "grad_norm": 0.08913805490526741, "learning_rate": 4.994880207481053e-06, "loss": 0.8031, "step": 125 }, { "epoch": 0.3080684596577017, "grad_norm": 0.11884806379374013, "learning_rate": 4.994797991668197e-06, "loss": 1.0872, "step": 126 }, { "epoch": 0.3105134474327628, "grad_norm": 0.11868854741395989, "learning_rate": 4.994715121662198e-06, "loss": 0.5876, "step": 127 }, { "epoch": 0.31295843520782396, "grad_norm": 0.08173481608418573, "learning_rate": 4.994631597484787e-06, "loss": 0.7205, "step": 128 }, { "epoch": 0.3154034229828851, "grad_norm": 0.08210949719491274, "learning_rate": 4.994547419157866e-06, "loss": 0.5876, "step": 129 }, { "epoch": 0.31784841075794623, "grad_norm": 0.15524404469106484, "learning_rate": 4.994462586703506e-06, "loss": 0.6807, "step": 130 }, { "epoch": 0.3202933985330073, "grad_norm": 0.06519568859034326, "learning_rate": 4.994377100143957e-06, "loss": 0.8641, "step": 131 }, { "epoch": 0.32273838630806845, "grad_norm": 0.08912855892712686, "learning_rate": 4.994290959501631e-06, "loss": 0.782, "step": 132 }, { "epoch": 0.3251833740831296, "grad_norm": 0.08772647938292119, "learning_rate": 4.994204164799118e-06, "loss": 0.6825, "step": 133 }, { "epoch": 0.3276283618581907, "grad_norm": 0.08539072589695366, "learning_rate": 4.994116716059177e-06, "loss": 0.6012, "step": 134 }, { "epoch": 0.33007334963325186, "grad_norm": 0.1137745075536628, "learning_rate": 4.99402861330474e-06, "loss": 1.1046, "step": 135 }, { "epoch": 0.33251833740831294, "grad_norm": 0.11184409656681345, "learning_rate": 4.993939856558908e-06, "loss": 0.8484, "step": 136 }, { "epoch": 0.33496332518337407, "grad_norm": 0.07911652357186882, "learning_rate": 4.993850445844956e-06, "loss": 0.6034, "step": 137 }, { "epoch": 0.3374083129584352, "grad_norm": 0.12022480170363364, "learning_rate": 4.99376038118633e-06, "loss": 0.673, "step": 138 }, { "epoch": 0.33985330073349634, "grad_norm": 0.12788232296499707, "learning_rate": 4.993669662606646e-06, "loss": 0.8366, "step": 139 }, { "epoch": 0.3422982885085575, "grad_norm": 0.08129035332772191, "learning_rate": 4.993578290129694e-06, "loss": 0.6278, "step": 140 }, { "epoch": 0.34474327628361856, "grad_norm": 0.12858988487215975, "learning_rate": 4.993486263779432e-06, "loss": 1.0111, "step": 141 }, { "epoch": 0.3471882640586797, "grad_norm": 0.08500035990172021, "learning_rate": 4.993393583579992e-06, "loss": 0.732, "step": 142 }, { "epoch": 0.34963325183374083, "grad_norm": 0.09539096134569047, "learning_rate": 4.993300249555678e-06, "loss": 0.5754, "step": 143 }, { "epoch": 0.35207823960880197, "grad_norm": 0.0651542161701372, "learning_rate": 4.993206261730963e-06, "loss": 0.5034, "step": 144 }, { "epoch": 0.3545232273838631, "grad_norm": 0.08111164431686171, "learning_rate": 4.993111620130493e-06, "loss": 0.7779, "step": 145 }, { "epoch": 0.3569682151589242, "grad_norm": 0.08996874840635453, "learning_rate": 4.993016324779086e-06, "loss": 1.1616, "step": 146 }, { "epoch": 0.3594132029339853, "grad_norm": 0.0879884319379973, "learning_rate": 4.9929203757017295e-06, "loss": 0.6133, "step": 147 }, { "epoch": 0.36185819070904646, "grad_norm": 0.10527125061550155, "learning_rate": 4.992823772923585e-06, "loss": 0.8257, "step": 148 }, { "epoch": 0.3643031784841076, "grad_norm": 0.06923364001443395, "learning_rate": 4.992726516469983e-06, "loss": 0.9113, "step": 149 }, { "epoch": 0.36674816625916873, "grad_norm": 0.0911528537182838, "learning_rate": 4.9926286063664265e-06, "loss": 1.377, "step": 150 }, { "epoch": 0.3691931540342298, "grad_norm": 0.06336974023058425, "learning_rate": 4.9925300426385896e-06, "loss": 0.5948, "step": 151 }, { "epoch": 0.37163814180929094, "grad_norm": 0.09224131658729902, "learning_rate": 4.992430825312317e-06, "loss": 0.5692, "step": 152 }, { "epoch": 0.3740831295843521, "grad_norm": 0.0678876532088674, "learning_rate": 4.992330954413628e-06, "loss": 0.9431, "step": 153 }, { "epoch": 0.3765281173594132, "grad_norm": 0.0674633387529636, "learning_rate": 4.99223042996871e-06, "loss": 0.5367, "step": 154 }, { "epoch": 0.37897310513447435, "grad_norm": 0.1611234007061063, "learning_rate": 4.9921292520039224e-06, "loss": 0.657, "step": 155 }, { "epoch": 0.38141809290953543, "grad_norm": 0.09120305182979466, "learning_rate": 4.9920274205457964e-06, "loss": 0.6761, "step": 156 }, { "epoch": 0.38386308068459657, "grad_norm": 0.10383633337877886, "learning_rate": 4.991924935621035e-06, "loss": 1.3519, "step": 157 }, { "epoch": 0.3863080684596577, "grad_norm": 0.09009436772559609, "learning_rate": 4.991821797256512e-06, "loss": 1.1781, "step": 158 }, { "epoch": 0.38875305623471884, "grad_norm": 0.19129170972605575, "learning_rate": 4.991718005479272e-06, "loss": 0.6583, "step": 159 }, { "epoch": 0.39119804400978, "grad_norm": 0.07005021371016842, "learning_rate": 4.991613560316533e-06, "loss": 0.7996, "step": 160 }, { "epoch": 0.39364303178484106, "grad_norm": 0.0996636908815799, "learning_rate": 4.991508461795682e-06, "loss": 0.6251, "step": 161 }, { "epoch": 0.3960880195599022, "grad_norm": 0.1549911052603041, "learning_rate": 4.991402709944277e-06, "loss": 0.7505, "step": 162 }, { "epoch": 0.39853300733496333, "grad_norm": 0.06689787692146222, "learning_rate": 4.9912963047900515e-06, "loss": 0.5236, "step": 163 }, { "epoch": 0.40097799511002447, "grad_norm": 0.08118830506393378, "learning_rate": 4.991189246360905e-06, "loss": 1.0172, "step": 164 }, { "epoch": 0.4034229828850856, "grad_norm": 0.06949004489354703, "learning_rate": 4.991081534684911e-06, "loss": 0.6692, "step": 165 }, { "epoch": 0.4058679706601467, "grad_norm": 0.09296671391886693, "learning_rate": 4.990973169790315e-06, "loss": 0.6944, "step": 166 }, { "epoch": 0.4083129584352078, "grad_norm": 0.07484237489228208, "learning_rate": 4.990864151705532e-06, "loss": 0.5256, "step": 167 }, { "epoch": 0.41075794621026895, "grad_norm": 0.10433853476942256, "learning_rate": 4.990754480459149e-06, "loss": 0.9094, "step": 168 }, { "epoch": 0.4132029339853301, "grad_norm": 0.12247812230834339, "learning_rate": 4.990644156079925e-06, "loss": 0.7113, "step": 169 }, { "epoch": 0.4156479217603912, "grad_norm": 0.05804892945521931, "learning_rate": 4.990533178596788e-06, "loss": 0.7857, "step": 170 }, { "epoch": 0.4180929095354523, "grad_norm": 0.07237411770989431, "learning_rate": 4.99042154803884e-06, "loss": 0.961, "step": 171 }, { "epoch": 0.42053789731051344, "grad_norm": 0.07228973210856648, "learning_rate": 4.990309264435354e-06, "loss": 0.6686, "step": 172 }, { "epoch": 0.4229828850855746, "grad_norm": 0.06901133513710181, "learning_rate": 4.9901963278157715e-06, "loss": 0.816, "step": 173 }, { "epoch": 0.4254278728606357, "grad_norm": 0.07174856426018103, "learning_rate": 4.990082738209707e-06, "loss": 0.595, "step": 174 }, { "epoch": 0.4278728606356968, "grad_norm": 0.06384119191550144, "learning_rate": 4.989968495646949e-06, "loss": 0.8963, "step": 175 }, { "epoch": 0.43031784841075793, "grad_norm": 0.060398178651819484, "learning_rate": 4.989853600157451e-06, "loss": 0.5991, "step": 176 }, { "epoch": 0.43276283618581907, "grad_norm": 0.0713760600484018, "learning_rate": 4.989738051771344e-06, "loss": 0.7191, "step": 177 }, { "epoch": 0.4352078239608802, "grad_norm": 0.08843620230170797, "learning_rate": 4.989621850518925e-06, "loss": 0.6627, "step": 178 }, { "epoch": 0.43765281173594134, "grad_norm": 0.051954527341017354, "learning_rate": 4.989504996430667e-06, "loss": 0.4925, "step": 179 }, { "epoch": 0.4400977995110024, "grad_norm": 0.0708987237498778, "learning_rate": 4.989387489537211e-06, "loss": 0.706, "step": 180 }, { "epoch": 0.44254278728606355, "grad_norm": 0.0801266850649354, "learning_rate": 4.98926932986937e-06, "loss": 0.6977, "step": 181 }, { "epoch": 0.4449877750611247, "grad_norm": 0.10512480393573406, "learning_rate": 4.989150517458127e-06, "loss": 0.521, "step": 182 }, { "epoch": 0.4474327628361858, "grad_norm": 0.06320368064017733, "learning_rate": 4.9890310523346386e-06, "loss": 0.7557, "step": 183 }, { "epoch": 0.44987775061124696, "grad_norm": 0.056159744095997874, "learning_rate": 4.988910934530231e-06, "loss": 0.5743, "step": 184 }, { "epoch": 0.45232273838630804, "grad_norm": 0.12744617399117472, "learning_rate": 4.988790164076401e-06, "loss": 0.6362, "step": 185 }, { "epoch": 0.4547677261613692, "grad_norm": 0.08930527880819754, "learning_rate": 4.98866874100482e-06, "loss": 0.5671, "step": 186 }, { "epoch": 0.4572127139364303, "grad_norm": 0.07149671636535629, "learning_rate": 4.988546665347325e-06, "loss": 0.8099, "step": 187 }, { "epoch": 0.45965770171149145, "grad_norm": 0.07794984955470813, "learning_rate": 4.988423937135927e-06, "loss": 0.5913, "step": 188 }, { "epoch": 0.4621026894865526, "grad_norm": 0.10918496095120744, "learning_rate": 4.988300556402811e-06, "loss": 1.0037, "step": 189 }, { "epoch": 0.46454767726161367, "grad_norm": 0.11961317163600049, "learning_rate": 4.988176523180328e-06, "loss": 0.5171, "step": 190 }, { "epoch": 0.4669926650366748, "grad_norm": 0.06838542971570936, "learning_rate": 4.9880518375010024e-06, "loss": 0.615, "step": 191 }, { "epoch": 0.46943765281173594, "grad_norm": 0.07856151694766761, "learning_rate": 4.98792649939753e-06, "loss": 0.8231, "step": 192 }, { "epoch": 0.4718826405867971, "grad_norm": 0.12075114188625571, "learning_rate": 4.987800508902778e-06, "loss": 0.7494, "step": 193 }, { "epoch": 0.4743276283618582, "grad_norm": 0.06830223665328845, "learning_rate": 4.987673866049784e-06, "loss": 0.6304, "step": 194 }, { "epoch": 0.4767726161369193, "grad_norm": 0.10675312491364994, "learning_rate": 4.987546570871755e-06, "loss": 0.7213, "step": 195 }, { "epoch": 0.4792176039119804, "grad_norm": 0.08088271092943289, "learning_rate": 4.9874186234020715e-06, "loss": 0.501, "step": 196 }, { "epoch": 0.48166259168704156, "grad_norm": 0.06901160602750281, "learning_rate": 4.987290023674285e-06, "loss": 0.6102, "step": 197 }, { "epoch": 0.4841075794621027, "grad_norm": 0.08330753598800536, "learning_rate": 4.987160771722117e-06, "loss": 0.9907, "step": 198 }, { "epoch": 0.48655256723716384, "grad_norm": 0.0642050181277398, "learning_rate": 4.9870308675794604e-06, "loss": 0.6304, "step": 199 }, { "epoch": 0.4889975550122249, "grad_norm": 0.05503037907417415, "learning_rate": 4.9869003112803785e-06, "loss": 0.803, "step": 200 }, { "epoch": 0.49144254278728605, "grad_norm": 0.05343579542611549, "learning_rate": 4.986769102859106e-06, "loss": 0.4103, "step": 201 }, { "epoch": 0.4938875305623472, "grad_norm": 0.08640774665689151, "learning_rate": 4.98663724235005e-06, "loss": 0.555, "step": 202 }, { "epoch": 0.4963325183374083, "grad_norm": 0.08602755799563155, "learning_rate": 4.986504729787786e-06, "loss": 0.7291, "step": 203 }, { "epoch": 0.49877750611246946, "grad_norm": 0.10140175293983117, "learning_rate": 4.986371565207062e-06, "loss": 0.9228, "step": 204 }, { "epoch": 0.5012224938875306, "grad_norm": 0.07644994488427348, "learning_rate": 4.986237748642798e-06, "loss": 0.7743, "step": 205 }, { "epoch": 0.5036674816625917, "grad_norm": 0.07819041605586327, "learning_rate": 4.986103280130083e-06, "loss": 0.6905, "step": 206 }, { "epoch": 0.5061124694376528, "grad_norm": 0.0791217176414588, "learning_rate": 4.985968159704176e-06, "loss": 0.7729, "step": 207 }, { "epoch": 0.508557457212714, "grad_norm": 0.06863020871294506, "learning_rate": 4.985832387400512e-06, "loss": 0.5923, "step": 208 }, { "epoch": 0.511002444987775, "grad_norm": 0.08929132493963589, "learning_rate": 4.985695963254692e-06, "loss": 0.6418, "step": 209 }, { "epoch": 0.5134474327628362, "grad_norm": 0.09899762276269189, "learning_rate": 4.985558887302489e-06, "loss": 0.9342, "step": 210 }, { "epoch": 0.5158924205378973, "grad_norm": 0.08193665403820839, "learning_rate": 4.985421159579847e-06, "loss": 1.0324, "step": 211 }, { "epoch": 0.5183374083129584, "grad_norm": 0.09959839296825133, "learning_rate": 4.9852827801228835e-06, "loss": 0.9113, "step": 212 }, { "epoch": 0.5207823960880196, "grad_norm": 0.1315051002524026, "learning_rate": 4.985143748967883e-06, "loss": 0.5565, "step": 213 }, { "epoch": 0.5232273838630807, "grad_norm": 0.06915030902244963, "learning_rate": 4.985004066151304e-06, "loss": 0.4822, "step": 214 }, { "epoch": 0.5256723716381418, "grad_norm": 0.10091503632779024, "learning_rate": 4.984863731709772e-06, "loss": 0.5046, "step": 215 }, { "epoch": 0.5281173594132029, "grad_norm": 0.06507643097510693, "learning_rate": 4.984722745680089e-06, "loss": 0.9793, "step": 216 }, { "epoch": 0.530562347188264, "grad_norm": 0.0879113975750124, "learning_rate": 4.984581108099223e-06, "loss": 0.6283, "step": 217 }, { "epoch": 0.5330073349633252, "grad_norm": 0.09582350126030938, "learning_rate": 4.984438819004315e-06, "loss": 0.5287, "step": 218 }, { "epoch": 0.5354523227383863, "grad_norm": 0.06701152242509115, "learning_rate": 4.984295878432677e-06, "loss": 0.7057, "step": 219 }, { "epoch": 0.5378973105134475, "grad_norm": 0.0718490412776238, "learning_rate": 4.98415228642179e-06, "loss": 0.6648, "step": 220 }, { "epoch": 0.5403422982885085, "grad_norm": 0.09307015191606556, "learning_rate": 4.984008043009308e-06, "loss": 0.7452, "step": 221 }, { "epoch": 0.5427872860635696, "grad_norm": 0.10303736607634002, "learning_rate": 4.983863148233055e-06, "loss": 0.7592, "step": 222 }, { "epoch": 0.5452322738386308, "grad_norm": 0.08702846146876625, "learning_rate": 4.983717602131025e-06, "loss": 0.5841, "step": 223 }, { "epoch": 0.5476772616136919, "grad_norm": 0.05850403452669591, "learning_rate": 4.983571404741384e-06, "loss": 0.5318, "step": 224 }, { "epoch": 0.5501222493887531, "grad_norm": 0.08068398717142776, "learning_rate": 4.983424556102469e-06, "loss": 0.717, "step": 225 }, { "epoch": 0.5525672371638142, "grad_norm": 0.07123259556915239, "learning_rate": 4.983277056252787e-06, "loss": 0.8039, "step": 226 }, { "epoch": 0.5550122249388753, "grad_norm": 0.09188810320332673, "learning_rate": 4.983128905231014e-06, "loss": 0.8955, "step": 227 }, { "epoch": 0.5574572127139364, "grad_norm": 0.07481109001543708, "learning_rate": 4.982980103076e-06, "loss": 0.6147, "step": 228 }, { "epoch": 0.5599022004889975, "grad_norm": 0.12774035377023493, "learning_rate": 4.982830649826764e-06, "loss": 0.5272, "step": 229 }, { "epoch": 0.5623471882640587, "grad_norm": 0.08134594677844227, "learning_rate": 4.982680545522497e-06, "loss": 0.5825, "step": 230 }, { "epoch": 0.5647921760391198, "grad_norm": 0.06308804044499798, "learning_rate": 4.982529790202559e-06, "loss": 0.7986, "step": 231 }, { "epoch": 0.5672371638141809, "grad_norm": 0.07309317085146093, "learning_rate": 4.982378383906481e-06, "loss": 0.8898, "step": 232 }, { "epoch": 0.5696821515892421, "grad_norm": 0.06734889017367603, "learning_rate": 4.982226326673966e-06, "loss": 0.6332, "step": 233 }, { "epoch": 0.5721271393643031, "grad_norm": 0.06932860707953248, "learning_rate": 4.982073618544886e-06, "loss": 1.1766, "step": 234 }, { "epoch": 0.5745721271393643, "grad_norm": 0.06996838757752463, "learning_rate": 4.981920259559285e-06, "loss": 0.7283, "step": 235 }, { "epoch": 0.5770171149144254, "grad_norm": 0.05357288768842464, "learning_rate": 4.981766249757378e-06, "loss": 0.5487, "step": 236 }, { "epoch": 0.5794621026894865, "grad_norm": 0.08579442726514055, "learning_rate": 4.981611589179549e-06, "loss": 0.9645, "step": 237 }, { "epoch": 0.5819070904645477, "grad_norm": 0.08745993239648804, "learning_rate": 4.981456277866355e-06, "loss": 0.9741, "step": 238 }, { "epoch": 0.5843520782396088, "grad_norm": 0.07579885675925718, "learning_rate": 4.98130031585852e-06, "loss": 0.5609, "step": 239 }, { "epoch": 0.58679706601467, "grad_norm": 0.09232934876594046, "learning_rate": 4.981143703196941e-06, "loss": 0.5401, "step": 240 }, { "epoch": 0.589242053789731, "grad_norm": 0.08730183070991888, "learning_rate": 4.980986439922688e-06, "loss": 0.7162, "step": 241 }, { "epoch": 0.5916870415647921, "grad_norm": 0.0874818677863648, "learning_rate": 4.980828526076996e-06, "loss": 0.6562, "step": 242 }, { "epoch": 0.5941320293398533, "grad_norm": 0.06531945462857051, "learning_rate": 4.980669961701276e-06, "loss": 0.6015, "step": 243 }, { "epoch": 0.5965770171149144, "grad_norm": 0.11538804730109553, "learning_rate": 4.9805107468371055e-06, "loss": 0.6362, "step": 244 }, { "epoch": 0.5990220048899756, "grad_norm": 0.08947315493090728, "learning_rate": 4.980350881526235e-06, "loss": 0.8897, "step": 245 }, { "epoch": 0.6014669926650367, "grad_norm": 0.07891063165167597, "learning_rate": 4.980190365810584e-06, "loss": 0.7953, "step": 246 }, { "epoch": 0.6039119804400978, "grad_norm": 0.09066564849285777, "learning_rate": 4.980029199732245e-06, "loss": 1.0367, "step": 247 }, { "epoch": 0.6063569682151589, "grad_norm": 0.09298654353803167, "learning_rate": 4.979867383333479e-06, "loss": 0.6467, "step": 248 }, { "epoch": 0.60880195599022, "grad_norm": 0.07442509371951192, "learning_rate": 4.979704916656717e-06, "loss": 0.6237, "step": 249 }, { "epoch": 0.6112469437652812, "grad_norm": 0.08031377511460987, "learning_rate": 4.9795417997445615e-06, "loss": 1.4601, "step": 250 }, { "epoch": 0.6136919315403423, "grad_norm": 0.07068617682262945, "learning_rate": 4.979378032639787e-06, "loss": 0.8154, "step": 251 }, { "epoch": 0.6161369193154034, "grad_norm": 0.08803651290720371, "learning_rate": 4.979213615385334e-06, "loss": 0.7019, "step": 252 }, { "epoch": 0.6185819070904646, "grad_norm": 0.09405352195827332, "learning_rate": 4.9790485480243205e-06, "loss": 0.6672, "step": 253 }, { "epoch": 0.6210268948655256, "grad_norm": 0.07611275328318817, "learning_rate": 4.9788828306000285e-06, "loss": 0.7606, "step": 254 }, { "epoch": 0.6234718826405868, "grad_norm": 0.054319948235132615, "learning_rate": 4.978716463155913e-06, "loss": 0.431, "step": 255 }, { "epoch": 0.6259168704156479, "grad_norm": 0.10166914464066547, "learning_rate": 4.9785494457356e-06, "loss": 0.9267, "step": 256 }, { "epoch": 0.628361858190709, "grad_norm": 0.07316616910070232, "learning_rate": 4.9783817783828845e-06, "loss": 1.1379, "step": 257 }, { "epoch": 0.6308068459657702, "grad_norm": 0.07284557902814078, "learning_rate": 4.978213461141733e-06, "loss": 0.8238, "step": 258 }, { "epoch": 0.6332518337408313, "grad_norm": 0.07729127701923229, "learning_rate": 4.978044494056282e-06, "loss": 0.6282, "step": 259 }, { "epoch": 0.6356968215158925, "grad_norm": 0.10965476253473438, "learning_rate": 4.97787487717084e-06, "loss": 0.7746, "step": 260 }, { "epoch": 0.6381418092909535, "grad_norm": 0.08328320785681154, "learning_rate": 4.977704610529883e-06, "loss": 0.7129, "step": 261 }, { "epoch": 0.6405867970660146, "grad_norm": 0.07096269237068077, "learning_rate": 4.977533694178058e-06, "loss": 0.8062, "step": 262 }, { "epoch": 0.6430317848410758, "grad_norm": 0.11117598057147879, "learning_rate": 4.977362128160186e-06, "loss": 0.6799, "step": 263 }, { "epoch": 0.6454767726161369, "grad_norm": 0.09735342376530616, "learning_rate": 4.9771899125212524e-06, "loss": 0.8234, "step": 264 }, { "epoch": 0.6479217603911981, "grad_norm": 0.09144966869987771, "learning_rate": 4.977017047306418e-06, "loss": 0.7613, "step": 265 }, { "epoch": 0.6503667481662592, "grad_norm": 0.09346114112765455, "learning_rate": 4.9768435325610106e-06, "loss": 0.7318, "step": 266 }, { "epoch": 0.6528117359413202, "grad_norm": 0.06779016560747965, "learning_rate": 4.9766693683305326e-06, "loss": 0.6908, "step": 267 }, { "epoch": 0.6552567237163814, "grad_norm": 0.13383732323227207, "learning_rate": 4.97649455466065e-06, "loss": 1.3706, "step": 268 }, { "epoch": 0.6577017114914425, "grad_norm": 0.0788712267195195, "learning_rate": 4.976319091597206e-06, "loss": 0.6005, "step": 269 }, { "epoch": 0.6601466992665037, "grad_norm": 0.06570350427696932, "learning_rate": 4.976142979186209e-06, "loss": 0.8911, "step": 270 }, { "epoch": 0.6625916870415648, "grad_norm": 0.08063591302532015, "learning_rate": 4.975966217473841e-06, "loss": 0.8505, "step": 271 }, { "epoch": 0.6650366748166259, "grad_norm": 0.0826057148632207, "learning_rate": 4.975788806506452e-06, "loss": 0.7124, "step": 272 }, { "epoch": 0.6674816625916871, "grad_norm": 0.07363801056784766, "learning_rate": 4.975610746330565e-06, "loss": 0.9228, "step": 273 }, { "epoch": 0.6699266503667481, "grad_norm": 0.08839135450743345, "learning_rate": 4.975432036992869e-06, "loss": 0.8497, "step": 274 }, { "epoch": 0.6723716381418093, "grad_norm": 0.06362767692976444, "learning_rate": 4.9752526785402276e-06, "loss": 0.5299, "step": 275 }, { "epoch": 0.6748166259168704, "grad_norm": 0.09506102741641534, "learning_rate": 4.975072671019671e-06, "loss": 0.6699, "step": 276 }, { "epoch": 0.6772616136919315, "grad_norm": 0.0732489846524302, "learning_rate": 4.974892014478403e-06, "loss": 0.5213, "step": 277 }, { "epoch": 0.6797066014669927, "grad_norm": 0.07792011711234956, "learning_rate": 4.974710708963795e-06, "loss": 0.7079, "step": 278 }, { "epoch": 0.6821515892420538, "grad_norm": 0.06443317774634497, "learning_rate": 4.9745287545233904e-06, "loss": 0.784, "step": 279 }, { "epoch": 0.684596577017115, "grad_norm": 0.08024122985918529, "learning_rate": 4.9743461512049e-06, "loss": 0.6112, "step": 280 }, { "epoch": 0.687041564792176, "grad_norm": 0.09010298542382467, "learning_rate": 4.974162899056207e-06, "loss": 1.0372, "step": 281 }, { "epoch": 0.6894865525672371, "grad_norm": 0.07100616876826107, "learning_rate": 4.973978998125366e-06, "loss": 1.023, "step": 282 }, { "epoch": 0.6919315403422983, "grad_norm": 0.09164189430137386, "learning_rate": 4.9737944484605995e-06, "loss": 1.0753, "step": 283 }, { "epoch": 0.6943765281173594, "grad_norm": 0.0718132356345249, "learning_rate": 4.9736092501102985e-06, "loss": 0.6648, "step": 284 }, { "epoch": 0.6968215158924206, "grad_norm": 0.07884549975722853, "learning_rate": 4.973423403123029e-06, "loss": 0.7576, "step": 285 }, { "epoch": 0.6992665036674817, "grad_norm": 0.10110247892652786, "learning_rate": 4.973236907547523e-06, "loss": 0.7716, "step": 286 }, { "epoch": 0.7017114914425427, "grad_norm": 0.07723692555839858, "learning_rate": 4.973049763432684e-06, "loss": 0.7622, "step": 287 }, { "epoch": 0.7041564792176039, "grad_norm": 0.07134818457945936, "learning_rate": 4.972861970827585e-06, "loss": 0.5653, "step": 288 }, { "epoch": 0.706601466992665, "grad_norm": 0.07619769017028144, "learning_rate": 4.97267352978147e-06, "loss": 0.9603, "step": 289 }, { "epoch": 0.7090464547677262, "grad_norm": 0.0839175935654035, "learning_rate": 4.972484440343754e-06, "loss": 0.8661, "step": 290 }, { "epoch": 0.7114914425427873, "grad_norm": 0.08203933567175961, "learning_rate": 4.9722947025640175e-06, "loss": 0.6861, "step": 291 }, { "epoch": 0.7139364303178484, "grad_norm": 0.08518667181270598, "learning_rate": 4.972104316492016e-06, "loss": 1.0165, "step": 292 }, { "epoch": 0.7163814180929096, "grad_norm": 0.10970590133781132, "learning_rate": 4.971913282177674e-06, "loss": 0.6495, "step": 293 }, { "epoch": 0.7188264058679706, "grad_norm": 0.06977398950464615, "learning_rate": 4.971721599671083e-06, "loss": 0.4463, "step": 294 }, { "epoch": 0.7212713936430318, "grad_norm": 0.0953089382370969, "learning_rate": 4.971529269022508e-06, "loss": 1.1599, "step": 295 }, { "epoch": 0.7237163814180929, "grad_norm": 0.07231474734901186, "learning_rate": 4.971336290282382e-06, "loss": 0.5773, "step": 296 }, { "epoch": 0.726161369193154, "grad_norm": 0.10325309775054034, "learning_rate": 4.971142663501308e-06, "loss": 0.5606, "step": 297 }, { "epoch": 0.7286063569682152, "grad_norm": 0.06493355868743536, "learning_rate": 4.970948388730059e-06, "loss": 0.5219, "step": 298 }, { "epoch": 0.7310513447432763, "grad_norm": 0.05275171607060085, "learning_rate": 4.970753466019581e-06, "loss": 0.4623, "step": 299 }, { "epoch": 0.7334963325183375, "grad_norm": 0.062003179052265364, "learning_rate": 4.970557895420984e-06, "loss": 0.8119, "step": 300 }, { "epoch": 0.7359413202933985, "grad_norm": 0.08765537692196579, "learning_rate": 4.970361676985553e-06, "loss": 1.2166, "step": 301 }, { "epoch": 0.7383863080684596, "grad_norm": 0.1057985963435635, "learning_rate": 4.97016481076474e-06, "loss": 0.6923, "step": 302 }, { "epoch": 0.7408312958435208, "grad_norm": 0.0715389858662257, "learning_rate": 4.969967296810168e-06, "loss": 0.535, "step": 303 }, { "epoch": 0.7432762836185819, "grad_norm": 0.09288946983927428, "learning_rate": 4.96976913517363e-06, "loss": 0.6887, "step": 304 }, { "epoch": 0.7457212713936431, "grad_norm": 0.06614264379965222, "learning_rate": 4.969570325907088e-06, "loss": 0.5541, "step": 305 }, { "epoch": 0.7481662591687042, "grad_norm": 0.07789423963719386, "learning_rate": 4.969370869062676e-06, "loss": 0.905, "step": 306 }, { "epoch": 0.7506112469437652, "grad_norm": 0.09710120230405622, "learning_rate": 4.969170764692693e-06, "loss": 0.573, "step": 307 }, { "epoch": 0.7530562347188264, "grad_norm": 0.0645825950181619, "learning_rate": 4.968970012849614e-06, "loss": 0.7413, "step": 308 }, { "epoch": 0.7555012224938875, "grad_norm": 0.08158264825180002, "learning_rate": 4.96876861358608e-06, "loss": 0.9351, "step": 309 }, { "epoch": 0.7579462102689487, "grad_norm": 0.0642139113205452, "learning_rate": 4.968566566954901e-06, "loss": 0.5706, "step": 310 }, { "epoch": 0.7603911980440098, "grad_norm": 0.08775438735010123, "learning_rate": 4.96836387300906e-06, "loss": 0.6897, "step": 311 }, { "epoch": 0.7628361858190709, "grad_norm": 0.06561298612751973, "learning_rate": 4.968160531801707e-06, "loss": 0.6126, "step": 312 }, { "epoch": 0.7652811735941321, "grad_norm": 0.15019058027061738, "learning_rate": 4.967956543386163e-06, "loss": 1.1087, "step": 313 }, { "epoch": 0.7677261613691931, "grad_norm": 0.10563230650201169, "learning_rate": 4.967751907815918e-06, "loss": 0.6726, "step": 314 }, { "epoch": 0.7701711491442543, "grad_norm": 0.13651356976880127, "learning_rate": 4.967546625144633e-06, "loss": 0.7918, "step": 315 }, { "epoch": 0.7726161369193154, "grad_norm": 0.08055315644947807, "learning_rate": 4.967340695426137e-06, "loss": 0.7152, "step": 316 }, { "epoch": 0.7750611246943765, "grad_norm": 0.10646718030189253, "learning_rate": 4.9671341187144305e-06, "loss": 1.0296, "step": 317 }, { "epoch": 0.7775061124694377, "grad_norm": 0.08035804242125767, "learning_rate": 4.966926895063682e-06, "loss": 0.4675, "step": 318 }, { "epoch": 0.7799511002444988, "grad_norm": 0.09080078591258836, "learning_rate": 4.966719024528229e-06, "loss": 0.5322, "step": 319 }, { "epoch": 0.78239608801956, "grad_norm": 0.07755784595559201, "learning_rate": 4.966510507162583e-06, "loss": 0.5181, "step": 320 }, { "epoch": 0.784841075794621, "grad_norm": 0.08574576983190237, "learning_rate": 4.9663013430214194e-06, "loss": 0.8412, "step": 321 }, { "epoch": 0.7872860635696821, "grad_norm": 0.07292180204112599, "learning_rate": 4.966091532159586e-06, "loss": 0.966, "step": 322 }, { "epoch": 0.7897310513447433, "grad_norm": 0.10982141476577932, "learning_rate": 4.965881074632102e-06, "loss": 1.0051, "step": 323 }, { "epoch": 0.7921760391198044, "grad_norm": 0.09431756479791206, "learning_rate": 4.965669970494151e-06, "loss": 0.8688, "step": 324 }, { "epoch": 0.7946210268948656, "grad_norm": 0.06966225643153073, "learning_rate": 4.965458219801093e-06, "loss": 0.543, "step": 325 }, { "epoch": 0.7970660146699267, "grad_norm": 0.09866879177917375, "learning_rate": 4.965245822608451e-06, "loss": 1.0296, "step": 326 }, { "epoch": 0.7995110024449877, "grad_norm": 0.07347140347158336, "learning_rate": 4.965032778971922e-06, "loss": 0.5063, "step": 327 }, { "epoch": 0.8019559902200489, "grad_norm": 0.19054475281388922, "learning_rate": 4.96481908894737e-06, "loss": 0.6894, "step": 328 }, { "epoch": 0.80440097799511, "grad_norm": 0.07284733096473339, "learning_rate": 4.96460475259083e-06, "loss": 0.5695, "step": 329 }, { "epoch": 0.8068459657701712, "grad_norm": 0.09008535141694302, "learning_rate": 4.964389769958506e-06, "loss": 1.3029, "step": 330 }, { "epoch": 0.8092909535452323, "grad_norm": 0.089770829238767, "learning_rate": 4.964174141106771e-06, "loss": 0.8726, "step": 331 }, { "epoch": 0.8117359413202934, "grad_norm": 0.06873500003410662, "learning_rate": 4.963957866092167e-06, "loss": 0.6982, "step": 332 }, { "epoch": 0.8141809290953546, "grad_norm": 0.07904686905540631, "learning_rate": 4.963740944971408e-06, "loss": 0.7142, "step": 333 }, { "epoch": 0.8166259168704156, "grad_norm": 0.08736362318728487, "learning_rate": 4.963523377801375e-06, "loss": 0.6935, "step": 334 }, { "epoch": 0.8190709046454768, "grad_norm": 0.0685205181370088, "learning_rate": 4.963305164639119e-06, "loss": 0.8552, "step": 335 }, { "epoch": 0.8215158924205379, "grad_norm": 0.07608190464825812, "learning_rate": 4.9630863055418586e-06, "loss": 1.0129, "step": 336 }, { "epoch": 0.823960880195599, "grad_norm": 0.08989916733561043, "learning_rate": 4.962866800566987e-06, "loss": 0.6522, "step": 337 }, { "epoch": 0.8264058679706602, "grad_norm": 0.08925436348801553, "learning_rate": 4.962646649772062e-06, "loss": 0.6081, "step": 338 }, { "epoch": 0.8288508557457213, "grad_norm": 0.08726652834221434, "learning_rate": 4.9624258532148116e-06, "loss": 0.4843, "step": 339 }, { "epoch": 0.8312958435207825, "grad_norm": 0.0719249289345701, "learning_rate": 4.962204410953135e-06, "loss": 0.8632, "step": 340 }, { "epoch": 0.8337408312958435, "grad_norm": 0.11864905600234564, "learning_rate": 4.961982323045097e-06, "loss": 0.478, "step": 341 }, { "epoch": 0.8361858190709046, "grad_norm": 0.09827018736440911, "learning_rate": 4.961759589548937e-06, "loss": 0.6793, "step": 342 }, { "epoch": 0.8386308068459658, "grad_norm": 0.09566327812901528, "learning_rate": 4.9615362105230595e-06, "loss": 0.6049, "step": 343 }, { "epoch": 0.8410757946210269, "grad_norm": 0.07070144260042295, "learning_rate": 4.96131218602604e-06, "loss": 0.6632, "step": 344 }, { "epoch": 0.843520782396088, "grad_norm": 0.0830820616569169, "learning_rate": 4.961087516116622e-06, "loss": 0.5394, "step": 345 }, { "epoch": 0.8459657701711492, "grad_norm": 0.11057233351152325, "learning_rate": 4.96086220085372e-06, "loss": 0.5714, "step": 346 }, { "epoch": 0.8484107579462102, "grad_norm": 0.06940174870366157, "learning_rate": 4.9606362402964155e-06, "loss": 0.5592, "step": 347 }, { "epoch": 0.8508557457212714, "grad_norm": 0.1000557755141698, "learning_rate": 4.960409634503962e-06, "loss": 0.8992, "step": 348 }, { "epoch": 0.8533007334963325, "grad_norm": 0.059909846258367434, "learning_rate": 4.96018238353578e-06, "loss": 0.7048, "step": 349 }, { "epoch": 0.8557457212713936, "grad_norm": 0.09166004278216504, "learning_rate": 4.959954487451461e-06, "loss": 0.762, "step": 350 }, { "epoch": 0.8581907090464548, "grad_norm": 0.07974581790684761, "learning_rate": 4.959725946310763e-06, "loss": 0.8647, "step": 351 }, { "epoch": 0.8606356968215159, "grad_norm": 0.07236511919690002, "learning_rate": 4.959496760173615e-06, "loss": 0.8792, "step": 352 }, { "epoch": 0.863080684596577, "grad_norm": 0.09899137329821, "learning_rate": 4.959266929100116e-06, "loss": 1.0709, "step": 353 }, { "epoch": 0.8655256723716381, "grad_norm": 0.11413955816663791, "learning_rate": 4.959036453150532e-06, "loss": 0.5481, "step": 354 }, { "epoch": 0.8679706601466992, "grad_norm": 0.07755742570827054, "learning_rate": 4.958805332385299e-06, "loss": 0.634, "step": 355 }, { "epoch": 0.8704156479217604, "grad_norm": 0.07744569811774757, "learning_rate": 4.9585735668650225e-06, "loss": 0.682, "step": 356 }, { "epoch": 0.8728606356968215, "grad_norm": 0.06647153060123005, "learning_rate": 4.958341156650477e-06, "loss": 0.4184, "step": 357 }, { "epoch": 0.8753056234718827, "grad_norm": 0.06996065107298906, "learning_rate": 4.958108101802605e-06, "loss": 0.5719, "step": 358 }, { "epoch": 0.8777506112469438, "grad_norm": 0.06901613460968989, "learning_rate": 4.95787440238252e-06, "loss": 0.601, "step": 359 }, { "epoch": 0.8801955990220048, "grad_norm": 0.10908189815492843, "learning_rate": 4.957640058451502e-06, "loss": 0.6224, "step": 360 }, { "epoch": 0.882640586797066, "grad_norm": 0.07519971673926353, "learning_rate": 4.957405070071001e-06, "loss": 0.4477, "step": 361 }, { "epoch": 0.8850855745721271, "grad_norm": 0.0941555591491303, "learning_rate": 4.957169437302639e-06, "loss": 0.7558, "step": 362 }, { "epoch": 0.8875305623471883, "grad_norm": 0.07441197948763678, "learning_rate": 4.9569331602082015e-06, "loss": 0.6555, "step": 363 }, { "epoch": 0.8899755501222494, "grad_norm": 0.10066309657301664, "learning_rate": 4.956696238849647e-06, "loss": 0.8644, "step": 364 }, { "epoch": 0.8924205378973105, "grad_norm": 0.060322858349314744, "learning_rate": 4.956458673289102e-06, "loss": 0.6688, "step": 365 }, { "epoch": 0.8948655256723717, "grad_norm": 0.07459723173599125, "learning_rate": 4.95622046358886e-06, "loss": 0.8178, "step": 366 }, { "epoch": 0.8973105134474327, "grad_norm": 0.07249063362436482, "learning_rate": 4.9559816098113874e-06, "loss": 0.7303, "step": 367 }, { "epoch": 0.8997555012224939, "grad_norm": 0.08828997974047631, "learning_rate": 4.955742112019316e-06, "loss": 0.5852, "step": 368 }, { "epoch": 0.902200488997555, "grad_norm": 0.06782498962008572, "learning_rate": 4.955501970275447e-06, "loss": 0.7972, "step": 369 }, { "epoch": 0.9046454767726161, "grad_norm": 0.09193553943494716, "learning_rate": 4.955261184642752e-06, "loss": 0.8673, "step": 370 }, { "epoch": 0.9070904645476773, "grad_norm": 0.0815775893964367, "learning_rate": 4.955019755184369e-06, "loss": 0.4362, "step": 371 }, { "epoch": 0.9095354523227384, "grad_norm": 0.0735163465256261, "learning_rate": 4.954777681963609e-06, "loss": 0.8682, "step": 372 }, { "epoch": 0.9119804400977995, "grad_norm": 0.0607622140169798, "learning_rate": 4.954534965043947e-06, "loss": 0.6241, "step": 373 }, { "epoch": 0.9144254278728606, "grad_norm": 0.09097140987050431, "learning_rate": 4.954291604489029e-06, "loss": 0.4739, "step": 374 }, { "epoch": 0.9168704156479217, "grad_norm": 0.09009782142347368, "learning_rate": 4.95404760036267e-06, "loss": 0.5854, "step": 375 }, { "epoch": 0.9193154034229829, "grad_norm": 0.10026963855738673, "learning_rate": 4.9538029527288545e-06, "loss": 0.5275, "step": 376 }, { "epoch": 0.921760391198044, "grad_norm": 0.06555658920279356, "learning_rate": 4.953557661651733e-06, "loss": 0.503, "step": 377 }, { "epoch": 0.9242053789731052, "grad_norm": 0.089032835475639, "learning_rate": 4.9533117271956286e-06, "loss": 1.0431, "step": 378 }, { "epoch": 0.9266503667481663, "grad_norm": 0.08650107379464002, "learning_rate": 4.953065149425029e-06, "loss": 0.9569, "step": 379 }, { "epoch": 0.9290953545232273, "grad_norm": 0.11733557999855201, "learning_rate": 4.952817928404593e-06, "loss": 0.5429, "step": 380 }, { "epoch": 0.9315403422982885, "grad_norm": 0.09000604566338713, "learning_rate": 4.952570064199148e-06, "loss": 0.8854, "step": 381 }, { "epoch": 0.9339853300733496, "grad_norm": 0.07898219178600274, "learning_rate": 4.952321556873689e-06, "loss": 0.4985, "step": 382 }, { "epoch": 0.9364303178484108, "grad_norm": 0.08131984444295022, "learning_rate": 4.952072406493382e-06, "loss": 0.7474, "step": 383 }, { "epoch": 0.9388753056234719, "grad_norm": 0.07659531144361392, "learning_rate": 4.951822613123558e-06, "loss": 0.5776, "step": 384 }, { "epoch": 0.941320293398533, "grad_norm": 0.10311268616339668, "learning_rate": 4.95157217682972e-06, "loss": 0.644, "step": 385 }, { "epoch": 0.9437652811735942, "grad_norm": 0.11823311556952831, "learning_rate": 4.951321097677536e-06, "loss": 0.7785, "step": 386 }, { "epoch": 0.9462102689486552, "grad_norm": 0.09248650345575686, "learning_rate": 4.9510693757328464e-06, "loss": 0.9739, "step": 387 }, { "epoch": 0.9486552567237164, "grad_norm": 0.05170086674421056, "learning_rate": 4.950817011061659e-06, "loss": 0.3465, "step": 388 }, { "epoch": 0.9511002444987775, "grad_norm": 0.07553393219105838, "learning_rate": 4.950564003730148e-06, "loss": 0.9992, "step": 389 }, { "epoch": 0.9535452322738386, "grad_norm": 0.1098893091484965, "learning_rate": 4.950310353804659e-06, "loss": 0.5216, "step": 390 }, { "epoch": 0.9559902200488998, "grad_norm": 0.08676183287703625, "learning_rate": 4.950056061351704e-06, "loss": 0.7807, "step": 391 }, { "epoch": 0.9584352078239609, "grad_norm": 0.0969873752285447, "learning_rate": 4.949801126437965e-06, "loss": 0.9221, "step": 392 }, { "epoch": 0.960880195599022, "grad_norm": 0.0854835249192429, "learning_rate": 4.94954554913029e-06, "loss": 0.8477, "step": 393 }, { "epoch": 0.9633251833740831, "grad_norm": 0.1207802459263373, "learning_rate": 4.9492893294957e-06, "loss": 0.61, "step": 394 }, { "epoch": 0.9657701711491442, "grad_norm": 0.09380341945695975, "learning_rate": 4.94903246760138e-06, "loss": 0.4529, "step": 395 }, { "epoch": 0.9682151589242054, "grad_norm": 0.09765648955425381, "learning_rate": 4.948774963514685e-06, "loss": 0.5131, "step": 396 }, { "epoch": 0.9706601466992665, "grad_norm": 0.09417538561326143, "learning_rate": 4.948516817303138e-06, "loss": 0.5324, "step": 397 }, { "epoch": 0.9731051344743277, "grad_norm": 0.0996378873170343, "learning_rate": 4.9482580290344325e-06, "loss": 0.5323, "step": 398 }, { "epoch": 0.9755501222493888, "grad_norm": 0.0888330405479318, "learning_rate": 4.947998598776429e-06, "loss": 0.5733, "step": 399 }, { "epoch": 0.9779951100244498, "grad_norm": 0.07794414942120746, "learning_rate": 4.9477385265971534e-06, "loss": 0.5127, "step": 400 }, { "epoch": 0.980440097799511, "grad_norm": 0.5580974031036784, "learning_rate": 4.947477812564804e-06, "loss": 0.9767, "step": 401 }, { "epoch": 0.9828850855745721, "grad_norm": 0.09663749476561667, "learning_rate": 4.9472164567477465e-06, "loss": 1.1788, "step": 402 }, { "epoch": 0.9853300733496333, "grad_norm": 0.09105593405749489, "learning_rate": 4.946954459214514e-06, "loss": 0.917, "step": 403 }, { "epoch": 0.9877750611246944, "grad_norm": 0.08597644599446, "learning_rate": 4.946691820033807e-06, "loss": 0.5118, "step": 404 }, { "epoch": 0.9902200488997555, "grad_norm": 0.09511456844930492, "learning_rate": 4.946428539274498e-06, "loss": 0.5629, "step": 405 }, { "epoch": 0.9926650366748166, "grad_norm": 0.0992683587423073, "learning_rate": 4.9461646170056235e-06, "loss": 1.0581, "step": 406 }, { "epoch": 0.9951100244498777, "grad_norm": 0.08940455356676082, "learning_rate": 4.94590005329639e-06, "loss": 0.4939, "step": 407 }, { "epoch": 0.9975550122249389, "grad_norm": 0.07104513302102258, "learning_rate": 4.945634848216173e-06, "loss": 0.6716, "step": 408 }, { "epoch": 1.0, "grad_norm": 0.06934468396267954, "learning_rate": 4.9453690018345144e-06, "loss": 0.5411, "step": 409 }, { "epoch": 1.0024449877750612, "grad_norm": 0.09105941539916199, "learning_rate": 4.945102514221126e-06, "loss": 0.7026, "step": 410 }, { "epoch": 1.0048899755501222, "grad_norm": 0.08924714864807391, "learning_rate": 4.944835385445886e-06, "loss": 0.5374, "step": 411 }, { "epoch": 1.0073349633251834, "grad_norm": 0.07192871393722046, "learning_rate": 4.944567615578844e-06, "loss": 0.656, "step": 412 }, { "epoch": 1.0097799511002445, "grad_norm": 0.0868990642167748, "learning_rate": 4.944299204690212e-06, "loss": 0.5648, "step": 413 }, { "epoch": 1.0122249388753057, "grad_norm": 0.07614894534555937, "learning_rate": 4.944030152850377e-06, "loss": 0.9752, "step": 414 }, { "epoch": 1.0146699266503667, "grad_norm": 0.0771935688711757, "learning_rate": 4.943760460129887e-06, "loss": 0.5689, "step": 415 }, { "epoch": 1.017114914425428, "grad_norm": 0.10450598696220409, "learning_rate": 4.943490126599463e-06, "loss": 0.8646, "step": 416 }, { "epoch": 1.019559902200489, "grad_norm": 0.09160376010054183, "learning_rate": 4.943219152329994e-06, "loss": 0.9548, "step": 417 }, { "epoch": 1.02200488997555, "grad_norm": 0.07862226233200668, "learning_rate": 4.942947537392535e-06, "loss": 0.4724, "step": 418 }, { "epoch": 1.0244498777506112, "grad_norm": 0.08603212800037052, "learning_rate": 4.942675281858308e-06, "loss": 0.6607, "step": 419 }, { "epoch": 1.0268948655256724, "grad_norm": 0.1251801147919071, "learning_rate": 4.942402385798707e-06, "loss": 0.9799, "step": 420 }, { "epoch": 1.0293398533007334, "grad_norm": 0.104080517337458, "learning_rate": 4.9421288492852905e-06, "loss": 0.5324, "step": 421 }, { "epoch": 1.0317848410757946, "grad_norm": 0.11161571103050913, "learning_rate": 4.941854672389786e-06, "loss": 0.6577, "step": 422 }, { "epoch": 1.0342298288508558, "grad_norm": 0.12021884481536534, "learning_rate": 4.941579855184089e-06, "loss": 0.9195, "step": 423 }, { "epoch": 1.0366748166259168, "grad_norm": 0.07288618577939406, "learning_rate": 4.941304397740262e-06, "loss": 0.5586, "step": 424 }, { "epoch": 1.039119804400978, "grad_norm": 0.10488610513725002, "learning_rate": 4.941028300130538e-06, "loss": 1.3365, "step": 425 }, { "epoch": 1.0415647921760391, "grad_norm": 0.11976183890406934, "learning_rate": 4.940751562427315e-06, "loss": 1.037, "step": 426 }, { "epoch": 1.0440097799511003, "grad_norm": 0.11851888661317067, "learning_rate": 4.94047418470316e-06, "loss": 0.5093, "step": 427 }, { "epoch": 1.0464547677261613, "grad_norm": 0.12888387709230967, "learning_rate": 4.940196167030808e-06, "loss": 0.6735, "step": 428 }, { "epoch": 1.0488997555012225, "grad_norm": 0.07891732461627987, "learning_rate": 4.939917509483162e-06, "loss": 0.5831, "step": 429 }, { "epoch": 1.0513447432762837, "grad_norm": 0.06648253960400075, "learning_rate": 4.939638212133292e-06, "loss": 0.5962, "step": 430 }, { "epoch": 1.0537897310513447, "grad_norm": 0.10996548239770788, "learning_rate": 4.9393582750544355e-06, "loss": 0.8285, "step": 431 }, { "epoch": 1.0562347188264058, "grad_norm": 0.08636236789345893, "learning_rate": 4.93907769832e-06, "loss": 0.5662, "step": 432 }, { "epoch": 1.058679706601467, "grad_norm": 0.08195324240018571, "learning_rate": 4.938796482003557e-06, "loss": 0.6888, "step": 433 }, { "epoch": 1.061124694376528, "grad_norm": 0.08104043991888836, "learning_rate": 4.93851462617885e-06, "loss": 0.7051, "step": 434 }, { "epoch": 1.0635696821515892, "grad_norm": 0.11111075681798355, "learning_rate": 4.938232130919786e-06, "loss": 0.7417, "step": 435 }, { "epoch": 1.0660146699266504, "grad_norm": 0.08823093697470814, "learning_rate": 4.9379489963004435e-06, "loss": 1.0322, "step": 436 }, { "epoch": 1.0684596577017116, "grad_norm": 0.0990873186490748, "learning_rate": 4.937665222395066e-06, "loss": 0.4236, "step": 437 }, { "epoch": 1.0709046454767726, "grad_norm": 0.07693311404671367, "learning_rate": 4.937380809278066e-06, "loss": 0.5643, "step": 438 }, { "epoch": 1.0733496332518337, "grad_norm": 0.11993647913661866, "learning_rate": 4.937095757024023e-06, "loss": 0.9462, "step": 439 }, { "epoch": 1.075794621026895, "grad_norm": 0.13528409042624007, "learning_rate": 4.936810065707684e-06, "loss": 0.8804, "step": 440 }, { "epoch": 1.078239608801956, "grad_norm": 0.09500668501845402, "learning_rate": 4.936523735403963e-06, "loss": 0.6563, "step": 441 }, { "epoch": 1.080684596577017, "grad_norm": 0.08985865165384338, "learning_rate": 4.936236766187944e-06, "loss": 0.5655, "step": 442 }, { "epoch": 1.0831295843520783, "grad_norm": 0.06979213353817815, "learning_rate": 4.935949158134876e-06, "loss": 0.4365, "step": 443 }, { "epoch": 1.0855745721271393, "grad_norm": 0.08587094419976782, "learning_rate": 4.935660911320176e-06, "loss": 0.5458, "step": 444 }, { "epoch": 1.0880195599022005, "grad_norm": 0.06459945875041573, "learning_rate": 4.9353720258194305e-06, "loss": 0.3973, "step": 445 }, { "epoch": 1.0904645476772616, "grad_norm": 0.06338378792968512, "learning_rate": 4.93508250170839e-06, "loss": 0.4729, "step": 446 }, { "epoch": 1.0929095354523228, "grad_norm": 0.08758125427073629, "learning_rate": 4.934792339062975e-06, "loss": 0.9112, "step": 447 }, { "epoch": 1.0953545232273838, "grad_norm": 0.08612473092257603, "learning_rate": 4.934501537959274e-06, "loss": 0.6525, "step": 448 }, { "epoch": 1.097799511002445, "grad_norm": 0.08764721761343358, "learning_rate": 4.93421009847354e-06, "loss": 0.7525, "step": 449 }, { "epoch": 1.1002444987775062, "grad_norm": 0.09500426925935404, "learning_rate": 4.933918020682196e-06, "loss": 0.7781, "step": 450 }, { "epoch": 1.1026894865525672, "grad_norm": 0.13051747246092554, "learning_rate": 4.933625304661831e-06, "loss": 0.5881, "step": 451 }, { "epoch": 1.1051344743276283, "grad_norm": 0.11079305835216474, "learning_rate": 4.933331950489202e-06, "loss": 0.6278, "step": 452 }, { "epoch": 1.1075794621026895, "grad_norm": 0.1226532214058774, "learning_rate": 4.933037958241233e-06, "loss": 1.0762, "step": 453 }, { "epoch": 1.1100244498777505, "grad_norm": 0.08619532896324963, "learning_rate": 4.932743327995017e-06, "loss": 0.7636, "step": 454 }, { "epoch": 1.1124694376528117, "grad_norm": 0.08527134420681212, "learning_rate": 4.932448059827811e-06, "loss": 0.5058, "step": 455 }, { "epoch": 1.114914425427873, "grad_norm": 0.08535253600194993, "learning_rate": 4.9321521538170416e-06, "loss": 0.9615, "step": 456 }, { "epoch": 1.117359413202934, "grad_norm": 0.08863816004879815, "learning_rate": 4.931855610040302e-06, "loss": 1.0531, "step": 457 }, { "epoch": 1.119804400977995, "grad_norm": 0.09487267031600953, "learning_rate": 4.931558428575353e-06, "loss": 0.8523, "step": 458 }, { "epoch": 1.1222493887530562, "grad_norm": 0.09118134439196415, "learning_rate": 4.931260609500122e-06, "loss": 0.5716, "step": 459 }, { "epoch": 1.1246943765281174, "grad_norm": 0.08850224361698028, "learning_rate": 4.930962152892705e-06, "loss": 0.5606, "step": 460 }, { "epoch": 1.1271393643031784, "grad_norm": 0.07368599896584542, "learning_rate": 4.930663058831363e-06, "loss": 0.6512, "step": 461 }, { "epoch": 1.1295843520782396, "grad_norm": 0.15835534489259462, "learning_rate": 4.930363327394526e-06, "loss": 0.483, "step": 462 }, { "epoch": 1.1320293398533008, "grad_norm": 0.09847635370746623, "learning_rate": 4.930062958660791e-06, "loss": 0.6378, "step": 463 }, { "epoch": 1.1344743276283618, "grad_norm": 0.12826222819889038, "learning_rate": 4.929761952708919e-06, "loss": 0.6869, "step": 464 }, { "epoch": 1.136919315403423, "grad_norm": 0.0984962566017247, "learning_rate": 4.929460309617844e-06, "loss": 0.6264, "step": 465 }, { "epoch": 1.1393643031784841, "grad_norm": 0.06699741722565294, "learning_rate": 4.92915802946666e-06, "loss": 0.5402, "step": 466 }, { "epoch": 1.141809290953545, "grad_norm": 0.06976404766756682, "learning_rate": 4.928855112334637e-06, "loss": 0.5558, "step": 467 }, { "epoch": 1.1442542787286063, "grad_norm": 0.09089804323140854, "learning_rate": 4.928551558301202e-06, "loss": 0.5615, "step": 468 }, { "epoch": 1.1466992665036675, "grad_norm": 0.0728901145580466, "learning_rate": 4.928247367445955e-06, "loss": 0.5883, "step": 469 }, { "epoch": 1.1491442542787287, "grad_norm": 0.07749424237991626, "learning_rate": 4.927942539848664e-06, "loss": 0.4626, "step": 470 }, { "epoch": 1.1515892420537897, "grad_norm": 0.10542078728716303, "learning_rate": 4.927637075589259e-06, "loss": 0.4384, "step": 471 }, { "epoch": 1.1540342298288508, "grad_norm": 0.09666276313939713, "learning_rate": 4.9273309747478405e-06, "loss": 0.8628, "step": 472 }, { "epoch": 1.156479217603912, "grad_norm": 0.08803531242886628, "learning_rate": 4.927024237404676e-06, "loss": 0.9808, "step": 473 }, { "epoch": 1.1589242053789732, "grad_norm": 0.06543209808717486, "learning_rate": 4.926716863640198e-06, "loss": 0.7855, "step": 474 }, { "epoch": 1.1613691931540342, "grad_norm": 0.07860052143999972, "learning_rate": 4.926408853535009e-06, "loss": 0.8213, "step": 475 }, { "epoch": 1.1638141809290954, "grad_norm": 0.10931352468951576, "learning_rate": 4.926100207169874e-06, "loss": 0.581, "step": 476 }, { "epoch": 1.1662591687041566, "grad_norm": 0.15010409191628882, "learning_rate": 4.925790924625727e-06, "loss": 0.6576, "step": 477 }, { "epoch": 1.1687041564792175, "grad_norm": 0.09910094501518846, "learning_rate": 4.925481005983671e-06, "loss": 0.5308, "step": 478 }, { "epoch": 1.1711491442542787, "grad_norm": 0.09770242061791468, "learning_rate": 4.925170451324973e-06, "loss": 0.9874, "step": 479 }, { "epoch": 1.17359413202934, "grad_norm": 0.08030153268201529, "learning_rate": 4.924859260731067e-06, "loss": 0.6364, "step": 480 }, { "epoch": 1.176039119804401, "grad_norm": 0.06928780822098991, "learning_rate": 4.924547434283553e-06, "loss": 0.5545, "step": 481 }, { "epoch": 1.178484107579462, "grad_norm": 0.09237649881859512, "learning_rate": 4.924234972064202e-06, "loss": 0.5735, "step": 482 }, { "epoch": 1.1809290953545233, "grad_norm": 0.08688040667426593, "learning_rate": 4.923921874154947e-06, "loss": 0.6035, "step": 483 }, { "epoch": 1.1833740831295843, "grad_norm": 0.11807118434345489, "learning_rate": 4.9236081406378896e-06, "loss": 0.828, "step": 484 }, { "epoch": 1.1858190709046454, "grad_norm": 0.08796747058149752, "learning_rate": 4.923293771595298e-06, "loss": 0.6264, "step": 485 }, { "epoch": 1.1882640586797066, "grad_norm": 0.09561220786072974, "learning_rate": 4.922978767109606e-06, "loss": 0.7769, "step": 486 }, { "epoch": 1.1907090464547676, "grad_norm": 0.09088930540866695, "learning_rate": 4.922663127263417e-06, "loss": 0.8633, "step": 487 }, { "epoch": 1.1931540342298288, "grad_norm": 0.10340377119760334, "learning_rate": 4.922346852139498e-06, "loss": 0.5747, "step": 488 }, { "epoch": 1.19559902200489, "grad_norm": 0.11309677235280163, "learning_rate": 4.922029941820784e-06, "loss": 0.7417, "step": 489 }, { "epoch": 1.1980440097799512, "grad_norm": 0.08953581187302126, "learning_rate": 4.921712396390375e-06, "loss": 0.7057, "step": 490 }, { "epoch": 1.2004889975550121, "grad_norm": 0.09542073194832858, "learning_rate": 4.92139421593154e-06, "loss": 1.0115, "step": 491 }, { "epoch": 1.2029339853300733, "grad_norm": 0.09435097212535767, "learning_rate": 4.921075400527712e-06, "loss": 0.8826, "step": 492 }, { "epoch": 1.2053789731051345, "grad_norm": 0.11780954250490948, "learning_rate": 4.920755950262491e-06, "loss": 0.6608, "step": 493 }, { "epoch": 1.2078239608801957, "grad_norm": 0.07779457888373696, "learning_rate": 4.920435865219647e-06, "loss": 0.4154, "step": 494 }, { "epoch": 1.2102689486552567, "grad_norm": 0.11527553832608461, "learning_rate": 4.920115145483112e-06, "loss": 0.5662, "step": 495 }, { "epoch": 1.2127139364303179, "grad_norm": 0.0888650857447355, "learning_rate": 4.919793791136986e-06, "loss": 0.9286, "step": 496 }, { "epoch": 1.215158924205379, "grad_norm": 0.07589598115227027, "learning_rate": 4.919471802265536e-06, "loss": 0.5071, "step": 497 }, { "epoch": 1.21760391198044, "grad_norm": 0.0964212187879595, "learning_rate": 4.919149178953193e-06, "loss": 0.952, "step": 498 }, { "epoch": 1.2200488997555012, "grad_norm": 0.0983068903375565, "learning_rate": 4.918825921284559e-06, "loss": 0.8473, "step": 499 }, { "epoch": 1.2224938875305624, "grad_norm": 0.11017880721530393, "learning_rate": 4.918502029344398e-06, "loss": 0.5354, "step": 500 }, { "epoch": 1.2249388753056234, "grad_norm": 0.08369505279974075, "learning_rate": 4.918177503217642e-06, "loss": 0.5319, "step": 501 }, { "epoch": 1.2273838630806846, "grad_norm": 0.08956559278845494, "learning_rate": 4.917852342989389e-06, "loss": 0.8995, "step": 502 }, { "epoch": 1.2298288508557458, "grad_norm": 0.07862102544640515, "learning_rate": 4.917526548744903e-06, "loss": 0.6915, "step": 503 }, { "epoch": 1.2322738386308068, "grad_norm": 0.11242447186742278, "learning_rate": 4.917200120569617e-06, "loss": 0.5762, "step": 504 }, { "epoch": 1.234718826405868, "grad_norm": 0.08847837399366476, "learning_rate": 4.916873058549125e-06, "loss": 0.5408, "step": 505 }, { "epoch": 1.2371638141809291, "grad_norm": 0.09974872244595148, "learning_rate": 4.916545362769193e-06, "loss": 0.7934, "step": 506 }, { "epoch": 1.23960880195599, "grad_norm": 0.0912495810073242, "learning_rate": 4.9162170333157475e-06, "loss": 0.6672, "step": 507 }, { "epoch": 1.2420537897310513, "grad_norm": 0.08109256305667999, "learning_rate": 4.915888070274886e-06, "loss": 0.5375, "step": 508 }, { "epoch": 1.2444987775061125, "grad_norm": 0.07759709958745548, "learning_rate": 4.915558473732869e-06, "loss": 0.6376, "step": 509 }, { "epoch": 1.2469437652811737, "grad_norm": 0.10773079697506899, "learning_rate": 4.915228243776125e-06, "loss": 0.7018, "step": 510 }, { "epoch": 1.2493887530562346, "grad_norm": 0.06967041925635098, "learning_rate": 4.914897380491248e-06, "loss": 0.3793, "step": 511 }, { "epoch": 1.2518337408312958, "grad_norm": 0.11057503599973109, "learning_rate": 4.914565883964997e-06, "loss": 0.5856, "step": 512 }, { "epoch": 1.254278728606357, "grad_norm": 0.14273770447725426, "learning_rate": 4.914233754284299e-06, "loss": 0.8803, "step": 513 }, { "epoch": 1.2567237163814182, "grad_norm": 0.11294192465425508, "learning_rate": 4.913900991536245e-06, "loss": 0.7196, "step": 514 }, { "epoch": 1.2591687041564792, "grad_norm": 0.107378850412448, "learning_rate": 4.913567595808094e-06, "loss": 0.9583, "step": 515 }, { "epoch": 1.2616136919315404, "grad_norm": 0.11842710303845422, "learning_rate": 4.913233567187269e-06, "loss": 0.5769, "step": 516 }, { "epoch": 1.2640586797066016, "grad_norm": 0.10554500812774531, "learning_rate": 4.91289890576136e-06, "loss": 1.0385, "step": 517 }, { "epoch": 1.2665036674816625, "grad_norm": 0.09190779802084394, "learning_rate": 4.912563611618125e-06, "loss": 0.5658, "step": 518 }, { "epoch": 1.2689486552567237, "grad_norm": 0.07373464166101175, "learning_rate": 4.912227684845483e-06, "loss": 0.5163, "step": 519 }, { "epoch": 1.271393643031785, "grad_norm": 0.1363402262737353, "learning_rate": 4.911891125531524e-06, "loss": 0.5468, "step": 520 }, { "epoch": 1.273838630806846, "grad_norm": 0.07531750339186945, "learning_rate": 4.911553933764499e-06, "loss": 0.5689, "step": 521 }, { "epoch": 1.276283618581907, "grad_norm": 0.11737246092971938, "learning_rate": 4.911216109632829e-06, "loss": 0.6143, "step": 522 }, { "epoch": 1.2787286063569683, "grad_norm": 0.10621515820557399, "learning_rate": 4.9108776532251e-06, "loss": 0.5884, "step": 523 }, { "epoch": 1.2811735941320292, "grad_norm": 0.10331919975435638, "learning_rate": 4.910538564630061e-06, "loss": 0.9745, "step": 524 }, { "epoch": 1.2836185819070904, "grad_norm": 0.05899657654770767, "learning_rate": 4.91019884393663e-06, "loss": 0.3933, "step": 525 }, { "epoch": 1.2860635696821516, "grad_norm": 0.09123789489194563, "learning_rate": 4.90985849123389e-06, "loss": 0.7584, "step": 526 }, { "epoch": 1.2885085574572126, "grad_norm": 0.1295666077133763, "learning_rate": 4.909517506611088e-06, "loss": 1.1315, "step": 527 }, { "epoch": 1.2909535452322738, "grad_norm": 0.08248827441123903, "learning_rate": 4.909175890157639e-06, "loss": 0.9952, "step": 528 }, { "epoch": 1.293398533007335, "grad_norm": 0.08797103703404537, "learning_rate": 4.908833641963122e-06, "loss": 0.6969, "step": 529 }, { "epoch": 1.295843520782396, "grad_norm": 0.1240972475850901, "learning_rate": 4.9084907621172826e-06, "loss": 0.6394, "step": 530 }, { "epoch": 1.2982885085574571, "grad_norm": 0.09819301167927469, "learning_rate": 4.908147250710032e-06, "loss": 0.4723, "step": 531 }, { "epoch": 1.3007334963325183, "grad_norm": 0.11175425875316981, "learning_rate": 4.907803107831446e-06, "loss": 0.9053, "step": 532 }, { "epoch": 1.3031784841075795, "grad_norm": 0.10047112342425647, "learning_rate": 4.907458333571767e-06, "loss": 0.5744, "step": 533 }, { "epoch": 1.3056234718826407, "grad_norm": 0.11503275720822878, "learning_rate": 4.907112928021404e-06, "loss": 0.5232, "step": 534 }, { "epoch": 1.3080684596577017, "grad_norm": 0.0715738413310953, "learning_rate": 4.906766891270929e-06, "loss": 0.5465, "step": 535 }, { "epoch": 1.3105134474327629, "grad_norm": 0.10741727799875068, "learning_rate": 4.906420223411081e-06, "loss": 0.5217, "step": 536 }, { "epoch": 1.312958435207824, "grad_norm": 0.10568838429719278, "learning_rate": 4.906072924532763e-06, "loss": 0.8003, "step": 537 }, { "epoch": 1.315403422982885, "grad_norm": 0.08231711039006366, "learning_rate": 4.905724994727047e-06, "loss": 0.4224, "step": 538 }, { "epoch": 1.3178484107579462, "grad_norm": 0.08272248563292553, "learning_rate": 4.905376434085166e-06, "loss": 0.5293, "step": 539 }, { "epoch": 1.3202933985330074, "grad_norm": 0.14225219728070146, "learning_rate": 4.905027242698523e-06, "loss": 0.7543, "step": 540 }, { "epoch": 1.3227383863080684, "grad_norm": 0.08937075513214697, "learning_rate": 4.904677420658681e-06, "loss": 0.7368, "step": 541 }, { "epoch": 1.3251833740831296, "grad_norm": 0.10592659035193994, "learning_rate": 4.904326968057375e-06, "loss": 0.9604, "step": 542 }, { "epoch": 1.3276283618581908, "grad_norm": 0.11904973404505163, "learning_rate": 4.903975884986498e-06, "loss": 0.6545, "step": 543 }, { "epoch": 1.3300733496332517, "grad_norm": 0.14959675224549268, "learning_rate": 4.903624171538114e-06, "loss": 0.528, "step": 544 }, { "epoch": 1.332518337408313, "grad_norm": 0.09726056784858919, "learning_rate": 4.90327182780445e-06, "loss": 0.9798, "step": 545 }, { "epoch": 1.3349633251833741, "grad_norm": 0.1177418094033225, "learning_rate": 4.902918853877899e-06, "loss": 0.929, "step": 546 }, { "epoch": 1.337408312958435, "grad_norm": 0.08955869555869381, "learning_rate": 4.902565249851018e-06, "loss": 0.5253, "step": 547 }, { "epoch": 1.3398533007334963, "grad_norm": 0.14758677449990193, "learning_rate": 4.9022110158165305e-06, "loss": 1.174, "step": 548 }, { "epoch": 1.3422982885085575, "grad_norm": 0.10098649490249025, "learning_rate": 4.901856151867325e-06, "loss": 1.0261, "step": 549 }, { "epoch": 1.3447432762836184, "grad_norm": 0.09395978178833167, "learning_rate": 4.901500658096454e-06, "loss": 1.0083, "step": 550 }, { "epoch": 1.3471882640586796, "grad_norm": 0.09996676907015078, "learning_rate": 4.901144534597137e-06, "loss": 0.6274, "step": 551 }, { "epoch": 1.3496332518337408, "grad_norm": 0.08130887512859405, "learning_rate": 4.900787781462759e-06, "loss": 0.5326, "step": 552 }, { "epoch": 1.352078239608802, "grad_norm": 0.10394170759786459, "learning_rate": 4.9004303987868666e-06, "loss": 1.2606, "step": 553 }, { "epoch": 1.3545232273838632, "grad_norm": 0.13384230535890962, "learning_rate": 4.900072386663174e-06, "loss": 0.397, "step": 554 }, { "epoch": 1.3569682151589242, "grad_norm": 0.09223993894602162, "learning_rate": 4.89971374518556e-06, "loss": 0.6086, "step": 555 }, { "epoch": 1.3594132029339854, "grad_norm": 0.14212086767132762, "learning_rate": 4.899354474448071e-06, "loss": 0.5452, "step": 556 }, { "epoch": 1.3618581907090466, "grad_norm": 0.12688109844870638, "learning_rate": 4.898994574544913e-06, "loss": 0.9662, "step": 557 }, { "epoch": 1.3643031784841075, "grad_norm": 0.10037622149272513, "learning_rate": 4.898634045570462e-06, "loss": 1.0087, "step": 558 }, { "epoch": 1.3667481662591687, "grad_norm": 0.09468301175713245, "learning_rate": 4.898272887619256e-06, "loss": 0.6379, "step": 559 }, { "epoch": 1.36919315403423, "grad_norm": 0.07645506116702744, "learning_rate": 4.897911100785999e-06, "loss": 0.51, "step": 560 }, { "epoch": 1.371638141809291, "grad_norm": 0.1022377367280307, "learning_rate": 4.89754868516556e-06, "loss": 0.7167, "step": 561 }, { "epoch": 1.374083129584352, "grad_norm": 0.08250956797244187, "learning_rate": 4.897185640852973e-06, "loss": 0.4682, "step": 562 }, { "epoch": 1.3765281173594133, "grad_norm": 0.10906077574384798, "learning_rate": 4.8968219679434355e-06, "loss": 0.5498, "step": 563 }, { "epoch": 1.3789731051344742, "grad_norm": 0.1050286030157204, "learning_rate": 4.896457666532312e-06, "loss": 0.9047, "step": 564 }, { "epoch": 1.3814180929095354, "grad_norm": 0.11088361772351074, "learning_rate": 4.89609273671513e-06, "loss": 0.4295, "step": 565 }, { "epoch": 1.3838630806845966, "grad_norm": 0.1169530036601329, "learning_rate": 4.895727178587582e-06, "loss": 0.483, "step": 566 }, { "epoch": 1.3863080684596576, "grad_norm": 0.10731107916852321, "learning_rate": 4.895360992245527e-06, "loss": 0.5084, "step": 567 }, { "epoch": 1.3887530562347188, "grad_norm": 0.08644998920942834, "learning_rate": 4.894994177784986e-06, "loss": 0.6619, "step": 568 }, { "epoch": 1.39119804400978, "grad_norm": 0.09434787955252864, "learning_rate": 4.894626735302148e-06, "loss": 0.9048, "step": 569 }, { "epoch": 1.393643031784841, "grad_norm": 0.13113408831860734, "learning_rate": 4.894258664893363e-06, "loss": 0.6551, "step": 570 }, { "epoch": 1.3960880195599021, "grad_norm": 0.09423882718762316, "learning_rate": 4.893889966655148e-06, "loss": 0.7766, "step": 571 }, { "epoch": 1.3985330073349633, "grad_norm": 0.09437923132162945, "learning_rate": 4.893520640684185e-06, "loss": 0.5685, "step": 572 }, { "epoch": 1.4009779951100245, "grad_norm": 0.11353033262784303, "learning_rate": 4.893150687077318e-06, "loss": 0.4965, "step": 573 }, { "epoch": 1.4034229828850857, "grad_norm": 0.08549231850722978, "learning_rate": 4.892780105931559e-06, "loss": 0.5272, "step": 574 }, { "epoch": 1.4058679706601467, "grad_norm": 0.11421060288616502, "learning_rate": 4.892408897344083e-06, "loss": 0.6486, "step": 575 }, { "epoch": 1.4083129584352079, "grad_norm": 0.1331043299439341, "learning_rate": 4.8920370614122285e-06, "loss": 0.5588, "step": 576 }, { "epoch": 1.410757946210269, "grad_norm": 0.1976210080602442, "learning_rate": 4.891664598233499e-06, "loss": 0.7242, "step": 577 }, { "epoch": 1.41320293398533, "grad_norm": 0.13434955998938355, "learning_rate": 4.891291507905562e-06, "loss": 0.4834, "step": 578 }, { "epoch": 1.4156479217603912, "grad_norm": 0.12494403475060947, "learning_rate": 4.8909177905262535e-06, "loss": 1.1105, "step": 579 }, { "epoch": 1.4180929095354524, "grad_norm": 0.11365029821295738, "learning_rate": 4.8905434461935685e-06, "loss": 0.4461, "step": 580 }, { "epoch": 1.4205378973105134, "grad_norm": 0.10077186189663338, "learning_rate": 4.890168475005669e-06, "loss": 0.8464, "step": 581 }, { "epoch": 1.4229828850855746, "grad_norm": 0.12555527740911893, "learning_rate": 4.889792877060881e-06, "loss": 0.5611, "step": 582 }, { "epoch": 1.4254278728606358, "grad_norm": 0.09187134597673266, "learning_rate": 4.889416652457695e-06, "loss": 0.4763, "step": 583 }, { "epoch": 1.4278728606356967, "grad_norm": 0.09586209515602012, "learning_rate": 4.889039801294766e-06, "loss": 0.6886, "step": 584 }, { "epoch": 1.430317848410758, "grad_norm": 0.10092738627949223, "learning_rate": 4.888662323670913e-06, "loss": 0.8033, "step": 585 }, { "epoch": 1.4327628361858191, "grad_norm": 0.11759176095909574, "learning_rate": 4.888284219685118e-06, "loss": 0.4628, "step": 586 }, { "epoch": 1.43520782396088, "grad_norm": 0.09084250273361362, "learning_rate": 4.8879054894365304e-06, "loss": 0.6345, "step": 587 }, { "epoch": 1.4376528117359413, "grad_norm": 0.11195802313909085, "learning_rate": 4.887526133024461e-06, "loss": 0.5431, "step": 588 }, { "epoch": 1.4400977995110025, "grad_norm": 0.07866952681745683, "learning_rate": 4.887146150548385e-06, "loss": 0.5937, "step": 589 }, { "epoch": 1.4425427872860634, "grad_norm": 0.1047830764367085, "learning_rate": 4.886765542107944e-06, "loss": 0.6097, "step": 590 }, { "epoch": 1.4449877750611246, "grad_norm": 0.11716144373604559, "learning_rate": 4.886384307802942e-06, "loss": 0.5488, "step": 591 }, { "epoch": 1.4474327628361858, "grad_norm": 0.10157336232284198, "learning_rate": 4.8860024477333454e-06, "loss": 0.4825, "step": 592 }, { "epoch": 1.449877750611247, "grad_norm": 0.09591034445057087, "learning_rate": 4.88561996199929e-06, "loss": 0.7181, "step": 593 }, { "epoch": 1.452322738386308, "grad_norm": 0.12073546731090626, "learning_rate": 4.885236850701069e-06, "loss": 0.6599, "step": 594 }, { "epoch": 1.4547677261613692, "grad_norm": 0.12359034156662388, "learning_rate": 4.8848531139391445e-06, "loss": 0.5033, "step": 595 }, { "epoch": 1.4572127139364304, "grad_norm": 0.11139842337838901, "learning_rate": 4.884468751814141e-06, "loss": 0.6316, "step": 596 }, { "epoch": 1.4596577017114916, "grad_norm": 0.09633096943614694, "learning_rate": 4.884083764426848e-06, "loss": 0.5619, "step": 597 }, { "epoch": 1.4621026894865525, "grad_norm": 0.12047753581913498, "learning_rate": 4.883698151878216e-06, "loss": 0.8111, "step": 598 }, { "epoch": 1.4645476772616137, "grad_norm": 0.08539043747485742, "learning_rate": 4.883311914269363e-06, "loss": 0.499, "step": 599 }, { "epoch": 1.466992665036675, "grad_norm": 0.0853362857331394, "learning_rate": 4.882925051701568e-06, "loss": 0.5415, "step": 600 }, { "epoch": 1.4694376528117359, "grad_norm": 0.07729888911689278, "learning_rate": 4.882537564276277e-06, "loss": 0.502, "step": 601 }, { "epoch": 1.471882640586797, "grad_norm": 0.119589983288633, "learning_rate": 4.882149452095098e-06, "loss": 0.5199, "step": 602 }, { "epoch": 1.4743276283618583, "grad_norm": 0.10830874637390676, "learning_rate": 4.8817607152598e-06, "loss": 0.7056, "step": 603 }, { "epoch": 1.4767726161369192, "grad_norm": 0.11304500383396528, "learning_rate": 4.881371353872321e-06, "loss": 0.5552, "step": 604 }, { "epoch": 1.4792176039119804, "grad_norm": 0.14761148604727087, "learning_rate": 4.88098136803476e-06, "loss": 0.4864, "step": 605 }, { "epoch": 1.4816625916870416, "grad_norm": 0.08616052223843693, "learning_rate": 4.8805907578493815e-06, "loss": 0.6932, "step": 606 }, { "epoch": 1.4841075794621026, "grad_norm": 0.0969487209157877, "learning_rate": 4.880199523418612e-06, "loss": 0.5311, "step": 607 }, { "epoch": 1.4865525672371638, "grad_norm": 0.1708784553402437, "learning_rate": 4.87980766484504e-06, "loss": 0.7008, "step": 608 }, { "epoch": 1.488997555012225, "grad_norm": 0.1516345812888604, "learning_rate": 4.879415182231421e-06, "loss": 0.9707, "step": 609 }, { "epoch": 1.491442542787286, "grad_norm": 0.10924570175359778, "learning_rate": 4.879022075680674e-06, "loss": 0.4253, "step": 610 }, { "epoch": 1.4938875305623471, "grad_norm": 0.09878909510845578, "learning_rate": 4.87862834529588e-06, "loss": 0.9475, "step": 611 }, { "epoch": 1.4963325183374083, "grad_norm": 0.10440136070009452, "learning_rate": 4.878233991180282e-06, "loss": 0.507, "step": 612 }, { "epoch": 1.4987775061124695, "grad_norm": 0.14019649232764822, "learning_rate": 4.877839013437292e-06, "loss": 0.5253, "step": 613 }, { "epoch": 1.5012224938875307, "grad_norm": 0.1657457171820653, "learning_rate": 4.87744341217048e-06, "loss": 0.9782, "step": 614 }, { "epoch": 1.5036674816625917, "grad_norm": 0.10151625603491048, "learning_rate": 4.877047187483583e-06, "loss": 0.6492, "step": 615 }, { "epoch": 1.5061124694376526, "grad_norm": 0.07992442002773023, "learning_rate": 4.8766503394804996e-06, "loss": 0.4593, "step": 616 }, { "epoch": 1.508557457212714, "grad_norm": 0.1881540397932995, "learning_rate": 4.876252868265292e-06, "loss": 0.5718, "step": 617 }, { "epoch": 1.511002444987775, "grad_norm": 0.14995548029453482, "learning_rate": 4.875854773942187e-06, "loss": 1.1579, "step": 618 }, { "epoch": 1.5134474327628362, "grad_norm": 0.1251722466892237, "learning_rate": 4.8754560566155736e-06, "loss": 0.5597, "step": 619 }, { "epoch": 1.5158924205378974, "grad_norm": 0.11206684872942402, "learning_rate": 4.875056716390004e-06, "loss": 0.5726, "step": 620 }, { "epoch": 1.5183374083129584, "grad_norm": 0.13186339836517458, "learning_rate": 4.874656753370196e-06, "loss": 0.4605, "step": 621 }, { "epoch": 1.5207823960880196, "grad_norm": 0.10769843328151026, "learning_rate": 4.874256167661028e-06, "loss": 0.4862, "step": 622 }, { "epoch": 1.5232273838630808, "grad_norm": 0.16229971977633795, "learning_rate": 4.873854959367543e-06, "loss": 0.5198, "step": 623 }, { "epoch": 1.5256723716381417, "grad_norm": 0.0718356191915513, "learning_rate": 4.873453128594947e-06, "loss": 0.4432, "step": 624 }, { "epoch": 1.528117359413203, "grad_norm": 0.14969965259276746, "learning_rate": 4.873050675448608e-06, "loss": 0.72, "step": 625 }, { "epoch": 1.5305623471882641, "grad_norm": 0.1438543773392057, "learning_rate": 4.8726476000340605e-06, "loss": 1.0206, "step": 626 }, { "epoch": 1.533007334963325, "grad_norm": 0.1090563727566657, "learning_rate": 4.872243902456999e-06, "loss": 0.8143, "step": 627 }, { "epoch": 1.5354523227383863, "grad_norm": 0.12047679246999657, "learning_rate": 4.8718395828232825e-06, "loss": 1.0677, "step": 628 }, { "epoch": 1.5378973105134475, "grad_norm": 0.18664156072454702, "learning_rate": 4.871434641238932e-06, "loss": 0.8253, "step": 629 }, { "epoch": 1.5403422982885084, "grad_norm": 0.09963534330499423, "learning_rate": 4.871029077810134e-06, "loss": 0.6614, "step": 630 }, { "epoch": 1.5427872860635696, "grad_norm": 0.11440225158856199, "learning_rate": 4.870622892643234e-06, "loss": 0.81, "step": 631 }, { "epoch": 1.5452322738386308, "grad_norm": 0.7695658222676224, "learning_rate": 4.870216085844747e-06, "loss": 1.0181, "step": 632 }, { "epoch": 1.5476772616136918, "grad_norm": 0.09633012470089278, "learning_rate": 4.869808657521344e-06, "loss": 0.5962, "step": 633 }, { "epoch": 1.5501222493887532, "grad_norm": 0.10346289251632876, "learning_rate": 4.869400607779863e-06, "loss": 0.6151, "step": 634 }, { "epoch": 1.5525672371638142, "grad_norm": 0.152883330338306, "learning_rate": 4.868991936727303e-06, "loss": 0.8945, "step": 635 }, { "epoch": 1.5550122249388751, "grad_norm": 0.10790987901824774, "learning_rate": 4.868582644470829e-06, "loss": 0.7651, "step": 636 }, { "epoch": 1.5574572127139366, "grad_norm": 0.12241414703705915, "learning_rate": 4.868172731117764e-06, "loss": 0.9482, "step": 637 }, { "epoch": 1.5599022004889975, "grad_norm": 0.10669500105409589, "learning_rate": 4.867762196775599e-06, "loss": 0.71, "step": 638 }, { "epoch": 1.5623471882640587, "grad_norm": 0.09792055398862386, "learning_rate": 4.867351041551984e-06, "loss": 0.6317, "step": 639 }, { "epoch": 1.56479217603912, "grad_norm": 0.09256610738995151, "learning_rate": 4.866939265554735e-06, "loss": 0.4865, "step": 640 }, { "epoch": 1.5672371638141809, "grad_norm": 0.15834405522303402, "learning_rate": 4.866526868891828e-06, "loss": 0.9382, "step": 641 }, { "epoch": 1.569682151589242, "grad_norm": 0.10862668475006454, "learning_rate": 4.866113851671403e-06, "loss": 0.596, "step": 642 }, { "epoch": 1.5721271393643033, "grad_norm": 0.10906456716259076, "learning_rate": 4.865700214001762e-06, "loss": 0.4508, "step": 643 }, { "epoch": 1.5745721271393642, "grad_norm": 0.10754506423305892, "learning_rate": 4.865285955991371e-06, "loss": 0.6482, "step": 644 }, { "epoch": 1.5770171149144254, "grad_norm": 0.10263451111083424, "learning_rate": 4.864871077748857e-06, "loss": 0.5227, "step": 645 }, { "epoch": 1.5794621026894866, "grad_norm": 0.09591355620951793, "learning_rate": 4.864455579383013e-06, "loss": 0.5327, "step": 646 }, { "epoch": 1.5819070904645476, "grad_norm": 0.10028149108562139, "learning_rate": 4.8640394610027895e-06, "loss": 0.716, "step": 647 }, { "epoch": 1.5843520782396088, "grad_norm": 0.12897076554561426, "learning_rate": 4.863622722717303e-06, "loss": 0.9392, "step": 648 }, { "epoch": 1.58679706601467, "grad_norm": 0.1370399105921739, "learning_rate": 4.863205364635833e-06, "loss": 0.4962, "step": 649 }, { "epoch": 1.589242053789731, "grad_norm": 0.13232056999970115, "learning_rate": 4.8627873868678175e-06, "loss": 0.4834, "step": 650 }, { "epoch": 1.5916870415647921, "grad_norm": 0.09440559206872451, "learning_rate": 4.862368789522864e-06, "loss": 0.652, "step": 651 }, { "epoch": 1.5941320293398533, "grad_norm": 0.11018987161159294, "learning_rate": 4.861949572710734e-06, "loss": 0.8007, "step": 652 }, { "epoch": 1.5965770171149143, "grad_norm": 0.08912551249463314, "learning_rate": 4.8615297365413585e-06, "loss": 0.4337, "step": 653 }, { "epoch": 1.5990220048899757, "grad_norm": 0.13250698685524578, "learning_rate": 4.861109281124828e-06, "loss": 0.4639, "step": 654 }, { "epoch": 1.6014669926650367, "grad_norm": 0.11707705855611844, "learning_rate": 4.8606882065713946e-06, "loss": 0.9195, "step": 655 }, { "epoch": 1.6039119804400976, "grad_norm": 0.19145744294286274, "learning_rate": 4.860266512991474e-06, "loss": 0.4865, "step": 656 }, { "epoch": 1.606356968215159, "grad_norm": 0.17011802967734546, "learning_rate": 4.859844200495644e-06, "loss": 0.5022, "step": 657 }, { "epoch": 1.60880195599022, "grad_norm": 0.0906458080351694, "learning_rate": 4.859421269194644e-06, "loss": 0.4341, "step": 658 }, { "epoch": 1.6112469437652812, "grad_norm": 0.099955504189599, "learning_rate": 4.8589977191993774e-06, "loss": 0.5138, "step": 659 }, { "epoch": 1.6136919315403424, "grad_norm": 0.12079758445894387, "learning_rate": 4.8585735506209085e-06, "loss": 0.5769, "step": 660 }, { "epoch": 1.6161369193154034, "grad_norm": 0.11353390882262827, "learning_rate": 4.858148763570463e-06, "loss": 0.8745, "step": 661 }, { "epoch": 1.6185819070904646, "grad_norm": 0.1038161883749101, "learning_rate": 4.8577233581594306e-06, "loss": 0.7222, "step": 662 }, { "epoch": 1.6210268948655258, "grad_norm": 0.14748619844276883, "learning_rate": 4.857297334499362e-06, "loss": 0.595, "step": 663 }, { "epoch": 1.6234718826405867, "grad_norm": 0.14399985929179457, "learning_rate": 4.856870692701972e-06, "loss": 0.8044, "step": 664 }, { "epoch": 1.625916870415648, "grad_norm": 0.09579065432453852, "learning_rate": 4.856443432879134e-06, "loss": 0.5055, "step": 665 }, { "epoch": 1.628361858190709, "grad_norm": 0.10542201469767118, "learning_rate": 4.856015555142884e-06, "loss": 0.6142, "step": 666 }, { "epoch": 1.63080684596577, "grad_norm": 0.1590815488156907, "learning_rate": 4.855587059605424e-06, "loss": 0.6915, "step": 667 }, { "epoch": 1.6332518337408313, "grad_norm": 0.1145729050653525, "learning_rate": 4.855157946379114e-06, "loss": 0.4971, "step": 668 }, { "epoch": 1.6356968215158925, "grad_norm": 0.12269124317193993, "learning_rate": 4.854728215576479e-06, "loss": 1.0122, "step": 669 }, { "epoch": 1.6381418092909534, "grad_norm": 0.11115980535243569, "learning_rate": 4.8542978673102015e-06, "loss": 0.9206, "step": 670 }, { "epoch": 1.6405867970660146, "grad_norm": 0.13340921198939854, "learning_rate": 4.853866901693131e-06, "loss": 1.0498, "step": 671 }, { "epoch": 1.6430317848410758, "grad_norm": 0.10597716474785111, "learning_rate": 4.853435318838274e-06, "loss": 0.5938, "step": 672 }, { "epoch": 1.6454767726161368, "grad_norm": 0.0963222664089259, "learning_rate": 4.853003118858805e-06, "loss": 0.5491, "step": 673 }, { "epoch": 1.6479217603911982, "grad_norm": 0.10252375914532312, "learning_rate": 4.852570301868053e-06, "loss": 0.4956, "step": 674 }, { "epoch": 1.6503667481662592, "grad_norm": 0.06516883895046549, "learning_rate": 4.852136867979515e-06, "loss": 0.3496, "step": 675 }, { "epoch": 1.6528117359413201, "grad_norm": 0.11019712518768765, "learning_rate": 4.851702817306848e-06, "loss": 0.5599, "step": 676 }, { "epoch": 1.6552567237163816, "grad_norm": 0.08023275962675919, "learning_rate": 4.851268149963867e-06, "loss": 0.456, "step": 677 }, { "epoch": 1.6577017114914425, "grad_norm": 0.12889745988554777, "learning_rate": 4.850832866064554e-06, "loss": 0.8283, "step": 678 }, { "epoch": 1.6601466992665037, "grad_norm": 0.10399112768826765, "learning_rate": 4.85039696572305e-06, "loss": 0.5162, "step": 679 }, { "epoch": 1.662591687041565, "grad_norm": 0.13260070586917, "learning_rate": 4.849960449053656e-06, "loss": 0.6224, "step": 680 }, { "epoch": 1.6650366748166259, "grad_norm": 0.13257132298385343, "learning_rate": 4.849523316170841e-06, "loss": 0.6098, "step": 681 }, { "epoch": 1.667481662591687, "grad_norm": 0.10854327220032024, "learning_rate": 4.849085567189227e-06, "loss": 0.6587, "step": 682 }, { "epoch": 1.6699266503667483, "grad_norm": 0.1336869953402008, "learning_rate": 4.848647202223604e-06, "loss": 0.531, "step": 683 }, { "epoch": 1.6723716381418092, "grad_norm": 0.12036566125205295, "learning_rate": 4.848208221388922e-06, "loss": 0.7005, "step": 684 }, { "epoch": 1.6748166259168704, "grad_norm": 0.17965483234072654, "learning_rate": 4.847768624800289e-06, "loss": 0.5561, "step": 685 }, { "epoch": 1.6772616136919316, "grad_norm": 0.15513040979227, "learning_rate": 4.847328412572979e-06, "loss": 0.7079, "step": 686 }, { "epoch": 1.6797066014669926, "grad_norm": 0.12647590000250364, "learning_rate": 4.846887584822426e-06, "loss": 0.596, "step": 687 }, { "epoch": 1.6821515892420538, "grad_norm": 0.12898244125643704, "learning_rate": 4.846446141664225e-06, "loss": 0.4973, "step": 688 }, { "epoch": 1.684596577017115, "grad_norm": 0.14708750883534086, "learning_rate": 4.846004083214132e-06, "loss": 0.5912, "step": 689 }, { "epoch": 1.687041564792176, "grad_norm": 0.08923262598268303, "learning_rate": 4.845561409588065e-06, "loss": 0.3206, "step": 690 }, { "epoch": 1.6894865525672371, "grad_norm": 0.10013333489447876, "learning_rate": 4.8451181209021046e-06, "loss": 0.656, "step": 691 }, { "epoch": 1.6919315403422983, "grad_norm": 0.08596983742011301, "learning_rate": 4.844674217272488e-06, "loss": 0.4029, "step": 692 }, { "epoch": 1.6943765281173593, "grad_norm": 0.15451618563853647, "learning_rate": 4.844229698815621e-06, "loss": 1.3626, "step": 693 }, { "epoch": 1.6968215158924207, "grad_norm": 0.13761627324353434, "learning_rate": 4.843784565648063e-06, "loss": 0.855, "step": 694 }, { "epoch": 1.6992665036674817, "grad_norm": 0.12673090169921572, "learning_rate": 4.843338817886539e-06, "loss": 0.8405, "step": 695 }, { "epoch": 1.7017114914425426, "grad_norm": 0.11822363087418883, "learning_rate": 4.842892455647935e-06, "loss": 0.5759, "step": 696 }, { "epoch": 1.704156479217604, "grad_norm": 0.11608723785933603, "learning_rate": 4.842445479049298e-06, "loss": 0.4765, "step": 697 }, { "epoch": 1.706601466992665, "grad_norm": 0.12431526901571709, "learning_rate": 4.841997888207834e-06, "loss": 0.5168, "step": 698 }, { "epoch": 1.7090464547677262, "grad_norm": 0.1511962223294636, "learning_rate": 4.841549683240912e-06, "loss": 0.5864, "step": 699 }, { "epoch": 1.7114914425427874, "grad_norm": 0.09940499078026584, "learning_rate": 4.841100864266063e-06, "loss": 0.5639, "step": 700 }, { "epoch": 1.7139364303178484, "grad_norm": 0.08216183752276986, "learning_rate": 4.840651431400975e-06, "loss": 0.4249, "step": 701 }, { "epoch": 1.7163814180929096, "grad_norm": 0.16990803333881208, "learning_rate": 4.840201384763501e-06, "loss": 0.5247, "step": 702 }, { "epoch": 1.7188264058679708, "grad_norm": 0.09956485159354872, "learning_rate": 4.839750724471654e-06, "loss": 0.76, "step": 703 }, { "epoch": 1.7212713936430317, "grad_norm": 0.09117376701958932, "learning_rate": 4.839299450643608e-06, "loss": 0.4501, "step": 704 }, { "epoch": 1.723716381418093, "grad_norm": 0.1495785025028801, "learning_rate": 4.838847563397694e-06, "loss": 0.5441, "step": 705 }, { "epoch": 1.726161369193154, "grad_norm": 0.1363291441701734, "learning_rate": 4.838395062852411e-06, "loss": 0.5831, "step": 706 }, { "epoch": 1.728606356968215, "grad_norm": 0.1342577920804179, "learning_rate": 4.837941949126413e-06, "loss": 0.5628, "step": 707 }, { "epoch": 1.7310513447432763, "grad_norm": 0.12277509365567624, "learning_rate": 4.837488222338517e-06, "loss": 0.5569, "step": 708 }, { "epoch": 1.7334963325183375, "grad_norm": 0.17138058565246797, "learning_rate": 4.8370338826077016e-06, "loss": 0.6677, "step": 709 }, { "epoch": 1.7359413202933984, "grad_norm": 0.1621979769406644, "learning_rate": 4.836578930053104e-06, "loss": 0.5889, "step": 710 }, { "epoch": 1.7383863080684596, "grad_norm": 0.20299917175052679, "learning_rate": 4.836123364794023e-06, "loss": 0.5004, "step": 711 }, { "epoch": 1.7408312958435208, "grad_norm": 0.18306590170774154, "learning_rate": 4.835667186949918e-06, "loss": 0.4694, "step": 712 }, { "epoch": 1.7432762836185818, "grad_norm": 0.11741062071005252, "learning_rate": 4.835210396640412e-06, "loss": 0.9465, "step": 713 }, { "epoch": 1.7457212713936432, "grad_norm": 0.11523234138292135, "learning_rate": 4.834752993985282e-06, "loss": 0.8555, "step": 714 }, { "epoch": 1.7481662591687042, "grad_norm": 0.10149021095616566, "learning_rate": 4.834294979104472e-06, "loss": 0.4816, "step": 715 }, { "epoch": 1.7506112469437651, "grad_norm": 0.17718400299002438, "learning_rate": 4.833836352118084e-06, "loss": 0.7864, "step": 716 }, { "epoch": 1.7530562347188265, "grad_norm": 0.10049599079078834, "learning_rate": 4.833377113146379e-06, "loss": 0.4811, "step": 717 }, { "epoch": 1.7555012224938875, "grad_norm": 0.1542851289951912, "learning_rate": 4.832917262309781e-06, "loss": 1.1234, "step": 718 }, { "epoch": 1.7579462102689487, "grad_norm": 0.12182326363857869, "learning_rate": 4.832456799728873e-06, "loss": 0.5557, "step": 719 }, { "epoch": 1.76039119804401, "grad_norm": 0.10784170758062846, "learning_rate": 4.831995725524399e-06, "loss": 0.5179, "step": 720 }, { "epoch": 1.7628361858190709, "grad_norm": 0.13804408323952488, "learning_rate": 4.831534039817263e-06, "loss": 0.4202, "step": 721 }, { "epoch": 1.765281173594132, "grad_norm": 0.11764823403614114, "learning_rate": 4.83107174272853e-06, "loss": 0.4983, "step": 722 }, { "epoch": 1.7677261613691932, "grad_norm": 0.14996593574917372, "learning_rate": 4.830608834379425e-06, "loss": 0.6088, "step": 723 }, { "epoch": 1.7701711491442542, "grad_norm": 0.1450566340673448, "learning_rate": 4.8301453148913315e-06, "loss": 1.0136, "step": 724 }, { "epoch": 1.7726161369193154, "grad_norm": 0.10361854705671132, "learning_rate": 4.829681184385797e-06, "loss": 0.6096, "step": 725 }, { "epoch": 1.7750611246943766, "grad_norm": 0.09918891795291579, "learning_rate": 4.829216442984527e-06, "loss": 0.7074, "step": 726 }, { "epoch": 1.7775061124694376, "grad_norm": 0.0973180623640307, "learning_rate": 4.828751090809386e-06, "loss": 0.5953, "step": 727 }, { "epoch": 1.7799511002444988, "grad_norm": 0.12789101585716287, "learning_rate": 4.828285127982401e-06, "loss": 0.7094, "step": 728 }, { "epoch": 1.78239608801956, "grad_norm": 0.07630922705512852, "learning_rate": 4.827818554625757e-06, "loss": 0.5477, "step": 729 }, { "epoch": 1.784841075794621, "grad_norm": 0.156995071857263, "learning_rate": 4.827351370861802e-06, "loss": 0.7341, "step": 730 }, { "epoch": 1.7872860635696821, "grad_norm": 0.1444320392801839, "learning_rate": 4.8268835768130416e-06, "loss": 0.4963, "step": 731 }, { "epoch": 1.7897310513447433, "grad_norm": 0.09681528053510027, "learning_rate": 4.826415172602141e-06, "loss": 0.5473, "step": 732 }, { "epoch": 1.7921760391198043, "grad_norm": 0.12584394861018136, "learning_rate": 4.8259461583519275e-06, "loss": 0.8137, "step": 733 }, { "epoch": 1.7946210268948657, "grad_norm": 0.1316575626084965, "learning_rate": 4.825476534185387e-06, "loss": 0.7689, "step": 734 }, { "epoch": 1.7970660146699267, "grad_norm": 0.13746165912076075, "learning_rate": 4.8250063002256665e-06, "loss": 1.1722, "step": 735 }, { "epoch": 1.7995110024449876, "grad_norm": 0.12278482239621036, "learning_rate": 4.8245354565960715e-06, "loss": 0.5363, "step": 736 }, { "epoch": 1.801955990220049, "grad_norm": 0.10831116479829957, "learning_rate": 4.824064003420067e-06, "loss": 0.4675, "step": 737 }, { "epoch": 1.80440097799511, "grad_norm": 0.13601052407661196, "learning_rate": 4.8235919408212805e-06, "loss": 0.6535, "step": 738 }, { "epoch": 1.8068459657701712, "grad_norm": 0.1410565398203603, "learning_rate": 4.823119268923496e-06, "loss": 0.4656, "step": 739 }, { "epoch": 1.8092909535452324, "grad_norm": 0.14323506634884267, "learning_rate": 4.822645987850662e-06, "loss": 0.4327, "step": 740 }, { "epoch": 1.8117359413202934, "grad_norm": 0.1551502268756037, "learning_rate": 4.82217209772688e-06, "loss": 0.5046, "step": 741 }, { "epoch": 1.8141809290953546, "grad_norm": 0.16024173573383538, "learning_rate": 4.821697598676416e-06, "loss": 0.5508, "step": 742 }, { "epoch": 1.8166259168704157, "grad_norm": 0.12314984515616777, "learning_rate": 4.821222490823695e-06, "loss": 0.5043, "step": 743 }, { "epoch": 1.8190709046454767, "grad_norm": 0.1307407136558782, "learning_rate": 4.820746774293301e-06, "loss": 0.9258, "step": 744 }, { "epoch": 1.821515892420538, "grad_norm": 0.1222848554854843, "learning_rate": 4.820270449209979e-06, "loss": 0.6961, "step": 745 }, { "epoch": 1.823960880195599, "grad_norm": 0.15014357077629403, "learning_rate": 4.819793515698629e-06, "loss": 0.6656, "step": 746 }, { "epoch": 1.82640586797066, "grad_norm": 0.10984734986693803, "learning_rate": 4.819315973884317e-06, "loss": 0.5384, "step": 747 }, { "epoch": 1.8288508557457213, "grad_norm": 0.12184988588281984, "learning_rate": 4.818837823892264e-06, "loss": 0.8951, "step": 748 }, { "epoch": 1.8312958435207825, "grad_norm": 0.11079195537156254, "learning_rate": 4.8183590658478516e-06, "loss": 0.7624, "step": 749 }, { "epoch": 1.8337408312958434, "grad_norm": 0.11008485332204336, "learning_rate": 4.817879699876623e-06, "loss": 0.5527, "step": 750 }, { "epoch": 1.8361858190709046, "grad_norm": 0.1415032346155075, "learning_rate": 4.817399726104277e-06, "loss": 0.577, "step": 751 }, { "epoch": 1.8386308068459658, "grad_norm": 0.1158999823098757, "learning_rate": 4.8169191446566746e-06, "loss": 0.6826, "step": 752 }, { "epoch": 1.8410757946210268, "grad_norm": 0.09991612935598868, "learning_rate": 4.816437955659835e-06, "loss": 0.3888, "step": 753 }, { "epoch": 1.843520782396088, "grad_norm": 0.14642084595183114, "learning_rate": 4.815956159239937e-06, "loss": 1.0093, "step": 754 }, { "epoch": 1.8459657701711492, "grad_norm": 0.12277287073197717, "learning_rate": 4.815473755523319e-06, "loss": 0.4162, "step": 755 }, { "epoch": 1.8484107579462101, "grad_norm": 0.15622600367136993, "learning_rate": 4.8149907446364774e-06, "loss": 0.3899, "step": 756 }, { "epoch": 1.8508557457212715, "grad_norm": 0.15908980682987184, "learning_rate": 4.81450712670607e-06, "loss": 0.8689, "step": 757 }, { "epoch": 1.8533007334963325, "grad_norm": 0.1910220045335047, "learning_rate": 4.81402290185891e-06, "loss": 1.0322, "step": 758 }, { "epoch": 1.8557457212713935, "grad_norm": 0.13781731232482605, "learning_rate": 4.813538070221976e-06, "loss": 0.6689, "step": 759 }, { "epoch": 1.858190709046455, "grad_norm": 0.12187660472100656, "learning_rate": 4.813052631922399e-06, "loss": 0.5701, "step": 760 }, { "epoch": 1.8606356968215159, "grad_norm": 0.10704190521669231, "learning_rate": 4.812566587087473e-06, "loss": 0.541, "step": 761 }, { "epoch": 1.863080684596577, "grad_norm": 0.12089889772706895, "learning_rate": 4.81207993584465e-06, "loss": 0.7169, "step": 762 }, { "epoch": 1.8655256723716382, "grad_norm": 0.17754599312710506, "learning_rate": 4.811592678321541e-06, "loss": 0.5148, "step": 763 }, { "epoch": 1.8679706601466992, "grad_norm": 0.10337394968121207, "learning_rate": 4.811104814645916e-06, "loss": 0.6708, "step": 764 }, { "epoch": 1.8704156479217604, "grad_norm": 0.15990897703899912, "learning_rate": 4.810616344945706e-06, "loss": 0.9321, "step": 765 }, { "epoch": 1.8728606356968216, "grad_norm": 0.12330790293414083, "learning_rate": 4.810127269348995e-06, "loss": 0.6211, "step": 766 }, { "epoch": 1.8753056234718826, "grad_norm": 0.09210771404422552, "learning_rate": 4.809637587984033e-06, "loss": 0.6974, "step": 767 }, { "epoch": 1.8777506112469438, "grad_norm": 0.168545999335004, "learning_rate": 4.809147300979225e-06, "loss": 1.1879, "step": 768 }, { "epoch": 1.880195599022005, "grad_norm": 0.11482535276554853, "learning_rate": 4.808656408463134e-06, "loss": 0.7129, "step": 769 }, { "epoch": 1.882640586797066, "grad_norm": 0.19798126409860214, "learning_rate": 4.808164910564486e-06, "loss": 0.8188, "step": 770 }, { "epoch": 1.885085574572127, "grad_norm": 0.15737377378099945, "learning_rate": 4.80767280741216e-06, "loss": 0.729, "step": 771 }, { "epoch": 1.8875305623471883, "grad_norm": 0.10255358358507011, "learning_rate": 4.807180099135199e-06, "loss": 0.5795, "step": 772 }, { "epoch": 1.8899755501222493, "grad_norm": 0.12399009514435624, "learning_rate": 4.806686785862801e-06, "loss": 0.5496, "step": 773 }, { "epoch": 1.8924205378973105, "grad_norm": 0.12451001518390688, "learning_rate": 4.8061928677243244e-06, "loss": 0.8979, "step": 774 }, { "epoch": 1.8948655256723717, "grad_norm": 0.16118155814169954, "learning_rate": 4.805698344849286e-06, "loss": 0.8007, "step": 775 }, { "epoch": 1.8973105134474326, "grad_norm": 0.11970416495016106, "learning_rate": 4.805203217367362e-06, "loss": 0.6259, "step": 776 }, { "epoch": 1.899755501222494, "grad_norm": 0.15685960511820304, "learning_rate": 4.804707485408384e-06, "loss": 0.8075, "step": 777 }, { "epoch": 1.902200488997555, "grad_norm": 0.1578579719899877, "learning_rate": 4.8042111491023455e-06, "loss": 0.4877, "step": 778 }, { "epoch": 1.904645476772616, "grad_norm": 0.11629835768054726, "learning_rate": 4.803714208579398e-06, "loss": 0.5222, "step": 779 }, { "epoch": 1.9070904645476774, "grad_norm": 0.11701472311495625, "learning_rate": 4.803216663969849e-06, "loss": 0.8609, "step": 780 }, { "epoch": 1.9095354523227384, "grad_norm": 0.10103024460163296, "learning_rate": 4.802718515404168e-06, "loss": 0.5263, "step": 781 }, { "epoch": 1.9119804400977995, "grad_norm": 0.1518993777182326, "learning_rate": 4.802219763012978e-06, "loss": 0.6717, "step": 782 }, { "epoch": 1.9144254278728607, "grad_norm": 0.13697745718100693, "learning_rate": 4.801720406927067e-06, "loss": 0.5797, "step": 783 }, { "epoch": 1.9168704156479217, "grad_norm": 0.1336832074610851, "learning_rate": 4.8012204472773745e-06, "loss": 0.607, "step": 784 }, { "epoch": 1.919315403422983, "grad_norm": 0.1286659830336251, "learning_rate": 4.800719884195003e-06, "loss": 0.3785, "step": 785 }, { "epoch": 1.921760391198044, "grad_norm": 0.13505916681429353, "learning_rate": 4.800218717811212e-06, "loss": 0.4645, "step": 786 }, { "epoch": 1.924205378973105, "grad_norm": 0.11679295475938653, "learning_rate": 4.799716948257418e-06, "loss": 0.7331, "step": 787 }, { "epoch": 1.9266503667481663, "grad_norm": 0.10508732261419586, "learning_rate": 4.799214575665196e-06, "loss": 0.5308, "step": 788 }, { "epoch": 1.9290953545232274, "grad_norm": 0.12161540145949726, "learning_rate": 4.79871160016628e-06, "loss": 0.8243, "step": 789 }, { "epoch": 1.9315403422982884, "grad_norm": 0.13688323244743186, "learning_rate": 4.798208021892562e-06, "loss": 0.6833, "step": 790 }, { "epoch": 1.9339853300733496, "grad_norm": 0.10398439966137216, "learning_rate": 4.797703840976091e-06, "loss": 0.5062, "step": 791 }, { "epoch": 1.9364303178484108, "grad_norm": 0.1408936864234124, "learning_rate": 4.797199057549076e-06, "loss": 0.5466, "step": 792 }, { "epoch": 1.9388753056234718, "grad_norm": 0.11584925478575149, "learning_rate": 4.796693671743882e-06, "loss": 0.4612, "step": 793 }, { "epoch": 1.941320293398533, "grad_norm": 0.08981749019854286, "learning_rate": 4.7961876836930324e-06, "loss": 0.4266, "step": 794 }, { "epoch": 1.9437652811735942, "grad_norm": 0.18994463221787466, "learning_rate": 4.79568109352921e-06, "loss": 0.9827, "step": 795 }, { "epoch": 1.9462102689486551, "grad_norm": 0.16703185893755695, "learning_rate": 4.795173901385253e-06, "loss": 0.4553, "step": 796 }, { "epoch": 1.9486552567237165, "grad_norm": 0.10413546765513156, "learning_rate": 4.794666107394159e-06, "loss": 0.6008, "step": 797 }, { "epoch": 1.9511002444987775, "grad_norm": 0.14401763284737917, "learning_rate": 4.794157711689084e-06, "loss": 0.6441, "step": 798 }, { "epoch": 1.9535452322738385, "grad_norm": 0.14523066687569391, "learning_rate": 4.7936487144033415e-06, "loss": 0.5983, "step": 799 }, { "epoch": 1.9559902200488999, "grad_norm": 0.11476848790727802, "learning_rate": 4.7931391156704e-06, "loss": 0.849, "step": 800 }, { "epoch": 1.9584352078239609, "grad_norm": 0.13416802946301673, "learning_rate": 4.792628915623889e-06, "loss": 0.8281, "step": 801 }, { "epoch": 1.960880195599022, "grad_norm": 0.14401738017231736, "learning_rate": 4.792118114397597e-06, "loss": 0.6789, "step": 802 }, { "epoch": 1.9633251833740832, "grad_norm": 0.11154655375586213, "learning_rate": 4.791606712125464e-06, "loss": 0.5504, "step": 803 }, { "epoch": 1.9657701711491442, "grad_norm": 0.17603785657545673, "learning_rate": 4.791094708941594e-06, "loss": 0.5241, "step": 804 }, { "epoch": 1.9682151589242054, "grad_norm": 0.10288638917315537, "learning_rate": 4.790582104980244e-06, "loss": 0.4565, "step": 805 }, { "epoch": 1.9706601466992666, "grad_norm": 0.11130311416716382, "learning_rate": 4.790068900375831e-06, "loss": 0.8371, "step": 806 }, { "epoch": 1.9731051344743276, "grad_norm": 0.12434686315986627, "learning_rate": 4.789555095262932e-06, "loss": 0.5618, "step": 807 }, { "epoch": 1.9755501222493888, "grad_norm": 0.1265132607261373, "learning_rate": 4.789040689776274e-06, "loss": 0.6798, "step": 808 }, { "epoch": 1.97799511002445, "grad_norm": 0.1792243095832928, "learning_rate": 4.788525684050749e-06, "loss": 0.6625, "step": 809 }, { "epoch": 1.980440097799511, "grad_norm": 0.11837845869571981, "learning_rate": 4.788010078221401e-06, "loss": 0.5964, "step": 810 }, { "epoch": 1.982885085574572, "grad_norm": 0.14304280258710575, "learning_rate": 4.787493872423435e-06, "loss": 0.8614, "step": 811 }, { "epoch": 1.9853300733496333, "grad_norm": 0.11049639213632467, "learning_rate": 4.786977066792212e-06, "loss": 0.4396, "step": 812 }, { "epoch": 1.9877750611246943, "grad_norm": 0.15406352563106712, "learning_rate": 4.786459661463251e-06, "loss": 0.9546, "step": 813 }, { "epoch": 1.9902200488997555, "grad_norm": 0.15202454763031117, "learning_rate": 4.785941656572225e-06, "loss": 1.1633, "step": 814 }, { "epoch": 1.9926650366748166, "grad_norm": 0.15407890432406673, "learning_rate": 4.78542305225497e-06, "loss": 0.6896, "step": 815 }, { "epoch": 1.9951100244498776, "grad_norm": 0.13348299433406074, "learning_rate": 4.784903848647473e-06, "loss": 0.6009, "step": 816 }, { "epoch": 1.997555012224939, "grad_norm": 0.10584937182336218, "learning_rate": 4.784384045885883e-06, "loss": 0.4955, "step": 817 }, { "epoch": 2.0, "grad_norm": 0.147434937419781, "learning_rate": 4.783863644106502e-06, "loss": 0.879, "step": 818 }, { "epoch": 2.002444987775061, "grad_norm": 0.14706158144953857, "learning_rate": 4.7833426434457945e-06, "loss": 0.5248, "step": 819 }, { "epoch": 2.0048899755501224, "grad_norm": 0.15922707965360527, "learning_rate": 4.782821044040376e-06, "loss": 0.4911, "step": 820 }, { "epoch": 2.0073349633251834, "grad_norm": 0.11953551149768692, "learning_rate": 4.782298846027024e-06, "loss": 0.8306, "step": 821 }, { "epoch": 2.0097799511002443, "grad_norm": 0.10233519689936803, "learning_rate": 4.781776049542669e-06, "loss": 0.5911, "step": 822 }, { "epoch": 2.0122249388753057, "grad_norm": 0.20332745990141157, "learning_rate": 4.781252654724401e-06, "loss": 0.5306, "step": 823 }, { "epoch": 2.0146699266503667, "grad_norm": 0.07298230084059151, "learning_rate": 4.780728661709465e-06, "loss": 0.3727, "step": 824 }, { "epoch": 2.0171149144254277, "grad_norm": 0.19146845969916104, "learning_rate": 4.7802040706352655e-06, "loss": 0.6415, "step": 825 }, { "epoch": 2.019559902200489, "grad_norm": 0.17102622997073183, "learning_rate": 4.779678881639363e-06, "loss": 0.7184, "step": 826 }, { "epoch": 2.02200488997555, "grad_norm": 0.13068934102332075, "learning_rate": 4.779153094859471e-06, "loss": 0.5366, "step": 827 }, { "epoch": 2.0244498777506115, "grad_norm": 0.16215713018093997, "learning_rate": 4.778626710433465e-06, "loss": 0.8962, "step": 828 }, { "epoch": 2.0268948655256724, "grad_norm": 0.13228487122095742, "learning_rate": 4.778099728499375e-06, "loss": 0.5542, "step": 829 }, { "epoch": 2.0293398533007334, "grad_norm": 0.12567736735023852, "learning_rate": 4.777572149195387e-06, "loss": 0.8489, "step": 830 }, { "epoch": 2.031784841075795, "grad_norm": 0.1278871828157041, "learning_rate": 4.777043972659845e-06, "loss": 0.7214, "step": 831 }, { "epoch": 2.034229828850856, "grad_norm": 0.12969716220662594, "learning_rate": 4.776515199031248e-06, "loss": 0.8372, "step": 832 }, { "epoch": 2.0366748166259168, "grad_norm": 0.12442518835234063, "learning_rate": 4.7759858284482544e-06, "loss": 0.7463, "step": 833 }, { "epoch": 2.039119804400978, "grad_norm": 0.12009626731653386, "learning_rate": 4.775455861049676e-06, "loss": 0.5872, "step": 834 }, { "epoch": 2.041564792176039, "grad_norm": 0.1029923305504545, "learning_rate": 4.774925296974482e-06, "loss": 0.4112, "step": 835 }, { "epoch": 2.0440097799511, "grad_norm": 0.19129615884375242, "learning_rate": 4.774394136361799e-06, "loss": 0.6626, "step": 836 }, { "epoch": 2.0464547677261615, "grad_norm": 0.1230039057494566, "learning_rate": 4.77386237935091e-06, "loss": 0.8634, "step": 837 }, { "epoch": 2.0488997555012225, "grad_norm": 0.12053378320852179, "learning_rate": 4.7733300260812525e-06, "loss": 1.0506, "step": 838 }, { "epoch": 2.0513447432762835, "grad_norm": 0.17000545122958616, "learning_rate": 4.7727970766924234e-06, "loss": 1.1707, "step": 839 }, { "epoch": 2.053789731051345, "grad_norm": 0.13108008962080583, "learning_rate": 4.772263531324173e-06, "loss": 0.5354, "step": 840 }, { "epoch": 2.056234718826406, "grad_norm": 0.14682324754138906, "learning_rate": 4.771729390116409e-06, "loss": 0.8349, "step": 841 }, { "epoch": 2.058679706601467, "grad_norm": 0.10463182967955202, "learning_rate": 4.771194653209197e-06, "loss": 0.4461, "step": 842 }, { "epoch": 2.0611246943765282, "grad_norm": 0.14846004991553632, "learning_rate": 4.770659320742755e-06, "loss": 0.548, "step": 843 }, { "epoch": 2.063569682151589, "grad_norm": 0.17413156200466165, "learning_rate": 4.770123392857461e-06, "loss": 0.5997, "step": 844 }, { "epoch": 2.06601466992665, "grad_norm": 0.10301084940909348, "learning_rate": 4.769586869693848e-06, "loss": 0.5678, "step": 845 }, { "epoch": 2.0684596577017116, "grad_norm": 0.15027344077997845, "learning_rate": 4.7690497513926025e-06, "loss": 0.4973, "step": 846 }, { "epoch": 2.0709046454767726, "grad_norm": 0.14456329482622987, "learning_rate": 4.768512038094571e-06, "loss": 0.8347, "step": 847 }, { "epoch": 2.0733496332518335, "grad_norm": 0.09476506032195829, "learning_rate": 4.767973729940754e-06, "loss": 0.5279, "step": 848 }, { "epoch": 2.075794621026895, "grad_norm": 0.1281703164245905, "learning_rate": 4.767434827072308e-06, "loss": 0.6836, "step": 849 }, { "epoch": 2.078239608801956, "grad_norm": 0.11499591039702096, "learning_rate": 4.766895329630545e-06, "loss": 0.5449, "step": 850 }, { "epoch": 2.0806845965770173, "grad_norm": 0.12725902925028126, "learning_rate": 4.766355237756936e-06, "loss": 0.5523, "step": 851 }, { "epoch": 2.0831295843520783, "grad_norm": 0.11967575642898146, "learning_rate": 4.765814551593102e-06, "loss": 0.5467, "step": 852 }, { "epoch": 2.0855745721271393, "grad_norm": 0.13449356042079105, "learning_rate": 4.765273271280825e-06, "loss": 0.7627, "step": 853 }, { "epoch": 2.0880195599022007, "grad_norm": 0.1545773457575999, "learning_rate": 4.7647313969620425e-06, "loss": 1.1005, "step": 854 }, { "epoch": 2.0904645476772616, "grad_norm": 0.13926844739111374, "learning_rate": 4.764188928778843e-06, "loss": 0.6687, "step": 855 }, { "epoch": 2.0929095354523226, "grad_norm": 0.13175102178710443, "learning_rate": 4.763645866873478e-06, "loss": 0.7133, "step": 856 }, { "epoch": 2.095354523227384, "grad_norm": 0.14983695092211466, "learning_rate": 4.7631022113883475e-06, "loss": 0.376, "step": 857 }, { "epoch": 2.097799511002445, "grad_norm": 0.15965998543574816, "learning_rate": 4.762557962466011e-06, "loss": 0.7313, "step": 858 }, { "epoch": 2.100244498777506, "grad_norm": 0.11394697192652084, "learning_rate": 4.762013120249185e-06, "loss": 0.7536, "step": 859 }, { "epoch": 2.1026894865525674, "grad_norm": 0.13115564134488228, "learning_rate": 4.761467684880738e-06, "loss": 0.4737, "step": 860 }, { "epoch": 2.1051344743276283, "grad_norm": 0.13628436164284463, "learning_rate": 4.760921656503696e-06, "loss": 0.8629, "step": 861 }, { "epoch": 2.1075794621026893, "grad_norm": 0.16944637192952391, "learning_rate": 4.76037503526124e-06, "loss": 0.4788, "step": 862 }, { "epoch": 2.1100244498777507, "grad_norm": 0.17113161308688937, "learning_rate": 4.759827821296706e-06, "loss": 0.6829, "step": 863 }, { "epoch": 2.1124694376528117, "grad_norm": 0.12499755808514822, "learning_rate": 4.7592800147535875e-06, "loss": 0.7053, "step": 864 }, { "epoch": 2.1149144254278727, "grad_norm": 0.1958786675928173, "learning_rate": 4.7587316157755305e-06, "loss": 1.2257, "step": 865 }, { "epoch": 2.117359413202934, "grad_norm": 0.17356508282630306, "learning_rate": 4.758182624506339e-06, "loss": 0.459, "step": 866 }, { "epoch": 2.119804400977995, "grad_norm": 0.09330789803328009, "learning_rate": 4.757633041089971e-06, "loss": 0.4654, "step": 867 }, { "epoch": 2.122249388753056, "grad_norm": 0.16023903984593896, "learning_rate": 4.757082865670538e-06, "loss": 0.5512, "step": 868 }, { "epoch": 2.1246943765281174, "grad_norm": 0.18351489392869802, "learning_rate": 4.75653209839231e-06, "loss": 0.5189, "step": 869 }, { "epoch": 2.1271393643031784, "grad_norm": 0.15640933622450706, "learning_rate": 4.755980739399712e-06, "loss": 0.3466, "step": 870 }, { "epoch": 2.12958435207824, "grad_norm": 0.1444257407646091, "learning_rate": 4.755428788837321e-06, "loss": 0.7519, "step": 871 }, { "epoch": 2.132029339853301, "grad_norm": 0.14595829245939346, "learning_rate": 4.754876246849871e-06, "loss": 0.4077, "step": 872 }, { "epoch": 2.1344743276283618, "grad_norm": 0.13451553604309138, "learning_rate": 4.754323113582253e-06, "loss": 0.8014, "step": 873 }, { "epoch": 2.136919315403423, "grad_norm": 0.1657308840208002, "learning_rate": 4.753769389179509e-06, "loss": 0.5881, "step": 874 }, { "epoch": 2.139364303178484, "grad_norm": 0.11880111012595808, "learning_rate": 4.75321507378684e-06, "loss": 0.7447, "step": 875 }, { "epoch": 2.141809290953545, "grad_norm": 0.10161063275077391, "learning_rate": 4.752660167549599e-06, "loss": 0.4295, "step": 876 }, { "epoch": 2.1442542787286065, "grad_norm": 0.14444496812933494, "learning_rate": 4.752104670613296e-06, "loss": 0.3627, "step": 877 }, { "epoch": 2.1466992665036675, "grad_norm": 0.10630406958696703, "learning_rate": 4.751548583123595e-06, "loss": 0.4577, "step": 878 }, { "epoch": 2.1491442542787285, "grad_norm": 0.09966748882034411, "learning_rate": 4.750991905226313e-06, "loss": 0.652, "step": 879 }, { "epoch": 2.15158924205379, "grad_norm": 0.1293482303798882, "learning_rate": 4.750434637067427e-06, "loss": 0.9405, "step": 880 }, { "epoch": 2.154034229828851, "grad_norm": 0.1529528807983986, "learning_rate": 4.749876778793062e-06, "loss": 0.4113, "step": 881 }, { "epoch": 2.156479217603912, "grad_norm": 0.21978250281734155, "learning_rate": 4.749318330549504e-06, "loss": 1.009, "step": 882 }, { "epoch": 2.1589242053789732, "grad_norm": 0.24391905542317935, "learning_rate": 4.748759292483189e-06, "loss": 0.775, "step": 883 }, { "epoch": 2.161369193154034, "grad_norm": 0.1337623060502161, "learning_rate": 4.748199664740711e-06, "loss": 0.4591, "step": 884 }, { "epoch": 2.163814180929095, "grad_norm": 0.1581243389135384, "learning_rate": 4.747639447468816e-06, "loss": 0.7211, "step": 885 }, { "epoch": 2.1662591687041566, "grad_norm": 0.16282780623445856, "learning_rate": 4.747078640814407e-06, "loss": 0.4574, "step": 886 }, { "epoch": 2.1687041564792175, "grad_norm": 0.11872462173956766, "learning_rate": 4.7465172449245395e-06, "loss": 0.4549, "step": 887 }, { "epoch": 2.1711491442542785, "grad_norm": 0.1499580439233291, "learning_rate": 4.745955259946424e-06, "loss": 0.9132, "step": 888 }, { "epoch": 2.17359413202934, "grad_norm": 0.12523862110866715, "learning_rate": 4.745392686027427e-06, "loss": 0.5837, "step": 889 }, { "epoch": 2.176039119804401, "grad_norm": 0.1787126704314975, "learning_rate": 4.744829523315067e-06, "loss": 0.623, "step": 890 }, { "epoch": 2.178484107579462, "grad_norm": 0.1568240826353576, "learning_rate": 4.744265771957019e-06, "loss": 0.5474, "step": 891 }, { "epoch": 2.1809290953545233, "grad_norm": 0.18830288131446615, "learning_rate": 4.743701432101111e-06, "loss": 0.8836, "step": 892 }, { "epoch": 2.1833740831295843, "grad_norm": 0.14104979392400585, "learning_rate": 4.743136503895327e-06, "loss": 0.6751, "step": 893 }, { "epoch": 2.1858190709046457, "grad_norm": 0.17488735388963428, "learning_rate": 4.742570987487802e-06, "loss": 0.5186, "step": 894 }, { "epoch": 2.1882640586797066, "grad_norm": 0.12165539735442335, "learning_rate": 4.7420048830268295e-06, "loss": 0.7064, "step": 895 }, { "epoch": 2.1907090464547676, "grad_norm": 0.16492081834271125, "learning_rate": 4.741438190660854e-06, "loss": 0.7279, "step": 896 }, { "epoch": 2.193154034229829, "grad_norm": 0.16121870729834303, "learning_rate": 4.740870910538475e-06, "loss": 0.803, "step": 897 }, { "epoch": 2.19559902200489, "grad_norm": 0.1620792840653181, "learning_rate": 4.740303042808447e-06, "loss": 0.5295, "step": 898 }, { "epoch": 2.198044009779951, "grad_norm": 0.11649367121019195, "learning_rate": 4.739734587619677e-06, "loss": 0.3964, "step": 899 }, { "epoch": 2.2004889975550124, "grad_norm": 0.13711991196472734, "learning_rate": 4.739165545121228e-06, "loss": 0.5327, "step": 900 }, { "epoch": 2.2029339853300733, "grad_norm": 0.08865038921741594, "learning_rate": 4.738595915462316e-06, "loss": 0.4779, "step": 901 }, { "epoch": 2.2053789731051343, "grad_norm": 0.12785264003277, "learning_rate": 4.7380256987923104e-06, "loss": 0.5592, "step": 902 }, { "epoch": 2.2078239608801957, "grad_norm": 0.1294170638533336, "learning_rate": 4.737454895260734e-06, "loss": 0.5868, "step": 903 }, { "epoch": 2.2102689486552567, "grad_norm": 0.11187402251584612, "learning_rate": 4.736883505017267e-06, "loss": 0.776, "step": 904 }, { "epoch": 2.2127139364303177, "grad_norm": 0.10857150942150316, "learning_rate": 4.736311528211738e-06, "loss": 0.4057, "step": 905 }, { "epoch": 2.215158924205379, "grad_norm": 0.1819830411382696, "learning_rate": 4.735738964994135e-06, "loss": 0.4603, "step": 906 }, { "epoch": 2.21760391198044, "grad_norm": 0.1769365870947913, "learning_rate": 4.735165815514596e-06, "loss": 1.2002, "step": 907 }, { "epoch": 2.220048899755501, "grad_norm": 0.10794380648866674, "learning_rate": 4.734592079923414e-06, "loss": 0.7213, "step": 908 }, { "epoch": 2.2224938875305624, "grad_norm": 0.15834685157815057, "learning_rate": 4.734017758371037e-06, "loss": 0.439, "step": 909 }, { "epoch": 2.2249388753056234, "grad_norm": 0.15262814891961574, "learning_rate": 4.733442851008063e-06, "loss": 0.6084, "step": 910 }, { "epoch": 2.227383863080685, "grad_norm": 0.15008262053687232, "learning_rate": 4.732867357985248e-06, "loss": 0.3913, "step": 911 }, { "epoch": 2.229828850855746, "grad_norm": 0.13663814374154887, "learning_rate": 4.732291279453497e-06, "loss": 0.7689, "step": 912 }, { "epoch": 2.2322738386308068, "grad_norm": 0.21862191649934534, "learning_rate": 4.7317146155638725e-06, "loss": 0.4013, "step": 913 }, { "epoch": 2.234718826405868, "grad_norm": 0.11337906385319835, "learning_rate": 4.73113736646759e-06, "loss": 0.4518, "step": 914 }, { "epoch": 2.237163814180929, "grad_norm": 0.12287967933339534, "learning_rate": 4.730559532316015e-06, "loss": 0.6515, "step": 915 }, { "epoch": 2.23960880195599, "grad_norm": 0.20033366922416965, "learning_rate": 4.72998111326067e-06, "loss": 0.6552, "step": 916 }, { "epoch": 2.2420537897310515, "grad_norm": 0.15508185938460795, "learning_rate": 4.72940210945323e-06, "loss": 0.4936, "step": 917 }, { "epoch": 2.2444987775061125, "grad_norm": 0.11590757225801478, "learning_rate": 4.728822521045523e-06, "loss": 0.599, "step": 918 }, { "epoch": 2.2469437652811735, "grad_norm": 0.11662840836827894, "learning_rate": 4.728242348189531e-06, "loss": 0.5455, "step": 919 }, { "epoch": 2.249388753056235, "grad_norm": 0.1226509694511489, "learning_rate": 4.727661591037387e-06, "loss": 0.3578, "step": 920 }, { "epoch": 2.251833740831296, "grad_norm": 0.12986041073194354, "learning_rate": 4.727080249741379e-06, "loss": 0.5854, "step": 921 }, { "epoch": 2.254278728606357, "grad_norm": 0.1450672009254647, "learning_rate": 4.726498324453949e-06, "loss": 0.4936, "step": 922 }, { "epoch": 2.256723716381418, "grad_norm": 0.1458859513236837, "learning_rate": 4.725915815327692e-06, "loss": 0.4962, "step": 923 }, { "epoch": 2.259168704156479, "grad_norm": 0.11725198765875824, "learning_rate": 4.725332722515352e-06, "loss": 0.408, "step": 924 }, { "epoch": 2.26161369193154, "grad_norm": 0.13243842444222276, "learning_rate": 4.724749046169833e-06, "loss": 0.7271, "step": 925 }, { "epoch": 2.2640586797066016, "grad_norm": 0.13751507762577303, "learning_rate": 4.724164786444185e-06, "loss": 0.6217, "step": 926 }, { "epoch": 2.2665036674816625, "grad_norm": 0.14330890183748815, "learning_rate": 4.723579943491618e-06, "loss": 0.706, "step": 927 }, { "epoch": 2.2689486552567235, "grad_norm": 0.14210027436540335, "learning_rate": 4.722994517465487e-06, "loss": 0.7109, "step": 928 }, { "epoch": 2.271393643031785, "grad_norm": 0.21532180195609618, "learning_rate": 4.722408508519306e-06, "loss": 0.6747, "step": 929 }, { "epoch": 2.273838630806846, "grad_norm": 0.16318400866093985, "learning_rate": 4.721821916806741e-06, "loss": 0.7447, "step": 930 }, { "epoch": 2.276283618581907, "grad_norm": 0.12780152224285085, "learning_rate": 4.721234742481609e-06, "loss": 0.4632, "step": 931 }, { "epoch": 2.2787286063569683, "grad_norm": 0.1792340002645962, "learning_rate": 4.72064698569788e-06, "loss": 1.0131, "step": 932 }, { "epoch": 2.2811735941320292, "grad_norm": 0.19779190322598758, "learning_rate": 4.720058646609677e-06, "loss": 0.5975, "step": 933 }, { "epoch": 2.28361858190709, "grad_norm": 0.22120562477086841, "learning_rate": 4.719469725371279e-06, "loss": 0.7179, "step": 934 }, { "epoch": 2.2860635696821516, "grad_norm": 0.12942999875436037, "learning_rate": 4.71888022213711e-06, "loss": 0.8739, "step": 935 }, { "epoch": 2.2885085574572126, "grad_norm": 0.175888536192299, "learning_rate": 4.718290137061755e-06, "loss": 0.5433, "step": 936 }, { "epoch": 2.290953545232274, "grad_norm": 0.1510107746602996, "learning_rate": 4.717699470299946e-06, "loss": 1.2176, "step": 937 }, { "epoch": 2.293398533007335, "grad_norm": 0.11915330226089885, "learning_rate": 4.717108222006569e-06, "loss": 0.5377, "step": 938 }, { "epoch": 2.295843520782396, "grad_norm": 0.13364831604716457, "learning_rate": 4.716516392336664e-06, "loss": 0.7308, "step": 939 }, { "epoch": 2.2982885085574574, "grad_norm": 0.17623058189430155, "learning_rate": 4.715923981445421e-06, "loss": 1.1305, "step": 940 }, { "epoch": 2.3007334963325183, "grad_norm": 0.15473015727939723, "learning_rate": 4.715330989488186e-06, "loss": 0.575, "step": 941 }, { "epoch": 2.3031784841075793, "grad_norm": 0.12013438697202018, "learning_rate": 4.714737416620453e-06, "loss": 0.5961, "step": 942 }, { "epoch": 2.3056234718826407, "grad_norm": 0.16772194657296288, "learning_rate": 4.714143262997871e-06, "loss": 0.823, "step": 943 }, { "epoch": 2.3080684596577017, "grad_norm": 0.15155374028723395, "learning_rate": 4.71354852877624e-06, "loss": 0.4549, "step": 944 }, { "epoch": 2.3105134474327627, "grad_norm": 0.17999504929451104, "learning_rate": 4.712953214111516e-06, "loss": 0.7118, "step": 945 }, { "epoch": 2.312958435207824, "grad_norm": 0.13059050021859978, "learning_rate": 4.712357319159799e-06, "loss": 0.5235, "step": 946 }, { "epoch": 2.315403422982885, "grad_norm": 0.09224310871861825, "learning_rate": 4.7117608440773496e-06, "loss": 0.399, "step": 947 }, { "epoch": 2.3178484107579465, "grad_norm": 0.11596373574185828, "learning_rate": 4.7111637890205775e-06, "loss": 0.5031, "step": 948 }, { "epoch": 2.3202933985330074, "grad_norm": 0.11438015839359254, "learning_rate": 4.710566154146044e-06, "loss": 0.7046, "step": 949 }, { "epoch": 2.3227383863080684, "grad_norm": 0.12551851594910046, "learning_rate": 4.7099679396104605e-06, "loss": 0.6393, "step": 950 }, { "epoch": 2.32518337408313, "grad_norm": 0.12653923848609436, "learning_rate": 4.709369145570694e-06, "loss": 0.7521, "step": 951 }, { "epoch": 2.3276283618581908, "grad_norm": 0.10619069024153853, "learning_rate": 4.708769772183763e-06, "loss": 0.7489, "step": 952 }, { "epoch": 2.3300733496332517, "grad_norm": 0.14512024517229347, "learning_rate": 4.7081698196068356e-06, "loss": 0.6433, "step": 953 }, { "epoch": 2.332518337408313, "grad_norm": 0.14017676886713798, "learning_rate": 4.707569287997233e-06, "loss": 0.5963, "step": 954 }, { "epoch": 2.334963325183374, "grad_norm": 0.18220370081871406, "learning_rate": 4.70696817751243e-06, "loss": 0.8583, "step": 955 }, { "epoch": 2.337408312958435, "grad_norm": 0.14414716774013517, "learning_rate": 4.706366488310049e-06, "loss": 0.8359, "step": 956 }, { "epoch": 2.3398533007334965, "grad_norm": 0.12027221973726152, "learning_rate": 4.705764220547868e-06, "loss": 0.5259, "step": 957 }, { "epoch": 2.3422982885085575, "grad_norm": 0.1431167847846667, "learning_rate": 4.705161374383816e-06, "loss": 0.5169, "step": 958 }, { "epoch": 2.3447432762836184, "grad_norm": 0.17822753630378646, "learning_rate": 4.704557949975973e-06, "loss": 0.6285, "step": 959 }, { "epoch": 2.34718826405868, "grad_norm": 0.11114969810413733, "learning_rate": 4.703953947482569e-06, "loss": 0.4389, "step": 960 }, { "epoch": 2.349633251833741, "grad_norm": 0.2697790285068632, "learning_rate": 4.70334936706199e-06, "loss": 0.6521, "step": 961 }, { "epoch": 2.352078239608802, "grad_norm": 0.1445098025634498, "learning_rate": 4.7027442088727685e-06, "loss": 0.626, "step": 962 }, { "epoch": 2.354523227383863, "grad_norm": 0.12773159187186936, "learning_rate": 4.702138473073591e-06, "loss": 0.419, "step": 963 }, { "epoch": 2.356968215158924, "grad_norm": 0.12429265324879119, "learning_rate": 4.701532159823296e-06, "loss": 0.5937, "step": 964 }, { "epoch": 2.359413202933985, "grad_norm": 0.11163996494575917, "learning_rate": 4.700925269280873e-06, "loss": 0.6526, "step": 965 }, { "epoch": 2.3618581907090466, "grad_norm": 0.16217787992982088, "learning_rate": 4.700317801605463e-06, "loss": 0.9111, "step": 966 }, { "epoch": 2.3643031784841075, "grad_norm": 0.16188261942479187, "learning_rate": 4.6997097569563556e-06, "loss": 0.7366, "step": 967 }, { "epoch": 2.3667481662591685, "grad_norm": 0.13222198974900093, "learning_rate": 4.699101135492996e-06, "loss": 0.4229, "step": 968 }, { "epoch": 2.36919315403423, "grad_norm": 0.13167972641871356, "learning_rate": 4.698491937374979e-06, "loss": 0.6368, "step": 969 }, { "epoch": 2.371638141809291, "grad_norm": 0.1918013332396153, "learning_rate": 4.697882162762048e-06, "loss": 0.528, "step": 970 }, { "epoch": 2.374083129584352, "grad_norm": 0.11124572883435012, "learning_rate": 4.697271811814104e-06, "loss": 0.7842, "step": 971 }, { "epoch": 2.3765281173594133, "grad_norm": 0.2558350325922806, "learning_rate": 4.69666088469119e-06, "loss": 0.8491, "step": 972 }, { "epoch": 2.3789731051344742, "grad_norm": 0.09707668449837198, "learning_rate": 4.696049381553509e-06, "loss": 0.5957, "step": 973 }, { "epoch": 2.381418092909535, "grad_norm": 0.13278229718999052, "learning_rate": 4.695437302561408e-06, "loss": 0.7013, "step": 974 }, { "epoch": 2.3838630806845966, "grad_norm": 0.15091163013536923, "learning_rate": 4.694824647875391e-06, "loss": 0.5969, "step": 975 }, { "epoch": 2.3863080684596576, "grad_norm": 0.1923126066806301, "learning_rate": 4.6942114176561084e-06, "loss": 0.531, "step": 976 }, { "epoch": 2.388753056234719, "grad_norm": 0.1271699189668431, "learning_rate": 4.693597612064363e-06, "loss": 0.8707, "step": 977 }, { "epoch": 2.39119804400978, "grad_norm": 0.14677604098361635, "learning_rate": 4.69298323126111e-06, "loss": 0.4934, "step": 978 }, { "epoch": 2.393643031784841, "grad_norm": 0.13290736420022167, "learning_rate": 4.692368275407454e-06, "loss": 1.0197, "step": 979 }, { "epoch": 2.3960880195599024, "grad_norm": 0.18127629850782703, "learning_rate": 4.691752744664649e-06, "loss": 0.924, "step": 980 }, { "epoch": 2.3985330073349633, "grad_norm": 0.12153102637028339, "learning_rate": 4.691136639194102e-06, "loss": 0.8316, "step": 981 }, { "epoch": 2.4009779951100243, "grad_norm": 0.146796926374947, "learning_rate": 4.6905199591573706e-06, "loss": 0.5197, "step": 982 }, { "epoch": 2.4034229828850857, "grad_norm": 0.12258372469337145, "learning_rate": 4.6899027047161615e-06, "loss": 0.408, "step": 983 }, { "epoch": 2.4058679706601467, "grad_norm": 0.13064284406398707, "learning_rate": 4.689284876032334e-06, "loss": 0.501, "step": 984 }, { "epoch": 2.4083129584352077, "grad_norm": 0.13816045644304425, "learning_rate": 4.688666473267896e-06, "loss": 0.5305, "step": 985 }, { "epoch": 2.410757946210269, "grad_norm": 0.1429045056588975, "learning_rate": 4.688047496585007e-06, "loss": 0.5871, "step": 986 }, { "epoch": 2.41320293398533, "grad_norm": 0.15697968858913497, "learning_rate": 4.687427946145976e-06, "loss": 0.7086, "step": 987 }, { "epoch": 2.4156479217603914, "grad_norm": 0.14853228297253498, "learning_rate": 4.686807822113265e-06, "loss": 0.7972, "step": 988 }, { "epoch": 2.4180929095354524, "grad_norm": 0.1187897684974105, "learning_rate": 4.686187124649485e-06, "loss": 0.7091, "step": 989 }, { "epoch": 2.4205378973105134, "grad_norm": 0.10890781120831121, "learning_rate": 4.685565853917395e-06, "loss": 0.5262, "step": 990 }, { "epoch": 2.422982885085575, "grad_norm": 0.19300820999198887, "learning_rate": 4.684944010079908e-06, "loss": 1.0204, "step": 991 }, { "epoch": 2.4254278728606358, "grad_norm": 0.13237191204695267, "learning_rate": 4.684321593300085e-06, "loss": 0.4703, "step": 992 }, { "epoch": 2.4278728606356967, "grad_norm": 0.12070992681263393, "learning_rate": 4.6836986037411384e-06, "loss": 0.4505, "step": 993 }, { "epoch": 2.430317848410758, "grad_norm": 0.11625392508908099, "learning_rate": 4.683075041566429e-06, "loss": 0.5003, "step": 994 }, { "epoch": 2.432762836185819, "grad_norm": 0.1438745412848109, "learning_rate": 4.682450906939472e-06, "loss": 0.7978, "step": 995 }, { "epoch": 2.43520782396088, "grad_norm": 0.17819620163152924, "learning_rate": 4.681826200023927e-06, "loss": 0.4904, "step": 996 }, { "epoch": 2.4376528117359415, "grad_norm": 0.15526300365703913, "learning_rate": 4.681200920983607e-06, "loss": 0.8612, "step": 997 }, { "epoch": 2.4400977995110025, "grad_norm": 0.13594017934348038, "learning_rate": 4.680575069982475e-06, "loss": 0.4693, "step": 998 }, { "epoch": 2.4425427872860634, "grad_norm": 0.13653953028606336, "learning_rate": 4.679948647184644e-06, "loss": 0.8228, "step": 999 }, { "epoch": 2.444987775061125, "grad_norm": 0.16892391473239243, "learning_rate": 4.679321652754375e-06, "loss": 0.6108, "step": 1000 }, { "epoch": 2.447432762836186, "grad_norm": 0.1801068582953552, "learning_rate": 4.678694086856081e-06, "loss": 0.4818, "step": 1001 }, { "epoch": 2.449877750611247, "grad_norm": 0.16708118134265817, "learning_rate": 4.678065949654324e-06, "loss": 0.5227, "step": 1002 }, { "epoch": 2.452322738386308, "grad_norm": 0.11393260708516915, "learning_rate": 4.677437241313816e-06, "loss": 0.5652, "step": 1003 }, { "epoch": 2.454767726161369, "grad_norm": 0.1404136854199711, "learning_rate": 4.6768079619994195e-06, "loss": 0.6254, "step": 1004 }, { "epoch": 2.45721271393643, "grad_norm": 0.1603197520447684, "learning_rate": 4.676178111876145e-06, "loss": 0.8834, "step": 1005 }, { "epoch": 2.4596577017114916, "grad_norm": 0.11340308063412807, "learning_rate": 4.675547691109154e-06, "loss": 0.4282, "step": 1006 }, { "epoch": 2.4621026894865525, "grad_norm": 0.1431421510508806, "learning_rate": 4.674916699863756e-06, "loss": 0.5582, "step": 1007 }, { "epoch": 2.4645476772616135, "grad_norm": 0.09815779582924551, "learning_rate": 4.674285138305414e-06, "loss": 0.3488, "step": 1008 }, { "epoch": 2.466992665036675, "grad_norm": 0.17378708860918254, "learning_rate": 4.6736530065997356e-06, "loss": 0.4958, "step": 1009 }, { "epoch": 2.469437652811736, "grad_norm": 0.2074167538585944, "learning_rate": 4.6730203049124814e-06, "loss": 0.476, "step": 1010 }, { "epoch": 2.471882640586797, "grad_norm": 0.15955371484594874, "learning_rate": 4.672387033409559e-06, "loss": 0.4671, "step": 1011 }, { "epoch": 2.4743276283618583, "grad_norm": 0.149707285495147, "learning_rate": 4.671753192257028e-06, "loss": 0.404, "step": 1012 }, { "epoch": 2.4767726161369192, "grad_norm": 0.12917962775273079, "learning_rate": 4.6711187816210965e-06, "loss": 0.6948, "step": 1013 }, { "epoch": 2.47921760391198, "grad_norm": 0.15545400159453354, "learning_rate": 4.67048380166812e-06, "loss": 0.7672, "step": 1014 }, { "epoch": 2.4816625916870416, "grad_norm": 0.11432517987596699, "learning_rate": 4.669848252564605e-06, "loss": 0.5068, "step": 1015 }, { "epoch": 2.4841075794621026, "grad_norm": 0.18610809818951876, "learning_rate": 4.669212134477208e-06, "loss": 0.5838, "step": 1016 }, { "epoch": 2.486552567237164, "grad_norm": 0.1603389560247637, "learning_rate": 4.6685754475727315e-06, "loss": 0.4616, "step": 1017 }, { "epoch": 2.488997555012225, "grad_norm": 0.17459829017724168, "learning_rate": 4.667938192018133e-06, "loss": 1.0443, "step": 1018 }, { "epoch": 2.491442542787286, "grad_norm": 0.15081780987153742, "learning_rate": 4.667300367980512e-06, "loss": 0.7724, "step": 1019 }, { "epoch": 2.4938875305623474, "grad_norm": 0.11953483673601678, "learning_rate": 4.666661975627123e-06, "loss": 0.541, "step": 1020 }, { "epoch": 2.4963325183374083, "grad_norm": 0.18042774162777217, "learning_rate": 4.666023015125366e-06, "loss": 0.5689, "step": 1021 }, { "epoch": 2.4987775061124693, "grad_norm": 0.22952574694175681, "learning_rate": 4.665383486642791e-06, "loss": 0.9075, "step": 1022 }, { "epoch": 2.5012224938875307, "grad_norm": 0.12555143621518156, "learning_rate": 4.664743390347098e-06, "loss": 0.7301, "step": 1023 }, { "epoch": 2.5036674816625917, "grad_norm": 0.14648773488525046, "learning_rate": 4.664102726406132e-06, "loss": 1.0722, "step": 1024 }, { "epoch": 2.5061124694376526, "grad_norm": 0.20242907172936675, "learning_rate": 4.663461494987894e-06, "loss": 0.6039, "step": 1025 }, { "epoch": 2.508557457212714, "grad_norm": 0.15438751303198492, "learning_rate": 4.662819696260527e-06, "loss": 0.5306, "step": 1026 }, { "epoch": 2.511002444987775, "grad_norm": 0.1552410765148403, "learning_rate": 4.662177330392324e-06, "loss": 0.4432, "step": 1027 }, { "epoch": 2.5134474327628364, "grad_norm": 0.23341387331635374, "learning_rate": 4.661534397551732e-06, "loss": 0.7674, "step": 1028 }, { "epoch": 2.5158924205378974, "grad_norm": 0.1761720270268955, "learning_rate": 4.66089089790734e-06, "loss": 0.9345, "step": 1029 }, { "epoch": 2.5183374083129584, "grad_norm": 0.21442705492206302, "learning_rate": 4.660246831627888e-06, "loss": 0.397, "step": 1030 }, { "epoch": 2.52078239608802, "grad_norm": 0.15063333108197974, "learning_rate": 4.659602198882266e-06, "loss": 0.6694, "step": 1031 }, { "epoch": 2.5232273838630808, "grad_norm": 0.13793425707722298, "learning_rate": 4.658956999839512e-06, "loss": 0.5444, "step": 1032 }, { "epoch": 2.5256723716381417, "grad_norm": 0.14919245311740761, "learning_rate": 4.658311234668811e-06, "loss": 0.5382, "step": 1033 }, { "epoch": 2.528117359413203, "grad_norm": 0.1552781122188146, "learning_rate": 4.657664903539497e-06, "loss": 0.7202, "step": 1034 }, { "epoch": 2.530562347188264, "grad_norm": 0.1361755501567377, "learning_rate": 4.657018006621054e-06, "loss": 0.6824, "step": 1035 }, { "epoch": 2.533007334963325, "grad_norm": 0.17129766514631103, "learning_rate": 4.656370544083112e-06, "loss": 0.5598, "step": 1036 }, { "epoch": 2.5354523227383865, "grad_norm": 0.16876796946532477, "learning_rate": 4.6557225160954506e-06, "loss": 0.9978, "step": 1037 }, { "epoch": 2.5378973105134475, "grad_norm": 0.16195199884778624, "learning_rate": 4.655073922827998e-06, "loss": 0.5292, "step": 1038 }, { "epoch": 2.5403422982885084, "grad_norm": 0.14150547349395673, "learning_rate": 4.654424764450832e-06, "loss": 0.6811, "step": 1039 }, { "epoch": 2.54278728606357, "grad_norm": 0.2344239648151507, "learning_rate": 4.653775041134174e-06, "loss": 0.4769, "step": 1040 }, { "epoch": 2.545232273838631, "grad_norm": 0.2007802792542489, "learning_rate": 4.653124753048397e-06, "loss": 0.661, "step": 1041 }, { "epoch": 2.547677261613692, "grad_norm": 0.16131630634799532, "learning_rate": 4.652473900364021e-06, "loss": 0.8791, "step": 1042 }, { "epoch": 2.550122249388753, "grad_norm": 0.14814968797381928, "learning_rate": 4.6518224832517174e-06, "loss": 0.4839, "step": 1043 }, { "epoch": 2.552567237163814, "grad_norm": 0.15743299706469807, "learning_rate": 4.651170501882299e-06, "loss": 0.5036, "step": 1044 }, { "epoch": 2.555012224938875, "grad_norm": 0.17773198221053674, "learning_rate": 4.650517956426732e-06, "loss": 0.4923, "step": 1045 }, { "epoch": 2.5574572127139366, "grad_norm": 0.24212561295642518, "learning_rate": 4.649864847056129e-06, "loss": 0.5449, "step": 1046 }, { "epoch": 2.5599022004889975, "grad_norm": 0.2316556041739059, "learning_rate": 4.64921117394175e-06, "loss": 0.8427, "step": 1047 }, { "epoch": 2.5623471882640585, "grad_norm": 0.09738370266496339, "learning_rate": 4.6485569372550025e-06, "loss": 0.4653, "step": 1048 }, { "epoch": 2.56479217603912, "grad_norm": 0.16681168705222116, "learning_rate": 4.647902137167443e-06, "loss": 0.5186, "step": 1049 }, { "epoch": 2.567237163814181, "grad_norm": 0.1499951678784469, "learning_rate": 4.647246773850773e-06, "loss": 0.4371, "step": 1050 }, { "epoch": 2.569682151589242, "grad_norm": 0.16475374429973771, "learning_rate": 4.646590847476847e-06, "loss": 0.5339, "step": 1051 }, { "epoch": 2.5721271393643033, "grad_norm": 0.19656045166488556, "learning_rate": 4.645934358217662e-06, "loss": 0.4902, "step": 1052 }, { "epoch": 2.5745721271393642, "grad_norm": 0.13851180565946436, "learning_rate": 4.645277306245365e-06, "loss": 0.5153, "step": 1053 }, { "epoch": 2.577017114914425, "grad_norm": 0.13941503480662112, "learning_rate": 4.64461969173225e-06, "loss": 0.7362, "step": 1054 }, { "epoch": 2.5794621026894866, "grad_norm": 0.1317585043398215, "learning_rate": 4.643961514850759e-06, "loss": 0.4666, "step": 1055 }, { "epoch": 2.5819070904645476, "grad_norm": 0.13810522049144833, "learning_rate": 4.64330277577348e-06, "loss": 0.3837, "step": 1056 }, { "epoch": 2.5843520782396086, "grad_norm": 0.13132952247335872, "learning_rate": 4.64264347467315e-06, "loss": 0.6076, "step": 1057 }, { "epoch": 2.58679706601467, "grad_norm": 0.21038598840295386, "learning_rate": 4.641983611722653e-06, "loss": 0.5085, "step": 1058 }, { "epoch": 2.589242053789731, "grad_norm": 0.1263050206040351, "learning_rate": 4.6413231870950204e-06, "loss": 0.9666, "step": 1059 }, { "epoch": 2.591687041564792, "grad_norm": 0.17622449682485733, "learning_rate": 4.64066220096343e-06, "loss": 0.4403, "step": 1060 }, { "epoch": 2.5941320293398533, "grad_norm": 0.13355327315207532, "learning_rate": 4.640000653501207e-06, "loss": 0.5231, "step": 1061 }, { "epoch": 2.5965770171149143, "grad_norm": 0.20810238917678098, "learning_rate": 4.639338544881826e-06, "loss": 0.6895, "step": 1062 }, { "epoch": 2.5990220048899757, "grad_norm": 0.16758404947862493, "learning_rate": 4.638675875278907e-06, "loss": 0.8839, "step": 1063 }, { "epoch": 2.6014669926650367, "grad_norm": 0.21635251669690372, "learning_rate": 4.638012644866216e-06, "loss": 1.1101, "step": 1064 }, { "epoch": 2.6039119804400976, "grad_norm": 0.19783844597685704, "learning_rate": 4.637348853817666e-06, "loss": 1.3538, "step": 1065 }, { "epoch": 2.606356968215159, "grad_norm": 0.20553761607024784, "learning_rate": 4.636684502307321e-06, "loss": 0.4355, "step": 1066 }, { "epoch": 2.60880195599022, "grad_norm": 0.1605473761595738, "learning_rate": 4.636019590509387e-06, "loss": 0.5234, "step": 1067 }, { "epoch": 2.6112469437652814, "grad_norm": 0.16229957508562184, "learning_rate": 4.635354118598221e-06, "loss": 0.71, "step": 1068 }, { "epoch": 2.6136919315403424, "grad_norm": 0.2628449579611908, "learning_rate": 4.634688086748323e-06, "loss": 0.7098, "step": 1069 }, { "epoch": 2.6161369193154034, "grad_norm": 0.11654349733319179, "learning_rate": 4.634021495134343e-06, "loss": 0.4696, "step": 1070 }, { "epoch": 2.618581907090465, "grad_norm": 0.18229759695237086, "learning_rate": 4.633354343931077e-06, "loss": 1.0007, "step": 1071 }, { "epoch": 2.6210268948655258, "grad_norm": 0.12092147841102768, "learning_rate": 4.6326866333134655e-06, "loss": 0.4601, "step": 1072 }, { "epoch": 2.6234718826405867, "grad_norm": 0.1198755721681329, "learning_rate": 4.6320183634565996e-06, "loss": 0.4213, "step": 1073 }, { "epoch": 2.625916870415648, "grad_norm": 0.1413613396448796, "learning_rate": 4.631349534535713e-06, "loss": 0.4565, "step": 1074 }, { "epoch": 2.628361858190709, "grad_norm": 0.15358467095315684, "learning_rate": 4.6306801467261895e-06, "loss": 0.5086, "step": 1075 }, { "epoch": 2.63080684596577, "grad_norm": 0.21728528200081973, "learning_rate": 4.6300102002035576e-06, "loss": 0.7187, "step": 1076 }, { "epoch": 2.6332518337408315, "grad_norm": 0.15114237598030636, "learning_rate": 4.629339695143492e-06, "loss": 0.5394, "step": 1077 }, { "epoch": 2.6356968215158925, "grad_norm": 0.21030276424096803, "learning_rate": 4.628668631721816e-06, "loss": 0.5601, "step": 1078 }, { "epoch": 2.6381418092909534, "grad_norm": 0.39810798703626527, "learning_rate": 4.627997010114495e-06, "loss": 0.923, "step": 1079 }, { "epoch": 2.640586797066015, "grad_norm": 0.1353956511691117, "learning_rate": 4.627324830497646e-06, "loss": 0.5467, "step": 1080 }, { "epoch": 2.643031784841076, "grad_norm": 0.1808890853882477, "learning_rate": 4.6266520930475286e-06, "loss": 0.7682, "step": 1081 }, { "epoch": 2.645476772616137, "grad_norm": 0.1256481379291353, "learning_rate": 4.62597879794055e-06, "loss": 0.6591, "step": 1082 }, { "epoch": 2.647921760391198, "grad_norm": 0.12709757847018727, "learning_rate": 4.625304945353265e-06, "loss": 0.7998, "step": 1083 }, { "epoch": 2.650366748166259, "grad_norm": 0.18860867427527078, "learning_rate": 4.6246305354623725e-06, "loss": 0.9052, "step": 1084 }, { "epoch": 2.65281173594132, "grad_norm": 0.10716849762092664, "learning_rate": 4.623955568444717e-06, "loss": 0.4708, "step": 1085 }, { "epoch": 2.6552567237163816, "grad_norm": 0.14974019719972007, "learning_rate": 4.623280044477291e-06, "loss": 1.0287, "step": 1086 }, { "epoch": 2.6577017114914425, "grad_norm": 0.1771362167640652, "learning_rate": 4.6226039637372334e-06, "loss": 0.7266, "step": 1087 }, { "epoch": 2.6601466992665035, "grad_norm": 0.17744649800600945, "learning_rate": 4.6219273264018265e-06, "loss": 0.5589, "step": 1088 }, { "epoch": 2.662591687041565, "grad_norm": 0.1414630445393385, "learning_rate": 4.621250132648501e-06, "loss": 0.4179, "step": 1089 }, { "epoch": 2.665036674816626, "grad_norm": 0.14623572467848908, "learning_rate": 4.620572382654833e-06, "loss": 0.5305, "step": 1090 }, { "epoch": 2.667481662591687, "grad_norm": 0.1627982292451145, "learning_rate": 4.619894076598543e-06, "loss": 0.5862, "step": 1091 }, { "epoch": 2.6699266503667483, "grad_norm": 0.19089055860318369, "learning_rate": 4.619215214657498e-06, "loss": 1.0751, "step": 1092 }, { "epoch": 2.6723716381418092, "grad_norm": 0.12607217717478977, "learning_rate": 4.618535797009713e-06, "loss": 0.6387, "step": 1093 }, { "epoch": 2.67481662591687, "grad_norm": 0.12552754367747399, "learning_rate": 4.6178558238333456e-06, "loss": 0.7387, "step": 1094 }, { "epoch": 2.6772616136919316, "grad_norm": 0.11711762571840323, "learning_rate": 4.617175295306701e-06, "loss": 0.4759, "step": 1095 }, { "epoch": 2.6797066014669926, "grad_norm": 0.16928081239958062, "learning_rate": 4.616494211608229e-06, "loss": 0.4545, "step": 1096 }, { "epoch": 2.6821515892420535, "grad_norm": 0.16753995890049916, "learning_rate": 4.615812572916525e-06, "loss": 0.5309, "step": 1097 }, { "epoch": 2.684596577017115, "grad_norm": 0.15483599059983444, "learning_rate": 4.615130379410329e-06, "loss": 0.6783, "step": 1098 }, { "epoch": 2.687041564792176, "grad_norm": 0.22104396671414372, "learning_rate": 4.614447631268531e-06, "loss": 0.6145, "step": 1099 }, { "epoch": 2.689486552567237, "grad_norm": 0.1786122056076665, "learning_rate": 4.613764328670161e-06, "loss": 0.9847, "step": 1100 }, { "epoch": 2.6919315403422983, "grad_norm": 0.1639304987605364, "learning_rate": 4.613080471794397e-06, "loss": 0.7752, "step": 1101 }, { "epoch": 2.6943765281173593, "grad_norm": 0.21666933161378235, "learning_rate": 4.612396060820563e-06, "loss": 0.4715, "step": 1102 }, { "epoch": 2.6968215158924207, "grad_norm": 0.1406384182804471, "learning_rate": 4.611711095928126e-06, "loss": 0.4199, "step": 1103 }, { "epoch": 2.6992665036674817, "grad_norm": 0.14032307272962788, "learning_rate": 4.611025577296699e-06, "loss": 0.4079, "step": 1104 }, { "epoch": 2.7017114914425426, "grad_norm": 0.1735388645254333, "learning_rate": 4.610339505106042e-06, "loss": 0.5341, "step": 1105 }, { "epoch": 2.704156479217604, "grad_norm": 0.13987110208172146, "learning_rate": 4.6096528795360575e-06, "loss": 0.7478, "step": 1106 }, { "epoch": 2.706601466992665, "grad_norm": 0.12405469393257719, "learning_rate": 4.608965700766795e-06, "loss": 0.5862, "step": 1107 }, { "epoch": 2.7090464547677264, "grad_norm": 0.13439305148331876, "learning_rate": 4.608277968978449e-06, "loss": 0.472, "step": 1108 }, { "epoch": 2.7114914425427874, "grad_norm": 0.22771469787635684, "learning_rate": 4.607589684351358e-06, "loss": 0.6159, "step": 1109 }, { "epoch": 2.7139364303178484, "grad_norm": 0.20777836853947337, "learning_rate": 4.606900847066007e-06, "loss": 0.9789, "step": 1110 }, { "epoch": 2.71638141809291, "grad_norm": 0.10579540328389136, "learning_rate": 4.606211457303023e-06, "loss": 0.4085, "step": 1111 }, { "epoch": 2.7188264058679708, "grad_norm": 0.12372849376020849, "learning_rate": 4.605521515243181e-06, "loss": 0.3852, "step": 1112 }, { "epoch": 2.7212713936430317, "grad_norm": 0.18214940924175324, "learning_rate": 4.604831021067399e-06, "loss": 0.7262, "step": 1113 }, { "epoch": 2.723716381418093, "grad_norm": 0.16508046445229, "learning_rate": 4.604139974956742e-06, "loss": 1.1808, "step": 1114 }, { "epoch": 2.726161369193154, "grad_norm": 0.14812349353016127, "learning_rate": 4.603448377092417e-06, "loss": 0.6326, "step": 1115 }, { "epoch": 2.728606356968215, "grad_norm": 0.1524642289589814, "learning_rate": 4.602756227655775e-06, "loss": 0.848, "step": 1116 }, { "epoch": 2.7310513447432765, "grad_norm": 0.14841344133862183, "learning_rate": 4.602063526828317e-06, "loss": 0.4519, "step": 1117 }, { "epoch": 2.7334963325183375, "grad_norm": 0.22498396643855229, "learning_rate": 4.601370274791683e-06, "loss": 0.6302, "step": 1118 }, { "epoch": 2.7359413202933984, "grad_norm": 0.17427705509509117, "learning_rate": 4.600676471727659e-06, "loss": 0.4823, "step": 1119 }, { "epoch": 2.73838630806846, "grad_norm": 0.11765019267284965, "learning_rate": 4.599982117818177e-06, "loss": 0.6129, "step": 1120 }, { "epoch": 2.740831295843521, "grad_norm": 0.12678908999019384, "learning_rate": 4.599287213245312e-06, "loss": 0.4959, "step": 1121 }, { "epoch": 2.743276283618582, "grad_norm": 0.17136478844031794, "learning_rate": 4.598591758191284e-06, "loss": 0.492, "step": 1122 }, { "epoch": 2.745721271393643, "grad_norm": 0.1619824238914578, "learning_rate": 4.597895752838458e-06, "loss": 1.0815, "step": 1123 }, { "epoch": 2.748166259168704, "grad_norm": 0.134354982306176, "learning_rate": 4.597199197369342e-06, "loss": 0.7212, "step": 1124 }, { "epoch": 2.750611246943765, "grad_norm": 0.1791360354603443, "learning_rate": 4.5965020919665874e-06, "loss": 0.5534, "step": 1125 }, { "epoch": 2.7530562347188265, "grad_norm": 0.16845668333686556, "learning_rate": 4.595804436812994e-06, "loss": 0.7317, "step": 1126 }, { "epoch": 2.7555012224938875, "grad_norm": 0.24891756650069807, "learning_rate": 4.595106232091501e-06, "loss": 0.3743, "step": 1127 }, { "epoch": 2.7579462102689485, "grad_norm": 0.12827044760700881, "learning_rate": 4.594407477985194e-06, "loss": 0.5523, "step": 1128 }, { "epoch": 2.76039119804401, "grad_norm": 0.1535366582898294, "learning_rate": 4.593708174677302e-06, "loss": 0.9273, "step": 1129 }, { "epoch": 2.762836185819071, "grad_norm": 0.13908411961849743, "learning_rate": 4.593008322351199e-06, "loss": 0.5235, "step": 1130 }, { "epoch": 2.765281173594132, "grad_norm": 0.15128583163765236, "learning_rate": 4.592307921190402e-06, "loss": 0.6772, "step": 1131 }, { "epoch": 2.7677261613691932, "grad_norm": 0.20381586142103972, "learning_rate": 4.591606971378573e-06, "loss": 0.4186, "step": 1132 }, { "epoch": 2.770171149144254, "grad_norm": 0.19922371363212993, "learning_rate": 4.590905473099515e-06, "loss": 0.4984, "step": 1133 }, { "epoch": 2.772616136919315, "grad_norm": 0.1656824219964035, "learning_rate": 4.590203426537179e-06, "loss": 0.8546, "step": 1134 }, { "epoch": 2.7750611246943766, "grad_norm": 0.18635990870109867, "learning_rate": 4.589500831875658e-06, "loss": 0.4809, "step": 1135 }, { "epoch": 2.7775061124694376, "grad_norm": 0.11909794478489337, "learning_rate": 4.588797689299185e-06, "loss": 0.5238, "step": 1136 }, { "epoch": 2.7799511002444985, "grad_norm": 0.09250628919307029, "learning_rate": 4.588093998992144e-06, "loss": 0.391, "step": 1137 }, { "epoch": 2.78239608801956, "grad_norm": 0.2038924536008576, "learning_rate": 4.587389761139057e-06, "loss": 0.8296, "step": 1138 }, { "epoch": 2.784841075794621, "grad_norm": 0.20391590091864686, "learning_rate": 4.586684975924591e-06, "loss": 1.0367, "step": 1139 }, { "epoch": 2.787286063569682, "grad_norm": 0.19815045096211195, "learning_rate": 4.585979643533558e-06, "loss": 0.9047, "step": 1140 }, { "epoch": 2.7897310513447433, "grad_norm": 0.180186434042982, "learning_rate": 4.585273764150911e-06, "loss": 0.8345, "step": 1141 }, { "epoch": 2.7921760391198043, "grad_norm": 0.18892634971252834, "learning_rate": 4.584567337961748e-06, "loss": 0.8406, "step": 1142 }, { "epoch": 2.7946210268948657, "grad_norm": 0.15912441421470258, "learning_rate": 4.583860365151311e-06, "loss": 0.4768, "step": 1143 }, { "epoch": 2.7970660146699267, "grad_norm": 0.1454674981094276, "learning_rate": 4.583152845904984e-06, "loss": 0.5008, "step": 1144 }, { "epoch": 2.7995110024449876, "grad_norm": 0.18786730789366027, "learning_rate": 4.582444780408294e-06, "loss": 1.0029, "step": 1145 }, { "epoch": 2.801955990220049, "grad_norm": 0.18217041426647468, "learning_rate": 4.581736168846913e-06, "loss": 0.8548, "step": 1146 }, { "epoch": 2.80440097799511, "grad_norm": 0.12938018093067463, "learning_rate": 4.581027011406654e-06, "loss": 0.4258, "step": 1147 }, { "epoch": 2.8068459657701714, "grad_norm": 0.21399634838588283, "learning_rate": 4.580317308273476e-06, "loss": 0.5438, "step": 1148 }, { "epoch": 2.8092909535452324, "grad_norm": 0.18062754053107197, "learning_rate": 4.579607059633478e-06, "loss": 1.2083, "step": 1149 }, { "epoch": 2.8117359413202934, "grad_norm": 0.13329013577527737, "learning_rate": 4.578896265672904e-06, "loss": 0.805, "step": 1150 }, { "epoch": 2.8141809290953548, "grad_norm": 0.1907589633661021, "learning_rate": 4.57818492657814e-06, "loss": 0.6109, "step": 1151 }, { "epoch": 2.8166259168704157, "grad_norm": 0.1955894405461504, "learning_rate": 4.577473042535716e-06, "loss": 0.4771, "step": 1152 }, { "epoch": 2.8190709046454767, "grad_norm": 0.16305278316672941, "learning_rate": 4.576760613732304e-06, "loss": 0.7663, "step": 1153 }, { "epoch": 2.821515892420538, "grad_norm": 0.18503576796762433, "learning_rate": 4.5760476403547195e-06, "loss": 0.54, "step": 1154 }, { "epoch": 2.823960880195599, "grad_norm": 0.13403390245576055, "learning_rate": 4.57533412258992e-06, "loss": 0.795, "step": 1155 }, { "epoch": 2.82640586797066, "grad_norm": 0.1058672590575123, "learning_rate": 4.574620060625007e-06, "loss": 0.4006, "step": 1156 }, { "epoch": 2.8288508557457215, "grad_norm": 0.2619085689200034, "learning_rate": 4.573905454647221e-06, "loss": 0.5499, "step": 1157 }, { "epoch": 2.8312958435207825, "grad_norm": 0.20133711206782226, "learning_rate": 4.573190304843953e-06, "loss": 0.4823, "step": 1158 }, { "epoch": 2.8337408312958434, "grad_norm": 0.176477271768202, "learning_rate": 4.572474611402727e-06, "loss": 0.4592, "step": 1159 }, { "epoch": 2.836185819070905, "grad_norm": 0.15256435679263647, "learning_rate": 4.571758374511217e-06, "loss": 0.6061, "step": 1160 }, { "epoch": 2.838630806845966, "grad_norm": 0.1578203928635288, "learning_rate": 4.571041594357236e-06, "loss": 0.7555, "step": 1161 }, { "epoch": 2.8410757946210268, "grad_norm": 0.18603253858672134, "learning_rate": 4.5703242711287394e-06, "loss": 0.4318, "step": 1162 }, { "epoch": 2.843520782396088, "grad_norm": 0.20892142538076658, "learning_rate": 4.569606405013827e-06, "loss": 1.1209, "step": 1163 }, { "epoch": 2.845965770171149, "grad_norm": 0.15200102966958237, "learning_rate": 4.568887996200738e-06, "loss": 0.6955, "step": 1164 }, { "epoch": 2.84841075794621, "grad_norm": 0.1435574938658663, "learning_rate": 4.568169044877857e-06, "loss": 0.4129, "step": 1165 }, { "epoch": 2.8508557457212715, "grad_norm": 0.13225672577335346, "learning_rate": 4.56744955123371e-06, "loss": 0.4034, "step": 1166 }, { "epoch": 2.8533007334963325, "grad_norm": 0.1323564354039993, "learning_rate": 4.5667295154569635e-06, "loss": 0.5591, "step": 1167 }, { "epoch": 2.8557457212713935, "grad_norm": 0.12474035239423872, "learning_rate": 4.566008937736428e-06, "loss": 0.723, "step": 1168 }, { "epoch": 2.858190709046455, "grad_norm": 0.15188252590376547, "learning_rate": 4.5652878182610545e-06, "loss": 0.5545, "step": 1169 }, { "epoch": 2.860635696821516, "grad_norm": 0.14490741605601506, "learning_rate": 4.564566157219938e-06, "loss": 0.4108, "step": 1170 }, { "epoch": 2.863080684596577, "grad_norm": 0.13716370715185708, "learning_rate": 4.563843954802315e-06, "loss": 0.5045, "step": 1171 }, { "epoch": 2.8655256723716382, "grad_norm": 0.14640593017626594, "learning_rate": 4.563121211197562e-06, "loss": 0.4527, "step": 1172 }, { "epoch": 2.867970660146699, "grad_norm": 0.15428150761732115, "learning_rate": 4.562397926595202e-06, "loss": 0.5177, "step": 1173 }, { "epoch": 2.87041564792176, "grad_norm": 0.17295647417142113, "learning_rate": 4.561674101184891e-06, "loss": 0.4766, "step": 1174 }, { "epoch": 2.8728606356968216, "grad_norm": 0.18327794991110657, "learning_rate": 4.560949735156438e-06, "loss": 0.3641, "step": 1175 }, { "epoch": 2.8753056234718826, "grad_norm": 0.16893337654175863, "learning_rate": 4.560224828699786e-06, "loss": 0.5531, "step": 1176 }, { "epoch": 2.8777506112469435, "grad_norm": 0.11770490149142304, "learning_rate": 4.5594993820050236e-06, "loss": 0.5493, "step": 1177 }, { "epoch": 2.880195599022005, "grad_norm": 0.13852478666511223, "learning_rate": 4.558773395262378e-06, "loss": 0.4518, "step": 1178 }, { "epoch": 2.882640586797066, "grad_norm": 0.16954234879886837, "learning_rate": 4.55804686866222e-06, "loss": 0.5094, "step": 1179 }, { "epoch": 2.885085574572127, "grad_norm": 0.1553310105636114, "learning_rate": 4.557319802395061e-06, "loss": 0.8109, "step": 1180 }, { "epoch": 2.8875305623471883, "grad_norm": 0.13436894812173883, "learning_rate": 4.556592196651556e-06, "loss": 0.5678, "step": 1181 }, { "epoch": 2.8899755501222493, "grad_norm": 0.11568213342939627, "learning_rate": 4.555864051622497e-06, "loss": 0.4483, "step": 1182 }, { "epoch": 2.8924205378973102, "grad_norm": 0.2173866815381029, "learning_rate": 4.555135367498824e-06, "loss": 0.6624, "step": 1183 }, { "epoch": 2.8948655256723717, "grad_norm": 0.11794853163201566, "learning_rate": 4.554406144471612e-06, "loss": 0.4613, "step": 1184 }, { "epoch": 2.8973105134474326, "grad_norm": 0.11660525071756107, "learning_rate": 4.553676382732081e-06, "loss": 0.6771, "step": 1185 }, { "epoch": 2.899755501222494, "grad_norm": 0.17638501613839083, "learning_rate": 4.552946082471591e-06, "loss": 0.5546, "step": 1186 }, { "epoch": 2.902200488997555, "grad_norm": 0.1388948745682831, "learning_rate": 4.5522152438816424e-06, "loss": 0.5451, "step": 1187 }, { "epoch": 2.904645476772616, "grad_norm": 0.19680729431463675, "learning_rate": 4.55148386715388e-06, "loss": 1.0006, "step": 1188 }, { "epoch": 2.9070904645476774, "grad_norm": 0.15430184633797764, "learning_rate": 4.550751952480086e-06, "loss": 0.4794, "step": 1189 }, { "epoch": 2.9095354523227384, "grad_norm": 0.19232521175241812, "learning_rate": 4.550019500052186e-06, "loss": 0.4749, "step": 1190 }, { "epoch": 2.9119804400977998, "grad_norm": 0.10232736836420402, "learning_rate": 4.549286510062245e-06, "loss": 0.2845, "step": 1191 }, { "epoch": 2.9144254278728607, "grad_norm": 0.12211782695807297, "learning_rate": 4.54855298270247e-06, "loss": 0.5865, "step": 1192 }, { "epoch": 2.9168704156479217, "grad_norm": 0.219019708747965, "learning_rate": 4.547818918165209e-06, "loss": 0.4852, "step": 1193 }, { "epoch": 2.919315403422983, "grad_norm": 0.21299814297361516, "learning_rate": 4.54708431664295e-06, "loss": 1.1898, "step": 1194 }, { "epoch": 2.921760391198044, "grad_norm": 0.26406013347393353, "learning_rate": 4.546349178328322e-06, "loss": 0.7653, "step": 1195 }, { "epoch": 2.924205378973105, "grad_norm": 0.26119771654190144, "learning_rate": 4.545613503414097e-06, "loss": 0.4015, "step": 1196 }, { "epoch": 2.9266503667481665, "grad_norm": 0.12431891239243097, "learning_rate": 4.5448772920931835e-06, "loss": 0.4894, "step": 1197 }, { "epoch": 2.9290953545232274, "grad_norm": 0.1393440146323019, "learning_rate": 4.544140544558634e-06, "loss": 0.6626, "step": 1198 }, { "epoch": 2.9315403422982884, "grad_norm": 0.15876534649796736, "learning_rate": 4.543403261003642e-06, "loss": 0.5137, "step": 1199 }, { "epoch": 2.93398533007335, "grad_norm": 0.20863670265453652, "learning_rate": 4.542665441621537e-06, "loss": 0.8357, "step": 1200 }, { "epoch": 2.936430317848411, "grad_norm": 0.17858106152611558, "learning_rate": 4.5419270866057944e-06, "loss": 0.7318, "step": 1201 }, { "epoch": 2.9388753056234718, "grad_norm": 0.17963165759212266, "learning_rate": 4.541188196150027e-06, "loss": 0.4322, "step": 1202 }, { "epoch": 2.941320293398533, "grad_norm": 0.0869756874887522, "learning_rate": 4.540448770447989e-06, "loss": 0.4098, "step": 1203 }, { "epoch": 2.943765281173594, "grad_norm": 0.2315108456952936, "learning_rate": 4.539708809693575e-06, "loss": 1.0613, "step": 1204 }, { "epoch": 2.946210268948655, "grad_norm": 0.22025547790682168, "learning_rate": 4.53896831408082e-06, "loss": 0.5298, "step": 1205 }, { "epoch": 2.9486552567237165, "grad_norm": 0.16179186752052616, "learning_rate": 4.538227283803897e-06, "loss": 0.6957, "step": 1206 }, { "epoch": 2.9511002444987775, "grad_norm": 0.2619071044019018, "learning_rate": 4.537485719057123e-06, "loss": 0.8555, "step": 1207 }, { "epoch": 2.9535452322738385, "grad_norm": 0.11835943841204094, "learning_rate": 4.536743620034952e-06, "loss": 0.4937, "step": 1208 }, { "epoch": 2.9559902200489, "grad_norm": 0.1297255155184526, "learning_rate": 4.536000986931981e-06, "loss": 0.7769, "step": 1209 }, { "epoch": 2.958435207823961, "grad_norm": 0.2595960675736413, "learning_rate": 4.535257819942944e-06, "loss": 0.542, "step": 1210 }, { "epoch": 2.960880195599022, "grad_norm": 0.19099562458836739, "learning_rate": 4.534514119262717e-06, "loss": 0.5471, "step": 1211 }, { "epoch": 2.9633251833740832, "grad_norm": 0.14805890495778595, "learning_rate": 4.533769885086315e-06, "loss": 0.556, "step": 1212 }, { "epoch": 2.965770171149144, "grad_norm": 0.2310952050169163, "learning_rate": 4.533025117608893e-06, "loss": 0.4327, "step": 1213 }, { "epoch": 2.968215158924205, "grad_norm": 0.13329291154467213, "learning_rate": 4.532279817025746e-06, "loss": 0.5053, "step": 1214 }, { "epoch": 2.9706601466992666, "grad_norm": 0.16297651212210837, "learning_rate": 4.53153398353231e-06, "loss": 0.4704, "step": 1215 }, { "epoch": 2.9731051344743276, "grad_norm": 0.15181748206516257, "learning_rate": 4.53078761732416e-06, "loss": 0.4626, "step": 1216 }, { "epoch": 2.9755501222493885, "grad_norm": 0.18555734990057685, "learning_rate": 4.530040718597008e-06, "loss": 0.4609, "step": 1217 }, { "epoch": 2.97799511002445, "grad_norm": 0.2179231804133384, "learning_rate": 4.529293287546709e-06, "loss": 0.7654, "step": 1218 }, { "epoch": 2.980440097799511, "grad_norm": 0.14695249721169837, "learning_rate": 4.528545324369257e-06, "loss": 0.6068, "step": 1219 }, { "epoch": 2.982885085574572, "grad_norm": 0.17154534442966718, "learning_rate": 4.527796829260784e-06, "loss": 0.5653, "step": 1220 }, { "epoch": 2.9853300733496333, "grad_norm": 0.1884027033641388, "learning_rate": 4.527047802417565e-06, "loss": 0.7649, "step": 1221 }, { "epoch": 2.9877750611246943, "grad_norm": 0.24625535458810247, "learning_rate": 4.526298244036009e-06, "loss": 0.5049, "step": 1222 }, { "epoch": 2.9902200488997552, "grad_norm": 0.11502702460276133, "learning_rate": 4.525548154312669e-06, "loss": 0.301, "step": 1223 }, { "epoch": 2.9926650366748166, "grad_norm": 0.1405433891124375, "learning_rate": 4.5247975334442354e-06, "loss": 0.4516, "step": 1224 }, { "epoch": 2.9951100244498776, "grad_norm": 0.12385925983270608, "learning_rate": 4.5240463816275386e-06, "loss": 0.6481, "step": 1225 }, { "epoch": 2.997555012224939, "grad_norm": 0.17135102010592884, "learning_rate": 4.523294699059546e-06, "loss": 0.4397, "step": 1226 }, { "epoch": 3.0, "grad_norm": 0.17556446279068616, "learning_rate": 4.522542485937369e-06, "loss": 0.4461, "step": 1227 }, { "epoch": 3.002444987775061, "grad_norm": 0.22017811012032287, "learning_rate": 4.521789742458252e-06, "loss": 0.7488, "step": 1228 }, { "epoch": 3.0048899755501224, "grad_norm": 0.14308431153355256, "learning_rate": 4.521036468819585e-06, "loss": 0.3856, "step": 1229 }, { "epoch": 3.0073349633251834, "grad_norm": 0.13413740267855528, "learning_rate": 4.520282665218889e-06, "loss": 0.5433, "step": 1230 }, { "epoch": 3.0097799511002443, "grad_norm": 0.13691744977686368, "learning_rate": 4.519528331853833e-06, "loss": 0.6822, "step": 1231 }, { "epoch": 3.0122249388753057, "grad_norm": 0.14155039512126036, "learning_rate": 4.518773468922219e-06, "loss": 0.6341, "step": 1232 }, { "epoch": 3.0146699266503667, "grad_norm": 0.17764974230942743, "learning_rate": 4.518018076621987e-06, "loss": 0.744, "step": 1233 }, { "epoch": 3.0171149144254277, "grad_norm": 0.11359776688505224, "learning_rate": 4.517262155151222e-06, "loss": 0.3236, "step": 1234 }, { "epoch": 3.019559902200489, "grad_norm": 0.17191503122410867, "learning_rate": 4.516505704708141e-06, "loss": 0.5528, "step": 1235 }, { "epoch": 3.02200488997555, "grad_norm": 0.18983494459128966, "learning_rate": 4.515748725491103e-06, "loss": 0.4683, "step": 1236 }, { "epoch": 3.0244498777506115, "grad_norm": 0.23179954830868868, "learning_rate": 4.514991217698606e-06, "loss": 0.4708, "step": 1237 }, { "epoch": 3.0268948655256724, "grad_norm": 0.12989589330009677, "learning_rate": 4.514233181529287e-06, "loss": 0.6847, "step": 1238 }, { "epoch": 3.0293398533007334, "grad_norm": 0.1625443865259557, "learning_rate": 4.51347461718192e-06, "loss": 0.7301, "step": 1239 }, { "epoch": 3.031784841075795, "grad_norm": 0.1443303118696757, "learning_rate": 4.512715524855415e-06, "loss": 0.6634, "step": 1240 }, { "epoch": 3.034229828850856, "grad_norm": 0.1800621731396222, "learning_rate": 4.511955904748827e-06, "loss": 0.8164, "step": 1241 }, { "epoch": 3.0366748166259168, "grad_norm": 0.1524683088085164, "learning_rate": 4.511195757061344e-06, "loss": 0.5134, "step": 1242 }, { "epoch": 3.039119804400978, "grad_norm": 0.23900869253038692, "learning_rate": 4.510435081992295e-06, "loss": 1.0166, "step": 1243 }, { "epoch": 3.041564792176039, "grad_norm": 0.1339810297782538, "learning_rate": 4.5096738797411445e-06, "loss": 0.3627, "step": 1244 }, { "epoch": 3.0440097799511, "grad_norm": 0.22190862069788, "learning_rate": 4.508912150507499e-06, "loss": 0.4595, "step": 1245 }, { "epoch": 3.0464547677261615, "grad_norm": 0.09981991541901061, "learning_rate": 4.508149894491102e-06, "loss": 0.4481, "step": 1246 }, { "epoch": 3.0488997555012225, "grad_norm": 0.12651881824722214, "learning_rate": 4.507387111891832e-06, "loss": 0.4101, "step": 1247 }, { "epoch": 3.0513447432762835, "grad_norm": 0.17516120702364987, "learning_rate": 4.506623802909711e-06, "loss": 0.6584, "step": 1248 }, { "epoch": 3.053789731051345, "grad_norm": 0.14734649247028397, "learning_rate": 4.505859967744893e-06, "loss": 0.3531, "step": 1249 }, { "epoch": 3.056234718826406, "grad_norm": 0.1609056089426988, "learning_rate": 4.505095606597675e-06, "loss": 0.5998, "step": 1250 }, { "epoch": 3.058679706601467, "grad_norm": 0.17704999441932256, "learning_rate": 4.504330719668489e-06, "loss": 0.7456, "step": 1251 }, { "epoch": 3.0611246943765282, "grad_norm": 0.2341202807189493, "learning_rate": 4.503565307157908e-06, "loss": 0.9408, "step": 1252 }, { "epoch": 3.063569682151589, "grad_norm": 0.15203998286845186, "learning_rate": 4.502799369266638e-06, "loss": 0.4612, "step": 1253 }, { "epoch": 3.06601466992665, "grad_norm": 0.16767054732418843, "learning_rate": 4.502032906195527e-06, "loss": 0.6231, "step": 1254 }, { "epoch": 3.0684596577017116, "grad_norm": 0.18127768128806845, "learning_rate": 4.501265918145559e-06, "loss": 0.4582, "step": 1255 }, { "epoch": 3.0709046454767726, "grad_norm": 0.11480184930677406, "learning_rate": 4.500498405317855e-06, "loss": 0.5081, "step": 1256 }, { "epoch": 3.0733496332518335, "grad_norm": 0.1903991347399298, "learning_rate": 4.499730367913674e-06, "loss": 0.4764, "step": 1257 }, { "epoch": 3.075794621026895, "grad_norm": 0.2165356089746921, "learning_rate": 4.498961806134416e-06, "loss": 0.7513, "step": 1258 }, { "epoch": 3.078239608801956, "grad_norm": 0.19112686737044238, "learning_rate": 4.498192720181613e-06, "loss": 0.5037, "step": 1259 }, { "epoch": 3.0806845965770173, "grad_norm": 0.16812609180766974, "learning_rate": 4.497423110256936e-06, "loss": 0.5316, "step": 1260 }, { "epoch": 3.0831295843520783, "grad_norm": 0.1739822648682703, "learning_rate": 4.496652976562197e-06, "loss": 0.7983, "step": 1261 }, { "epoch": 3.0855745721271393, "grad_norm": 0.08510611049043375, "learning_rate": 4.495882319299341e-06, "loss": 0.3519, "step": 1262 }, { "epoch": 3.0880195599022007, "grad_norm": 0.14984509142203617, "learning_rate": 4.495111138670453e-06, "loss": 0.4065, "step": 1263 }, { "epoch": 3.0904645476772616, "grad_norm": 0.13732612078852077, "learning_rate": 4.494339434877752e-06, "loss": 0.5618, "step": 1264 }, { "epoch": 3.0929095354523226, "grad_norm": 0.16662068437497599, "learning_rate": 4.493567208123599e-06, "loss": 0.875, "step": 1265 }, { "epoch": 3.095354523227384, "grad_norm": 0.21586131738190212, "learning_rate": 4.492794458610489e-06, "loss": 0.579, "step": 1266 }, { "epoch": 3.097799511002445, "grad_norm": 0.16120181201505424, "learning_rate": 4.492021186541054e-06, "loss": 0.3902, "step": 1267 }, { "epoch": 3.100244498777506, "grad_norm": 0.1710438386660066, "learning_rate": 4.491247392118064e-06, "loss": 0.7774, "step": 1268 }, { "epoch": 3.1026894865525674, "grad_norm": 0.1781472626823595, "learning_rate": 4.490473075544425e-06, "loss": 0.4266, "step": 1269 }, { "epoch": 3.1051344743276283, "grad_norm": 0.15279216474346197, "learning_rate": 4.4896982370231814e-06, "loss": 0.5912, "step": 1270 }, { "epoch": 3.1075794621026893, "grad_norm": 0.2513092610834609, "learning_rate": 4.4889228767575125e-06, "loss": 0.5008, "step": 1271 }, { "epoch": 3.1100244498777507, "grad_norm": 0.1334857450489098, "learning_rate": 4.488146994950736e-06, "loss": 0.4526, "step": 1272 }, { "epoch": 3.1124694376528117, "grad_norm": 0.15430155417138328, "learning_rate": 4.487370591806307e-06, "loss": 0.5617, "step": 1273 }, { "epoch": 3.1149144254278727, "grad_norm": 0.18054157010670074, "learning_rate": 4.4865936675278146e-06, "loss": 0.4552, "step": 1274 }, { "epoch": 3.117359413202934, "grad_norm": 0.15352619819594984, "learning_rate": 4.485816222318986e-06, "loss": 0.5506, "step": 1275 }, { "epoch": 3.119804400977995, "grad_norm": 0.16326556676056037, "learning_rate": 4.4850382563836866e-06, "loss": 0.5217, "step": 1276 }, { "epoch": 3.122249388753056, "grad_norm": 0.20044497673056416, "learning_rate": 4.484259769925917e-06, "loss": 0.9806, "step": 1277 }, { "epoch": 3.1246943765281174, "grad_norm": 0.21235811569160104, "learning_rate": 4.483480763149812e-06, "loss": 0.5584, "step": 1278 }, { "epoch": 3.1271393643031784, "grad_norm": 0.1276886046716671, "learning_rate": 4.482701236259647e-06, "loss": 0.4395, "step": 1279 }, { "epoch": 3.12958435207824, "grad_norm": 0.12622011128561542, "learning_rate": 4.48192118945983e-06, "loss": 0.4504, "step": 1280 }, { "epoch": 3.132029339853301, "grad_norm": 0.1577682178159191, "learning_rate": 4.48114062295491e-06, "loss": 0.9822, "step": 1281 }, { "epoch": 3.1344743276283618, "grad_norm": 0.15604296175203658, "learning_rate": 4.480359536949566e-06, "loss": 0.5444, "step": 1282 }, { "epoch": 3.136919315403423, "grad_norm": 0.20696327454044353, "learning_rate": 4.47957793164862e-06, "loss": 1.1878, "step": 1283 }, { "epoch": 3.139364303178484, "grad_norm": 0.2562488852692517, "learning_rate": 4.478795807257024e-06, "loss": 1.2973, "step": 1284 }, { "epoch": 3.141809290953545, "grad_norm": 0.1441785371357995, "learning_rate": 4.4780131639798695e-06, "loss": 0.6982, "step": 1285 }, { "epoch": 3.1442542787286065, "grad_norm": 0.1385038619084508, "learning_rate": 4.477230002022385e-06, "loss": 0.5676, "step": 1286 }, { "epoch": 3.1466992665036675, "grad_norm": 0.17215628950115078, "learning_rate": 4.476446321589932e-06, "loss": 0.4852, "step": 1287 }, { "epoch": 3.1491442542787285, "grad_norm": 0.18127179781756417, "learning_rate": 4.4756621228880105e-06, "loss": 0.4488, "step": 1288 }, { "epoch": 3.15158924205379, "grad_norm": 0.16891594311283806, "learning_rate": 4.474877406122254e-06, "loss": 0.5933, "step": 1289 }, { "epoch": 3.154034229828851, "grad_norm": 0.14345758158197966, "learning_rate": 4.4740921714984346e-06, "loss": 0.7536, "step": 1290 }, { "epoch": 3.156479217603912, "grad_norm": 0.1642820279512776, "learning_rate": 4.473306419222458e-06, "loss": 0.8624, "step": 1291 }, { "epoch": 3.1589242053789732, "grad_norm": 0.15664337588829666, "learning_rate": 4.472520149500365e-06, "loss": 0.5104, "step": 1292 }, { "epoch": 3.161369193154034, "grad_norm": 0.19938166287186623, "learning_rate": 4.471733362538337e-06, "loss": 0.9768, "step": 1293 }, { "epoch": 3.163814180929095, "grad_norm": 0.19656281672816017, "learning_rate": 4.470946058542685e-06, "loss": 0.6707, "step": 1294 }, { "epoch": 3.1662591687041566, "grad_norm": 0.18351179065718753, "learning_rate": 4.4701582377198594e-06, "loss": 0.4688, "step": 1295 }, { "epoch": 3.1687041564792175, "grad_norm": 0.15249767677811685, "learning_rate": 4.469369900276443e-06, "loss": 0.4937, "step": 1296 }, { "epoch": 3.1711491442542785, "grad_norm": 0.17664107501039134, "learning_rate": 4.4685810464191584e-06, "loss": 0.8433, "step": 1297 }, { "epoch": 3.17359413202934, "grad_norm": 0.17313713762115512, "learning_rate": 4.46779167635486e-06, "loss": 0.6598, "step": 1298 }, { "epoch": 3.176039119804401, "grad_norm": 0.1207350934618037, "learning_rate": 4.467001790290538e-06, "loss": 0.4687, "step": 1299 }, { "epoch": 3.178484107579462, "grad_norm": 0.16127282387547937, "learning_rate": 4.466211388433319e-06, "loss": 0.4434, "step": 1300 }, { "epoch": 3.1809290953545233, "grad_norm": 0.16511133323502677, "learning_rate": 4.465420470990466e-06, "loss": 0.6124, "step": 1301 }, { "epoch": 3.1833740831295843, "grad_norm": 0.13963470303586623, "learning_rate": 4.464629038169373e-06, "loss": 0.3694, "step": 1302 }, { "epoch": 3.1858190709046457, "grad_norm": 0.22506221024699372, "learning_rate": 4.463837090177573e-06, "loss": 1.0833, "step": 1303 }, { "epoch": 3.1882640586797066, "grad_norm": 0.21540378537989388, "learning_rate": 4.463044627222734e-06, "loss": 0.614, "step": 1304 }, { "epoch": 3.1907090464547676, "grad_norm": 0.1427782779960142, "learning_rate": 4.462251649512656e-06, "loss": 0.5039, "step": 1305 }, { "epoch": 3.193154034229829, "grad_norm": 0.20557753007232604, "learning_rate": 4.461458157255278e-06, "loss": 0.9929, "step": 1306 }, { "epoch": 3.19559902200489, "grad_norm": 0.18081999185390915, "learning_rate": 4.46066415065867e-06, "loss": 0.4553, "step": 1307 }, { "epoch": 3.198044009779951, "grad_norm": 0.11179310092695381, "learning_rate": 4.459869629931039e-06, "loss": 0.4595, "step": 1308 }, { "epoch": 3.2004889975550124, "grad_norm": 0.1900949783350105, "learning_rate": 4.4590745952807275e-06, "loss": 0.6683, "step": 1309 }, { "epoch": 3.2029339853300733, "grad_norm": 0.25915575506080024, "learning_rate": 4.4582790469162116e-06, "loss": 0.5207, "step": 1310 }, { "epoch": 3.2053789731051343, "grad_norm": 0.17578677687295075, "learning_rate": 4.457482985046101e-06, "loss": 0.7039, "step": 1311 }, { "epoch": 3.2078239608801957, "grad_norm": 0.1584048253811272, "learning_rate": 4.456686409879143e-06, "loss": 0.704, "step": 1312 }, { "epoch": 3.2102689486552567, "grad_norm": 0.10762403748793255, "learning_rate": 4.455889321624217e-06, "loss": 0.8104, "step": 1313 }, { "epoch": 3.2127139364303177, "grad_norm": 0.1385164779105021, "learning_rate": 4.4550917204903384e-06, "loss": 0.9023, "step": 1314 }, { "epoch": 3.215158924205379, "grad_norm": 0.26520266059550096, "learning_rate": 4.454293606686655e-06, "loss": 0.4816, "step": 1315 }, { "epoch": 3.21760391198044, "grad_norm": 0.18442310939643727, "learning_rate": 4.4534949804224516e-06, "loss": 0.463, "step": 1316 }, { "epoch": 3.220048899755501, "grad_norm": 0.15599800908581243, "learning_rate": 4.452695841907146e-06, "loss": 0.4833, "step": 1317 }, { "epoch": 3.2224938875305624, "grad_norm": 0.14586960791085962, "learning_rate": 4.451896191350292e-06, "loss": 0.6932, "step": 1318 }, { "epoch": 3.2249388753056234, "grad_norm": 0.12815509749195497, "learning_rate": 4.451096028961574e-06, "loss": 0.5276, "step": 1319 }, { "epoch": 3.227383863080685, "grad_norm": 0.19605937567101867, "learning_rate": 4.450295354950814e-06, "loss": 0.4353, "step": 1320 }, { "epoch": 3.229828850855746, "grad_norm": 0.21004343334280354, "learning_rate": 4.449494169527967e-06, "loss": 0.444, "step": 1321 }, { "epoch": 3.2322738386308068, "grad_norm": 0.13561312683289953, "learning_rate": 4.448692472903123e-06, "loss": 0.7136, "step": 1322 }, { "epoch": 3.234718826405868, "grad_norm": 0.15234110978656384, "learning_rate": 4.447890265286504e-06, "loss": 0.445, "step": 1323 }, { "epoch": 3.237163814180929, "grad_norm": 0.24344737661470825, "learning_rate": 4.447087546888467e-06, "loss": 0.4139, "step": 1324 }, { "epoch": 3.23960880195599, "grad_norm": 0.18254591203805187, "learning_rate": 4.446284317919503e-06, "loss": 0.7247, "step": 1325 }, { "epoch": 3.2420537897310515, "grad_norm": 0.23369416459868653, "learning_rate": 4.445480578590239e-06, "loss": 0.7487, "step": 1326 }, { "epoch": 3.2444987775061125, "grad_norm": 0.15565741297396668, "learning_rate": 4.444676329111432e-06, "loss": 0.7647, "step": 1327 }, { "epoch": 3.2469437652811735, "grad_norm": 0.1189697691681801, "learning_rate": 4.443871569693976e-06, "loss": 0.4374, "step": 1328 }, { "epoch": 3.249388753056235, "grad_norm": 0.14522022741890428, "learning_rate": 4.443066300548896e-06, "loss": 0.4156, "step": 1329 }, { "epoch": 3.251833740831296, "grad_norm": 0.16398341849098805, "learning_rate": 4.442260521887353e-06, "loss": 0.8357, "step": 1330 }, { "epoch": 3.254278728606357, "grad_norm": 0.1762782723475786, "learning_rate": 4.44145423392064e-06, "loss": 0.5304, "step": 1331 }, { "epoch": 3.256723716381418, "grad_norm": 0.1922090385737338, "learning_rate": 4.440647436860184e-06, "loss": 0.5482, "step": 1332 }, { "epoch": 3.259168704156479, "grad_norm": 0.12490568420733943, "learning_rate": 4.439840130917546e-06, "loss": 0.4291, "step": 1333 }, { "epoch": 3.26161369193154, "grad_norm": 0.15498586251161028, "learning_rate": 4.43903231630442e-06, "loss": 0.4267, "step": 1334 }, { "epoch": 3.2640586797066016, "grad_norm": 0.20029005869070945, "learning_rate": 4.4382239932326344e-06, "loss": 1.0601, "step": 1335 }, { "epoch": 3.2665036674816625, "grad_norm": 0.15807953353238713, "learning_rate": 4.437415161914148e-06, "loss": 0.4628, "step": 1336 }, { "epoch": 3.2689486552567235, "grad_norm": 0.1333337603647761, "learning_rate": 4.436605822561056e-06, "loss": 0.4927, "step": 1337 }, { "epoch": 3.271393643031785, "grad_norm": 0.17488025046141334, "learning_rate": 4.435795975385587e-06, "loss": 0.4456, "step": 1338 }, { "epoch": 3.273838630806846, "grad_norm": 0.21241879714365483, "learning_rate": 4.4349856206001e-06, "loss": 0.4399, "step": 1339 }, { "epoch": 3.276283618581907, "grad_norm": 0.1749284252524475, "learning_rate": 4.434174758417088e-06, "loss": 0.4009, "step": 1340 }, { "epoch": 3.2787286063569683, "grad_norm": 0.3018037054563145, "learning_rate": 4.4333633890491786e-06, "loss": 1.102, "step": 1341 }, { "epoch": 3.2811735941320292, "grad_norm": 0.15889734041520165, "learning_rate": 4.432551512709132e-06, "loss": 0.4829, "step": 1342 }, { "epoch": 3.28361858190709, "grad_norm": 0.1820426500272768, "learning_rate": 4.4317391296098385e-06, "loss": 0.5921, "step": 1343 }, { "epoch": 3.2860635696821516, "grad_norm": 0.15985652528319985, "learning_rate": 4.430926239964327e-06, "loss": 0.7949, "step": 1344 }, { "epoch": 3.2885085574572126, "grad_norm": 0.21200904226124137, "learning_rate": 4.430112843985753e-06, "loss": 0.542, "step": 1345 }, { "epoch": 3.290953545232274, "grad_norm": 0.2019499855173706, "learning_rate": 4.429298941887408e-06, "loss": 0.3833, "step": 1346 }, { "epoch": 3.293398533007335, "grad_norm": 0.20393170602186828, "learning_rate": 4.4284845338827165e-06, "loss": 0.5783, "step": 1347 }, { "epoch": 3.295843520782396, "grad_norm": 0.28244006800033244, "learning_rate": 4.4276696201852346e-06, "loss": 0.4809, "step": 1348 }, { "epoch": 3.2982885085574574, "grad_norm": 0.1301736319744728, "learning_rate": 4.426854201008651e-06, "loss": 0.4721, "step": 1349 }, { "epoch": 3.3007334963325183, "grad_norm": 0.20962482316826106, "learning_rate": 4.426038276566788e-06, "loss": 0.3922, "step": 1350 }, { "epoch": 3.3031784841075793, "grad_norm": 0.16291490072675338, "learning_rate": 4.4252218470735984e-06, "loss": 0.3659, "step": 1351 }, { "epoch": 3.3056234718826407, "grad_norm": 0.29003709329839317, "learning_rate": 4.424404912743171e-06, "loss": 1.108, "step": 1352 }, { "epoch": 3.3080684596577017, "grad_norm": 0.13014565347766072, "learning_rate": 4.423587473789722e-06, "loss": 0.4388, "step": 1353 }, { "epoch": 3.3105134474327627, "grad_norm": 0.15602013644278834, "learning_rate": 4.422769530427604e-06, "loss": 0.4812, "step": 1354 }, { "epoch": 3.312958435207824, "grad_norm": 0.17247109398420665, "learning_rate": 4.421951082871301e-06, "loss": 0.5419, "step": 1355 }, { "epoch": 3.315403422982885, "grad_norm": 0.24803878066460436, "learning_rate": 4.421132131335427e-06, "loss": 0.7113, "step": 1356 }, { "epoch": 3.3178484107579465, "grad_norm": 0.21826640030444933, "learning_rate": 4.420312676034731e-06, "loss": 0.5719, "step": 1357 }, { "epoch": 3.3202933985330074, "grad_norm": 0.1518586778362871, "learning_rate": 4.419492717184094e-06, "loss": 0.7889, "step": 1358 }, { "epoch": 3.3227383863080684, "grad_norm": 0.20829789104766844, "learning_rate": 4.418672254998525e-06, "loss": 0.9048, "step": 1359 }, { "epoch": 3.32518337408313, "grad_norm": 0.1768582651782789, "learning_rate": 4.417851289693171e-06, "loss": 0.8189, "step": 1360 }, { "epoch": 3.3276283618581908, "grad_norm": 0.1534543141728541, "learning_rate": 4.417029821483306e-06, "loss": 0.4563, "step": 1361 }, { "epoch": 3.3300733496332517, "grad_norm": 0.1055761917647611, "learning_rate": 4.416207850584339e-06, "loss": 0.4424, "step": 1362 }, { "epoch": 3.332518337408313, "grad_norm": 0.1940248670005149, "learning_rate": 4.415385377211808e-06, "loss": 0.7029, "step": 1363 }, { "epoch": 3.334963325183374, "grad_norm": 0.18171624743522638, "learning_rate": 4.414562401581386e-06, "loss": 0.6187, "step": 1364 }, { "epoch": 3.337408312958435, "grad_norm": 0.22204221011849848, "learning_rate": 4.4137389239088744e-06, "loss": 0.5287, "step": 1365 }, { "epoch": 3.3398533007334965, "grad_norm": 0.2805067684006206, "learning_rate": 4.41291494441021e-06, "loss": 0.7874, "step": 1366 }, { "epoch": 3.3422982885085575, "grad_norm": 0.11870399869303958, "learning_rate": 4.412090463301458e-06, "loss": 0.5957, "step": 1367 }, { "epoch": 3.3447432762836184, "grad_norm": 0.20919219321658356, "learning_rate": 4.411265480798815e-06, "loss": 0.6323, "step": 1368 }, { "epoch": 3.34718826405868, "grad_norm": 0.15914712958604574, "learning_rate": 4.410439997118612e-06, "loss": 0.5404, "step": 1369 }, { "epoch": 3.349633251833741, "grad_norm": 0.17930234472139145, "learning_rate": 4.4096140124773095e-06, "loss": 0.7741, "step": 1370 }, { "epoch": 3.352078239608802, "grad_norm": 0.17758547784514836, "learning_rate": 4.408787527091499e-06, "loss": 0.9643, "step": 1371 }, { "epoch": 3.354523227383863, "grad_norm": 0.1785453425744948, "learning_rate": 4.407960541177905e-06, "loss": 0.4483, "step": 1372 }, { "epoch": 3.356968215158924, "grad_norm": 0.16515287143659235, "learning_rate": 4.40713305495338e-06, "loss": 1.0549, "step": 1373 }, { "epoch": 3.359413202933985, "grad_norm": 0.16622841635302463, "learning_rate": 4.406305068634912e-06, "loss": 0.5289, "step": 1374 }, { "epoch": 3.3618581907090466, "grad_norm": 0.18395560324980878, "learning_rate": 4.405476582439616e-06, "loss": 0.7516, "step": 1375 }, { "epoch": 3.3643031784841075, "grad_norm": 0.21607564782044592, "learning_rate": 4.404647596584742e-06, "loss": 0.4733, "step": 1376 }, { "epoch": 3.3667481662591685, "grad_norm": 0.1526065309931781, "learning_rate": 4.403818111287668e-06, "loss": 0.7207, "step": 1377 }, { "epoch": 3.36919315403423, "grad_norm": 0.19394622016390767, "learning_rate": 4.4029881267659045e-06, "loss": 0.921, "step": 1378 }, { "epoch": 3.371638141809291, "grad_norm": 0.20355571954244783, "learning_rate": 4.402157643237092e-06, "loss": 0.9869, "step": 1379 }, { "epoch": 3.374083129584352, "grad_norm": 0.12071461333941827, "learning_rate": 4.401326660919002e-06, "loss": 0.4069, "step": 1380 }, { "epoch": 3.3765281173594133, "grad_norm": 0.16293483783139062, "learning_rate": 4.400495180029537e-06, "loss": 0.4753, "step": 1381 }, { "epoch": 3.3789731051344742, "grad_norm": 0.18484211920235352, "learning_rate": 4.399663200786733e-06, "loss": 0.6018, "step": 1382 }, { "epoch": 3.381418092909535, "grad_norm": 0.10182079121309406, "learning_rate": 4.39883072340875e-06, "loss": 0.486, "step": 1383 }, { "epoch": 3.3838630806845966, "grad_norm": 0.17739761366535337, "learning_rate": 4.397997748113884e-06, "loss": 0.88, "step": 1384 }, { "epoch": 3.3863080684596576, "grad_norm": 0.20190398997845, "learning_rate": 4.3971642751205626e-06, "loss": 0.6489, "step": 1385 }, { "epoch": 3.388753056234719, "grad_norm": 0.20790052550815696, "learning_rate": 4.3963303046473385e-06, "loss": 0.7959, "step": 1386 }, { "epoch": 3.39119804400978, "grad_norm": 0.20335046330718368, "learning_rate": 4.395495836912899e-06, "loss": 0.6075, "step": 1387 }, { "epoch": 3.393643031784841, "grad_norm": 0.18343092134458344, "learning_rate": 4.394660872136061e-06, "loss": 0.9653, "step": 1388 }, { "epoch": 3.3960880195599024, "grad_norm": 0.2260283261789901, "learning_rate": 4.393825410535771e-06, "loss": 0.5119, "step": 1389 }, { "epoch": 3.3985330073349633, "grad_norm": 0.22771194568008446, "learning_rate": 4.392989452331106e-06, "loss": 0.3854, "step": 1390 }, { "epoch": 3.4009779951100243, "grad_norm": 0.23458777399375302, "learning_rate": 4.392152997741273e-06, "loss": 0.5985, "step": 1391 }, { "epoch": 3.4034229828850857, "grad_norm": 0.14535138495432257, "learning_rate": 4.391316046985611e-06, "loss": 0.3749, "step": 1392 }, { "epoch": 3.4058679706601467, "grad_norm": 0.18522645316667558, "learning_rate": 4.390478600283585e-06, "loss": 0.8911, "step": 1393 }, { "epoch": 3.4083129584352077, "grad_norm": 0.22187619657363306, "learning_rate": 4.389640657854796e-06, "loss": 0.6332, "step": 1394 }, { "epoch": 3.410757946210269, "grad_norm": 0.12981434073399714, "learning_rate": 4.388802219918969e-06, "loss": 0.3989, "step": 1395 }, { "epoch": 3.41320293398533, "grad_norm": 0.17414209909242584, "learning_rate": 4.387963286695963e-06, "loss": 0.6862, "step": 1396 }, { "epoch": 3.4156479217603914, "grad_norm": 0.16355705303495935, "learning_rate": 4.387123858405765e-06, "loss": 0.6819, "step": 1397 }, { "epoch": 3.4180929095354524, "grad_norm": 0.21839448707169096, "learning_rate": 4.386283935268493e-06, "loss": 0.6606, "step": 1398 }, { "epoch": 3.4205378973105134, "grad_norm": 0.2005445773900621, "learning_rate": 4.385443517504393e-06, "loss": 0.762, "step": 1399 }, { "epoch": 3.422982885085575, "grad_norm": 0.14701066238117477, "learning_rate": 4.384602605333843e-06, "loss": 0.4893, "step": 1400 }, { "epoch": 3.4254278728606358, "grad_norm": 0.1590540649426254, "learning_rate": 4.3837611989773485e-06, "loss": 0.5255, "step": 1401 }, { "epoch": 3.4278728606356967, "grad_norm": 0.20913179942040158, "learning_rate": 4.3829192986555445e-06, "loss": 0.3938, "step": 1402 }, { "epoch": 3.430317848410758, "grad_norm": 0.1958948456807502, "learning_rate": 4.3820769045891995e-06, "loss": 0.9963, "step": 1403 }, { "epoch": 3.432762836185819, "grad_norm": 0.2076553151679762, "learning_rate": 4.381234016999206e-06, "loss": 0.4865, "step": 1404 }, { "epoch": 3.43520782396088, "grad_norm": 0.17323831054579475, "learning_rate": 4.38039063610659e-06, "loss": 0.699, "step": 1405 }, { "epoch": 3.4376528117359415, "grad_norm": 0.20331146573962627, "learning_rate": 4.379546762132504e-06, "loss": 0.376, "step": 1406 }, { "epoch": 3.4400977995110025, "grad_norm": 0.1267542509895385, "learning_rate": 4.378702395298231e-06, "loss": 0.617, "step": 1407 }, { "epoch": 3.4425427872860634, "grad_norm": 0.1640927808711629, "learning_rate": 4.3778575358251844e-06, "loss": 0.3967, "step": 1408 }, { "epoch": 3.444987775061125, "grad_norm": 0.1281459689292886, "learning_rate": 4.3770121839349054e-06, "loss": 0.5522, "step": 1409 }, { "epoch": 3.447432762836186, "grad_norm": 0.20616684405351077, "learning_rate": 4.376166339849064e-06, "loss": 0.5627, "step": 1410 }, { "epoch": 3.449877750611247, "grad_norm": 0.24047842981071962, "learning_rate": 4.37532000378946e-06, "loss": 0.4852, "step": 1411 }, { "epoch": 3.452322738386308, "grad_norm": 0.11413467377691602, "learning_rate": 4.374473175978023e-06, "loss": 0.4811, "step": 1412 }, { "epoch": 3.454767726161369, "grad_norm": 0.12112066177938999, "learning_rate": 4.37362585663681e-06, "loss": 0.5288, "step": 1413 }, { "epoch": 3.45721271393643, "grad_norm": 0.23508199707061983, "learning_rate": 4.372778045988008e-06, "loss": 0.6079, "step": 1414 }, { "epoch": 3.4596577017114916, "grad_norm": 0.20379458878680404, "learning_rate": 4.37192974425393e-06, "loss": 0.5999, "step": 1415 }, { "epoch": 3.4621026894865525, "grad_norm": 0.17966541922919488, "learning_rate": 4.371080951657024e-06, "loss": 0.736, "step": 1416 }, { "epoch": 3.4645476772616135, "grad_norm": 0.18645930734314087, "learning_rate": 4.37023166841986e-06, "loss": 0.8258, "step": 1417 }, { "epoch": 3.466992665036675, "grad_norm": 0.27363349253423064, "learning_rate": 4.369381894765141e-06, "loss": 0.4445, "step": 1418 }, { "epoch": 3.469437652811736, "grad_norm": 0.21538590190623094, "learning_rate": 4.3685316309156955e-06, "loss": 0.7469, "step": 1419 }, { "epoch": 3.471882640586797, "grad_norm": 0.20406088705742215, "learning_rate": 4.367680877094484e-06, "loss": 0.4606, "step": 1420 }, { "epoch": 3.4743276283618583, "grad_norm": 0.21520826602228588, "learning_rate": 4.366829633524592e-06, "loss": 0.4542, "step": 1421 }, { "epoch": 3.4767726161369192, "grad_norm": 0.16909965092155824, "learning_rate": 4.3659779004292366e-06, "loss": 0.4435, "step": 1422 }, { "epoch": 3.47921760391198, "grad_norm": 0.17452577866518698, "learning_rate": 4.3651256780317605e-06, "loss": 0.7729, "step": 1423 }, { "epoch": 3.4816625916870416, "grad_norm": 0.12607983488825988, "learning_rate": 4.364272966555636e-06, "loss": 0.7839, "step": 1424 }, { "epoch": 3.4841075794621026, "grad_norm": 0.15244942520596705, "learning_rate": 4.363419766224465e-06, "loss": 0.644, "step": 1425 }, { "epoch": 3.486552567237164, "grad_norm": 0.15320036776048773, "learning_rate": 4.362566077261974e-06, "loss": 0.6334, "step": 1426 }, { "epoch": 3.488997555012225, "grad_norm": 0.19369292462420964, "learning_rate": 4.361711899892021e-06, "loss": 0.8962, "step": 1427 }, { "epoch": 3.491442542787286, "grad_norm": 0.24324497766246367, "learning_rate": 4.360857234338591e-06, "loss": 0.5653, "step": 1428 }, { "epoch": 3.4938875305623474, "grad_norm": 0.15193539668516265, "learning_rate": 4.360002080825796e-06, "loss": 0.4622, "step": 1429 }, { "epoch": 3.4963325183374083, "grad_norm": 0.15479630971448796, "learning_rate": 4.359146439577878e-06, "loss": 0.4407, "step": 1430 }, { "epoch": 3.4987775061124693, "grad_norm": 0.13791943750594668, "learning_rate": 4.3582903108192054e-06, "loss": 0.6699, "step": 1431 }, { "epoch": 3.5012224938875307, "grad_norm": 0.14650506014495693, "learning_rate": 4.357433694774274e-06, "loss": 0.499, "step": 1432 }, { "epoch": 3.5036674816625917, "grad_norm": 0.1871743483050339, "learning_rate": 4.356576591667709e-06, "loss": 0.4916, "step": 1433 }, { "epoch": 3.5061124694376526, "grad_norm": 0.17092922306443895, "learning_rate": 4.355719001724262e-06, "loss": 0.7783, "step": 1434 }, { "epoch": 3.508557457212714, "grad_norm": 0.1954119187353881, "learning_rate": 4.354860925168812e-06, "loss": 0.5707, "step": 1435 }, { "epoch": 3.511002444987775, "grad_norm": 0.21981753743401947, "learning_rate": 4.354002362226368e-06, "loss": 1.6729, "step": 1436 }, { "epoch": 3.5134474327628364, "grad_norm": 0.26154835211336896, "learning_rate": 4.353143313122064e-06, "loss": 0.9926, "step": 1437 }, { "epoch": 3.5158924205378974, "grad_norm": 0.14187237410065198, "learning_rate": 4.352283778081162e-06, "loss": 0.7691, "step": 1438 }, { "epoch": 3.5183374083129584, "grad_norm": 0.18628350673981284, "learning_rate": 4.351423757329052e-06, "loss": 0.8613, "step": 1439 }, { "epoch": 3.52078239608802, "grad_norm": 0.12831278154072343, "learning_rate": 4.350563251091252e-06, "loss": 0.8102, "step": 1440 }, { "epoch": 3.5232273838630808, "grad_norm": 0.2233819312228632, "learning_rate": 4.349702259593405e-06, "loss": 0.427, "step": 1441 }, { "epoch": 3.5256723716381417, "grad_norm": 0.2875855231881401, "learning_rate": 4.348840783061284e-06, "loss": 1.0445, "step": 1442 }, { "epoch": 3.528117359413203, "grad_norm": 0.2569706177012072, "learning_rate": 4.347978821720786e-06, "loss": 0.4943, "step": 1443 }, { "epoch": 3.530562347188264, "grad_norm": 0.20567981083111123, "learning_rate": 4.347116375797938e-06, "loss": 1.0003, "step": 1444 }, { "epoch": 3.533007334963325, "grad_norm": 0.21219055670552875, "learning_rate": 4.346253445518894e-06, "loss": 0.7194, "step": 1445 }, { "epoch": 3.5354523227383865, "grad_norm": 0.2748775990899456, "learning_rate": 4.345390031109934e-06, "loss": 0.7129, "step": 1446 }, { "epoch": 3.5378973105134475, "grad_norm": 0.15102110166666918, "learning_rate": 4.344526132797463e-06, "loss": 0.5414, "step": 1447 }, { "epoch": 3.5403422982885084, "grad_norm": 0.18082100796509593, "learning_rate": 4.343661750808017e-06, "loss": 0.8485, "step": 1448 }, { "epoch": 3.54278728606357, "grad_norm": 0.13442170133735099, "learning_rate": 4.342796885368256e-06, "loss": 0.6917, "step": 1449 }, { "epoch": 3.545232273838631, "grad_norm": 0.19256661819102777, "learning_rate": 4.341931536704968e-06, "loss": 0.5569, "step": 1450 }, { "epoch": 3.547677261613692, "grad_norm": 0.16459209263237912, "learning_rate": 4.341065705045066e-06, "loss": 0.4166, "step": 1451 }, { "epoch": 3.550122249388753, "grad_norm": 0.183932147676097, "learning_rate": 4.3401993906155914e-06, "loss": 0.6807, "step": 1452 }, { "epoch": 3.552567237163814, "grad_norm": 0.1900098132990076, "learning_rate": 4.339332593643713e-06, "loss": 0.8729, "step": 1453 }, { "epoch": 3.555012224938875, "grad_norm": 0.20799056978977354, "learning_rate": 4.338465314356722e-06, "loss": 0.3431, "step": 1454 }, { "epoch": 3.5574572127139366, "grad_norm": 0.1287927261339702, "learning_rate": 4.337597552982042e-06, "loss": 0.6634, "step": 1455 }, { "epoch": 3.5599022004889975, "grad_norm": 0.1769892306400033, "learning_rate": 4.3367293097472176e-06, "loss": 0.9828, "step": 1456 }, { "epoch": 3.5623471882640585, "grad_norm": 0.22673138798462789, "learning_rate": 4.335860584879924e-06, "loss": 0.4221, "step": 1457 }, { "epoch": 3.56479217603912, "grad_norm": 0.19840215749516363, "learning_rate": 4.334991378607959e-06, "loss": 0.4405, "step": 1458 }, { "epoch": 3.567237163814181, "grad_norm": 0.171654173913462, "learning_rate": 4.3341216911592496e-06, "loss": 0.5189, "step": 1459 }, { "epoch": 3.569682151589242, "grad_norm": 0.16664268588769726, "learning_rate": 4.333251522761847e-06, "loss": 0.5313, "step": 1460 }, { "epoch": 3.5721271393643033, "grad_norm": 0.2067790030698619, "learning_rate": 4.332380873643929e-06, "loss": 0.5817, "step": 1461 }, { "epoch": 3.5745721271393642, "grad_norm": 0.1486872688804778, "learning_rate": 4.3315097440338006e-06, "loss": 0.8625, "step": 1462 }, { "epoch": 3.577017114914425, "grad_norm": 0.2499377110971087, "learning_rate": 4.330638134159891e-06, "loss": 0.446, "step": 1463 }, { "epoch": 3.5794621026894866, "grad_norm": 0.2094586283328521, "learning_rate": 4.329766044250757e-06, "loss": 0.8141, "step": 1464 }, { "epoch": 3.5819070904645476, "grad_norm": 0.11548997375784534, "learning_rate": 4.328893474535081e-06, "loss": 0.5048, "step": 1465 }, { "epoch": 3.5843520782396086, "grad_norm": 0.1619720756009413, "learning_rate": 4.328020425241669e-06, "loss": 0.5402, "step": 1466 }, { "epoch": 3.58679706601467, "grad_norm": 0.1884784234051161, "learning_rate": 4.3271468965994566e-06, "loss": 0.8024, "step": 1467 }, { "epoch": 3.589242053789731, "grad_norm": 0.25480939961537113, "learning_rate": 4.3262728888375024e-06, "loss": 0.3462, "step": 1468 }, { "epoch": 3.591687041564792, "grad_norm": 0.18468870483292496, "learning_rate": 4.32539840218499e-06, "loss": 0.4549, "step": 1469 }, { "epoch": 3.5941320293398533, "grad_norm": 0.21589180748163994, "learning_rate": 4.324523436871231e-06, "loss": 0.6731, "step": 1470 }, { "epoch": 3.5965770171149143, "grad_norm": 0.15857567521917573, "learning_rate": 4.323647993125661e-06, "loss": 0.552, "step": 1471 }, { "epoch": 3.5990220048899757, "grad_norm": 0.2198527865014562, "learning_rate": 4.322772071177842e-06, "loss": 0.672, "step": 1472 }, { "epoch": 3.6014669926650367, "grad_norm": 0.21414227829709978, "learning_rate": 4.321895671257459e-06, "loss": 0.5018, "step": 1473 }, { "epoch": 3.6039119804400976, "grad_norm": 0.17079431168079148, "learning_rate": 4.321018793594326e-06, "loss": 0.5151, "step": 1474 }, { "epoch": 3.606356968215159, "grad_norm": 0.14659845528289905, "learning_rate": 4.320141438418379e-06, "loss": 0.4638, "step": 1475 }, { "epoch": 3.60880195599022, "grad_norm": 0.18733563204604847, "learning_rate": 4.319263605959683e-06, "loss": 0.469, "step": 1476 }, { "epoch": 3.6112469437652814, "grad_norm": 0.17218919258897608, "learning_rate": 4.318385296448422e-06, "loss": 0.5544, "step": 1477 }, { "epoch": 3.6136919315403424, "grad_norm": 0.14140450197392046, "learning_rate": 4.317506510114913e-06, "loss": 0.5253, "step": 1478 }, { "epoch": 3.6161369193154034, "grad_norm": 0.1396760457464423, "learning_rate": 4.316627247189589e-06, "loss": 0.4702, "step": 1479 }, { "epoch": 3.618581907090465, "grad_norm": 0.12013345575866503, "learning_rate": 4.315747507903018e-06, "loss": 0.4641, "step": 1480 }, { "epoch": 3.6210268948655258, "grad_norm": 0.15888546585628682, "learning_rate": 4.3148672924858855e-06, "loss": 0.7532, "step": 1481 }, { "epoch": 3.6234718826405867, "grad_norm": 0.16868996795096083, "learning_rate": 4.313986601169003e-06, "loss": 0.4134, "step": 1482 }, { "epoch": 3.625916870415648, "grad_norm": 0.28188336926929525, "learning_rate": 4.31310543418331e-06, "loss": 0.4692, "step": 1483 }, { "epoch": 3.628361858190709, "grad_norm": 0.1863411648408465, "learning_rate": 4.312223791759868e-06, "loss": 0.3785, "step": 1484 }, { "epoch": 3.63080684596577, "grad_norm": 0.20316204591819487, "learning_rate": 4.311341674129862e-06, "loss": 0.4637, "step": 1485 }, { "epoch": 3.6332518337408315, "grad_norm": 0.18107639605614553, "learning_rate": 4.310459081524606e-06, "loss": 0.8133, "step": 1486 }, { "epoch": 3.6356968215158925, "grad_norm": 0.19032454987562653, "learning_rate": 4.309576014175535e-06, "loss": 0.4337, "step": 1487 }, { "epoch": 3.6381418092909534, "grad_norm": 0.19454702114028524, "learning_rate": 4.308692472314209e-06, "loss": 0.467, "step": 1488 }, { "epoch": 3.640586797066015, "grad_norm": 0.13856707864158854, "learning_rate": 4.3078084561723134e-06, "loss": 0.4939, "step": 1489 }, { "epoch": 3.643031784841076, "grad_norm": 0.18387632838792337, "learning_rate": 4.306923965981657e-06, "loss": 0.3799, "step": 1490 }, { "epoch": 3.645476772616137, "grad_norm": 0.24807917153353443, "learning_rate": 4.3060390019741746e-06, "loss": 1.0492, "step": 1491 }, { "epoch": 3.647921760391198, "grad_norm": 0.18846077517423573, "learning_rate": 4.3051535643819225e-06, "loss": 0.381, "step": 1492 }, { "epoch": 3.650366748166259, "grad_norm": 0.25066795363480215, "learning_rate": 4.304267653437083e-06, "loss": 0.5585, "step": 1493 }, { "epoch": 3.65281173594132, "grad_norm": 0.22392544902309183, "learning_rate": 4.303381269371964e-06, "loss": 0.4853, "step": 1494 }, { "epoch": 3.6552567237163816, "grad_norm": 0.15902572689501818, "learning_rate": 4.302494412418994e-06, "loss": 0.7105, "step": 1495 }, { "epoch": 3.6577017114914425, "grad_norm": 0.23361247651338574, "learning_rate": 4.301607082810727e-06, "loss": 0.6046, "step": 1496 }, { "epoch": 3.6601466992665035, "grad_norm": 0.19385865405435335, "learning_rate": 4.300719280779842e-06, "loss": 1.009, "step": 1497 }, { "epoch": 3.662591687041565, "grad_norm": 0.19916281625821344, "learning_rate": 4.29983100655914e-06, "loss": 0.5657, "step": 1498 }, { "epoch": 3.665036674816626, "grad_norm": 0.207806164119596, "learning_rate": 4.298942260381548e-06, "loss": 0.6715, "step": 1499 }, { "epoch": 3.667481662591687, "grad_norm": 0.1942795111829436, "learning_rate": 4.2980530424801145e-06, "loss": 0.444, "step": 1500 }, { "epoch": 3.6699266503667483, "grad_norm": 0.1270088727281754, "learning_rate": 4.297163353088014e-06, "loss": 0.4151, "step": 1501 }, { "epoch": 3.6723716381418092, "grad_norm": 0.1700411507143267, "learning_rate": 4.296273192438543e-06, "loss": 0.5041, "step": 1502 }, { "epoch": 3.67481662591687, "grad_norm": 0.2903104934413432, "learning_rate": 4.295382560765121e-06, "loss": 0.4802, "step": 1503 }, { "epoch": 3.6772616136919316, "grad_norm": 0.18035000692924077, "learning_rate": 4.294491458301293e-06, "loss": 0.9768, "step": 1504 }, { "epoch": 3.6797066014669926, "grad_norm": 0.27120066828265305, "learning_rate": 4.293599885280727e-06, "loss": 0.7155, "step": 1505 }, { "epoch": 3.6821515892420535, "grad_norm": 0.15064437492867336, "learning_rate": 4.292707841937213e-06, "loss": 0.4737, "step": 1506 }, { "epoch": 3.684596577017115, "grad_norm": 0.31568396148302774, "learning_rate": 4.291815328504665e-06, "loss": 0.6019, "step": 1507 }, { "epoch": 3.687041564792176, "grad_norm": 0.20728697775484933, "learning_rate": 4.29092234521712e-06, "loss": 0.7537, "step": 1508 }, { "epoch": 3.689486552567237, "grad_norm": 0.2290461934726355, "learning_rate": 4.29002889230874e-06, "loss": 0.4736, "step": 1509 }, { "epoch": 3.6919315403422983, "grad_norm": 0.17725863292730443, "learning_rate": 4.289134970013809e-06, "loss": 0.4207, "step": 1510 }, { "epoch": 3.6943765281173593, "grad_norm": 0.21184428169776348, "learning_rate": 4.288240578566732e-06, "loss": 0.3837, "step": 1511 }, { "epoch": 3.6968215158924207, "grad_norm": 0.19281424747423478, "learning_rate": 4.287345718202041e-06, "loss": 0.7442, "step": 1512 }, { "epoch": 3.6992665036674817, "grad_norm": 0.20792668583419546, "learning_rate": 4.286450389154387e-06, "loss": 0.878, "step": 1513 }, { "epoch": 3.7017114914425426, "grad_norm": 0.16574980397941944, "learning_rate": 4.285554591658546e-06, "loss": 0.4383, "step": 1514 }, { "epoch": 3.704156479217604, "grad_norm": 0.16584324283169455, "learning_rate": 4.284658325949419e-06, "loss": 0.5184, "step": 1515 }, { "epoch": 3.706601466992665, "grad_norm": 0.2141895469207335, "learning_rate": 4.283761592262025e-06, "loss": 0.8406, "step": 1516 }, { "epoch": 3.7090464547677264, "grad_norm": 0.2120433571963209, "learning_rate": 4.282864390831508e-06, "loss": 0.6474, "step": 1517 }, { "epoch": 3.7114914425427874, "grad_norm": 0.20612972130386734, "learning_rate": 4.2819667218931375e-06, "loss": 0.5167, "step": 1518 }, { "epoch": 3.7139364303178484, "grad_norm": 0.23358016053152944, "learning_rate": 4.2810685856823e-06, "loss": 0.4651, "step": 1519 }, { "epoch": 3.71638141809291, "grad_norm": 0.26729679763270525, "learning_rate": 4.2801699824345074e-06, "loss": 1.0619, "step": 1520 }, { "epoch": 3.7188264058679708, "grad_norm": 0.18570088279451943, "learning_rate": 4.279270912385395e-06, "loss": 0.6004, "step": 1521 }, { "epoch": 3.7212713936430317, "grad_norm": 0.18522366998539547, "learning_rate": 4.27837137577072e-06, "loss": 1.0134, "step": 1522 }, { "epoch": 3.723716381418093, "grad_norm": 0.21074140330349975, "learning_rate": 4.27747137282636e-06, "loss": 0.507, "step": 1523 }, { "epoch": 3.726161369193154, "grad_norm": 0.26696686775835965, "learning_rate": 4.276570903788318e-06, "loss": 0.6401, "step": 1524 }, { "epoch": 3.728606356968215, "grad_norm": 0.16500629376413725, "learning_rate": 4.2756699688927176e-06, "loss": 0.6937, "step": 1525 }, { "epoch": 3.7310513447432765, "grad_norm": 0.13921129365137652, "learning_rate": 4.2747685683758036e-06, "loss": 0.7923, "step": 1526 }, { "epoch": 3.7334963325183375, "grad_norm": 0.1824163193267647, "learning_rate": 4.273866702473943e-06, "loss": 0.4964, "step": 1527 }, { "epoch": 3.7359413202933984, "grad_norm": 0.20847943266358018, "learning_rate": 4.272964371423628e-06, "loss": 0.6435, "step": 1528 }, { "epoch": 3.73838630806846, "grad_norm": 0.17606801285512352, "learning_rate": 4.272061575461469e-06, "loss": 0.9311, "step": 1529 }, { "epoch": 3.740831295843521, "grad_norm": 0.2299405760726726, "learning_rate": 4.2711583148241995e-06, "loss": 0.9356, "step": 1530 }, { "epoch": 3.743276283618582, "grad_norm": 0.16092293121770962, "learning_rate": 4.270254589748675e-06, "loss": 0.5201, "step": 1531 }, { "epoch": 3.745721271393643, "grad_norm": 0.18385487811520876, "learning_rate": 4.269350400471874e-06, "loss": 1.1429, "step": 1532 }, { "epoch": 3.748166259168704, "grad_norm": 0.2607026126241708, "learning_rate": 4.268445747230896e-06, "loss": 1.1623, "step": 1533 }, { "epoch": 3.750611246943765, "grad_norm": 0.14452905365906135, "learning_rate": 4.267540630262961e-06, "loss": 0.5079, "step": 1534 }, { "epoch": 3.7530562347188265, "grad_norm": 0.17244045236361516, "learning_rate": 4.26663504980541e-06, "loss": 0.3563, "step": 1535 }, { "epoch": 3.7555012224938875, "grad_norm": 0.18368827771342403, "learning_rate": 4.265729006095709e-06, "loss": 0.4127, "step": 1536 }, { "epoch": 3.7579462102689485, "grad_norm": 0.12539516835635212, "learning_rate": 4.264822499371442e-06, "loss": 0.4719, "step": 1537 }, { "epoch": 3.76039119804401, "grad_norm": 0.17665950638073558, "learning_rate": 4.263915529870319e-06, "loss": 0.4709, "step": 1538 }, { "epoch": 3.762836185819071, "grad_norm": 0.19173385021846664, "learning_rate": 4.263008097830162e-06, "loss": 0.4406, "step": 1539 }, { "epoch": 3.765281173594132, "grad_norm": 0.1658000759947098, "learning_rate": 4.262100203488927e-06, "loss": 0.7088, "step": 1540 }, { "epoch": 3.7677261613691932, "grad_norm": 0.22846945384107484, "learning_rate": 4.26119184708468e-06, "loss": 0.623, "step": 1541 }, { "epoch": 3.770171149144254, "grad_norm": 0.19206946579087833, "learning_rate": 4.2602830288556155e-06, "loss": 0.5145, "step": 1542 }, { "epoch": 3.772616136919315, "grad_norm": 0.24032548699199668, "learning_rate": 4.259373749040045e-06, "loss": 0.4808, "step": 1543 }, { "epoch": 3.7750611246943766, "grad_norm": 0.1404873040384191, "learning_rate": 4.2584640078764025e-06, "loss": 0.5355, "step": 1544 }, { "epoch": 3.7775061124694376, "grad_norm": 0.16166719602009264, "learning_rate": 4.257553805603245e-06, "loss": 0.6952, "step": 1545 }, { "epoch": 3.7799511002444985, "grad_norm": 0.22202483071601753, "learning_rate": 4.256643142459246e-06, "loss": 0.6393, "step": 1546 }, { "epoch": 3.78239608801956, "grad_norm": 0.20887696487231108, "learning_rate": 4.255732018683204e-06, "loss": 0.8471, "step": 1547 }, { "epoch": 3.784841075794621, "grad_norm": 0.15497906654704485, "learning_rate": 4.254820434514034e-06, "loss": 0.5881, "step": 1548 }, { "epoch": 3.787286063569682, "grad_norm": 0.12984721685974215, "learning_rate": 4.253908390190777e-06, "loss": 0.5935, "step": 1549 }, { "epoch": 3.7897310513447433, "grad_norm": 0.21225482954029468, "learning_rate": 4.25299588595259e-06, "loss": 0.5, "step": 1550 }, { "epoch": 3.7921760391198043, "grad_norm": 0.20013195653995045, "learning_rate": 4.2520829220387535e-06, "loss": 0.3674, "step": 1551 }, { "epoch": 3.7946210268948657, "grad_norm": 0.20720309807292175, "learning_rate": 4.251169498688668e-06, "loss": 0.3565, "step": 1552 }, { "epoch": 3.7970660146699267, "grad_norm": 0.14003994861678892, "learning_rate": 4.2502556161418515e-06, "loss": 0.5815, "step": 1553 }, { "epoch": 3.7995110024449876, "grad_norm": 0.22282677147792743, "learning_rate": 4.2493412746379485e-06, "loss": 0.4971, "step": 1554 }, { "epoch": 3.801955990220049, "grad_norm": 0.1399559531940187, "learning_rate": 4.248426474416718e-06, "loss": 0.5482, "step": 1555 }, { "epoch": 3.80440097799511, "grad_norm": 0.24546203825299318, "learning_rate": 4.247511215718042e-06, "loss": 0.734, "step": 1556 }, { "epoch": 3.8068459657701714, "grad_norm": 0.18488732525386353, "learning_rate": 4.246595498781922e-06, "loss": 0.8227, "step": 1557 }, { "epoch": 3.8092909535452324, "grad_norm": 0.1622969373666004, "learning_rate": 4.24567932384848e-06, "loss": 0.516, "step": 1558 }, { "epoch": 3.8117359413202934, "grad_norm": 0.23809265583221384, "learning_rate": 4.244762691157959e-06, "loss": 0.6903, "step": 1559 }, { "epoch": 3.8141809290953548, "grad_norm": 0.19879265661555484, "learning_rate": 4.24384560095072e-06, "loss": 0.8092, "step": 1560 }, { "epoch": 3.8166259168704157, "grad_norm": 0.20594885494352225, "learning_rate": 4.242928053467246e-06, "loss": 0.7784, "step": 1561 }, { "epoch": 3.8190709046454767, "grad_norm": 0.17180817524216707, "learning_rate": 4.242010048948138e-06, "loss": 0.4345, "step": 1562 }, { "epoch": 3.821515892420538, "grad_norm": 0.17764534261214263, "learning_rate": 4.241091587634118e-06, "loss": 0.4455, "step": 1563 }, { "epoch": 3.823960880195599, "grad_norm": 0.19811064694886915, "learning_rate": 4.240172669766027e-06, "loss": 0.8163, "step": 1564 }, { "epoch": 3.82640586797066, "grad_norm": 0.14952906767358026, "learning_rate": 4.239253295584828e-06, "loss": 0.5155, "step": 1565 }, { "epoch": 3.8288508557457215, "grad_norm": 0.19168341891658247, "learning_rate": 4.238333465331601e-06, "loss": 0.8454, "step": 1566 }, { "epoch": 3.8312958435207825, "grad_norm": 0.30231206111676673, "learning_rate": 4.237413179247546e-06, "loss": 0.4726, "step": 1567 }, { "epoch": 3.8337408312958434, "grad_norm": 0.13233568890046865, "learning_rate": 4.236492437573985e-06, "loss": 0.2259, "step": 1568 }, { "epoch": 3.836185819070905, "grad_norm": 0.17937110840805964, "learning_rate": 4.235571240552355e-06, "loss": 0.6933, "step": 1569 }, { "epoch": 3.838630806845966, "grad_norm": 0.16919566440693068, "learning_rate": 4.234649588424216e-06, "loss": 0.789, "step": 1570 }, { "epoch": 3.8410757946210268, "grad_norm": 0.15430861584473152, "learning_rate": 4.233727481431247e-06, "loss": 0.5344, "step": 1571 }, { "epoch": 3.843520782396088, "grad_norm": 0.22215341790533846, "learning_rate": 4.232804919815245e-06, "loss": 1.3468, "step": 1572 }, { "epoch": 3.845965770171149, "grad_norm": 0.1902720520035249, "learning_rate": 4.2318819038181264e-06, "loss": 0.4194, "step": 1573 }, { "epoch": 3.84841075794621, "grad_norm": 0.23471616108777327, "learning_rate": 4.230958433681928e-06, "loss": 0.4626, "step": 1574 }, { "epoch": 3.8508557457212715, "grad_norm": 0.1949645465012799, "learning_rate": 4.230034509648803e-06, "loss": 0.7755, "step": 1575 }, { "epoch": 3.8533007334963325, "grad_norm": 0.1422097616776362, "learning_rate": 4.229110131961029e-06, "loss": 0.5084, "step": 1576 }, { "epoch": 3.8557457212713935, "grad_norm": 0.23504625714161964, "learning_rate": 4.228185300860994e-06, "loss": 0.521, "step": 1577 }, { "epoch": 3.858190709046455, "grad_norm": 0.17536469979245362, "learning_rate": 4.2272600165912146e-06, "loss": 0.5521, "step": 1578 }, { "epoch": 3.860635696821516, "grad_norm": 0.2015376955427226, "learning_rate": 4.2263342793943184e-06, "loss": 0.471, "step": 1579 }, { "epoch": 3.863080684596577, "grad_norm": 0.1796933714389003, "learning_rate": 4.225408089513057e-06, "loss": 0.6722, "step": 1580 }, { "epoch": 3.8655256723716382, "grad_norm": 0.293466873610754, "learning_rate": 4.224481447190296e-06, "loss": 0.5212, "step": 1581 }, { "epoch": 3.867970660146699, "grad_norm": 0.19800107930789396, "learning_rate": 4.223554352669026e-06, "loss": 0.4922, "step": 1582 }, { "epoch": 3.87041564792176, "grad_norm": 0.25473286406229834, "learning_rate": 4.222626806192348e-06, "loss": 0.4795, "step": 1583 }, { "epoch": 3.8728606356968216, "grad_norm": 0.16630353602020626, "learning_rate": 4.2216988080034894e-06, "loss": 0.776, "step": 1584 }, { "epoch": 3.8753056234718826, "grad_norm": 0.2119233515040073, "learning_rate": 4.220770358345791e-06, "loss": 0.5058, "step": 1585 }, { "epoch": 3.8777506112469435, "grad_norm": 0.1903820401212916, "learning_rate": 4.2198414574627135e-06, "loss": 0.4093, "step": 1586 }, { "epoch": 3.880195599022005, "grad_norm": 0.2262167579208167, "learning_rate": 4.218912105597838e-06, "loss": 0.3843, "step": 1587 }, { "epoch": 3.882640586797066, "grad_norm": 0.15087880821374797, "learning_rate": 4.2179823029948585e-06, "loss": 0.3799, "step": 1588 }, { "epoch": 3.885085574572127, "grad_norm": 0.14002032366094444, "learning_rate": 4.217052049897592e-06, "loss": 0.7532, "step": 1589 }, { "epoch": 3.8875305623471883, "grad_norm": 0.1027091568738711, "learning_rate": 4.216121346549974e-06, "loss": 0.392, "step": 1590 }, { "epoch": 3.8899755501222493, "grad_norm": 0.16200987371775113, "learning_rate": 4.215190193196053e-06, "loss": 0.5896, "step": 1591 }, { "epoch": 3.8924205378973102, "grad_norm": 0.1284166160874503, "learning_rate": 4.21425859008e-06, "loss": 0.6224, "step": 1592 }, { "epoch": 3.8948655256723717, "grad_norm": 0.21395288701094411, "learning_rate": 4.213326537446104e-06, "loss": 0.7552, "step": 1593 }, { "epoch": 3.8973105134474326, "grad_norm": 0.12981866925141236, "learning_rate": 4.212394035538769e-06, "loss": 0.614, "step": 1594 }, { "epoch": 3.899755501222494, "grad_norm": 0.19813644440009445, "learning_rate": 4.211461084602519e-06, "loss": 0.5935, "step": 1595 }, { "epoch": 3.902200488997555, "grad_norm": 0.2708515972073224, "learning_rate": 4.210527684881996e-06, "loss": 0.9627, "step": 1596 }, { "epoch": 3.904645476772616, "grad_norm": 0.23673999506549334, "learning_rate": 4.209593836621957e-06, "loss": 1.0825, "step": 1597 }, { "epoch": 3.9070904645476774, "grad_norm": 0.16888179114165103, "learning_rate": 4.208659540067278e-06, "loss": 0.7053, "step": 1598 }, { "epoch": 3.9095354523227384, "grad_norm": 0.2120324456302641, "learning_rate": 4.207724795462955e-06, "loss": 0.5489, "step": 1599 }, { "epoch": 3.9119804400977998, "grad_norm": 0.3608041373714785, "learning_rate": 4.206789603054099e-06, "loss": 1.1897, "step": 1600 }, { "epoch": 3.9144254278728607, "grad_norm": 0.23182534157722756, "learning_rate": 4.205853963085937e-06, "loss": 0.5483, "step": 1601 }, { "epoch": 3.9168704156479217, "grad_norm": 0.23511133959491076, "learning_rate": 4.204917875803818e-06, "loss": 0.4389, "step": 1602 }, { "epoch": 3.919315403422983, "grad_norm": 0.15927780478854633, "learning_rate": 4.203981341453202e-06, "loss": 0.7396, "step": 1603 }, { "epoch": 3.921760391198044, "grad_norm": 0.22949192444221084, "learning_rate": 4.203044360279674e-06, "loss": 0.5043, "step": 1604 }, { "epoch": 3.924205378973105, "grad_norm": 0.20490558445871518, "learning_rate": 4.202106932528929e-06, "loss": 0.6175, "step": 1605 }, { "epoch": 3.9266503667481665, "grad_norm": 0.18046069191629607, "learning_rate": 4.201169058446783e-06, "loss": 0.399, "step": 1606 }, { "epoch": 3.9290953545232274, "grad_norm": 0.15374193138552703, "learning_rate": 4.200230738279167e-06, "loss": 0.4391, "step": 1607 }, { "epoch": 3.9315403422982884, "grad_norm": 0.20078915755772558, "learning_rate": 4.199291972272131e-06, "loss": 0.287, "step": 1608 }, { "epoch": 3.93398533007335, "grad_norm": 0.22147256550198563, "learning_rate": 4.198352760671841e-06, "loss": 0.492, "step": 1609 }, { "epoch": 3.936430317848411, "grad_norm": 0.16574030049361732, "learning_rate": 4.197413103724579e-06, "loss": 0.5324, "step": 1610 }, { "epoch": 3.9388753056234718, "grad_norm": 0.2103753507382826, "learning_rate": 4.1964730016767465e-06, "loss": 0.7365, "step": 1611 }, { "epoch": 3.941320293398533, "grad_norm": 0.1518267580262356, "learning_rate": 4.1955324547748565e-06, "loss": 0.8013, "step": 1612 }, { "epoch": 3.943765281173594, "grad_norm": 0.20154510576343865, "learning_rate": 4.1945914632655445e-06, "loss": 0.7363, "step": 1613 }, { "epoch": 3.946210268948655, "grad_norm": 0.15890745318421953, "learning_rate": 4.1936500273955604e-06, "loss": 0.3269, "step": 1614 }, { "epoch": 3.9486552567237165, "grad_norm": 0.16580897328637217, "learning_rate": 4.192708147411768e-06, "loss": 0.7251, "step": 1615 }, { "epoch": 3.9511002444987775, "grad_norm": 0.1795353181782988, "learning_rate": 4.191765823561151e-06, "loss": 0.4769, "step": 1616 }, { "epoch": 3.9535452322738385, "grad_norm": 0.16144794702069534, "learning_rate": 4.190823056090808e-06, "loss": 0.5483, "step": 1617 }, { "epoch": 3.9559902200489, "grad_norm": 0.16557851212954114, "learning_rate": 4.189879845247955e-06, "loss": 0.8062, "step": 1618 }, { "epoch": 3.958435207823961, "grad_norm": 0.2464569364178437, "learning_rate": 4.1889361912799236e-06, "loss": 0.4014, "step": 1619 }, { "epoch": 3.960880195599022, "grad_norm": 0.14265732039902237, "learning_rate": 4.18799209443416e-06, "loss": 0.4374, "step": 1620 }, { "epoch": 3.9633251833740832, "grad_norm": 0.22339158613662402, "learning_rate": 4.187047554958229e-06, "loss": 0.5925, "step": 1621 }, { "epoch": 3.965770171149144, "grad_norm": 0.16802303875551336, "learning_rate": 4.18610257309981e-06, "loss": 0.6077, "step": 1622 }, { "epoch": 3.968215158924205, "grad_norm": 0.19415483104611594, "learning_rate": 4.185157149106699e-06, "loss": 0.6042, "step": 1623 }, { "epoch": 3.9706601466992666, "grad_norm": 0.19962803492883538, "learning_rate": 4.184211283226808e-06, "loss": 0.5957, "step": 1624 }, { "epoch": 3.9731051344743276, "grad_norm": 0.18497252565632344, "learning_rate": 4.183264975708165e-06, "loss": 0.5733, "step": 1625 }, { "epoch": 3.9755501222493885, "grad_norm": 0.14586139866041184, "learning_rate": 4.1823182267989134e-06, "loss": 0.7121, "step": 1626 }, { "epoch": 3.97799511002445, "grad_norm": 0.22799748060588776, "learning_rate": 4.181371036747312e-06, "loss": 0.4812, "step": 1627 }, { "epoch": 3.980440097799511, "grad_norm": 0.1820323704700886, "learning_rate": 4.180423405801734e-06, "loss": 0.516, "step": 1628 }, { "epoch": 3.982885085574572, "grad_norm": 0.20330517608379037, "learning_rate": 4.179475334210674e-06, "loss": 0.3732, "step": 1629 }, { "epoch": 3.9853300733496333, "grad_norm": 0.15813036453233392, "learning_rate": 4.178526822222736e-06, "loss": 0.6198, "step": 1630 }, { "epoch": 3.9877750611246943, "grad_norm": 0.15778655137904618, "learning_rate": 4.177577870086641e-06, "loss": 0.8369, "step": 1631 }, { "epoch": 3.9902200488997552, "grad_norm": 0.3010724095591919, "learning_rate": 4.176628478051225e-06, "loss": 0.4517, "step": 1632 }, { "epoch": 3.9926650366748166, "grad_norm": 0.20983829947240318, "learning_rate": 4.175678646365443e-06, "loss": 0.5051, "step": 1633 }, { "epoch": 3.9951100244498776, "grad_norm": 0.2001018715260085, "learning_rate": 4.1747283752783614e-06, "loss": 0.5244, "step": 1634 }, { "epoch": 3.997555012224939, "grad_norm": 0.21021590865627596, "learning_rate": 4.173777665039163e-06, "loss": 0.6324, "step": 1635 }, { "epoch": 4.0, "grad_norm": 0.3458555843589508, "learning_rate": 4.172826515897146e-06, "loss": 0.736, "step": 1636 }, { "epoch": 4.002444987775061, "grad_norm": 0.1518077119406904, "learning_rate": 4.171874928101723e-06, "loss": 0.792, "step": 1637 }, { "epoch": 4.004889975550122, "grad_norm": 0.1421089320209172, "learning_rate": 4.170922901902422e-06, "loss": 0.5526, "step": 1638 }, { "epoch": 4.007334963325183, "grad_norm": 0.1388042758702329, "learning_rate": 4.169970437548887e-06, "loss": 0.6591, "step": 1639 }, { "epoch": 4.009779951100245, "grad_norm": 0.20370445584125405, "learning_rate": 4.1690175352908745e-06, "loss": 0.4398, "step": 1640 }, { "epoch": 4.012224938875305, "grad_norm": 0.1708535653334295, "learning_rate": 4.1680641953782584e-06, "loss": 0.4895, "step": 1641 }, { "epoch": 4.014669926650367, "grad_norm": 0.18480442299047597, "learning_rate": 4.167110418061027e-06, "loss": 0.6549, "step": 1642 }, { "epoch": 4.017114914425428, "grad_norm": 0.24273002542773492, "learning_rate": 4.1661562035892805e-06, "loss": 0.4971, "step": 1643 }, { "epoch": 4.019559902200489, "grad_norm": 0.17093116699432037, "learning_rate": 4.165201552213237e-06, "loss": 0.5171, "step": 1644 }, { "epoch": 4.02200488997555, "grad_norm": 0.17132151584278324, "learning_rate": 4.1642464641832275e-06, "loss": 0.4778, "step": 1645 }, { "epoch": 4.0244498777506115, "grad_norm": 0.1905381437255112, "learning_rate": 4.163290939749699e-06, "loss": 0.8282, "step": 1646 }, { "epoch": 4.026894865525672, "grad_norm": 0.18941523453512035, "learning_rate": 4.16233497916321e-06, "loss": 0.5959, "step": 1647 }, { "epoch": 4.029339853300733, "grad_norm": 0.15737630515843087, "learning_rate": 4.161378582674438e-06, "loss": 0.4567, "step": 1648 }, { "epoch": 4.031784841075795, "grad_norm": 0.22320279396684584, "learning_rate": 4.160421750534169e-06, "loss": 0.46, "step": 1649 }, { "epoch": 4.034229828850855, "grad_norm": 0.15903555038672185, "learning_rate": 4.159464482993308e-06, "loss": 0.4813, "step": 1650 }, { "epoch": 4.036674816625917, "grad_norm": 0.1594686881803847, "learning_rate": 4.158506780302872e-06, "loss": 0.7616, "step": 1651 }, { "epoch": 4.039119804400978, "grad_norm": 0.25305541297434053, "learning_rate": 4.157548642713993e-06, "loss": 1.1283, "step": 1652 }, { "epoch": 4.041564792176039, "grad_norm": 0.2789869036025608, "learning_rate": 4.156590070477916e-06, "loss": 0.4036, "step": 1653 }, { "epoch": 4.0440097799511, "grad_norm": 0.14314037616237907, "learning_rate": 4.1556310638460015e-06, "loss": 0.4155, "step": 1654 }, { "epoch": 4.0464547677261615, "grad_norm": 0.20900868903680211, "learning_rate": 4.154671623069723e-06, "loss": 0.849, "step": 1655 }, { "epoch": 4.048899755501223, "grad_norm": 0.19439041231223247, "learning_rate": 4.153711748400666e-06, "loss": 0.7261, "step": 1656 }, { "epoch": 4.0513447432762835, "grad_norm": 0.17187170420442915, "learning_rate": 4.152751440090533e-06, "loss": 0.6516, "step": 1657 }, { "epoch": 4.053789731051345, "grad_norm": 0.17641294394021761, "learning_rate": 4.151790698391139e-06, "loss": 0.8281, "step": 1658 }, { "epoch": 4.056234718826406, "grad_norm": 0.14973407190393406, "learning_rate": 4.150829523554412e-06, "loss": 0.4672, "step": 1659 }, { "epoch": 4.058679706601467, "grad_norm": 0.17623868288621014, "learning_rate": 4.149867915832394e-06, "loss": 0.6132, "step": 1660 }, { "epoch": 4.061124694376528, "grad_norm": 0.18759592893698873, "learning_rate": 4.148905875477241e-06, "loss": 0.5966, "step": 1661 }, { "epoch": 4.06356968215159, "grad_norm": 0.213524041968869, "learning_rate": 4.147943402741221e-06, "loss": 0.4755, "step": 1662 }, { "epoch": 4.06601466992665, "grad_norm": 0.13691961930828037, "learning_rate": 4.146980497876718e-06, "loss": 0.654, "step": 1663 }, { "epoch": 4.068459657701712, "grad_norm": 0.1382214264948139, "learning_rate": 4.146017161136226e-06, "loss": 0.4311, "step": 1664 }, { "epoch": 4.070904645476773, "grad_norm": 0.1579008065563269, "learning_rate": 4.145053392772357e-06, "loss": 0.5373, "step": 1665 }, { "epoch": 4.0733496332518335, "grad_norm": 0.2314255331933333, "learning_rate": 4.144089193037828e-06, "loss": 1.5379, "step": 1666 }, { "epoch": 4.075794621026895, "grad_norm": 0.14734056300997364, "learning_rate": 4.143124562185479e-06, "loss": 0.5351, "step": 1667 }, { "epoch": 4.078239608801956, "grad_norm": 0.16489353002071347, "learning_rate": 4.142159500468256e-06, "loss": 0.5031, "step": 1668 }, { "epoch": 4.080684596577017, "grad_norm": 0.18067862871934579, "learning_rate": 4.14119400813922e-06, "loss": 0.7994, "step": 1669 }, { "epoch": 4.083129584352078, "grad_norm": 0.2007292764285633, "learning_rate": 4.140228085451547e-06, "loss": 0.4887, "step": 1670 }, { "epoch": 4.08557457212714, "grad_norm": 0.13194276336074412, "learning_rate": 4.1392617326585215e-06, "loss": 0.3955, "step": 1671 }, { "epoch": 4.0880195599022, "grad_norm": 0.17421264694468763, "learning_rate": 4.138294950013546e-06, "loss": 0.3632, "step": 1672 }, { "epoch": 4.090464547677262, "grad_norm": 0.15223383470903787, "learning_rate": 4.137327737770131e-06, "loss": 0.3526, "step": 1673 }, { "epoch": 4.092909535452323, "grad_norm": 0.15296515038485883, "learning_rate": 4.136360096181904e-06, "loss": 0.509, "step": 1674 }, { "epoch": 4.095354523227384, "grad_norm": 0.17636666370307155, "learning_rate": 4.1353920255026e-06, "loss": 0.8884, "step": 1675 }, { "epoch": 4.097799511002445, "grad_norm": 0.21033548482512168, "learning_rate": 4.13442352598607e-06, "loss": 0.8026, "step": 1676 }, { "epoch": 4.100244498777506, "grad_norm": 0.1768482151374831, "learning_rate": 4.133454597886278e-06, "loss": 0.4585, "step": 1677 }, { "epoch": 4.102689486552567, "grad_norm": 0.18983964173445708, "learning_rate": 4.132485241457299e-06, "loss": 0.721, "step": 1678 }, { "epoch": 4.105134474327628, "grad_norm": 0.1591509256280448, "learning_rate": 4.131515456953319e-06, "loss": 0.5851, "step": 1679 }, { "epoch": 4.10757946210269, "grad_norm": 0.12720504523789747, "learning_rate": 4.130545244628639e-06, "loss": 0.4402, "step": 1680 }, { "epoch": 4.11002444987775, "grad_norm": 0.3026554285814813, "learning_rate": 4.129574604737671e-06, "loss": 0.7289, "step": 1681 }, { "epoch": 4.112469437652812, "grad_norm": 0.16282817183258416, "learning_rate": 4.128603537534936e-06, "loss": 0.6524, "step": 1682 }, { "epoch": 4.114914425427873, "grad_norm": 0.18469978114627153, "learning_rate": 4.127632043275075e-06, "loss": 0.3755, "step": 1683 }, { "epoch": 4.117359413202934, "grad_norm": 0.24026038701837185, "learning_rate": 4.126660122212833e-06, "loss": 0.6782, "step": 1684 }, { "epoch": 4.119804400977995, "grad_norm": 0.15762160010421056, "learning_rate": 4.125687774603071e-06, "loss": 0.3026, "step": 1685 }, { "epoch": 4.1222493887530565, "grad_norm": 0.22776534075144217, "learning_rate": 4.124715000700759e-06, "loss": 0.7973, "step": 1686 }, { "epoch": 4.124694376528117, "grad_norm": 0.17985068128101256, "learning_rate": 4.123741800760982e-06, "loss": 0.3162, "step": 1687 }, { "epoch": 4.127139364303178, "grad_norm": 0.18723261069880706, "learning_rate": 4.1227681750389356e-06, "loss": 1.0402, "step": 1688 }, { "epoch": 4.12958435207824, "grad_norm": 0.15943760525014972, "learning_rate": 4.121794123789926e-06, "loss": 0.681, "step": 1689 }, { "epoch": 4.1320293398533, "grad_norm": 0.14188143452333535, "learning_rate": 4.120819647269373e-06, "loss": 0.4526, "step": 1690 }, { "epoch": 4.134474327628362, "grad_norm": 0.22423321228533125, "learning_rate": 4.119844745732805e-06, "loss": 0.9314, "step": 1691 }, { "epoch": 4.136919315403423, "grad_norm": 0.1983247750301809, "learning_rate": 4.118869419435864e-06, "loss": 0.4461, "step": 1692 }, { "epoch": 4.139364303178484, "grad_norm": 0.13899832920549743, "learning_rate": 4.1178936686343035e-06, "loss": 0.4981, "step": 1693 }, { "epoch": 4.141809290953545, "grad_norm": 0.2773548838963686, "learning_rate": 4.116917493583987e-06, "loss": 0.369, "step": 1694 }, { "epoch": 4.1442542787286065, "grad_norm": 0.16370477728609112, "learning_rate": 4.115940894540891e-06, "loss": 0.675, "step": 1695 }, { "epoch": 4.146699266503667, "grad_norm": 0.19750941908226533, "learning_rate": 4.114963871761101e-06, "loss": 1.0292, "step": 1696 }, { "epoch": 4.1491442542787285, "grad_norm": 0.17229118212678296, "learning_rate": 4.113986425500814e-06, "loss": 0.5258, "step": 1697 }, { "epoch": 4.15158924205379, "grad_norm": 0.2563897044610342, "learning_rate": 4.113008556016341e-06, "loss": 1.0661, "step": 1698 }, { "epoch": 4.15403422982885, "grad_norm": 0.2079464054693061, "learning_rate": 4.1120302635641e-06, "loss": 0.6705, "step": 1699 }, { "epoch": 4.156479217603912, "grad_norm": 0.21156203157025513, "learning_rate": 4.111051548400623e-06, "loss": 0.7136, "step": 1700 }, { "epoch": 4.158924205378973, "grad_norm": 0.24742298365123258, "learning_rate": 4.11007241078255e-06, "loss": 0.8015, "step": 1701 }, { "epoch": 4.161369193154035, "grad_norm": 0.2352102139484552, "learning_rate": 4.109092850966635e-06, "loss": 0.3871, "step": 1702 }, { "epoch": 4.163814180929095, "grad_norm": 0.32180295269890397, "learning_rate": 4.10811286920974e-06, "loss": 0.6764, "step": 1703 }, { "epoch": 4.166259168704157, "grad_norm": 0.2982592472054046, "learning_rate": 4.107132465768839e-06, "loss": 0.5148, "step": 1704 }, { "epoch": 4.168704156479218, "grad_norm": 0.17478824715792302, "learning_rate": 4.106151640901015e-06, "loss": 0.528, "step": 1705 }, { "epoch": 4.1711491442542785, "grad_norm": 0.16932053451028353, "learning_rate": 4.105170394863465e-06, "loss": 0.6537, "step": 1706 }, { "epoch": 4.17359413202934, "grad_norm": 0.2728418361198744, "learning_rate": 4.104188727913492e-06, "loss": 0.9982, "step": 1707 }, { "epoch": 4.176039119804401, "grad_norm": 0.2484005884172726, "learning_rate": 4.103206640308513e-06, "loss": 0.4627, "step": 1708 }, { "epoch": 4.178484107579462, "grad_norm": 0.1624321113208904, "learning_rate": 4.102224132306053e-06, "loss": 0.4261, "step": 1709 }, { "epoch": 4.180929095354523, "grad_norm": 0.18311486500358004, "learning_rate": 4.101241204163748e-06, "loss": 0.7649, "step": 1710 }, { "epoch": 4.183374083129585, "grad_norm": 0.1878438592262303, "learning_rate": 4.100257856139345e-06, "loss": 0.5934, "step": 1711 }, { "epoch": 4.185819070904645, "grad_norm": 0.134180268795672, "learning_rate": 4.0992740884907005e-06, "loss": 0.3716, "step": 1712 }, { "epoch": 4.188264058679707, "grad_norm": 0.19271192195398162, "learning_rate": 4.098289901475781e-06, "loss": 1.0834, "step": 1713 }, { "epoch": 4.190709046454768, "grad_norm": 0.1801323161805221, "learning_rate": 4.0973052953526604e-06, "loss": 0.8863, "step": 1714 }, { "epoch": 4.193154034229829, "grad_norm": 0.19938665333273717, "learning_rate": 4.096320270379528e-06, "loss": 0.8924, "step": 1715 }, { "epoch": 4.19559902200489, "grad_norm": 0.18916681893391316, "learning_rate": 4.095334826814678e-06, "loss": 0.8121, "step": 1716 }, { "epoch": 4.198044009779951, "grad_norm": 0.1939948039664555, "learning_rate": 4.094348964916518e-06, "loss": 0.5514, "step": 1717 }, { "epoch": 4.200488997555012, "grad_norm": 0.24374525762095, "learning_rate": 4.093362684943563e-06, "loss": 0.5261, "step": 1718 }, { "epoch": 4.202933985330073, "grad_norm": 0.10104504785597727, "learning_rate": 4.092375987154436e-06, "loss": 0.2721, "step": 1719 }, { "epoch": 4.205378973105135, "grad_norm": 0.2090159826655009, "learning_rate": 4.091388871807875e-06, "loss": 0.5686, "step": 1720 }, { "epoch": 4.207823960880195, "grad_norm": 0.24106039689704994, "learning_rate": 4.0904013391627215e-06, "loss": 0.6725, "step": 1721 }, { "epoch": 4.210268948655257, "grad_norm": 0.1711164393090134, "learning_rate": 4.089413389477932e-06, "loss": 0.6855, "step": 1722 }, { "epoch": 4.212713936430318, "grad_norm": 0.20650419400661307, "learning_rate": 4.088425023012567e-06, "loss": 0.3486, "step": 1723 }, { "epoch": 4.215158924205379, "grad_norm": 0.19761023162015257, "learning_rate": 4.087436240025799e-06, "loss": 0.5559, "step": 1724 }, { "epoch": 4.21760391198044, "grad_norm": 0.19256799855029696, "learning_rate": 4.086447040776911e-06, "loss": 0.518, "step": 1725 }, { "epoch": 4.2200488997555015, "grad_norm": 0.24045062378784746, "learning_rate": 4.0854574255252936e-06, "loss": 0.5368, "step": 1726 }, { "epoch": 4.222493887530562, "grad_norm": 0.19615511642049846, "learning_rate": 4.084467394530446e-06, "loss": 0.6047, "step": 1727 }, { "epoch": 4.224938875305623, "grad_norm": 0.19862164662345816, "learning_rate": 4.083476948051975e-06, "loss": 0.4562, "step": 1728 }, { "epoch": 4.227383863080685, "grad_norm": 0.24290472576999142, "learning_rate": 4.082486086349603e-06, "loss": 0.4321, "step": 1729 }, { "epoch": 4.229828850855745, "grad_norm": 0.24874969368093675, "learning_rate": 4.0814948096831515e-06, "loss": 0.9137, "step": 1730 }, { "epoch": 4.232273838630807, "grad_norm": 0.2004430668876246, "learning_rate": 4.080503118312559e-06, "loss": 0.4105, "step": 1731 }, { "epoch": 4.234718826405868, "grad_norm": 0.30042391552745845, "learning_rate": 4.07951101249787e-06, "loss": 0.4581, "step": 1732 }, { "epoch": 4.237163814180929, "grad_norm": 0.15593661106081655, "learning_rate": 4.078518492499235e-06, "loss": 0.5084, "step": 1733 }, { "epoch": 4.23960880195599, "grad_norm": 0.13460768923376595, "learning_rate": 4.077525558576918e-06, "loss": 0.3965, "step": 1734 }, { "epoch": 4.2420537897310515, "grad_norm": 0.2806401285595934, "learning_rate": 4.076532210991287e-06, "loss": 0.4495, "step": 1735 }, { "epoch": 4.244498777506112, "grad_norm": 0.17670674889130006, "learning_rate": 4.075538450002822e-06, "loss": 0.5333, "step": 1736 }, { "epoch": 4.2469437652811735, "grad_norm": 0.1669677734742625, "learning_rate": 4.074544275872109e-06, "loss": 1.1966, "step": 1737 }, { "epoch": 4.249388753056235, "grad_norm": 0.1980139139445526, "learning_rate": 4.0735496888598425e-06, "loss": 0.9163, "step": 1738 }, { "epoch": 4.251833740831296, "grad_norm": 0.24699550494926706, "learning_rate": 4.072554689226828e-06, "loss": 0.5292, "step": 1739 }, { "epoch": 4.254278728606357, "grad_norm": 0.12767823783425794, "learning_rate": 4.071559277233975e-06, "loss": 0.4056, "step": 1740 }, { "epoch": 4.256723716381418, "grad_norm": 0.18410892576079202, "learning_rate": 4.070563453142305e-06, "loss": 0.5237, "step": 1741 }, { "epoch": 4.25916870415648, "grad_norm": 0.1264051982372328, "learning_rate": 4.069567217212945e-06, "loss": 0.3116, "step": 1742 }, { "epoch": 4.26161369193154, "grad_norm": 0.18251878071261948, "learning_rate": 4.068570569707132e-06, "loss": 0.5484, "step": 1743 }, { "epoch": 4.264058679706602, "grad_norm": 0.13456310387728304, "learning_rate": 4.067573510886208e-06, "loss": 0.379, "step": 1744 }, { "epoch": 4.266503667481663, "grad_norm": 0.17153722104478095, "learning_rate": 4.066576041011626e-06, "loss": 0.562, "step": 1745 }, { "epoch": 4.2689486552567235, "grad_norm": 0.16674707226468977, "learning_rate": 4.065578160344944e-06, "loss": 0.4714, "step": 1746 }, { "epoch": 4.271393643031785, "grad_norm": 0.18346192347526105, "learning_rate": 4.064579869147831e-06, "loss": 0.5058, "step": 1747 }, { "epoch": 4.273838630806846, "grad_norm": 0.15110590133481344, "learning_rate": 4.063581167682062e-06, "loss": 0.4163, "step": 1748 }, { "epoch": 4.276283618581907, "grad_norm": 0.18313462366880673, "learning_rate": 4.062582056209516e-06, "loss": 0.5325, "step": 1749 }, { "epoch": 4.278728606356968, "grad_norm": 0.2365198025369478, "learning_rate": 4.061582534992187e-06, "loss": 0.5257, "step": 1750 }, { "epoch": 4.28117359413203, "grad_norm": 0.1604786642045765, "learning_rate": 4.060582604292169e-06, "loss": 0.8026, "step": 1751 }, { "epoch": 4.28361858190709, "grad_norm": 0.1863887245925862, "learning_rate": 4.05958226437167e-06, "loss": 0.4638, "step": 1752 }, { "epoch": 4.286063569682152, "grad_norm": 0.23044308543333392, "learning_rate": 4.058581515492999e-06, "loss": 0.4691, "step": 1753 }, { "epoch": 4.288508557457213, "grad_norm": 0.1892552082385137, "learning_rate": 4.057580357918577e-06, "loss": 0.4189, "step": 1754 }, { "epoch": 4.290953545232274, "grad_norm": 0.21843822922905395, "learning_rate": 4.056578791910931e-06, "loss": 0.6979, "step": 1755 }, { "epoch": 4.293398533007335, "grad_norm": 0.20821046782495747, "learning_rate": 4.055576817732693e-06, "loss": 0.4588, "step": 1756 }, { "epoch": 4.295843520782396, "grad_norm": 0.1719093538131664, "learning_rate": 4.054574435646604e-06, "loss": 0.4315, "step": 1757 }, { "epoch": 4.298288508557457, "grad_norm": 0.3132694618291291, "learning_rate": 4.0535716459155105e-06, "loss": 0.5837, "step": 1758 }, { "epoch": 4.300733496332518, "grad_norm": 0.20402567179431644, "learning_rate": 4.052568448802369e-06, "loss": 0.9718, "step": 1759 }, { "epoch": 4.30317848410758, "grad_norm": 0.14336884048577295, "learning_rate": 4.051564844570239e-06, "loss": 0.4744, "step": 1760 }, { "epoch": 4.30562347188264, "grad_norm": 0.22774705291664313, "learning_rate": 4.050560833482291e-06, "loss": 0.4027, "step": 1761 }, { "epoch": 4.308068459657702, "grad_norm": 0.6390468195830807, "learning_rate": 4.0495564158017964e-06, "loss": 0.8346, "step": 1762 }, { "epoch": 4.310513447432763, "grad_norm": 0.2645342446436882, "learning_rate": 4.048551591792139e-06, "loss": 0.6561, "step": 1763 }, { "epoch": 4.312958435207824, "grad_norm": 0.21334544502180697, "learning_rate": 4.047546361716806e-06, "loss": 0.3755, "step": 1764 }, { "epoch": 4.315403422982885, "grad_norm": 0.1988327070302909, "learning_rate": 4.0465407258393905e-06, "loss": 0.4531, "step": 1765 }, { "epoch": 4.3178484107579465, "grad_norm": 0.20237744161343457, "learning_rate": 4.045534684423596e-06, "loss": 0.6549, "step": 1766 }, { "epoch": 4.320293398533007, "grad_norm": 0.24120107619189118, "learning_rate": 4.044528237733226e-06, "loss": 0.5196, "step": 1767 }, { "epoch": 4.322738386308068, "grad_norm": 0.17158185891352729, "learning_rate": 4.043521386032196e-06, "loss": 0.4657, "step": 1768 }, { "epoch": 4.32518337408313, "grad_norm": 0.22935870136502498, "learning_rate": 4.042514129584526e-06, "loss": 0.5189, "step": 1769 }, { "epoch": 4.32762836185819, "grad_norm": 0.14087522812290942, "learning_rate": 4.04150646865434e-06, "loss": 0.6564, "step": 1770 }, { "epoch": 4.330073349633252, "grad_norm": 0.17686764150528303, "learning_rate": 4.040498403505872e-06, "loss": 0.6632, "step": 1771 }, { "epoch": 4.332518337408313, "grad_norm": 0.19924420889254563, "learning_rate": 4.039489934403458e-06, "loss": 0.725, "step": 1772 }, { "epoch": 4.334963325183374, "grad_norm": 0.2177841923449248, "learning_rate": 4.038481061611542e-06, "loss": 0.5316, "step": 1773 }, { "epoch": 4.337408312958435, "grad_norm": 0.201290162040878, "learning_rate": 4.037471785394673e-06, "loss": 1.0606, "step": 1774 }, { "epoch": 4.3398533007334965, "grad_norm": 0.18398617109739293, "learning_rate": 4.036462106017508e-06, "loss": 0.383, "step": 1775 }, { "epoch": 4.342298288508557, "grad_norm": 0.3274435806789136, "learning_rate": 4.035452023744807e-06, "loss": 0.9509, "step": 1776 }, { "epoch": 4.3447432762836184, "grad_norm": 0.22585707793721407, "learning_rate": 4.034441538841437e-06, "loss": 0.4751, "step": 1777 }, { "epoch": 4.34718826405868, "grad_norm": 0.17996294049670775, "learning_rate": 4.033430651572369e-06, "loss": 0.6846, "step": 1778 }, { "epoch": 4.34963325183374, "grad_norm": 0.1579337410408365, "learning_rate": 4.032419362202683e-06, "loss": 0.3962, "step": 1779 }, { "epoch": 4.352078239608802, "grad_norm": 0.18352501604993987, "learning_rate": 4.03140767099756e-06, "loss": 0.7938, "step": 1780 }, { "epoch": 4.354523227383863, "grad_norm": 0.3073501777722171, "learning_rate": 4.0303955782222905e-06, "loss": 0.4474, "step": 1781 }, { "epoch": 4.356968215158924, "grad_norm": 0.20519229629056135, "learning_rate": 4.029383084142268e-06, "loss": 0.4718, "step": 1782 }, { "epoch": 4.359413202933985, "grad_norm": 0.17426326249792826, "learning_rate": 4.02837018902299e-06, "loss": 0.4655, "step": 1783 }, { "epoch": 4.361858190709047, "grad_norm": 0.20204797164590707, "learning_rate": 4.027356893130062e-06, "loss": 0.451, "step": 1784 }, { "epoch": 4.364303178484108, "grad_norm": 0.24220725507855714, "learning_rate": 4.026343196729194e-06, "loss": 0.5278, "step": 1785 }, { "epoch": 4.3667481662591685, "grad_norm": 0.16983783158252388, "learning_rate": 4.0253291000861995e-06, "loss": 0.6553, "step": 1786 }, { "epoch": 4.36919315403423, "grad_norm": 0.2216676257208036, "learning_rate": 4.0243146034669975e-06, "loss": 0.9118, "step": 1787 }, { "epoch": 4.371638141809291, "grad_norm": 0.19887613832894624, "learning_rate": 4.023299707137613e-06, "loss": 0.4599, "step": 1788 }, { "epoch": 4.374083129584352, "grad_norm": 0.17733841855995897, "learning_rate": 4.022284411364176e-06, "loss": 0.4989, "step": 1789 }, { "epoch": 4.376528117359413, "grad_norm": 0.2700709655444306, "learning_rate": 4.021268716412918e-06, "loss": 0.4295, "step": 1790 }, { "epoch": 4.378973105134475, "grad_norm": 0.18641684424913657, "learning_rate": 4.020252622550179e-06, "loss": 0.8216, "step": 1791 }, { "epoch": 4.381418092909535, "grad_norm": 0.25579577790537766, "learning_rate": 4.0192361300424024e-06, "loss": 0.5524, "step": 1792 }, { "epoch": 4.383863080684597, "grad_norm": 0.1704170266308227, "learning_rate": 4.018219239156134e-06, "loss": 0.4501, "step": 1793 }, { "epoch": 4.386308068459658, "grad_norm": 0.30986540950700525, "learning_rate": 4.0172019501580275e-06, "loss": 0.5034, "step": 1794 }, { "epoch": 4.388753056234719, "grad_norm": 0.19879134398740553, "learning_rate": 4.016184263314839e-06, "loss": 0.4996, "step": 1795 }, { "epoch": 4.39119804400978, "grad_norm": 0.15689007778639283, "learning_rate": 4.015166178893428e-06, "loss": 0.5767, "step": 1796 }, { "epoch": 4.393643031784841, "grad_norm": 0.22665268849645154, "learning_rate": 4.0141476971607626e-06, "loss": 0.5678, "step": 1797 }, { "epoch": 4.396088019559902, "grad_norm": 0.1612728689316623, "learning_rate": 4.013128818383909e-06, "loss": 0.5074, "step": 1798 }, { "epoch": 4.398533007334963, "grad_norm": 0.19454716973470845, "learning_rate": 4.012109542830042e-06, "loss": 0.4193, "step": 1799 }, { "epoch": 4.400977995110025, "grad_norm": 0.2900809894034894, "learning_rate": 4.011089870766437e-06, "loss": 0.7492, "step": 1800 }, { "epoch": 4.403422982885085, "grad_norm": 0.17017235076254456, "learning_rate": 4.010069802460479e-06, "loss": 0.4003, "step": 1801 }, { "epoch": 4.405867970660147, "grad_norm": 0.11863350690988382, "learning_rate": 4.00904933817965e-06, "loss": 0.4585, "step": 1802 }, { "epoch": 4.408312958435208, "grad_norm": 0.201116212157392, "learning_rate": 4.008028478191541e-06, "loss": 0.3439, "step": 1803 }, { "epoch": 4.410757946210269, "grad_norm": 0.18935881941306143, "learning_rate": 4.007007222763844e-06, "loss": 0.5309, "step": 1804 }, { "epoch": 4.41320293398533, "grad_norm": 0.21402870084554498, "learning_rate": 4.005985572164356e-06, "loss": 0.9903, "step": 1805 }, { "epoch": 4.4156479217603914, "grad_norm": 0.2234688076653641, "learning_rate": 4.004963526660977e-06, "loss": 0.3975, "step": 1806 }, { "epoch": 4.418092909535452, "grad_norm": 0.20426539988315862, "learning_rate": 4.00394108652171e-06, "loss": 0.6047, "step": 1807 }, { "epoch": 4.420537897310513, "grad_norm": 0.20070267367364727, "learning_rate": 4.002918252014664e-06, "loss": 0.5787, "step": 1808 }, { "epoch": 4.422982885085575, "grad_norm": 0.2610563973291742, "learning_rate": 4.001895023408047e-06, "loss": 1.1268, "step": 1809 }, { "epoch": 4.425427872860635, "grad_norm": 0.20834543458077015, "learning_rate": 4.000871400970175e-06, "loss": 0.5251, "step": 1810 }, { "epoch": 4.427872860635697, "grad_norm": 0.27337758605500256, "learning_rate": 3.999847384969465e-06, "loss": 0.7662, "step": 1811 }, { "epoch": 4.430317848410758, "grad_norm": 0.2397016683189441, "learning_rate": 3.998822975674437e-06, "loss": 0.6643, "step": 1812 }, { "epoch": 4.432762836185819, "grad_norm": 0.1684599967440021, "learning_rate": 3.997798173353714e-06, "loss": 0.7524, "step": 1813 }, { "epoch": 4.43520782396088, "grad_norm": 0.1904418463849341, "learning_rate": 3.996772978276025e-06, "loss": 0.4726, "step": 1814 }, { "epoch": 4.4376528117359415, "grad_norm": 0.22925054435124603, "learning_rate": 3.995747390710196e-06, "loss": 0.5277, "step": 1815 }, { "epoch": 4.440097799511002, "grad_norm": 0.16079024094680633, "learning_rate": 3.994721410925162e-06, "loss": 0.6566, "step": 1816 }, { "epoch": 4.442542787286063, "grad_norm": 0.2463820434037348, "learning_rate": 3.993695039189957e-06, "loss": 0.7487, "step": 1817 }, { "epoch": 4.444987775061125, "grad_norm": 0.22008504254238753, "learning_rate": 3.992668275773722e-06, "loss": 0.6086, "step": 1818 }, { "epoch": 4.447432762836186, "grad_norm": 0.1828612602387857, "learning_rate": 3.9916411209456955e-06, "loss": 0.5678, "step": 1819 }, { "epoch": 4.449877750611247, "grad_norm": 0.1686064422524913, "learning_rate": 3.99061357497522e-06, "loss": 0.3714, "step": 1820 }, { "epoch": 4.452322738386308, "grad_norm": 0.23980047548549416, "learning_rate": 3.989585638131744e-06, "loss": 0.609, "step": 1821 }, { "epoch": 4.45476772616137, "grad_norm": 0.17846990185641398, "learning_rate": 3.988557310684815e-06, "loss": 0.5094, "step": 1822 }, { "epoch": 4.45721271393643, "grad_norm": 0.21079232393269018, "learning_rate": 3.987528592904084e-06, "loss": 0.6343, "step": 1823 }, { "epoch": 4.459657701711492, "grad_norm": 0.14001613922297487, "learning_rate": 3.986499485059304e-06, "loss": 0.5633, "step": 1824 }, { "epoch": 4.462102689486553, "grad_norm": 0.19152436557825264, "learning_rate": 3.985469987420331e-06, "loss": 0.6092, "step": 1825 }, { "epoch": 4.4645476772616135, "grad_norm": 0.1980740598537105, "learning_rate": 3.984440100257123e-06, "loss": 0.4678, "step": 1826 }, { "epoch": 4.466992665036675, "grad_norm": 0.2710528962094933, "learning_rate": 3.9834098238397396e-06, "loss": 1.1581, "step": 1827 }, { "epoch": 4.469437652811736, "grad_norm": 0.22415951174185758, "learning_rate": 3.982379158438344e-06, "loss": 0.4822, "step": 1828 }, { "epoch": 4.471882640586797, "grad_norm": 0.16679266462959624, "learning_rate": 3.9813481043231985e-06, "loss": 0.421, "step": 1829 }, { "epoch": 4.474327628361858, "grad_norm": 0.2112976244305874, "learning_rate": 3.980316661764671e-06, "loss": 1.1342, "step": 1830 }, { "epoch": 4.47677261613692, "grad_norm": 0.24289323730858625, "learning_rate": 3.9792848310332264e-06, "loss": 0.5676, "step": 1831 }, { "epoch": 4.47921760391198, "grad_norm": 0.22967380686068448, "learning_rate": 3.978252612399437e-06, "loss": 0.509, "step": 1832 }, { "epoch": 4.481662591687042, "grad_norm": 0.1821182498724709, "learning_rate": 3.977220006133974e-06, "loss": 0.4698, "step": 1833 }, { "epoch": 4.484107579462103, "grad_norm": 0.14355172592759233, "learning_rate": 3.97618701250761e-06, "loss": 0.4379, "step": 1834 }, { "epoch": 4.486552567237164, "grad_norm": 0.19990549910478775, "learning_rate": 3.975153631791219e-06, "loss": 0.5842, "step": 1835 }, { "epoch": 4.488997555012225, "grad_norm": 0.16541384726173608, "learning_rate": 3.9741198642557775e-06, "loss": 0.4023, "step": 1836 }, { "epoch": 4.491442542787286, "grad_norm": 0.1588940364784366, "learning_rate": 3.973085710172363e-06, "loss": 0.902, "step": 1837 }, { "epoch": 4.493887530562347, "grad_norm": 0.15597887789605894, "learning_rate": 3.9720511698121545e-06, "loss": 0.4805, "step": 1838 }, { "epoch": 4.496332518337408, "grad_norm": 0.29329667943109233, "learning_rate": 3.971016243446432e-06, "loss": 1.0666, "step": 1839 }, { "epoch": 4.49877750611247, "grad_norm": 0.22005140845318127, "learning_rate": 3.969980931346577e-06, "loss": 0.6153, "step": 1840 }, { "epoch": 4.50122249388753, "grad_norm": 0.1852103371068952, "learning_rate": 3.968945233784071e-06, "loss": 0.465, "step": 1841 }, { "epoch": 4.503667481662592, "grad_norm": 0.2446345860628552, "learning_rate": 3.967909151030499e-06, "loss": 0.5681, "step": 1842 }, { "epoch": 4.506112469437653, "grad_norm": 0.24009461048981703, "learning_rate": 3.9668726833575455e-06, "loss": 0.7541, "step": 1843 }, { "epoch": 4.508557457212714, "grad_norm": 0.14015557083908609, "learning_rate": 3.965835831036995e-06, "loss": 0.6138, "step": 1844 }, { "epoch": 4.511002444987775, "grad_norm": 0.22249708612642274, "learning_rate": 3.9647985943407354e-06, "loss": 0.3902, "step": 1845 }, { "epoch": 4.513447432762836, "grad_norm": 0.3066803834758657, "learning_rate": 3.963760973540753e-06, "loss": 0.5112, "step": 1846 }, { "epoch": 4.515892420537897, "grad_norm": 0.1606742484773916, "learning_rate": 3.962722968909136e-06, "loss": 0.4571, "step": 1847 }, { "epoch": 4.518337408312958, "grad_norm": 0.18699561638523465, "learning_rate": 3.961684580718072e-06, "loss": 0.5717, "step": 1848 }, { "epoch": 4.52078239608802, "grad_norm": 0.20952396515852445, "learning_rate": 3.960645809239851e-06, "loss": 0.6622, "step": 1849 }, { "epoch": 4.52322738386308, "grad_norm": 0.2017567125761346, "learning_rate": 3.959606654746864e-06, "loss": 0.6138, "step": 1850 }, { "epoch": 4.525672371638142, "grad_norm": 0.1642699895413477, "learning_rate": 3.958567117511598e-06, "loss": 0.6044, "step": 1851 }, { "epoch": 4.528117359413203, "grad_norm": 0.23514934733691917, "learning_rate": 3.957527197806647e-06, "loss": 0.9963, "step": 1852 }, { "epoch": 4.530562347188264, "grad_norm": 0.2635508464939336, "learning_rate": 3.956486895904699e-06, "loss": 0.803, "step": 1853 }, { "epoch": 4.533007334963325, "grad_norm": 0.30614670873905137, "learning_rate": 3.955446212078545e-06, "loss": 0.9068, "step": 1854 }, { "epoch": 4.5354523227383865, "grad_norm": 0.24150155579032728, "learning_rate": 3.954405146601079e-06, "loss": 0.6967, "step": 1855 }, { "epoch": 4.537897310513447, "grad_norm": 0.16581157252024892, "learning_rate": 3.953363699745289e-06, "loss": 0.5162, "step": 1856 }, { "epoch": 4.540342298288508, "grad_norm": 0.2387610117554493, "learning_rate": 3.9523218717842675e-06, "loss": 0.6121, "step": 1857 }, { "epoch": 4.54278728606357, "grad_norm": 0.17651754729692953, "learning_rate": 3.951279662991206e-06, "loss": 0.4259, "step": 1858 }, { "epoch": 4.54523227383863, "grad_norm": 0.14499539782092472, "learning_rate": 3.950237073639395e-06, "loss": 0.4067, "step": 1859 }, { "epoch": 4.547677261613692, "grad_norm": 0.23292429591774189, "learning_rate": 3.949194104002224e-06, "loss": 0.5699, "step": 1860 }, { "epoch": 4.550122249388753, "grad_norm": 0.1339464462617127, "learning_rate": 3.948150754353187e-06, "loss": 0.3574, "step": 1861 }, { "epoch": 4.552567237163814, "grad_norm": 0.18376054811835527, "learning_rate": 3.947107024965872e-06, "loss": 0.7236, "step": 1862 }, { "epoch": 4.555012224938875, "grad_norm": 0.21717306051206267, "learning_rate": 3.946062916113967e-06, "loss": 0.8475, "step": 1863 }, { "epoch": 4.557457212713937, "grad_norm": 0.216445086430067, "learning_rate": 3.945018428071264e-06, "loss": 0.5497, "step": 1864 }, { "epoch": 4.559902200488997, "grad_norm": 0.1862245673034554, "learning_rate": 3.943973561111651e-06, "loss": 0.6632, "step": 1865 }, { "epoch": 4.5623471882640585, "grad_norm": 0.1458842636032291, "learning_rate": 3.942928315509115e-06, "loss": 0.4285, "step": 1866 }, { "epoch": 4.56479217603912, "grad_norm": 0.14425197581722418, "learning_rate": 3.941882691537745e-06, "loss": 0.5793, "step": 1867 }, { "epoch": 4.56723716381418, "grad_norm": 0.2811144462792972, "learning_rate": 3.940836689471726e-06, "loss": 0.3542, "step": 1868 }, { "epoch": 4.569682151589242, "grad_norm": 0.16165048847183508, "learning_rate": 3.939790309585344e-06, "loss": 0.5952, "step": 1869 }, { "epoch": 4.572127139364303, "grad_norm": 0.3270001859298466, "learning_rate": 3.938743552152985e-06, "loss": 0.4275, "step": 1870 }, { "epoch": 4.574572127139365, "grad_norm": 0.20009795157423668, "learning_rate": 3.937696417449131e-06, "loss": 0.5025, "step": 1871 }, { "epoch": 4.577017114914425, "grad_norm": 0.30726422777285173, "learning_rate": 3.936648905748366e-06, "loss": 0.4029, "step": 1872 }, { "epoch": 4.579462102689487, "grad_norm": 0.1451535724279507, "learning_rate": 3.9356010173253714e-06, "loss": 0.4649, "step": 1873 }, { "epoch": 4.581907090464548, "grad_norm": 0.17192979659205962, "learning_rate": 3.934552752454926e-06, "loss": 0.7784, "step": 1874 }, { "epoch": 4.5843520782396086, "grad_norm": 0.23242553728874452, "learning_rate": 3.93350411141191e-06, "loss": 0.525, "step": 1875 }, { "epoch": 4.58679706601467, "grad_norm": 0.22615294912869327, "learning_rate": 3.932455094471303e-06, "loss": 0.4938, "step": 1876 }, { "epoch": 4.589242053789731, "grad_norm": 0.16984893052173727, "learning_rate": 3.931405701908177e-06, "loss": 0.5479, "step": 1877 }, { "epoch": 4.591687041564792, "grad_norm": 0.19651716525114596, "learning_rate": 3.9303559339977114e-06, "loss": 0.7397, "step": 1878 }, { "epoch": 4.594132029339853, "grad_norm": 0.1595045025638794, "learning_rate": 3.9293057910151765e-06, "loss": 0.7098, "step": 1879 }, { "epoch": 4.596577017114915, "grad_norm": 0.2809518858234133, "learning_rate": 3.928255273235943e-06, "loss": 0.6728, "step": 1880 }, { "epoch": 4.599022004889975, "grad_norm": 0.3305036299706136, "learning_rate": 3.927204380935483e-06, "loss": 0.5246, "step": 1881 }, { "epoch": 4.601466992665037, "grad_norm": 0.22271749803837831, "learning_rate": 3.926153114389364e-06, "loss": 0.782, "step": 1882 }, { "epoch": 4.603911980440098, "grad_norm": 0.20451195584354562, "learning_rate": 3.925101473873252e-06, "loss": 0.6166, "step": 1883 }, { "epoch": 4.606356968215159, "grad_norm": 0.19031764205979998, "learning_rate": 3.924049459662911e-06, "loss": 0.4109, "step": 1884 }, { "epoch": 4.60880195599022, "grad_norm": 0.23794814895868074, "learning_rate": 3.922997072034202e-06, "loss": 0.3295, "step": 1885 }, { "epoch": 4.611246943765281, "grad_norm": 0.1779918921973669, "learning_rate": 3.921944311263088e-06, "loss": 0.4323, "step": 1886 }, { "epoch": 4.613691931540342, "grad_norm": 0.20230169512440785, "learning_rate": 3.920891177625623e-06, "loss": 0.441, "step": 1887 }, { "epoch": 4.616136919315403, "grad_norm": 0.20682415426651563, "learning_rate": 3.9198376713979675e-06, "loss": 0.8208, "step": 1888 }, { "epoch": 4.618581907090465, "grad_norm": 0.21065620274465463, "learning_rate": 3.9187837928563725e-06, "loss": 0.9154, "step": 1889 }, { "epoch": 4.621026894865525, "grad_norm": 0.257671479348446, "learning_rate": 3.917729542277188e-06, "loss": 0.5919, "step": 1890 }, { "epoch": 4.623471882640587, "grad_norm": 0.1594716575405407, "learning_rate": 3.916674919936864e-06, "loss": 0.4284, "step": 1891 }, { "epoch": 4.625916870415648, "grad_norm": 0.1444027691432101, "learning_rate": 3.9156199261119455e-06, "loss": 0.6073, "step": 1892 }, { "epoch": 4.628361858190709, "grad_norm": 0.18558313884434222, "learning_rate": 3.914564561079078e-06, "loss": 0.898, "step": 1893 }, { "epoch": 4.63080684596577, "grad_norm": 0.3672970379679832, "learning_rate": 3.913508825115001e-06, "loss": 0.3401, "step": 1894 }, { "epoch": 4.6332518337408315, "grad_norm": 0.18765576347420218, "learning_rate": 3.912452718496554e-06, "loss": 0.5971, "step": 1895 }, { "epoch": 4.635696821515893, "grad_norm": 0.18746003819447513, "learning_rate": 3.911396241500669e-06, "loss": 0.5445, "step": 1896 }, { "epoch": 4.638141809290953, "grad_norm": 0.15935878787748933, "learning_rate": 3.910339394404382e-06, "loss": 0.4754, "step": 1897 }, { "epoch": 4.640586797066015, "grad_norm": 0.2236216344747909, "learning_rate": 3.909282177484821e-06, "loss": 0.4526, "step": 1898 }, { "epoch": 4.643031784841076, "grad_norm": 0.22629157971195515, "learning_rate": 3.908224591019213e-06, "loss": 0.4919, "step": 1899 }, { "epoch": 4.645476772616137, "grad_norm": 0.1903794903127352, "learning_rate": 3.907166635284881e-06, "loss": 0.9723, "step": 1900 }, { "epoch": 4.647921760391198, "grad_norm": 0.2997041759211504, "learning_rate": 3.9061083105592455e-06, "loss": 0.7076, "step": 1901 }, { "epoch": 4.65036674816626, "grad_norm": 0.15998817367315762, "learning_rate": 3.905049617119822e-06, "loss": 0.9578, "step": 1902 }, { "epoch": 4.65281173594132, "grad_norm": 0.2623170736236951, "learning_rate": 3.903990555244227e-06, "loss": 0.5527, "step": 1903 }, { "epoch": 4.6552567237163816, "grad_norm": 0.2216462606613719, "learning_rate": 3.902931125210168e-06, "loss": 0.8426, "step": 1904 }, { "epoch": 4.657701711491443, "grad_norm": 0.20229204987759317, "learning_rate": 3.901871327295453e-06, "loss": 0.4246, "step": 1905 }, { "epoch": 4.6601466992665035, "grad_norm": 0.26263612350646326, "learning_rate": 3.9008111617779855e-06, "loss": 0.7432, "step": 1906 }, { "epoch": 4.662591687041565, "grad_norm": 0.2947146796861689, "learning_rate": 3.899750628935764e-06, "loss": 0.8694, "step": 1907 }, { "epoch": 4.665036674816626, "grad_norm": 0.18774811915838951, "learning_rate": 3.8986897290468875e-06, "loss": 0.5394, "step": 1908 }, { "epoch": 4.667481662591687, "grad_norm": 0.21467511827136018, "learning_rate": 3.897628462389544e-06, "loss": 0.6262, "step": 1909 }, { "epoch": 4.669926650366748, "grad_norm": 0.1456336371248034, "learning_rate": 3.896566829242025e-06, "loss": 0.8374, "step": 1910 }, { "epoch": 4.67237163814181, "grad_norm": 0.2131324176040151, "learning_rate": 3.895504829882715e-06, "loss": 0.3847, "step": 1911 }, { "epoch": 4.67481662591687, "grad_norm": 0.15910066798828792, "learning_rate": 3.894442464590092e-06, "loss": 0.39, "step": 1912 }, { "epoch": 4.677261613691932, "grad_norm": 0.18842476919132325, "learning_rate": 3.8933797336427345e-06, "loss": 0.5324, "step": 1913 }, { "epoch": 4.679706601466993, "grad_norm": 0.2005519473645647, "learning_rate": 3.8923166373193146e-06, "loss": 0.5339, "step": 1914 }, { "epoch": 4.6821515892420535, "grad_norm": 0.1973795478010137, "learning_rate": 3.891253175898601e-06, "loss": 0.7942, "step": 1915 }, { "epoch": 4.684596577017115, "grad_norm": 0.2510740378252299, "learning_rate": 3.8901893496594565e-06, "loss": 0.5394, "step": 1916 }, { "epoch": 4.687041564792176, "grad_norm": 0.23551798025366968, "learning_rate": 3.889125158880841e-06, "loss": 0.4284, "step": 1917 }, { "epoch": 4.689486552567237, "grad_norm": 0.24952782997015438, "learning_rate": 3.88806060384181e-06, "loss": 0.4171, "step": 1918 }, { "epoch": 4.691931540342298, "grad_norm": 0.2479449163759259, "learning_rate": 3.886995684821514e-06, "loss": 0.4515, "step": 1919 }, { "epoch": 4.69437652811736, "grad_norm": 0.1931792617800815, "learning_rate": 3.8859304020992e-06, "loss": 0.6077, "step": 1920 }, { "epoch": 4.69682151589242, "grad_norm": 0.19937564088534307, "learning_rate": 3.884864755954207e-06, "loss": 0.8667, "step": 1921 }, { "epoch": 4.699266503667482, "grad_norm": 0.21468890912171396, "learning_rate": 3.883798746665975e-06, "loss": 0.4879, "step": 1922 }, { "epoch": 4.701711491442543, "grad_norm": 0.21647069204607092, "learning_rate": 3.882732374514033e-06, "loss": 0.4702, "step": 1923 }, { "epoch": 4.704156479217604, "grad_norm": 0.18320323555742968, "learning_rate": 3.881665639778012e-06, "loss": 0.4358, "step": 1924 }, { "epoch": 4.706601466992665, "grad_norm": 0.2699457173555042, "learning_rate": 3.88059854273763e-06, "loss": 0.6853, "step": 1925 }, { "epoch": 4.709046454767726, "grad_norm": 0.14644264747846136, "learning_rate": 3.879531083672708e-06, "loss": 0.4755, "step": 1926 }, { "epoch": 4.711491442542787, "grad_norm": 0.23900530595740624, "learning_rate": 3.8784632628631555e-06, "loss": 0.6236, "step": 1927 }, { "epoch": 4.713936430317848, "grad_norm": 0.22967985501134436, "learning_rate": 3.87739508058898e-06, "loss": 0.8071, "step": 1928 }, { "epoch": 4.71638141809291, "grad_norm": 0.16133319269491517, "learning_rate": 3.876326537130285e-06, "loss": 0.4893, "step": 1929 }, { "epoch": 4.71882640586797, "grad_norm": 0.20475268684471065, "learning_rate": 3.875257632767267e-06, "loss": 0.4853, "step": 1930 }, { "epoch": 4.721271393643032, "grad_norm": 0.2417358906988055, "learning_rate": 3.874188367780214e-06, "loss": 0.3741, "step": 1931 }, { "epoch": 4.723716381418093, "grad_norm": 0.1880112607770472, "learning_rate": 3.873118742449515e-06, "loss": 0.5762, "step": 1932 }, { "epoch": 4.726161369193154, "grad_norm": 0.1882466959111549, "learning_rate": 3.8720487570556485e-06, "loss": 0.6927, "step": 1933 }, { "epoch": 4.728606356968215, "grad_norm": 0.14870883943225105, "learning_rate": 3.87097841187919e-06, "loss": 0.3591, "step": 1934 }, { "epoch": 4.7310513447432765, "grad_norm": 0.20881363835655278, "learning_rate": 3.869907707200809e-06, "loss": 0.5028, "step": 1935 }, { "epoch": 4.733496332518337, "grad_norm": 0.17361856461676922, "learning_rate": 3.868836643301268e-06, "loss": 0.5445, "step": 1936 }, { "epoch": 4.735941320293398, "grad_norm": 0.16035092661048211, "learning_rate": 3.867765220461424e-06, "loss": 0.6175, "step": 1937 }, { "epoch": 4.73838630806846, "grad_norm": 0.23681248214655376, "learning_rate": 3.866693438962229e-06, "loss": 0.7777, "step": 1938 }, { "epoch": 4.74083129584352, "grad_norm": 0.16443145304661844, "learning_rate": 3.86562129908473e-06, "loss": 0.7213, "step": 1939 }, { "epoch": 4.743276283618582, "grad_norm": 0.23362913143257044, "learning_rate": 3.864548801110064e-06, "loss": 0.697, "step": 1940 }, { "epoch": 4.745721271393643, "grad_norm": 0.17138185489462232, "learning_rate": 3.863475945319467e-06, "loss": 0.5086, "step": 1941 }, { "epoch": 4.748166259168704, "grad_norm": 0.27202414886424636, "learning_rate": 3.8624027319942646e-06, "loss": 0.4491, "step": 1942 }, { "epoch": 4.750611246943765, "grad_norm": 0.22256250420000104, "learning_rate": 3.86132916141588e-06, "loss": 0.6502, "step": 1943 }, { "epoch": 4.7530562347188265, "grad_norm": 0.21566192500470766, "learning_rate": 3.860255233865825e-06, "loss": 0.7884, "step": 1944 }, { "epoch": 4.755501222493887, "grad_norm": 0.19536353880397456, "learning_rate": 3.8591809496257105e-06, "loss": 0.484, "step": 1945 }, { "epoch": 4.7579462102689485, "grad_norm": 0.21020061158050243, "learning_rate": 3.858106308977236e-06, "loss": 0.3889, "step": 1946 }, { "epoch": 4.76039119804401, "grad_norm": 0.2123902171862699, "learning_rate": 3.857031312202201e-06, "loss": 0.7079, "step": 1947 }, { "epoch": 4.76283618581907, "grad_norm": 0.15343300741797006, "learning_rate": 3.855955959582489e-06, "loss": 0.5779, "step": 1948 }, { "epoch": 4.765281173594132, "grad_norm": 0.2455924513677434, "learning_rate": 3.8548802514000865e-06, "loss": 0.9137, "step": 1949 }, { "epoch": 4.767726161369193, "grad_norm": 0.15175211624113477, "learning_rate": 3.853804187937066e-06, "loss": 0.6061, "step": 1950 }, { "epoch": 4.770171149144255, "grad_norm": 0.2901670525217353, "learning_rate": 3.852727769475598e-06, "loss": 0.4708, "step": 1951 }, { "epoch": 4.772616136919315, "grad_norm": 0.2189426443335556, "learning_rate": 3.851650996297942e-06, "loss": 0.4331, "step": 1952 }, { "epoch": 4.775061124694377, "grad_norm": 0.21269033874983664, "learning_rate": 3.850573868686455e-06, "loss": 0.4559, "step": 1953 }, { "epoch": 4.777506112469438, "grad_norm": 0.20700304799388738, "learning_rate": 3.849496386923581e-06, "loss": 0.8294, "step": 1954 }, { "epoch": 4.7799511002444985, "grad_norm": 0.2575916673336584, "learning_rate": 3.848418551291863e-06, "loss": 0.6847, "step": 1955 }, { "epoch": 4.78239608801956, "grad_norm": 0.19198169268911439, "learning_rate": 3.847340362073936e-06, "loss": 1.0064, "step": 1956 }, { "epoch": 4.784841075794621, "grad_norm": 0.21795549797020738, "learning_rate": 3.846261819552522e-06, "loss": 0.4946, "step": 1957 }, { "epoch": 4.787286063569682, "grad_norm": 0.26371608561823984, "learning_rate": 3.845182924010442e-06, "loss": 0.4436, "step": 1958 }, { "epoch": 4.789731051344743, "grad_norm": 0.2056800555113254, "learning_rate": 3.844103675730605e-06, "loss": 0.8141, "step": 1959 }, { "epoch": 4.792176039119805, "grad_norm": 0.11243758951674035, "learning_rate": 3.843024074996016e-06, "loss": 0.3891, "step": 1960 }, { "epoch": 4.794621026894865, "grad_norm": 0.26235840690026363, "learning_rate": 3.841944122089771e-06, "loss": 0.4628, "step": 1961 }, { "epoch": 4.797066014669927, "grad_norm": 0.12164715714691905, "learning_rate": 3.840863817295058e-06, "loss": 0.5938, "step": 1962 }, { "epoch": 4.799511002444988, "grad_norm": 0.21038809430988087, "learning_rate": 3.83978316089516e-06, "loss": 0.6312, "step": 1963 }, { "epoch": 4.801955990220049, "grad_norm": 0.3013022160166754, "learning_rate": 3.838702153173445e-06, "loss": 0.5416, "step": 1964 }, { "epoch": 4.80440097799511, "grad_norm": 0.261301971960218, "learning_rate": 3.837620794413382e-06, "loss": 0.5817, "step": 1965 }, { "epoch": 4.806845965770171, "grad_norm": 0.2614020785175297, "learning_rate": 3.836539084898526e-06, "loss": 0.4594, "step": 1966 }, { "epoch": 4.809290953545232, "grad_norm": 0.19587007139984464, "learning_rate": 3.835457024912528e-06, "loss": 0.6377, "step": 1967 }, { "epoch": 4.811735941320293, "grad_norm": 0.17601643732403502, "learning_rate": 3.8343746147391266e-06, "loss": 0.9869, "step": 1968 }, { "epoch": 4.814180929095355, "grad_norm": 0.22464474292812306, "learning_rate": 3.833291854662156e-06, "loss": 0.6477, "step": 1969 }, { "epoch": 4.816625916870415, "grad_norm": 0.21213694340223635, "learning_rate": 3.8322087449655395e-06, "loss": 1.0335, "step": 1970 }, { "epoch": 4.819070904645477, "grad_norm": 0.2874830695096747, "learning_rate": 3.8311252859332945e-06, "loss": 0.5767, "step": 1971 }, { "epoch": 4.821515892420538, "grad_norm": 0.20176036799039174, "learning_rate": 3.8300414778495285e-06, "loss": 0.5538, "step": 1972 }, { "epoch": 4.823960880195599, "grad_norm": 0.10918902686791593, "learning_rate": 3.828957320998441e-06, "loss": 0.3323, "step": 1973 }, { "epoch": 4.82640586797066, "grad_norm": 0.15742143109472015, "learning_rate": 3.8278728156643216e-06, "loss": 0.4958, "step": 1974 }, { "epoch": 4.8288508557457215, "grad_norm": 0.17326474685642712, "learning_rate": 3.826787962131553e-06, "loss": 0.4217, "step": 1975 }, { "epoch": 4.831295843520783, "grad_norm": 0.22096924192095757, "learning_rate": 3.82570276068461e-06, "loss": 1.1507, "step": 1976 }, { "epoch": 4.833740831295843, "grad_norm": 0.16739412855243058, "learning_rate": 3.824617211608054e-06, "loss": 0.6572, "step": 1977 }, { "epoch": 4.836185819070905, "grad_norm": 0.19427156035346388, "learning_rate": 3.823531315186544e-06, "loss": 0.6499, "step": 1978 }, { "epoch": 4.838630806845966, "grad_norm": 0.2020682508750726, "learning_rate": 3.822445071704826e-06, "loss": 0.6886, "step": 1979 }, { "epoch": 4.841075794621027, "grad_norm": 0.21305664021715248, "learning_rate": 3.8213584814477365e-06, "loss": 0.4027, "step": 1980 }, { "epoch": 4.843520782396088, "grad_norm": 0.19729619654737715, "learning_rate": 3.8202715447002055e-06, "loss": 0.7101, "step": 1981 }, { "epoch": 4.84596577017115, "grad_norm": 0.32905442981255023, "learning_rate": 3.8191842617472535e-06, "loss": 0.6772, "step": 1982 }, { "epoch": 4.84841075794621, "grad_norm": 0.14138651999902113, "learning_rate": 3.81809663287399e-06, "loss": 0.494, "step": 1983 }, { "epoch": 4.8508557457212715, "grad_norm": 0.25353772888397397, "learning_rate": 3.817008658365617e-06, "loss": 0.438, "step": 1984 }, { "epoch": 4.853300733496333, "grad_norm": 0.21700538081063245, "learning_rate": 3.815920338507424e-06, "loss": 0.7767, "step": 1985 }, { "epoch": 4.8557457212713935, "grad_norm": 0.24209812091480867, "learning_rate": 3.814831673584796e-06, "loss": 0.6428, "step": 1986 }, { "epoch": 4.858190709046455, "grad_norm": 0.2539413766704324, "learning_rate": 3.813742663883204e-06, "loss": 0.618, "step": 1987 }, { "epoch": 4.860635696821516, "grad_norm": 0.15096846661600963, "learning_rate": 3.8126533096882133e-06, "loss": 0.4192, "step": 1988 }, { "epoch": 4.863080684596577, "grad_norm": 0.17490886731874755, "learning_rate": 3.811563611285476e-06, "loss": 0.4625, "step": 1989 }, { "epoch": 4.865525672371638, "grad_norm": 0.31723157663273943, "learning_rate": 3.8104735689607366e-06, "loss": 1.1832, "step": 1990 }, { "epoch": 4.8679706601467, "grad_norm": 0.22811169507031812, "learning_rate": 3.8093831829998285e-06, "loss": 0.714, "step": 1991 }, { "epoch": 4.87041564792176, "grad_norm": 0.241903334113044, "learning_rate": 3.808292453688677e-06, "loss": 0.8457, "step": 1992 }, { "epoch": 4.872860635696822, "grad_norm": 0.2608841564707084, "learning_rate": 3.8072013813132946e-06, "loss": 1.2257, "step": 1993 }, { "epoch": 4.875305623471883, "grad_norm": 0.20358449819275207, "learning_rate": 3.8061099661597862e-06, "loss": 1.1289, "step": 1994 }, { "epoch": 4.8777506112469435, "grad_norm": 0.19425777775823166, "learning_rate": 3.8050182085143468e-06, "loss": 0.6455, "step": 1995 }, { "epoch": 4.880195599022005, "grad_norm": 0.17592971668666774, "learning_rate": 3.8039261086632594e-06, "loss": 0.5491, "step": 1996 }, { "epoch": 4.882640586797066, "grad_norm": 0.1708452357251292, "learning_rate": 3.802833666892898e-06, "loss": 0.5469, "step": 1997 }, { "epoch": 4.885085574572127, "grad_norm": 0.2092071207069566, "learning_rate": 3.801740883489724e-06, "loss": 0.5994, "step": 1998 }, { "epoch": 4.887530562347188, "grad_norm": 0.1986278689043537, "learning_rate": 3.8006477587402923e-06, "loss": 0.6345, "step": 1999 }, { "epoch": 4.88997555012225, "grad_norm": 0.3686688065604612, "learning_rate": 3.7995542929312456e-06, "loss": 0.5568, "step": 2000 }, { "epoch": 4.89242053789731, "grad_norm": 0.1557762142140868, "learning_rate": 3.798460486349314e-06, "loss": 0.6145, "step": 2001 }, { "epoch": 4.894865525672372, "grad_norm": 0.15066400384436882, "learning_rate": 3.79736633928132e-06, "loss": 0.4007, "step": 2002 }, { "epoch": 4.897310513447433, "grad_norm": 0.33131998490164694, "learning_rate": 3.796271852014172e-06, "loss": 0.4809, "step": 2003 }, { "epoch": 4.899755501222494, "grad_norm": 0.23293494050755983, "learning_rate": 3.795177024834872e-06, "loss": 0.4936, "step": 2004 }, { "epoch": 4.902200488997555, "grad_norm": 0.2000899059101259, "learning_rate": 3.794081858030507e-06, "loss": 0.8239, "step": 2005 }, { "epoch": 4.904645476772616, "grad_norm": 0.307332489168978, "learning_rate": 3.792986351888257e-06, "loss": 0.3519, "step": 2006 }, { "epoch": 4.907090464547677, "grad_norm": 0.30272302421376335, "learning_rate": 3.791890506695386e-06, "loss": 0.648, "step": 2007 }, { "epoch": 4.909535452322738, "grad_norm": 0.19490624767310388, "learning_rate": 3.790794322739251e-06, "loss": 0.5777, "step": 2008 }, { "epoch": 4.9119804400978, "grad_norm": 0.26806768413369647, "learning_rate": 3.789697800307297e-06, "loss": 0.6835, "step": 2009 }, { "epoch": 4.91442542787286, "grad_norm": 0.20663552404159946, "learning_rate": 3.788600939687057e-06, "loss": 0.8178, "step": 2010 }, { "epoch": 4.916870415647922, "grad_norm": 0.19750992003682344, "learning_rate": 3.7875037411661524e-06, "loss": 0.587, "step": 2011 }, { "epoch": 4.919315403422983, "grad_norm": 0.21485029330342917, "learning_rate": 3.7864062050322946e-06, "loss": 0.5488, "step": 2012 }, { "epoch": 4.921760391198044, "grad_norm": 0.26926451689765024, "learning_rate": 3.785308331573282e-06, "loss": 0.654, "step": 2013 }, { "epoch": 4.924205378973105, "grad_norm": 0.18553515270981613, "learning_rate": 3.7842101210770015e-06, "loss": 0.7023, "step": 2014 }, { "epoch": 4.9266503667481665, "grad_norm": 0.21393325805719873, "learning_rate": 3.7831115738314304e-06, "loss": 0.3538, "step": 2015 }, { "epoch": 4.929095354523227, "grad_norm": 0.23201438187328913, "learning_rate": 3.7820126901246333e-06, "loss": 0.6286, "step": 2016 }, { "epoch": 4.931540342298288, "grad_norm": 0.21047620000930423, "learning_rate": 3.7809134702447612e-06, "loss": 0.3687, "step": 2017 }, { "epoch": 4.93398533007335, "grad_norm": 0.406652349391441, "learning_rate": 3.7798139144800556e-06, "loss": 0.5549, "step": 2018 }, { "epoch": 4.93643031784841, "grad_norm": 0.17129572422517605, "learning_rate": 3.7787140231188447e-06, "loss": 0.6089, "step": 2019 }, { "epoch": 4.938875305623472, "grad_norm": 0.262309863813515, "learning_rate": 3.777613796449545e-06, "loss": 0.5803, "step": 2020 }, { "epoch": 4.941320293398533, "grad_norm": 0.17737294663134487, "learning_rate": 3.7765132347606624e-06, "loss": 0.9221, "step": 2021 }, { "epoch": 4.943765281173594, "grad_norm": 0.23111015373190685, "learning_rate": 3.775412338340788e-06, "loss": 0.8123, "step": 2022 }, { "epoch": 4.946210268948655, "grad_norm": 0.18297921749076518, "learning_rate": 3.774311107478603e-06, "loss": 0.4031, "step": 2023 }, { "epoch": 4.9486552567237165, "grad_norm": 0.2586420993427602, "learning_rate": 3.7732095424628745e-06, "loss": 0.3515, "step": 2024 }, { "epoch": 4.951100244498777, "grad_norm": 0.1539691933071811, "learning_rate": 3.772107643582459e-06, "loss": 0.5956, "step": 2025 }, { "epoch": 4.9535452322738385, "grad_norm": 0.24451172645146627, "learning_rate": 3.7710054111262983e-06, "loss": 0.5956, "step": 2026 }, { "epoch": 4.9559902200489, "grad_norm": 0.27555798407071436, "learning_rate": 3.769902845383424e-06, "loss": 0.452, "step": 2027 }, { "epoch": 4.95843520782396, "grad_norm": 0.17389396108444852, "learning_rate": 3.7687999466429537e-06, "loss": 0.5294, "step": 2028 }, { "epoch": 4.960880195599022, "grad_norm": 0.16614482374299225, "learning_rate": 3.7676967151940934e-06, "loss": 0.578, "step": 2029 }, { "epoch": 4.963325183374083, "grad_norm": 0.21186167741105447, "learning_rate": 3.7665931513261343e-06, "loss": 0.6392, "step": 2030 }, { "epoch": 4.965770171149144, "grad_norm": 0.15095146052310937, "learning_rate": 3.7654892553284572e-06, "loss": 0.5462, "step": 2031 }, { "epoch": 4.968215158924205, "grad_norm": 0.19698066864023028, "learning_rate": 3.764385027490529e-06, "loss": 0.9083, "step": 2032 }, { "epoch": 4.970660146699267, "grad_norm": 0.21849600590034604, "learning_rate": 3.7632804681019026e-06, "loss": 0.8427, "step": 2033 }, { "epoch": 4.973105134474328, "grad_norm": 0.1444160031193256, "learning_rate": 3.7621755774522194e-06, "loss": 0.5039, "step": 2034 }, { "epoch": 4.9755501222493885, "grad_norm": 0.1708550844064277, "learning_rate": 3.7610703558312066e-06, "loss": 0.6111, "step": 2035 }, { "epoch": 4.97799511002445, "grad_norm": 0.16775642790513573, "learning_rate": 3.7599648035286793e-06, "loss": 0.5445, "step": 2036 }, { "epoch": 4.980440097799511, "grad_norm": 0.1707128075406305, "learning_rate": 3.758858920834538e-06, "loss": 0.6843, "step": 2037 }, { "epoch": 4.982885085574572, "grad_norm": 0.1677513351740558, "learning_rate": 3.757752708038771e-06, "loss": 0.5329, "step": 2038 }, { "epoch": 4.985330073349633, "grad_norm": 0.18443435104057523, "learning_rate": 3.7566461654314517e-06, "loss": 0.5066, "step": 2039 }, { "epoch": 4.987775061124695, "grad_norm": 0.20616232490108505, "learning_rate": 3.7555392933027423e-06, "loss": 0.3955, "step": 2040 }, { "epoch": 4.990220048899755, "grad_norm": 0.17286930939597117, "learning_rate": 3.754432091942888e-06, "loss": 0.4773, "step": 2041 }, { "epoch": 4.992665036674817, "grad_norm": 0.19628475350325839, "learning_rate": 3.7533245616422244e-06, "loss": 0.8423, "step": 2042 }, { "epoch": 4.995110024449878, "grad_norm": 0.2452645530827968, "learning_rate": 3.75221670269117e-06, "loss": 0.7313, "step": 2043 }, { "epoch": 4.997555012224939, "grad_norm": 0.2615524126775721, "learning_rate": 3.7511085153802314e-06, "loss": 0.3855, "step": 2044 }, { "epoch": 5.0, "grad_norm": 0.2434355251250328, "learning_rate": 3.7500000000000005e-06, "loss": 0.7845, "step": 2045 }, { "epoch": 5.002444987775061, "grad_norm": 0.16109563114864184, "learning_rate": 3.7488911568411553e-06, "loss": 0.7569, "step": 2046 }, { "epoch": 5.004889975550122, "grad_norm": 0.15735684935296101, "learning_rate": 3.7477819861944597e-06, "loss": 0.3686, "step": 2047 }, { "epoch": 5.007334963325183, "grad_norm": 0.21194573241183948, "learning_rate": 3.746672488350764e-06, "loss": 0.4379, "step": 2048 }, { "epoch": 5.009779951100245, "grad_norm": 0.17339994080447857, "learning_rate": 3.7455626636010046e-06, "loss": 0.8235, "step": 2049 }, { "epoch": 5.012224938875305, "grad_norm": 0.31944516023074276, "learning_rate": 3.7444525122362015e-06, "loss": 0.7996, "step": 2050 }, { "epoch": 5.014669926650367, "grad_norm": 0.22939504986346795, "learning_rate": 3.7433420345474624e-06, "loss": 0.7048, "step": 2051 }, { "epoch": 5.017114914425428, "grad_norm": 0.1618363442223097, "learning_rate": 3.74223123082598e-06, "loss": 1.008, "step": 2052 }, { "epoch": 5.019559902200489, "grad_norm": 0.14721689164750812, "learning_rate": 3.741120101363033e-06, "loss": 0.5197, "step": 2053 }, { "epoch": 5.02200488997555, "grad_norm": 0.1694875910171043, "learning_rate": 3.740008646449984e-06, "loss": 0.3399, "step": 2054 }, { "epoch": 5.0244498777506115, "grad_norm": 0.2092040762878438, "learning_rate": 3.7388968663782825e-06, "loss": 0.5061, "step": 2055 }, { "epoch": 5.026894865525672, "grad_norm": 0.1808907995750573, "learning_rate": 3.737784761439463e-06, "loss": 0.6191, "step": 2056 }, { "epoch": 5.029339853300733, "grad_norm": 0.2167372295433262, "learning_rate": 3.736672331925144e-06, "loss": 0.903, "step": 2057 }, { "epoch": 5.031784841075795, "grad_norm": 0.19735001348887968, "learning_rate": 3.7355595781270304e-06, "loss": 0.8945, "step": 2058 }, { "epoch": 5.034229828850855, "grad_norm": 0.16513299651852423, "learning_rate": 3.734446500336912e-06, "loss": 0.4226, "step": 2059 }, { "epoch": 5.036674816625917, "grad_norm": 0.20150106643164067, "learning_rate": 3.733333098846663e-06, "loss": 0.6902, "step": 2060 }, { "epoch": 5.039119804400978, "grad_norm": 0.255453117342702, "learning_rate": 3.732219373948243e-06, "loss": 0.5388, "step": 2061 }, { "epoch": 5.041564792176039, "grad_norm": 0.21904539621790262, "learning_rate": 3.7311053259336953e-06, "loss": 0.3778, "step": 2062 }, { "epoch": 5.0440097799511, "grad_norm": 0.2536145338282659, "learning_rate": 3.7299909550951495e-06, "loss": 0.8119, "step": 2063 }, { "epoch": 5.0464547677261615, "grad_norm": 0.18498787151540202, "learning_rate": 3.72887626172482e-06, "loss": 0.4255, "step": 2064 }, { "epoch": 5.048899755501223, "grad_norm": 0.18836031374252818, "learning_rate": 3.7277612461150037e-06, "loss": 0.4448, "step": 2065 }, { "epoch": 5.0513447432762835, "grad_norm": 0.18298151989927058, "learning_rate": 3.7266459085580834e-06, "loss": 0.6381, "step": 2066 }, { "epoch": 5.053789731051345, "grad_norm": 0.25343862114536697, "learning_rate": 3.7255302493465264e-06, "loss": 0.5411, "step": 2067 }, { "epoch": 5.056234718826406, "grad_norm": 0.2842673456838669, "learning_rate": 3.7244142687728836e-06, "loss": 0.8084, "step": 2068 }, { "epoch": 5.058679706601467, "grad_norm": 0.23676455103581842, "learning_rate": 3.7232979671297908e-06, "loss": 1.305, "step": 2069 }, { "epoch": 5.061124694376528, "grad_norm": 0.1812856147831502, "learning_rate": 3.7221813447099693e-06, "loss": 0.769, "step": 2070 }, { "epoch": 5.06356968215159, "grad_norm": 0.22498980848527977, "learning_rate": 3.721064401806221e-06, "loss": 0.764, "step": 2071 }, { "epoch": 5.06601466992665, "grad_norm": 0.22482236418218207, "learning_rate": 3.7199471387114343e-06, "loss": 0.6532, "step": 2072 }, { "epoch": 5.068459657701712, "grad_norm": 0.18125859229578006, "learning_rate": 3.7188295557185817e-06, "loss": 0.3887, "step": 2073 }, { "epoch": 5.070904645476773, "grad_norm": 0.23672909655585825, "learning_rate": 3.717711653120719e-06, "loss": 0.465, "step": 2074 }, { "epoch": 5.0733496332518335, "grad_norm": 0.14252514902839977, "learning_rate": 3.7165934312109863e-06, "loss": 0.7082, "step": 2075 }, { "epoch": 5.075794621026895, "grad_norm": 0.1619753205830732, "learning_rate": 3.7154748902826066e-06, "loss": 0.9648, "step": 2076 }, { "epoch": 5.078239608801956, "grad_norm": 0.16355380726896754, "learning_rate": 3.7143560306288863e-06, "loss": 0.4784, "step": 2077 }, { "epoch": 5.080684596577017, "grad_norm": 0.19496912463798421, "learning_rate": 3.713236852543216e-06, "loss": 0.7877, "step": 2078 }, { "epoch": 5.083129584352078, "grad_norm": 0.17263534827473018, "learning_rate": 3.7121173563190704e-06, "loss": 0.4282, "step": 2079 }, { "epoch": 5.08557457212714, "grad_norm": 0.28495006712161103, "learning_rate": 3.7109975422500076e-06, "loss": 0.6145, "step": 2080 }, { "epoch": 5.0880195599022, "grad_norm": 0.211043385597544, "learning_rate": 3.709877410629668e-06, "loss": 0.7415, "step": 2081 }, { "epoch": 5.090464547677262, "grad_norm": 0.21493335041067174, "learning_rate": 3.7087569617517748e-06, "loss": 0.602, "step": 2082 }, { "epoch": 5.092909535452323, "grad_norm": 0.24835588154747737, "learning_rate": 3.707636195910137e-06, "loss": 0.8523, "step": 2083 }, { "epoch": 5.095354523227384, "grad_norm": 0.1714890789052701, "learning_rate": 3.7065151133986434e-06, "loss": 0.9308, "step": 2084 }, { "epoch": 5.097799511002445, "grad_norm": 0.16482953190126545, "learning_rate": 3.7053937145112682e-06, "loss": 0.5792, "step": 2085 }, { "epoch": 5.100244498777506, "grad_norm": 0.24576765578946383, "learning_rate": 3.704271999542069e-06, "loss": 0.6846, "step": 2086 }, { "epoch": 5.102689486552567, "grad_norm": 0.2284261127763889, "learning_rate": 3.7031499687851828e-06, "loss": 0.5299, "step": 2087 }, { "epoch": 5.105134474327628, "grad_norm": 0.1697789895836933, "learning_rate": 3.702027622534834e-06, "loss": 0.6674, "step": 2088 }, { "epoch": 5.10757946210269, "grad_norm": 0.20660629650210033, "learning_rate": 3.700904961085326e-06, "loss": 1.0055, "step": 2089 }, { "epoch": 5.11002444987775, "grad_norm": 0.23226771914771305, "learning_rate": 3.6997819847310475e-06, "loss": 0.6362, "step": 2090 }, { "epoch": 5.112469437652812, "grad_norm": 0.18380249217489605, "learning_rate": 3.6986586937664672e-06, "loss": 0.4715, "step": 2091 }, { "epoch": 5.114914425427873, "grad_norm": 0.12450233399837017, "learning_rate": 3.6975350884861384e-06, "loss": 0.7729, "step": 2092 }, { "epoch": 5.117359413202934, "grad_norm": 0.21849488853786664, "learning_rate": 3.6964111691846973e-06, "loss": 0.4097, "step": 2093 }, { "epoch": 5.119804400977995, "grad_norm": 0.18452563369418593, "learning_rate": 3.6952869361568587e-06, "loss": 0.8212, "step": 2094 }, { "epoch": 5.1222493887530565, "grad_norm": 0.29796345709010247, "learning_rate": 3.6941623896974245e-06, "loss": 0.4151, "step": 2095 }, { "epoch": 5.124694376528117, "grad_norm": 0.229999026641762, "learning_rate": 3.6930375301012755e-06, "loss": 0.6615, "step": 2096 }, { "epoch": 5.127139364303178, "grad_norm": 0.21413008497904962, "learning_rate": 3.691912357663376e-06, "loss": 0.6919, "step": 2097 }, { "epoch": 5.12958435207824, "grad_norm": 0.15737818578336213, "learning_rate": 3.6907868726787717e-06, "loss": 0.4261, "step": 2098 }, { "epoch": 5.1320293398533, "grad_norm": 0.28068837012746667, "learning_rate": 3.689661075442591e-06, "loss": 0.5471, "step": 2099 }, { "epoch": 5.134474327628362, "grad_norm": 0.33290811587452424, "learning_rate": 3.6885349662500423e-06, "loss": 0.5975, "step": 2100 }, { "epoch": 5.136919315403423, "grad_norm": 0.1786004918209996, "learning_rate": 3.687408545396419e-06, "loss": 0.7505, "step": 2101 }, { "epoch": 5.139364303178484, "grad_norm": 0.20152443049657645, "learning_rate": 3.686281813177094e-06, "loss": 0.7161, "step": 2102 }, { "epoch": 5.141809290953545, "grad_norm": 0.259211002387912, "learning_rate": 3.6851547698875216e-06, "loss": 0.5705, "step": 2103 }, { "epoch": 5.1442542787286065, "grad_norm": 0.26085245813654273, "learning_rate": 3.684027415823239e-06, "loss": 0.4421, "step": 2104 }, { "epoch": 5.146699266503667, "grad_norm": 0.22527074714918777, "learning_rate": 3.6828997512798635e-06, "loss": 0.5591, "step": 2105 }, { "epoch": 5.1491442542787285, "grad_norm": 0.1854647928967856, "learning_rate": 3.6817717765530953e-06, "loss": 1.0413, "step": 2106 }, { "epoch": 5.15158924205379, "grad_norm": 0.2058749798127868, "learning_rate": 3.680643491938715e-06, "loss": 0.5927, "step": 2107 }, { "epoch": 5.15403422982885, "grad_norm": 0.11999034936399257, "learning_rate": 3.6795148977325846e-06, "loss": 0.3803, "step": 2108 }, { "epoch": 5.156479217603912, "grad_norm": 0.21600662594533374, "learning_rate": 3.678385994230647e-06, "loss": 0.4288, "step": 2109 }, { "epoch": 5.158924205378973, "grad_norm": 0.25378512829289, "learning_rate": 3.6772567817289275e-06, "loss": 0.5874, "step": 2110 }, { "epoch": 5.161369193154035, "grad_norm": 0.24987066388456966, "learning_rate": 3.67612726052353e-06, "loss": 0.4971, "step": 2111 }, { "epoch": 5.163814180929095, "grad_norm": 0.14943094983788496, "learning_rate": 3.674997430910642e-06, "loss": 0.5019, "step": 2112 }, { "epoch": 5.166259168704157, "grad_norm": 0.1704947679311722, "learning_rate": 3.6738672931865297e-06, "loss": 0.9122, "step": 2113 }, { "epoch": 5.168704156479218, "grad_norm": 0.24898697634952338, "learning_rate": 3.6727368476475427e-06, "loss": 0.6099, "step": 2114 }, { "epoch": 5.1711491442542785, "grad_norm": 0.2667313247007851, "learning_rate": 3.6716060945901084e-06, "loss": 0.9603, "step": 2115 }, { "epoch": 5.17359413202934, "grad_norm": 0.18768121598491, "learning_rate": 3.670475034310736e-06, "loss": 0.5143, "step": 2116 }, { "epoch": 5.176039119804401, "grad_norm": 0.15997224837678883, "learning_rate": 3.6693436671060158e-06, "loss": 0.6646, "step": 2117 }, { "epoch": 5.178484107579462, "grad_norm": 0.2172855730645409, "learning_rate": 3.668211993272619e-06, "loss": 0.6451, "step": 2118 }, { "epoch": 5.180929095354523, "grad_norm": 0.1901992657389288, "learning_rate": 3.6670800131072948e-06, "loss": 0.4664, "step": 2119 }, { "epoch": 5.183374083129585, "grad_norm": 0.19258303832498572, "learning_rate": 3.6659477269068754e-06, "loss": 0.5509, "step": 2120 }, { "epoch": 5.185819070904645, "grad_norm": 0.2276976577830496, "learning_rate": 3.664815134968272e-06, "loss": 0.4729, "step": 2121 }, { "epoch": 5.188264058679707, "grad_norm": 0.1614653089989081, "learning_rate": 3.6636822375884757e-06, "loss": 0.5779, "step": 2122 }, { "epoch": 5.190709046454768, "grad_norm": 0.4340789950657901, "learning_rate": 3.662549035064559e-06, "loss": 0.4417, "step": 2123 }, { "epoch": 5.193154034229829, "grad_norm": 0.22815455022546718, "learning_rate": 3.6614155276936724e-06, "loss": 0.3982, "step": 2124 }, { "epoch": 5.19559902200489, "grad_norm": 0.1579487667006239, "learning_rate": 3.6602817157730486e-06, "loss": 0.4922, "step": 2125 }, { "epoch": 5.198044009779951, "grad_norm": 0.2209374313725165, "learning_rate": 3.659147599599998e-06, "loss": 0.4823, "step": 2126 }, { "epoch": 5.200488997555012, "grad_norm": 0.19335759246157247, "learning_rate": 3.658013179471912e-06, "loss": 0.4809, "step": 2127 }, { "epoch": 5.202933985330073, "grad_norm": 0.2272717050360955, "learning_rate": 3.6568784556862622e-06, "loss": 0.8961, "step": 2128 }, { "epoch": 5.205378973105135, "grad_norm": 0.23036034916843798, "learning_rate": 3.655743428540599e-06, "loss": 0.573, "step": 2129 }, { "epoch": 5.207823960880195, "grad_norm": 0.2362346540137818, "learning_rate": 3.654608098332553e-06, "loss": 0.4361, "step": 2130 }, { "epoch": 5.210268948655257, "grad_norm": 0.21135474452709654, "learning_rate": 3.6534724653598312e-06, "loss": 0.4775, "step": 2131 }, { "epoch": 5.212713936430318, "grad_norm": 0.23048788040404924, "learning_rate": 3.6523365299202255e-06, "loss": 0.603, "step": 2132 }, { "epoch": 5.215158924205379, "grad_norm": 0.29023952096835604, "learning_rate": 3.651200292311603e-06, "loss": 0.716, "step": 2133 }, { "epoch": 5.21760391198044, "grad_norm": 0.22837550170067364, "learning_rate": 3.6500637528319115e-06, "loss": 0.4567, "step": 2134 }, { "epoch": 5.2200488997555015, "grad_norm": 0.22903019299457172, "learning_rate": 3.6489269117791782e-06, "loss": 1.0954, "step": 2135 }, { "epoch": 5.222493887530562, "grad_norm": 0.1909403867781725, "learning_rate": 3.6477897694515063e-06, "loss": 0.659, "step": 2136 }, { "epoch": 5.224938875305623, "grad_norm": 0.21263450208630374, "learning_rate": 3.6466523261470837e-06, "loss": 0.7687, "step": 2137 }, { "epoch": 5.227383863080685, "grad_norm": 0.21628289821966418, "learning_rate": 3.6455145821641726e-06, "loss": 0.5646, "step": 2138 }, { "epoch": 5.229828850855745, "grad_norm": 0.21116576678531623, "learning_rate": 3.6443765378011154e-06, "loss": 0.3851, "step": 2139 }, { "epoch": 5.232273838630807, "grad_norm": 0.17493305020936142, "learning_rate": 3.6432381933563337e-06, "loss": 0.5168, "step": 2140 }, { "epoch": 5.234718826405868, "grad_norm": 0.27078908826322523, "learning_rate": 3.6420995491283274e-06, "loss": 1.2432, "step": 2141 }, { "epoch": 5.237163814180929, "grad_norm": 0.20666443048296643, "learning_rate": 3.6409606054156747e-06, "loss": 0.6603, "step": 2142 }, { "epoch": 5.23960880195599, "grad_norm": 0.1859391395608223, "learning_rate": 3.639821362517033e-06, "loss": 0.6153, "step": 2143 }, { "epoch": 5.2420537897310515, "grad_norm": 0.16753322963300563, "learning_rate": 3.638681820731139e-06, "loss": 0.5872, "step": 2144 }, { "epoch": 5.244498777506112, "grad_norm": 0.2258477684421268, "learning_rate": 3.637541980356805e-06, "loss": 0.5507, "step": 2145 }, { "epoch": 5.2469437652811735, "grad_norm": 0.1784607257063599, "learning_rate": 3.6364018416929236e-06, "loss": 0.5821, "step": 2146 }, { "epoch": 5.249388753056235, "grad_norm": 0.15719212308487598, "learning_rate": 3.6352614050384665e-06, "loss": 0.438, "step": 2147 }, { "epoch": 5.251833740831296, "grad_norm": 0.31697331904796755, "learning_rate": 3.6341206706924802e-06, "loss": 1.1717, "step": 2148 }, { "epoch": 5.254278728606357, "grad_norm": 0.22757953455097363, "learning_rate": 3.6329796389540935e-06, "loss": 0.6083, "step": 2149 }, { "epoch": 5.256723716381418, "grad_norm": 0.25453999116716636, "learning_rate": 3.6318383101225097e-06, "loss": 0.6664, "step": 2150 }, { "epoch": 5.25916870415648, "grad_norm": 0.27613469474801966, "learning_rate": 3.630696684497011e-06, "loss": 0.8694, "step": 2151 }, { "epoch": 5.26161369193154, "grad_norm": 0.21425127867120816, "learning_rate": 3.629554762376959e-06, "loss": 0.8508, "step": 2152 }, { "epoch": 5.264058679706602, "grad_norm": 0.15046011626706285, "learning_rate": 3.628412544061791e-06, "loss": 0.8085, "step": 2153 }, { "epoch": 5.266503667481663, "grad_norm": 0.17180039915197945, "learning_rate": 3.6272700298510233e-06, "loss": 0.8843, "step": 2154 }, { "epoch": 5.2689486552567235, "grad_norm": 0.18419989808008455, "learning_rate": 3.626127220044249e-06, "loss": 0.5409, "step": 2155 }, { "epoch": 5.271393643031785, "grad_norm": 0.29349251179461483, "learning_rate": 3.6249841149411385e-06, "loss": 0.4038, "step": 2156 }, { "epoch": 5.273838630806846, "grad_norm": 0.19601519037995702, "learning_rate": 3.6238407148414413e-06, "loss": 0.4002, "step": 2157 }, { "epoch": 5.276283618581907, "grad_norm": 0.21844047622123686, "learning_rate": 3.6226970200449824e-06, "loss": 0.5217, "step": 2158 }, { "epoch": 5.278728606356968, "grad_norm": 0.22321897660964074, "learning_rate": 3.621553030851664e-06, "loss": 0.6151, "step": 2159 }, { "epoch": 5.28117359413203, "grad_norm": 0.20922811172847908, "learning_rate": 3.620408747561468e-06, "loss": 0.3649, "step": 2160 }, { "epoch": 5.28361858190709, "grad_norm": 0.23511713245526447, "learning_rate": 3.6192641704744495e-06, "loss": 0.569, "step": 2161 }, { "epoch": 5.286063569682152, "grad_norm": 0.21968237424631146, "learning_rate": 3.6181192998907453e-06, "loss": 0.3507, "step": 2162 }, { "epoch": 5.288508557457213, "grad_norm": 0.18599691794295706, "learning_rate": 3.616974136110565e-06, "loss": 0.6965, "step": 2163 }, { "epoch": 5.290953545232274, "grad_norm": 0.25258341891239905, "learning_rate": 3.615828679434197e-06, "loss": 0.6335, "step": 2164 }, { "epoch": 5.293398533007335, "grad_norm": 0.1727851775207428, "learning_rate": 3.614682930162006e-06, "loss": 0.4426, "step": 2165 }, { "epoch": 5.295843520782396, "grad_norm": 0.21024083727583648, "learning_rate": 3.6135368885944344e-06, "loss": 0.5258, "step": 2166 }, { "epoch": 5.298288508557457, "grad_norm": 0.14742386508035366, "learning_rate": 3.6123905550320006e-06, "loss": 0.4863, "step": 2167 }, { "epoch": 5.300733496332518, "grad_norm": 0.1802107637301803, "learning_rate": 3.611243929775299e-06, "loss": 0.6268, "step": 2168 }, { "epoch": 5.30317848410758, "grad_norm": 0.2764852618844023, "learning_rate": 3.6100970131250013e-06, "loss": 0.6348, "step": 2169 }, { "epoch": 5.30562347188264, "grad_norm": 0.2430505816519341, "learning_rate": 3.6089498053818543e-06, "loss": 0.4831, "step": 2170 }, { "epoch": 5.308068459657702, "grad_norm": 0.232395854986584, "learning_rate": 3.6078023068466832e-06, "loss": 0.4745, "step": 2171 }, { "epoch": 5.310513447432763, "grad_norm": 0.20798738703477299, "learning_rate": 3.606654517820389e-06, "loss": 0.4821, "step": 2172 }, { "epoch": 5.312958435207824, "grad_norm": 0.30277682647660076, "learning_rate": 3.6055064386039467e-06, "loss": 0.3754, "step": 2173 }, { "epoch": 5.315403422982885, "grad_norm": 0.19687269394567342, "learning_rate": 3.60435806949841e-06, "loss": 0.4324, "step": 2174 }, { "epoch": 5.3178484107579465, "grad_norm": 0.24015875000730597, "learning_rate": 3.6032094108049063e-06, "loss": 0.4981, "step": 2175 }, { "epoch": 5.320293398533007, "grad_norm": 0.18127294191034132, "learning_rate": 3.6020604628246418e-06, "loss": 0.5015, "step": 2176 }, { "epoch": 5.322738386308068, "grad_norm": 0.28965796298554247, "learning_rate": 3.600911225858896e-06, "loss": 0.449, "step": 2177 }, { "epoch": 5.32518337408313, "grad_norm": 0.2591795274334654, "learning_rate": 3.5997617002090264e-06, "loss": 0.6724, "step": 2178 }, { "epoch": 5.32762836185819, "grad_norm": 0.22570884879404238, "learning_rate": 3.598611886176463e-06, "loss": 0.4894, "step": 2179 }, { "epoch": 5.330073349633252, "grad_norm": 0.13608595120491174, "learning_rate": 3.5974617840627145e-06, "loss": 0.382, "step": 2180 }, { "epoch": 5.332518337408313, "grad_norm": 0.21459449247210435, "learning_rate": 3.596311394169363e-06, "loss": 0.5245, "step": 2181 }, { "epoch": 5.334963325183374, "grad_norm": 0.1630110178679035, "learning_rate": 3.5951607167980695e-06, "loss": 0.4677, "step": 2182 }, { "epoch": 5.337408312958435, "grad_norm": 0.20975214105427356, "learning_rate": 3.5940097522505655e-06, "loss": 0.631, "step": 2183 }, { "epoch": 5.3398533007334965, "grad_norm": 0.19356520671954475, "learning_rate": 3.5928585008286614e-06, "loss": 0.9404, "step": 2184 }, { "epoch": 5.342298288508557, "grad_norm": 0.21928050038486538, "learning_rate": 3.591706962834241e-06, "loss": 0.6218, "step": 2185 }, { "epoch": 5.3447432762836184, "grad_norm": 0.1862271096654394, "learning_rate": 3.590555138569264e-06, "loss": 0.3539, "step": 2186 }, { "epoch": 5.34718826405868, "grad_norm": 0.18313391349123137, "learning_rate": 3.5894030283357663e-06, "loss": 0.3692, "step": 2187 }, { "epoch": 5.34963325183374, "grad_norm": 0.20052656395762344, "learning_rate": 3.5882506324358557e-06, "loss": 0.4359, "step": 2188 }, { "epoch": 5.352078239608802, "grad_norm": 0.2085014969718404, "learning_rate": 3.5870979511717184e-06, "loss": 0.6994, "step": 2189 }, { "epoch": 5.354523227383863, "grad_norm": 0.2930754367421236, "learning_rate": 3.585944984845613e-06, "loss": 0.6871, "step": 2190 }, { "epoch": 5.356968215158924, "grad_norm": 0.29896306658933597, "learning_rate": 3.584791733759873e-06, "loss": 0.4254, "step": 2191 }, { "epoch": 5.359413202933985, "grad_norm": 0.18337083746557345, "learning_rate": 3.5836381982169088e-06, "loss": 0.4719, "step": 2192 }, { "epoch": 5.361858190709047, "grad_norm": 0.20541684678338798, "learning_rate": 3.582484378519203e-06, "loss": 0.3929, "step": 2193 }, { "epoch": 5.364303178484108, "grad_norm": 0.20922361278730478, "learning_rate": 3.5813302749693143e-06, "loss": 0.7873, "step": 2194 }, { "epoch": 5.3667481662591685, "grad_norm": 0.17489750910655175, "learning_rate": 3.5801758878698735e-06, "loss": 0.5274, "step": 2195 }, { "epoch": 5.36919315403423, "grad_norm": 0.13452382413424466, "learning_rate": 3.5790212175235885e-06, "loss": 0.4157, "step": 2196 }, { "epoch": 5.371638141809291, "grad_norm": 0.14172307205017176, "learning_rate": 3.5778662642332395e-06, "loss": 0.4088, "step": 2197 }, { "epoch": 5.374083129584352, "grad_norm": 0.33251437361504943, "learning_rate": 3.5767110283016824e-06, "loss": 0.4245, "step": 2198 }, { "epoch": 5.376528117359413, "grad_norm": 0.2661282517138751, "learning_rate": 3.5755555100318466e-06, "loss": 0.4009, "step": 2199 }, { "epoch": 5.378973105134475, "grad_norm": 0.32569982209414106, "learning_rate": 3.5743997097267345e-06, "loss": 0.8575, "step": 2200 }, { "epoch": 5.381418092909535, "grad_norm": 0.2802546545487847, "learning_rate": 3.573243627689425e-06, "loss": 0.7675, "step": 2201 }, { "epoch": 5.383863080684597, "grad_norm": 0.2809570042766992, "learning_rate": 3.5720872642230674e-06, "loss": 0.6308, "step": 2202 }, { "epoch": 5.386308068459658, "grad_norm": 0.13076090260361758, "learning_rate": 3.5709306196308885e-06, "loss": 0.4768, "step": 2203 }, { "epoch": 5.388753056234719, "grad_norm": 0.21494050203518575, "learning_rate": 3.569773694216186e-06, "loss": 0.4779, "step": 2204 }, { "epoch": 5.39119804400978, "grad_norm": 0.18163821000288635, "learning_rate": 3.5686164882823314e-06, "loss": 0.6205, "step": 2205 }, { "epoch": 5.393643031784841, "grad_norm": 0.290270247174451, "learning_rate": 3.5674590021327723e-06, "loss": 0.4559, "step": 2206 }, { "epoch": 5.396088019559902, "grad_norm": 0.2746632414942217, "learning_rate": 3.5663012360710266e-06, "loss": 0.4281, "step": 2207 }, { "epoch": 5.398533007334963, "grad_norm": 0.3135986818857818, "learning_rate": 3.565143190400688e-06, "loss": 0.4237, "step": 2208 }, { "epoch": 5.400977995110025, "grad_norm": 0.22444419257332893, "learning_rate": 3.563984865425422e-06, "loss": 0.7947, "step": 2209 }, { "epoch": 5.403422982885085, "grad_norm": 0.2714304106414512, "learning_rate": 3.5628262614489684e-06, "loss": 1.6441, "step": 2210 }, { "epoch": 5.405867970660147, "grad_norm": 0.17090607325258253, "learning_rate": 3.56166737877514e-06, "loss": 0.382, "step": 2211 }, { "epoch": 5.408312958435208, "grad_norm": 0.2359439768148487, "learning_rate": 3.5605082177078215e-06, "loss": 0.5672, "step": 2212 }, { "epoch": 5.410757946210269, "grad_norm": 0.2053767338420804, "learning_rate": 3.5593487785509717e-06, "loss": 0.493, "step": 2213 }, { "epoch": 5.41320293398533, "grad_norm": 0.3124420531563946, "learning_rate": 3.5581890616086225e-06, "loss": 0.4293, "step": 2214 }, { "epoch": 5.4156479217603914, "grad_norm": 0.1826239169923515, "learning_rate": 3.5570290671848773e-06, "loss": 0.7154, "step": 2215 }, { "epoch": 5.418092909535452, "grad_norm": 0.27511127399033364, "learning_rate": 3.5558687955839146e-06, "loss": 0.5191, "step": 2216 }, { "epoch": 5.420537897310513, "grad_norm": 0.2598960192566311, "learning_rate": 3.554708247109984e-06, "loss": 0.648, "step": 2217 }, { "epoch": 5.422982885085575, "grad_norm": 0.24429128901864908, "learning_rate": 3.5535474220674054e-06, "loss": 0.3936, "step": 2218 }, { "epoch": 5.425427872860635, "grad_norm": 0.22226652699452956, "learning_rate": 3.5523863207605767e-06, "loss": 0.6975, "step": 2219 }, { "epoch": 5.427872860635697, "grad_norm": 0.20384416918918755, "learning_rate": 3.5512249434939637e-06, "loss": 0.9031, "step": 2220 }, { "epoch": 5.430317848410758, "grad_norm": 0.2581983614004652, "learning_rate": 3.5500632905721074e-06, "loss": 1.2965, "step": 2221 }, { "epoch": 5.432762836185819, "grad_norm": 0.15815602385438232, "learning_rate": 3.548901362299618e-06, "loss": 0.4571, "step": 2222 }, { "epoch": 5.43520782396088, "grad_norm": 0.23083962056025792, "learning_rate": 3.5477391589811817e-06, "loss": 1.0567, "step": 2223 }, { "epoch": 5.4376528117359415, "grad_norm": 0.1943489918572033, "learning_rate": 3.546576680921553e-06, "loss": 0.4071, "step": 2224 }, { "epoch": 5.440097799511002, "grad_norm": 0.19311107817517553, "learning_rate": 3.5454139284255605e-06, "loss": 0.5258, "step": 2225 }, { "epoch": 5.442542787286063, "grad_norm": 0.24669553035443006, "learning_rate": 3.544250901798106e-06, "loss": 0.6146, "step": 2226 }, { "epoch": 5.444987775061125, "grad_norm": 0.2187002169037261, "learning_rate": 3.5430876013441607e-06, "loss": 1.167, "step": 2227 }, { "epoch": 5.447432762836186, "grad_norm": 0.24262848916232485, "learning_rate": 3.5419240273687684e-06, "loss": 1.0638, "step": 2228 }, { "epoch": 5.449877750611247, "grad_norm": 0.275059167435587, "learning_rate": 3.5407601801770443e-06, "loss": 0.7149, "step": 2229 }, { "epoch": 5.452322738386308, "grad_norm": 0.23862185311846906, "learning_rate": 3.539596060074177e-06, "loss": 1.1739, "step": 2230 }, { "epoch": 5.45476772616137, "grad_norm": 0.30645879840030904, "learning_rate": 3.538431667365425e-06, "loss": 0.5851, "step": 2231 }, { "epoch": 5.45721271393643, "grad_norm": 0.16890222894354093, "learning_rate": 3.537267002356119e-06, "loss": 0.5203, "step": 2232 }, { "epoch": 5.459657701711492, "grad_norm": 0.17209024615549923, "learning_rate": 3.536102065351661e-06, "loss": 0.8368, "step": 2233 }, { "epoch": 5.462102689486553, "grad_norm": 0.18832669033785032, "learning_rate": 3.5349368566575226e-06, "loss": 0.4282, "step": 2234 }, { "epoch": 5.4645476772616135, "grad_norm": 0.23389315248914386, "learning_rate": 3.5337713765792493e-06, "loss": 0.5663, "step": 2235 }, { "epoch": 5.466992665036675, "grad_norm": 0.21298358173669918, "learning_rate": 3.532605625422457e-06, "loss": 0.652, "step": 2236 }, { "epoch": 5.469437652811736, "grad_norm": 0.22462721438526495, "learning_rate": 3.531439603492832e-06, "loss": 0.6973, "step": 2237 }, { "epoch": 5.471882640586797, "grad_norm": 0.20557470482948537, "learning_rate": 3.5302733110961316e-06, "loss": 0.4573, "step": 2238 }, { "epoch": 5.474327628361858, "grad_norm": 0.23949063014426908, "learning_rate": 3.5291067485381854e-06, "loss": 0.413, "step": 2239 }, { "epoch": 5.47677261613692, "grad_norm": 0.22312051049735845, "learning_rate": 3.5279399161248904e-06, "loss": 0.3989, "step": 2240 }, { "epoch": 5.47921760391198, "grad_norm": 0.16551310206286843, "learning_rate": 3.5267728141622205e-06, "loss": 0.7857, "step": 2241 }, { "epoch": 5.481662591687042, "grad_norm": 0.19896641660693515, "learning_rate": 3.5256054429562136e-06, "loss": 0.7117, "step": 2242 }, { "epoch": 5.484107579462103, "grad_norm": 0.2758334907879952, "learning_rate": 3.524437802812982e-06, "loss": 0.3938, "step": 2243 }, { "epoch": 5.486552567237164, "grad_norm": 0.20712906497050476, "learning_rate": 3.5232698940387078e-06, "loss": 0.4382, "step": 2244 }, { "epoch": 5.488997555012225, "grad_norm": 0.19328280667723455, "learning_rate": 3.5221017169396433e-06, "loss": 0.3808, "step": 2245 }, { "epoch": 5.491442542787286, "grad_norm": 0.25045253035904314, "learning_rate": 3.5209332718221112e-06, "loss": 0.8375, "step": 2246 }, { "epoch": 5.493887530562347, "grad_norm": 0.22663208500839396, "learning_rate": 3.519764558992505e-06, "loss": 0.7346, "step": 2247 }, { "epoch": 5.496332518337408, "grad_norm": 0.18821415071283903, "learning_rate": 3.518595578757288e-06, "loss": 0.3688, "step": 2248 }, { "epoch": 5.49877750611247, "grad_norm": 0.2773602702366634, "learning_rate": 3.5174263314229916e-06, "loss": 0.6316, "step": 2249 }, { "epoch": 5.50122249388753, "grad_norm": 0.1532534544817341, "learning_rate": 3.516256817296222e-06, "loss": 0.3511, "step": 2250 }, { "epoch": 5.503667481662592, "grad_norm": 0.17489945834887297, "learning_rate": 3.5150870366836513e-06, "loss": 0.7129, "step": 2251 }, { "epoch": 5.506112469437653, "grad_norm": 0.21826621247297684, "learning_rate": 3.513916989892023e-06, "loss": 0.4151, "step": 2252 }, { "epoch": 5.508557457212714, "grad_norm": 0.21626687541904638, "learning_rate": 3.51274667722815e-06, "loss": 0.6042, "step": 2253 }, { "epoch": 5.511002444987775, "grad_norm": 0.21136335403517106, "learning_rate": 3.5115760989989143e-06, "loss": 0.5028, "step": 2254 }, { "epoch": 5.513447432762836, "grad_norm": 0.22664864938328647, "learning_rate": 3.51040525551127e-06, "loss": 0.5551, "step": 2255 }, { "epoch": 5.515892420537897, "grad_norm": 0.2856589393689792, "learning_rate": 3.509234147072238e-06, "loss": 0.4004, "step": 2256 }, { "epoch": 5.518337408312958, "grad_norm": 0.16216507863084234, "learning_rate": 3.5080627739889096e-06, "loss": 0.3514, "step": 2257 }, { "epoch": 5.52078239608802, "grad_norm": 0.18189162316301738, "learning_rate": 3.5068911365684462e-06, "loss": 0.4052, "step": 2258 }, { "epoch": 5.52322738386308, "grad_norm": 0.22215664723077658, "learning_rate": 3.505719235118077e-06, "loss": 0.6929, "step": 2259 }, { "epoch": 5.525672371638142, "grad_norm": 0.2012573269615039, "learning_rate": 3.504547069945103e-06, "loss": 0.6347, "step": 2260 }, { "epoch": 5.528117359413203, "grad_norm": 0.16888406794469107, "learning_rate": 3.503374641356891e-06, "loss": 0.6199, "step": 2261 }, { "epoch": 5.530562347188264, "grad_norm": 0.22464729008898413, "learning_rate": 3.50220194966088e-06, "loss": 0.6904, "step": 2262 }, { "epoch": 5.533007334963325, "grad_norm": 0.19609021693176726, "learning_rate": 3.501028995164576e-06, "loss": 0.4726, "step": 2263 }, { "epoch": 5.5354523227383865, "grad_norm": 0.2156451675238026, "learning_rate": 3.499855778175554e-06, "loss": 0.339, "step": 2264 }, { "epoch": 5.537897310513447, "grad_norm": 0.2347227027702509, "learning_rate": 3.4986822990014598e-06, "loss": 0.692, "step": 2265 }, { "epoch": 5.540342298288508, "grad_norm": 0.1787402255681198, "learning_rate": 3.4975085579500046e-06, "loss": 0.4092, "step": 2266 }, { "epoch": 5.54278728606357, "grad_norm": 0.21335863348260645, "learning_rate": 3.496334555328972e-06, "loss": 0.4477, "step": 2267 }, { "epoch": 5.54523227383863, "grad_norm": 0.33863929079874755, "learning_rate": 3.495160291446211e-06, "loss": 0.4267, "step": 2268 }, { "epoch": 5.547677261613692, "grad_norm": 0.19779648068455502, "learning_rate": 3.493985766609641e-06, "loss": 0.4418, "step": 2269 }, { "epoch": 5.550122249388753, "grad_norm": 0.2354920358820682, "learning_rate": 3.492810981127249e-06, "loss": 0.5292, "step": 2270 }, { "epoch": 5.552567237163814, "grad_norm": 0.21692553317605812, "learning_rate": 3.4916359353070924e-06, "loss": 1.1664, "step": 2271 }, { "epoch": 5.555012224938875, "grad_norm": 0.19653955060153305, "learning_rate": 3.490460629457292e-06, "loss": 0.7786, "step": 2272 }, { "epoch": 5.557457212713937, "grad_norm": 0.19393087272557635, "learning_rate": 3.4892850638860425e-06, "loss": 0.5937, "step": 2273 }, { "epoch": 5.559902200488997, "grad_norm": 0.2249400587845235, "learning_rate": 3.4881092389016026e-06, "loss": 0.4538, "step": 2274 }, { "epoch": 5.5623471882640585, "grad_norm": 0.21143317314017557, "learning_rate": 3.4869331548123014e-06, "loss": 0.4862, "step": 2275 }, { "epoch": 5.56479217603912, "grad_norm": 0.3041515457346303, "learning_rate": 3.485756811926535e-06, "loss": 1.0056, "step": 2276 }, { "epoch": 5.56723716381418, "grad_norm": 0.13574752049043348, "learning_rate": 3.484580210552767e-06, "loss": 0.4825, "step": 2277 }, { "epoch": 5.569682151589242, "grad_norm": 0.15425320873268222, "learning_rate": 3.483403350999529e-06, "loss": 0.469, "step": 2278 }, { "epoch": 5.572127139364303, "grad_norm": 0.1868166851290588, "learning_rate": 3.482226233575421e-06, "loss": 0.3887, "step": 2279 }, { "epoch": 5.574572127139365, "grad_norm": 0.3065152082082507, "learning_rate": 3.4810488585891108e-06, "loss": 0.6598, "step": 2280 }, { "epoch": 5.577017114914425, "grad_norm": 0.18584043951062992, "learning_rate": 3.4798712263493316e-06, "loss": 0.3838, "step": 2281 }, { "epoch": 5.579462102689487, "grad_norm": 0.18302250257697428, "learning_rate": 3.478693337164887e-06, "loss": 0.4542, "step": 2282 }, { "epoch": 5.581907090464548, "grad_norm": 0.2198460604033387, "learning_rate": 3.477515191344645e-06, "loss": 0.4596, "step": 2283 }, { "epoch": 5.5843520782396086, "grad_norm": 0.2589073095649969, "learning_rate": 3.476336789197543e-06, "loss": 0.6148, "step": 2284 }, { "epoch": 5.58679706601467, "grad_norm": 0.25995060385363744, "learning_rate": 3.475158131032586e-06, "loss": 0.5695, "step": 2285 }, { "epoch": 5.589242053789731, "grad_norm": 0.19766335673366794, "learning_rate": 3.4739792171588445e-06, "loss": 0.4749, "step": 2286 }, { "epoch": 5.591687041564792, "grad_norm": 0.2783649648695533, "learning_rate": 3.4728000478854564e-06, "loss": 0.6517, "step": 2287 }, { "epoch": 5.594132029339853, "grad_norm": 0.2536550589136312, "learning_rate": 3.471620623521626e-06, "loss": 0.4347, "step": 2288 }, { "epoch": 5.596577017114915, "grad_norm": 0.2733002774272313, "learning_rate": 3.470440944376626e-06, "loss": 0.697, "step": 2289 }, { "epoch": 5.599022004889975, "grad_norm": 0.18664770028280567, "learning_rate": 3.4692610107597972e-06, "loss": 0.4176, "step": 2290 }, { "epoch": 5.601466992665037, "grad_norm": 0.17207840144638994, "learning_rate": 3.4680808229805435e-06, "loss": 0.4735, "step": 2291 }, { "epoch": 5.603911980440098, "grad_norm": 0.16852075324321733, "learning_rate": 3.466900381348337e-06, "loss": 0.4496, "step": 2292 }, { "epoch": 5.606356968215159, "grad_norm": 0.2238344788411242, "learning_rate": 3.4657196861727165e-06, "loss": 0.518, "step": 2293 }, { "epoch": 5.60880195599022, "grad_norm": 0.19941653823702699, "learning_rate": 3.464538737763288e-06, "loss": 0.5941, "step": 2294 }, { "epoch": 5.611246943765281, "grad_norm": 0.205900444381756, "learning_rate": 3.463357536429723e-06, "loss": 0.4351, "step": 2295 }, { "epoch": 5.613691931540342, "grad_norm": 0.23487522768185495, "learning_rate": 3.4621760824817596e-06, "loss": 0.9335, "step": 2296 }, { "epoch": 5.616136919315403, "grad_norm": 0.17050435137936504, "learning_rate": 3.4609943762292015e-06, "loss": 0.6002, "step": 2297 }, { "epoch": 5.618581907090465, "grad_norm": 0.1711676849426343, "learning_rate": 3.459812417981919e-06, "loss": 0.6122, "step": 2298 }, { "epoch": 5.621026894865525, "grad_norm": 0.22733038272996806, "learning_rate": 3.458630208049849e-06, "loss": 0.822, "step": 2299 }, { "epoch": 5.623471882640587, "grad_norm": 0.19948333506378357, "learning_rate": 3.457447746742995e-06, "loss": 0.6581, "step": 2300 }, { "epoch": 5.625916870415648, "grad_norm": 0.1325592647462057, "learning_rate": 3.4562650343714242e-06, "loss": 0.6559, "step": 2301 }, { "epoch": 5.628361858190709, "grad_norm": 0.31221786315430433, "learning_rate": 3.4550820712452716e-06, "loss": 0.5888, "step": 2302 }, { "epoch": 5.63080684596577, "grad_norm": 0.22947079647496033, "learning_rate": 3.4538988576747355e-06, "loss": 0.5025, "step": 2303 }, { "epoch": 5.6332518337408315, "grad_norm": 0.15431899545948138, "learning_rate": 3.4527153939700838e-06, "loss": 0.5115, "step": 2304 }, { "epoch": 5.635696821515893, "grad_norm": 0.2189846114351366, "learning_rate": 3.4515316804416464e-06, "loss": 0.5722, "step": 2305 }, { "epoch": 5.638141809290953, "grad_norm": 0.2011928578909688, "learning_rate": 3.450347717399821e-06, "loss": 0.398, "step": 2306 }, { "epoch": 5.640586797066015, "grad_norm": 0.16117320531938067, "learning_rate": 3.4491635051550697e-06, "loss": 0.3987, "step": 2307 }, { "epoch": 5.643031784841076, "grad_norm": 0.3559818285440626, "learning_rate": 3.447979044017918e-06, "loss": 0.3819, "step": 2308 }, { "epoch": 5.645476772616137, "grad_norm": 0.17453653781201287, "learning_rate": 3.4467943342989617e-06, "loss": 0.7916, "step": 2309 }, { "epoch": 5.647921760391198, "grad_norm": 0.22425214333448937, "learning_rate": 3.4456093763088577e-06, "loss": 0.8948, "step": 2310 }, { "epoch": 5.65036674816626, "grad_norm": 0.20817775833429783, "learning_rate": 3.444424170358329e-06, "loss": 0.365, "step": 2311 }, { "epoch": 5.65281173594132, "grad_norm": 0.3149536892096343, "learning_rate": 3.4432387167581637e-06, "loss": 0.3978, "step": 2312 }, { "epoch": 5.6552567237163816, "grad_norm": 0.23229871051620887, "learning_rate": 3.442053015819214e-06, "loss": 0.4871, "step": 2313 }, { "epoch": 5.657701711491443, "grad_norm": 0.23252590902316755, "learning_rate": 3.4408670678523994e-06, "loss": 0.5863, "step": 2314 }, { "epoch": 5.6601466992665035, "grad_norm": 0.24001265838313057, "learning_rate": 3.439680873168701e-06, "loss": 0.6851, "step": 2315 }, { "epoch": 5.662591687041565, "grad_norm": 0.15895325123358917, "learning_rate": 3.438494432079168e-06, "loss": 0.5993, "step": 2316 }, { "epoch": 5.665036674816626, "grad_norm": 0.357380995846018, "learning_rate": 3.437307744894911e-06, "loss": 0.4389, "step": 2317 }, { "epoch": 5.667481662591687, "grad_norm": 0.15978268934021794, "learning_rate": 3.4361208119271074e-06, "loss": 0.4446, "step": 2318 }, { "epoch": 5.669926650366748, "grad_norm": 0.19611973995521184, "learning_rate": 3.4349336334869976e-06, "loss": 0.4373, "step": 2319 }, { "epoch": 5.67237163814181, "grad_norm": 0.18070571348711656, "learning_rate": 3.433746209885887e-06, "loss": 0.4942, "step": 2320 }, { "epoch": 5.67481662591687, "grad_norm": 0.281780647619913, "learning_rate": 3.432558541435146e-06, "loss": 0.8002, "step": 2321 }, { "epoch": 5.677261613691932, "grad_norm": 0.2535031880235516, "learning_rate": 3.4313706284462077e-06, "loss": 0.3914, "step": 2322 }, { "epoch": 5.679706601466993, "grad_norm": 0.24631698730373758, "learning_rate": 3.43018247123057e-06, "loss": 0.4547, "step": 2323 }, { "epoch": 5.6821515892420535, "grad_norm": 0.2523638451053032, "learning_rate": 3.4289940700997953e-06, "loss": 0.4387, "step": 2324 }, { "epoch": 5.684596577017115, "grad_norm": 0.1785939487744636, "learning_rate": 3.4278054253655092e-06, "loss": 0.8406, "step": 2325 }, { "epoch": 5.687041564792176, "grad_norm": 0.18286204645114806, "learning_rate": 3.4266165373394024e-06, "loss": 0.3072, "step": 2326 }, { "epoch": 5.689486552567237, "grad_norm": 0.23259953466410532, "learning_rate": 3.425427406333228e-06, "loss": 0.4584, "step": 2327 }, { "epoch": 5.691931540342298, "grad_norm": 0.1034310274458396, "learning_rate": 3.424238032658803e-06, "loss": 0.2987, "step": 2328 }, { "epoch": 5.69437652811736, "grad_norm": 0.29110742561310504, "learning_rate": 3.423048416628009e-06, "loss": 0.4139, "step": 2329 }, { "epoch": 5.69682151589242, "grad_norm": 0.30323392285102774, "learning_rate": 3.42185855855279e-06, "loss": 0.4489, "step": 2330 }, { "epoch": 5.699266503667482, "grad_norm": 0.2699791636125737, "learning_rate": 3.4206684587451542e-06, "loss": 0.6126, "step": 2331 }, { "epoch": 5.701711491442543, "grad_norm": 0.21679526599851004, "learning_rate": 3.419478117517174e-06, "loss": 0.4635, "step": 2332 }, { "epoch": 5.704156479217604, "grad_norm": 0.28562832309040437, "learning_rate": 3.418287535180982e-06, "loss": 0.7509, "step": 2333 }, { "epoch": 5.706601466992665, "grad_norm": 0.20567200023337673, "learning_rate": 3.4170967120487787e-06, "loss": 0.5643, "step": 2334 }, { "epoch": 5.709046454767726, "grad_norm": 0.16727561433355023, "learning_rate": 3.4159056484328236e-06, "loss": 0.5666, "step": 2335 }, { "epoch": 5.711491442542787, "grad_norm": 0.21633055378235785, "learning_rate": 3.41471434464544e-06, "loss": 0.5114, "step": 2336 }, { "epoch": 5.713936430317848, "grad_norm": 0.22522864376940144, "learning_rate": 3.4135228009990175e-06, "loss": 0.4444, "step": 2337 }, { "epoch": 5.71638141809291, "grad_norm": 0.1595446128902458, "learning_rate": 3.4123310178060045e-06, "loss": 0.4456, "step": 2338 }, { "epoch": 5.71882640586797, "grad_norm": 0.2513658426774408, "learning_rate": 3.4111389953789138e-06, "loss": 0.4686, "step": 2339 }, { "epoch": 5.721271393643032, "grad_norm": 0.22831297887049987, "learning_rate": 3.4099467340303217e-06, "loss": 0.6991, "step": 2340 }, { "epoch": 5.723716381418093, "grad_norm": 0.25064053731396574, "learning_rate": 3.4087542340728657e-06, "loss": 0.6031, "step": 2341 }, { "epoch": 5.726161369193154, "grad_norm": 0.2348242841412849, "learning_rate": 3.4075614958192464e-06, "loss": 0.9422, "step": 2342 }, { "epoch": 5.728606356968215, "grad_norm": 0.32432060976957644, "learning_rate": 3.4063685195822284e-06, "loss": 0.3888, "step": 2343 }, { "epoch": 5.7310513447432765, "grad_norm": 0.25680401587189, "learning_rate": 3.405175305674637e-06, "loss": 0.5747, "step": 2344 }, { "epoch": 5.733496332518337, "grad_norm": 0.16128843814426544, "learning_rate": 3.4039818544093587e-06, "loss": 0.5815, "step": 2345 }, { "epoch": 5.735941320293398, "grad_norm": 0.19500002871723887, "learning_rate": 3.4027881660993454e-06, "loss": 0.6205, "step": 2346 }, { "epoch": 5.73838630806846, "grad_norm": 0.16938211466865477, "learning_rate": 3.401594241057608e-06, "loss": 0.6935, "step": 2347 }, { "epoch": 5.74083129584352, "grad_norm": 0.27400129801822554, "learning_rate": 3.400400079597223e-06, "loss": 0.6052, "step": 2348 }, { "epoch": 5.743276283618582, "grad_norm": 0.19611389587401065, "learning_rate": 3.399205682031326e-06, "loss": 0.5777, "step": 2349 }, { "epoch": 5.745721271393643, "grad_norm": 0.17681117528036347, "learning_rate": 3.398011048673114e-06, "loss": 0.4115, "step": 2350 }, { "epoch": 5.748166259168704, "grad_norm": 0.2375043940666574, "learning_rate": 3.396816179835849e-06, "loss": 0.446, "step": 2351 }, { "epoch": 5.750611246943765, "grad_norm": 0.20185581146427406, "learning_rate": 3.395621075832851e-06, "loss": 0.6884, "step": 2352 }, { "epoch": 5.7530562347188265, "grad_norm": 0.24129838127533926, "learning_rate": 3.3944257369775063e-06, "loss": 0.4558, "step": 2353 }, { "epoch": 5.755501222493887, "grad_norm": 0.16402889712190388, "learning_rate": 3.3932301635832586e-06, "loss": 0.5296, "step": 2354 }, { "epoch": 5.7579462102689485, "grad_norm": 0.2113663175776669, "learning_rate": 3.392034355963615e-06, "loss": 0.5456, "step": 2355 }, { "epoch": 5.76039119804401, "grad_norm": 0.2918791855462405, "learning_rate": 3.390838314432143e-06, "loss": 0.5995, "step": 2356 }, { "epoch": 5.76283618581907, "grad_norm": 0.24424203231717198, "learning_rate": 3.3896420393024715e-06, "loss": 0.8061, "step": 2357 }, { "epoch": 5.765281173594132, "grad_norm": 0.18538240695553068, "learning_rate": 3.3884455308882934e-06, "loss": 0.4448, "step": 2358 }, { "epoch": 5.767726161369193, "grad_norm": 0.2745739035076842, "learning_rate": 3.387248789503359e-06, "loss": 0.8291, "step": 2359 }, { "epoch": 5.770171149144255, "grad_norm": 0.27461866254046174, "learning_rate": 3.386051815461482e-06, "loss": 0.5828, "step": 2360 }, { "epoch": 5.772616136919315, "grad_norm": 0.23155057276696558, "learning_rate": 3.384854609076536e-06, "loss": 0.6884, "step": 2361 }, { "epoch": 5.775061124694377, "grad_norm": 0.19867924418971772, "learning_rate": 3.3836571706624553e-06, "loss": 0.4309, "step": 2362 }, { "epoch": 5.777506112469438, "grad_norm": 0.2259343515177491, "learning_rate": 3.3824595005332367e-06, "loss": 0.267, "step": 2363 }, { "epoch": 5.7799511002444985, "grad_norm": 0.2029600668067855, "learning_rate": 3.381261599002936e-06, "loss": 0.5574, "step": 2364 }, { "epoch": 5.78239608801956, "grad_norm": 0.18730033859746603, "learning_rate": 3.3800634663856718e-06, "loss": 0.3505, "step": 2365 }, { "epoch": 5.784841075794621, "grad_norm": 0.21496789442092193, "learning_rate": 3.3788651029956205e-06, "loss": 0.6177, "step": 2366 }, { "epoch": 5.787286063569682, "grad_norm": 0.2821557809849861, "learning_rate": 3.37766650914702e-06, "loss": 0.775, "step": 2367 }, { "epoch": 5.789731051344743, "grad_norm": 0.20019550564283553, "learning_rate": 3.37646768515417e-06, "loss": 0.4221, "step": 2368 }, { "epoch": 5.792176039119805, "grad_norm": 0.2764469422138519, "learning_rate": 3.37526863133143e-06, "loss": 0.3791, "step": 2369 }, { "epoch": 5.794621026894865, "grad_norm": 0.15641115271093253, "learning_rate": 3.3740693479932184e-06, "loss": 0.5918, "step": 2370 }, { "epoch": 5.797066014669927, "grad_norm": 0.14546444337694192, "learning_rate": 3.372869835454015e-06, "loss": 0.4162, "step": 2371 }, { "epoch": 5.799511002444988, "grad_norm": 0.33527827936283466, "learning_rate": 3.37167009402836e-06, "loss": 0.6958, "step": 2372 }, { "epoch": 5.801955990220049, "grad_norm": 0.27792526545278035, "learning_rate": 3.3704701240308525e-06, "loss": 1.1211, "step": 2373 }, { "epoch": 5.80440097799511, "grad_norm": 0.14290643840602424, "learning_rate": 3.3692699257761517e-06, "loss": 0.433, "step": 2374 }, { "epoch": 5.806845965770171, "grad_norm": 0.19447193625277287, "learning_rate": 3.3680694995789788e-06, "loss": 0.5248, "step": 2375 }, { "epoch": 5.809290953545232, "grad_norm": 0.24049011102098605, "learning_rate": 3.3668688457541103e-06, "loss": 0.6636, "step": 2376 }, { "epoch": 5.811735941320293, "grad_norm": 0.2144616902603737, "learning_rate": 3.3656679646163876e-06, "loss": 0.7337, "step": 2377 }, { "epoch": 5.814180929095355, "grad_norm": 0.28839858018112746, "learning_rate": 3.3644668564807083e-06, "loss": 0.4367, "step": 2378 }, { "epoch": 5.816625916870415, "grad_norm": 0.22087501383301614, "learning_rate": 3.3632655216620298e-06, "loss": 0.9032, "step": 2379 }, { "epoch": 5.819070904645477, "grad_norm": 0.23609888178573912, "learning_rate": 3.3620639604753702e-06, "loss": 0.4583, "step": 2380 }, { "epoch": 5.821515892420538, "grad_norm": 0.3296580145779637, "learning_rate": 3.3608621732358065e-06, "loss": 0.4745, "step": 2381 }, { "epoch": 5.823960880195599, "grad_norm": 0.19935878199940832, "learning_rate": 3.359660160258475e-06, "loss": 0.5053, "step": 2382 }, { "epoch": 5.82640586797066, "grad_norm": 0.3043803220681292, "learning_rate": 3.3584579218585706e-06, "loss": 0.3901, "step": 2383 }, { "epoch": 5.8288508557457215, "grad_norm": 0.2235664326070787, "learning_rate": 3.357255458351347e-06, "loss": 0.7055, "step": 2384 }, { "epoch": 5.831295843520783, "grad_norm": 0.22455293351263333, "learning_rate": 3.3560527700521196e-06, "loss": 0.4853, "step": 2385 }, { "epoch": 5.833740831295843, "grad_norm": 0.1839662483921858, "learning_rate": 3.354849857276259e-06, "loss": 0.4479, "step": 2386 }, { "epoch": 5.836185819070905, "grad_norm": 0.16714724607182893, "learning_rate": 3.3536467203391978e-06, "loss": 0.4347, "step": 2387 }, { "epoch": 5.838630806845966, "grad_norm": 0.20050619843367853, "learning_rate": 3.3524433595564254e-06, "loss": 0.8465, "step": 2388 }, { "epoch": 5.841075794621027, "grad_norm": 0.20325125777278805, "learning_rate": 3.35123977524349e-06, "loss": 0.9144, "step": 2389 }, { "epoch": 5.843520782396088, "grad_norm": 0.18782007021554803, "learning_rate": 3.3500359677159993e-06, "loss": 0.6118, "step": 2390 }, { "epoch": 5.84596577017115, "grad_norm": 0.31675184819799923, "learning_rate": 3.3488319372896193e-06, "loss": 0.6424, "step": 2391 }, { "epoch": 5.84841075794621, "grad_norm": 0.21448991271551274, "learning_rate": 3.347627684280075e-06, "loss": 0.6228, "step": 2392 }, { "epoch": 5.8508557457212715, "grad_norm": 0.2339446009528034, "learning_rate": 3.3464232090031485e-06, "loss": 0.2827, "step": 2393 }, { "epoch": 5.853300733496333, "grad_norm": 0.26442900891524973, "learning_rate": 3.3452185117746803e-06, "loss": 0.7434, "step": 2394 }, { "epoch": 5.8557457212713935, "grad_norm": 0.23474950026903882, "learning_rate": 3.3440135929105704e-06, "loss": 0.6813, "step": 2395 }, { "epoch": 5.858190709046455, "grad_norm": 0.2750358149936989, "learning_rate": 3.3428084527267756e-06, "loss": 0.6497, "step": 2396 }, { "epoch": 5.860635696821516, "grad_norm": 0.28262994192473495, "learning_rate": 3.341603091539312e-06, "loss": 0.3396, "step": 2397 }, { "epoch": 5.863080684596577, "grad_norm": 0.3141255163866676, "learning_rate": 3.340397509664252e-06, "loss": 0.706, "step": 2398 }, { "epoch": 5.865525672371638, "grad_norm": 0.1986453908934641, "learning_rate": 3.3391917074177277e-06, "loss": 0.4916, "step": 2399 }, { "epoch": 5.8679706601467, "grad_norm": 0.39451721311496163, "learning_rate": 3.3379856851159263e-06, "loss": 0.4414, "step": 2400 }, { "epoch": 5.87041564792176, "grad_norm": 0.30057604698705626, "learning_rate": 3.336779443075096e-06, "loss": 0.5294, "step": 2401 }, { "epoch": 5.872860635696822, "grad_norm": 0.2523414290743578, "learning_rate": 3.3355729816115413e-06, "loss": 1.1975, "step": 2402 }, { "epoch": 5.875305623471883, "grad_norm": 0.21586962034332774, "learning_rate": 3.3343663010416235e-06, "loss": 0.4718, "step": 2403 }, { "epoch": 5.8777506112469435, "grad_norm": 0.23634880431323999, "learning_rate": 3.3331594016817613e-06, "loss": 0.7717, "step": 2404 }, { "epoch": 5.880195599022005, "grad_norm": 0.25680720406637997, "learning_rate": 3.331952283848432e-06, "loss": 0.4945, "step": 2405 }, { "epoch": 5.882640586797066, "grad_norm": 0.1953450736442293, "learning_rate": 3.3307449478581687e-06, "loss": 0.7775, "step": 2406 }, { "epoch": 5.885085574572127, "grad_norm": 0.22571656809171378, "learning_rate": 3.3295373940275637e-06, "loss": 0.5714, "step": 2407 }, { "epoch": 5.887530562347188, "grad_norm": 0.1916419682898357, "learning_rate": 3.3283296226732655e-06, "loss": 0.7468, "step": 2408 }, { "epoch": 5.88997555012225, "grad_norm": 0.2269492305225127, "learning_rate": 3.3271216341119775e-06, "loss": 0.4385, "step": 2409 }, { "epoch": 5.89242053789731, "grad_norm": 0.26957986247964194, "learning_rate": 3.325913428660463e-06, "loss": 0.7777, "step": 2410 }, { "epoch": 5.894865525672372, "grad_norm": 0.205830739107848, "learning_rate": 3.3247050066355414e-06, "loss": 0.6831, "step": 2411 }, { "epoch": 5.897310513447433, "grad_norm": 0.27129806025979647, "learning_rate": 3.323496368354089e-06, "loss": 0.9132, "step": 2412 }, { "epoch": 5.899755501222494, "grad_norm": 0.25437160611557746, "learning_rate": 3.322287514133038e-06, "loss": 0.4542, "step": 2413 }, { "epoch": 5.902200488997555, "grad_norm": 0.22100498812937824, "learning_rate": 3.3210784442893774e-06, "loss": 0.7659, "step": 2414 }, { "epoch": 5.904645476772616, "grad_norm": 0.21827314939109838, "learning_rate": 3.319869159140152e-06, "loss": 0.5557, "step": 2415 }, { "epoch": 5.907090464547677, "grad_norm": 0.21848556949574652, "learning_rate": 3.318659659002466e-06, "loss": 0.4069, "step": 2416 }, { "epoch": 5.909535452322738, "grad_norm": 0.24273525461267176, "learning_rate": 3.3174499441934773e-06, "loss": 0.6805, "step": 2417 }, { "epoch": 5.9119804400978, "grad_norm": 0.2248193183476636, "learning_rate": 3.316240015030401e-06, "loss": 1.1249, "step": 2418 }, { "epoch": 5.91442542787286, "grad_norm": 0.30453235129947015, "learning_rate": 3.3150298718305074e-06, "loss": 0.4837, "step": 2419 }, { "epoch": 5.916870415647922, "grad_norm": 0.21179115452327044, "learning_rate": 3.313819514911124e-06, "loss": 1.205, "step": 2420 }, { "epoch": 5.919315403422983, "grad_norm": 0.3438382236206335, "learning_rate": 3.312608944589635e-06, "loss": 0.5912, "step": 2421 }, { "epoch": 5.921760391198044, "grad_norm": 0.27379111103155923, "learning_rate": 3.3113981611834777e-06, "loss": 0.5731, "step": 2422 }, { "epoch": 5.924205378973105, "grad_norm": 0.2283416150203339, "learning_rate": 3.3101871650101504e-06, "loss": 0.5076, "step": 2423 }, { "epoch": 5.9266503667481665, "grad_norm": 0.13232700038212505, "learning_rate": 3.308975956387201e-06, "loss": 0.3935, "step": 2424 }, { "epoch": 5.929095354523227, "grad_norm": 0.1916659782669626, "learning_rate": 3.3077645356322374e-06, "loss": 0.5252, "step": 2425 }, { "epoch": 5.931540342298288, "grad_norm": 0.1995413766740241, "learning_rate": 3.3065529030629216e-06, "loss": 0.3891, "step": 2426 }, { "epoch": 5.93398533007335, "grad_norm": 0.22823590683835818, "learning_rate": 3.3053410589969713e-06, "loss": 0.6494, "step": 2427 }, { "epoch": 5.93643031784841, "grad_norm": 0.28847211065857165, "learning_rate": 3.3041290037521607e-06, "loss": 1.2323, "step": 2428 }, { "epoch": 5.938875305623472, "grad_norm": 0.20218801050946034, "learning_rate": 3.3029167376463177e-06, "loss": 0.4479, "step": 2429 }, { "epoch": 5.941320293398533, "grad_norm": 0.1497490467100729, "learning_rate": 3.3017042609973253e-06, "loss": 0.4546, "step": 2430 }, { "epoch": 5.943765281173594, "grad_norm": 0.35081121438773644, "learning_rate": 3.300491574123125e-06, "loss": 0.6256, "step": 2431 }, { "epoch": 5.946210268948655, "grad_norm": 0.16232737418186707, "learning_rate": 3.2992786773417083e-06, "loss": 0.5296, "step": 2432 }, { "epoch": 5.9486552567237165, "grad_norm": 0.24893341466363325, "learning_rate": 3.2980655709711263e-06, "loss": 0.53, "step": 2433 }, { "epoch": 5.951100244498777, "grad_norm": 0.20435792058729257, "learning_rate": 3.296852255329483e-06, "loss": 0.4565, "step": 2434 }, { "epoch": 5.9535452322738385, "grad_norm": 0.25615497450801267, "learning_rate": 3.295638730734937e-06, "loss": 0.4393, "step": 2435 }, { "epoch": 5.9559902200489, "grad_norm": 0.25130964805989187, "learning_rate": 3.2944249975057025e-06, "loss": 0.5311, "step": 2436 }, { "epoch": 5.95843520782396, "grad_norm": 0.24047857468989378, "learning_rate": 3.2932110559600484e-06, "loss": 0.7289, "step": 2437 }, { "epoch": 5.960880195599022, "grad_norm": 0.23370984597690975, "learning_rate": 3.2919969064162976e-06, "loss": 0.47, "step": 2438 }, { "epoch": 5.963325183374083, "grad_norm": 0.1543545392566495, "learning_rate": 3.290782549192828e-06, "loss": 0.4325, "step": 2439 }, { "epoch": 5.965770171149144, "grad_norm": 0.31025893700427754, "learning_rate": 3.289567984608072e-06, "loss": 0.4936, "step": 2440 }, { "epoch": 5.968215158924205, "grad_norm": 0.25288069852872647, "learning_rate": 3.2883532129805163e-06, "loss": 1.4045, "step": 2441 }, { "epoch": 5.970660146699267, "grad_norm": 0.1238719909116039, "learning_rate": 3.2871382346287016e-06, "loss": 0.4157, "step": 2442 }, { "epoch": 5.973105134474328, "grad_norm": 0.21633176972154994, "learning_rate": 3.285923049871223e-06, "loss": 0.4001, "step": 2443 }, { "epoch": 5.9755501222493885, "grad_norm": 0.22771966866569052, "learning_rate": 3.2847076590267303e-06, "loss": 0.45, "step": 2444 }, { "epoch": 5.97799511002445, "grad_norm": 0.2841526951188325, "learning_rate": 3.2834920624139253e-06, "loss": 0.4622, "step": 2445 }, { "epoch": 5.980440097799511, "grad_norm": 0.32736960767062323, "learning_rate": 3.282276260351567e-06, "loss": 0.3976, "step": 2446 }, { "epoch": 5.982885085574572, "grad_norm": 0.183518306178426, "learning_rate": 3.2810602531584663e-06, "loss": 0.4949, "step": 2447 }, { "epoch": 5.985330073349633, "grad_norm": 0.16518993210911348, "learning_rate": 3.2798440411534878e-06, "loss": 0.3782, "step": 2448 }, { "epoch": 5.987775061124695, "grad_norm": 0.2560545753932088, "learning_rate": 3.2786276246555494e-06, "loss": 0.8431, "step": 2449 }, { "epoch": 5.990220048899755, "grad_norm": 0.23437409028110592, "learning_rate": 3.2774110039836233e-06, "loss": 0.6213, "step": 2450 }, { "epoch": 5.992665036674817, "grad_norm": 0.19194742058526684, "learning_rate": 3.276194179456737e-06, "loss": 0.4524, "step": 2451 }, { "epoch": 5.995110024449878, "grad_norm": 0.2782643495525404, "learning_rate": 3.274977151393969e-06, "loss": 0.9042, "step": 2452 }, { "epoch": 5.997555012224939, "grad_norm": 0.3026044596217569, "learning_rate": 3.273759920114451e-06, "loss": 0.5031, "step": 2453 }, { "epoch": 6.0, "grad_norm": 0.23036025024613652, "learning_rate": 3.272542485937369e-06, "loss": 0.5018, "step": 2454 }, { "epoch": 6.002444987775061, "grad_norm": 0.2513976929482409, "learning_rate": 3.2713248491819627e-06, "loss": 0.4424, "step": 2455 }, { "epoch": 6.004889975550122, "grad_norm": 0.22148987313550889, "learning_rate": 3.2701070101675245e-06, "loss": 0.8285, "step": 2456 }, { "epoch": 6.007334963325183, "grad_norm": 0.2279291324243467, "learning_rate": 3.268888969213401e-06, "loss": 0.4299, "step": 2457 }, { "epoch": 6.009779951100245, "grad_norm": 0.2758919141498375, "learning_rate": 3.2676707266389872e-06, "loss": 0.6349, "step": 2458 }, { "epoch": 6.012224938875305, "grad_norm": 0.28653927076381486, "learning_rate": 3.2664522827637356e-06, "loss": 0.762, "step": 2459 }, { "epoch": 6.014669926650367, "grad_norm": 0.20640339419662412, "learning_rate": 3.2652336379071507e-06, "loss": 0.4848, "step": 2460 }, { "epoch": 6.017114914425428, "grad_norm": 0.2532503963981896, "learning_rate": 3.264014792388789e-06, "loss": 0.4262, "step": 2461 }, { "epoch": 6.019559902200489, "grad_norm": 0.19932666924854608, "learning_rate": 3.2627957465282597e-06, "loss": 0.7791, "step": 2462 }, { "epoch": 6.02200488997555, "grad_norm": 0.2758667543455121, "learning_rate": 3.2615765006452237e-06, "loss": 0.8319, "step": 2463 }, { "epoch": 6.0244498777506115, "grad_norm": 0.26867954379582326, "learning_rate": 3.2603570550593956e-06, "loss": 0.7497, "step": 2464 }, { "epoch": 6.026894865525672, "grad_norm": 0.21691323347737473, "learning_rate": 3.259137410090542e-06, "loss": 0.5305, "step": 2465 }, { "epoch": 6.029339853300733, "grad_norm": 0.22995714216288998, "learning_rate": 3.2579175660584823e-06, "loss": 0.7501, "step": 2466 }, { "epoch": 6.031784841075795, "grad_norm": 0.14620688339538562, "learning_rate": 3.2566975232830866e-06, "loss": 0.3846, "step": 2467 }, { "epoch": 6.034229828850855, "grad_norm": 0.2284254270817203, "learning_rate": 3.2554772820842793e-06, "loss": 0.9083, "step": 2468 }, { "epoch": 6.036674816625917, "grad_norm": 0.2693249352597879, "learning_rate": 3.2542568427820336e-06, "loss": 1.1761, "step": 2469 }, { "epoch": 6.039119804400978, "grad_norm": 0.3425602955655935, "learning_rate": 3.2530362056963775e-06, "loss": 0.3438, "step": 2470 }, { "epoch": 6.041564792176039, "grad_norm": 0.17441389457322984, "learning_rate": 3.2518153711473917e-06, "loss": 0.619, "step": 2471 }, { "epoch": 6.0440097799511, "grad_norm": 0.18718678685061632, "learning_rate": 3.2505943394552046e-06, "loss": 0.9092, "step": 2472 }, { "epoch": 6.0464547677261615, "grad_norm": 0.29704072016157496, "learning_rate": 3.24937311094e-06, "loss": 0.4548, "step": 2473 }, { "epoch": 6.048899755501223, "grad_norm": 0.2865220371401877, "learning_rate": 3.2481516859220107e-06, "loss": 0.5146, "step": 2474 }, { "epoch": 6.0513447432762835, "grad_norm": 0.2783430806274797, "learning_rate": 3.2469300647215234e-06, "loss": 0.3906, "step": 2475 }, { "epoch": 6.053789731051345, "grad_norm": 0.32744867983122616, "learning_rate": 3.2457082476588752e-06, "loss": 0.4666, "step": 2476 }, { "epoch": 6.056234718826406, "grad_norm": 0.24895326915751598, "learning_rate": 3.2444862350544546e-06, "loss": 0.8376, "step": 2477 }, { "epoch": 6.058679706601467, "grad_norm": 0.1849687168234643, "learning_rate": 3.243264027228701e-06, "loss": 0.8888, "step": 2478 }, { "epoch": 6.061124694376528, "grad_norm": 0.25822622680820934, "learning_rate": 3.2420416245021036e-06, "loss": 0.4463, "step": 2479 }, { "epoch": 6.06356968215159, "grad_norm": 0.23066420052494863, "learning_rate": 3.2408190271952073e-06, "loss": 0.5821, "step": 2480 }, { "epoch": 6.06601466992665, "grad_norm": 0.22955840671648725, "learning_rate": 3.2395962356286025e-06, "loss": 0.4257, "step": 2481 }, { "epoch": 6.068459657701712, "grad_norm": 0.1915321510386415, "learning_rate": 3.2383732501229358e-06, "loss": 0.6004, "step": 2482 }, { "epoch": 6.070904645476773, "grad_norm": 0.18455112464336287, "learning_rate": 3.2371500709989e-06, "loss": 0.4722, "step": 2483 }, { "epoch": 6.0733496332518335, "grad_norm": 0.166267426214357, "learning_rate": 3.235926698577241e-06, "loss": 0.37, "step": 2484 }, { "epoch": 6.075794621026895, "grad_norm": 0.20769920510701356, "learning_rate": 3.234703133178755e-06, "loss": 0.8453, "step": 2485 }, { "epoch": 6.078239608801956, "grad_norm": 0.27994606697243635, "learning_rate": 3.2334793751242892e-06, "loss": 0.7242, "step": 2486 }, { "epoch": 6.080684596577017, "grad_norm": 0.250495774082677, "learning_rate": 3.2322554247347414e-06, "loss": 0.5065, "step": 2487 }, { "epoch": 6.083129584352078, "grad_norm": 0.18909120342728641, "learning_rate": 3.2310312823310584e-06, "loss": 0.3028, "step": 2488 }, { "epoch": 6.08557457212714, "grad_norm": 0.2518570256791643, "learning_rate": 3.229806948234239e-06, "loss": 0.4713, "step": 2489 }, { "epoch": 6.0880195599022, "grad_norm": 0.15464732727439603, "learning_rate": 3.228582422765332e-06, "loss": 0.3873, "step": 2490 }, { "epoch": 6.090464547677262, "grad_norm": 0.27351013014371445, "learning_rate": 3.2273577062454353e-06, "loss": 0.5432, "step": 2491 }, { "epoch": 6.092909535452323, "grad_norm": 0.3104069600799555, "learning_rate": 3.226132798995697e-06, "loss": 0.3664, "step": 2492 }, { "epoch": 6.095354523227384, "grad_norm": 0.2443331767595374, "learning_rate": 3.2249077013373174e-06, "loss": 0.4585, "step": 2493 }, { "epoch": 6.097799511002445, "grad_norm": 0.2530429237483783, "learning_rate": 3.2236824135915447e-06, "loss": 0.3253, "step": 2494 }, { "epoch": 6.100244498777506, "grad_norm": 0.18790324363781535, "learning_rate": 3.2224569360796763e-06, "loss": 0.5985, "step": 2495 }, { "epoch": 6.102689486552567, "grad_norm": 0.19242122029131728, "learning_rate": 3.221231269123062e-06, "loss": 0.4852, "step": 2496 }, { "epoch": 6.105134474327628, "grad_norm": 0.33081058547196296, "learning_rate": 3.220005413043098e-06, "loss": 0.5803, "step": 2497 }, { "epoch": 6.10757946210269, "grad_norm": 0.29377223958527715, "learning_rate": 3.2187793681612333e-06, "loss": 0.3906, "step": 2498 }, { "epoch": 6.11002444987775, "grad_norm": 0.19995748238701366, "learning_rate": 3.217553134798964e-06, "loss": 0.41, "step": 2499 }, { "epoch": 6.112469437652812, "grad_norm": 0.32221375834820093, "learning_rate": 3.2163267132778376e-06, "loss": 0.4924, "step": 2500 }, { "epoch": 6.114914425427873, "grad_norm": 0.2574256412043207, "learning_rate": 3.2151001039194496e-06, "loss": 0.5619, "step": 2501 }, { "epoch": 6.117359413202934, "grad_norm": 0.22303369012575647, "learning_rate": 3.213873307045444e-06, "loss": 0.9861, "step": 2502 }, { "epoch": 6.119804400977995, "grad_norm": 0.22929064114826783, "learning_rate": 3.2126463229775163e-06, "loss": 1.2961, "step": 2503 }, { "epoch": 6.1222493887530565, "grad_norm": 0.2554991718668828, "learning_rate": 3.2114191520374082e-06, "loss": 0.6175, "step": 2504 }, { "epoch": 6.124694376528117, "grad_norm": 0.17705901702793223, "learning_rate": 3.2101917945469142e-06, "loss": 0.7001, "step": 2505 }, { "epoch": 6.127139364303178, "grad_norm": 0.1564626973162439, "learning_rate": 3.208964250827874e-06, "loss": 0.4367, "step": 2506 }, { "epoch": 6.12958435207824, "grad_norm": 0.2090933123625577, "learning_rate": 3.2077365212021783e-06, "loss": 0.4469, "step": 2507 }, { "epoch": 6.1320293398533, "grad_norm": 0.2725937049814604, "learning_rate": 3.206508605991765e-06, "loss": 0.7854, "step": 2508 }, { "epoch": 6.134474327628362, "grad_norm": 0.2181447435168512, "learning_rate": 3.2052805055186225e-06, "loss": 0.9504, "step": 2509 }, { "epoch": 6.136919315403423, "grad_norm": 0.21896244225451889, "learning_rate": 3.2040522201047876e-06, "loss": 0.5894, "step": 2510 }, { "epoch": 6.139364303178484, "grad_norm": 0.22276593939660236, "learning_rate": 3.202823750072343e-06, "loss": 0.5784, "step": 2511 }, { "epoch": 6.141809290953545, "grad_norm": 0.23124239978473743, "learning_rate": 3.2015950957434227e-06, "loss": 0.3995, "step": 2512 }, { "epoch": 6.1442542787286065, "grad_norm": 0.300247092769625, "learning_rate": 3.200366257440208e-06, "loss": 0.5032, "step": 2513 }, { "epoch": 6.146699266503667, "grad_norm": 0.17729510603123494, "learning_rate": 3.199137235484927e-06, "loss": 0.8367, "step": 2514 }, { "epoch": 6.1491442542787285, "grad_norm": 0.3328879061878735, "learning_rate": 3.19790803019986e-06, "loss": 0.8704, "step": 2515 }, { "epoch": 6.15158924205379, "grad_norm": 0.18416947990908444, "learning_rate": 3.1966786419073313e-06, "loss": 0.4846, "step": 2516 }, { "epoch": 6.15403422982885, "grad_norm": 0.22164864846257565, "learning_rate": 3.1954490709297146e-06, "loss": 0.4616, "step": 2517 }, { "epoch": 6.156479217603912, "grad_norm": 0.17376515725261277, "learning_rate": 3.194219317589432e-06, "loss": 0.6963, "step": 2518 }, { "epoch": 6.158924205378973, "grad_norm": 0.24220704878122706, "learning_rate": 3.192989382208952e-06, "loss": 0.7876, "step": 2519 }, { "epoch": 6.161369193154035, "grad_norm": 0.19372422328941805, "learning_rate": 3.191759265110793e-06, "loss": 0.6141, "step": 2520 }, { "epoch": 6.163814180929095, "grad_norm": 0.24604473150736486, "learning_rate": 3.1905289666175192e-06, "loss": 0.3585, "step": 2521 }, { "epoch": 6.166259168704157, "grad_norm": 0.3211571738257054, "learning_rate": 3.189298487051744e-06, "loss": 0.6562, "step": 2522 }, { "epoch": 6.168704156479218, "grad_norm": 0.24270946462886525, "learning_rate": 3.1880678267361254e-06, "loss": 0.3482, "step": 2523 }, { "epoch": 6.1711491442542785, "grad_norm": 0.2959602281507325, "learning_rate": 3.186836985993372e-06, "loss": 0.7735, "step": 2524 }, { "epoch": 6.17359413202934, "grad_norm": 0.17218137517657722, "learning_rate": 3.1856059651462385e-06, "loss": 0.7214, "step": 2525 }, { "epoch": 6.176039119804401, "grad_norm": 0.25829475154854653, "learning_rate": 3.1843747645175264e-06, "loss": 0.8829, "step": 2526 }, { "epoch": 6.178484107579462, "grad_norm": 0.21937994400205987, "learning_rate": 3.183143384430085e-06, "loss": 0.4261, "step": 2527 }, { "epoch": 6.180929095354523, "grad_norm": 0.30233322799547563, "learning_rate": 3.18191182520681e-06, "loss": 0.4959, "step": 2528 }, { "epoch": 6.183374083129585, "grad_norm": 0.13198622409451818, "learning_rate": 3.180680087170644e-06, "loss": 0.3255, "step": 2529 }, { "epoch": 6.185819070904645, "grad_norm": 0.2093383653349948, "learning_rate": 3.1794481706445777e-06, "loss": 0.4383, "step": 2530 }, { "epoch": 6.188264058679707, "grad_norm": 0.16627778733948947, "learning_rate": 3.178216075951648e-06, "loss": 0.9405, "step": 2531 }, { "epoch": 6.190709046454768, "grad_norm": 0.1900166950557964, "learning_rate": 3.176983803414938e-06, "loss": 0.5988, "step": 2532 }, { "epoch": 6.193154034229829, "grad_norm": 0.25613220311532364, "learning_rate": 3.175751353357577e-06, "loss": 0.7824, "step": 2533 }, { "epoch": 6.19559902200489, "grad_norm": 0.33933603601668194, "learning_rate": 3.1745187261027424e-06, "loss": 0.4298, "step": 2534 }, { "epoch": 6.198044009779951, "grad_norm": 0.18346956908614717, "learning_rate": 3.1732859219736574e-06, "loss": 0.9447, "step": 2535 }, { "epoch": 6.200488997555012, "grad_norm": 0.23632754281393847, "learning_rate": 3.1720529412935913e-06, "loss": 0.7936, "step": 2536 }, { "epoch": 6.202933985330073, "grad_norm": 0.18023655928255033, "learning_rate": 3.1708197843858603e-06, "loss": 0.7261, "step": 2537 }, { "epoch": 6.205378973105135, "grad_norm": 0.16676162520201446, "learning_rate": 3.1695864515738244e-06, "loss": 0.6606, "step": 2538 }, { "epoch": 6.207823960880195, "grad_norm": 0.17347837612809358, "learning_rate": 3.1683529431808946e-06, "loss": 0.6643, "step": 2539 }, { "epoch": 6.210268948655257, "grad_norm": 0.1548626809353046, "learning_rate": 3.1671192595305224e-06, "loss": 0.4237, "step": 2540 }, { "epoch": 6.212713936430318, "grad_norm": 0.22254059636503765, "learning_rate": 3.16588540094621e-06, "loss": 0.3548, "step": 2541 }, { "epoch": 6.215158924205379, "grad_norm": 0.25483058080035254, "learning_rate": 3.1646513677515023e-06, "loss": 0.6653, "step": 2542 }, { "epoch": 6.21760391198044, "grad_norm": 0.22700774966605214, "learning_rate": 3.1634171602699905e-06, "loss": 0.379, "step": 2543 }, { "epoch": 6.2200488997555015, "grad_norm": 0.193380285790183, "learning_rate": 3.1621827788253133e-06, "loss": 0.6955, "step": 2544 }, { "epoch": 6.222493887530562, "grad_norm": 0.226972848027893, "learning_rate": 3.1609482237411528e-06, "loss": 0.8486, "step": 2545 }, { "epoch": 6.224938875305623, "grad_norm": 0.2227604730664185, "learning_rate": 3.1597134953412386e-06, "loss": 0.5108, "step": 2546 }, { "epoch": 6.227383863080685, "grad_norm": 0.27975335020918013, "learning_rate": 3.158478593949344e-06, "loss": 0.4804, "step": 2547 }, { "epoch": 6.229828850855745, "grad_norm": 0.20594765239204993, "learning_rate": 3.157243519889288e-06, "loss": 0.9913, "step": 2548 }, { "epoch": 6.232273838630807, "grad_norm": 0.11914015953993708, "learning_rate": 3.1560082734849367e-06, "loss": 0.3124, "step": 2549 }, { "epoch": 6.234718826405868, "grad_norm": 0.22108511080294047, "learning_rate": 3.154772855060199e-06, "loss": 0.628, "step": 2550 }, { "epoch": 6.237163814180929, "grad_norm": 0.18121199155274048, "learning_rate": 3.15353726493903e-06, "loss": 0.5399, "step": 2551 }, { "epoch": 6.23960880195599, "grad_norm": 0.20882863961173595, "learning_rate": 3.1523015034454296e-06, "loss": 0.7777, "step": 2552 }, { "epoch": 6.2420537897310515, "grad_norm": 0.17963013134617542, "learning_rate": 3.151065570903443e-06, "loss": 0.3443, "step": 2553 }, { "epoch": 6.244498777506112, "grad_norm": 0.2973909004383695, "learning_rate": 3.14982946763716e-06, "loss": 0.9612, "step": 2554 }, { "epoch": 6.2469437652811735, "grad_norm": 0.2756763464726199, "learning_rate": 3.1485931939707154e-06, "loss": 1.2303, "step": 2555 }, { "epoch": 6.249388753056235, "grad_norm": 0.20623850445049904, "learning_rate": 3.1473567502282874e-06, "loss": 0.3758, "step": 2556 }, { "epoch": 6.251833740831296, "grad_norm": 0.19694382871989244, "learning_rate": 3.146120136734101e-06, "loss": 0.4014, "step": 2557 }, { "epoch": 6.254278728606357, "grad_norm": 0.25355379159144675, "learning_rate": 3.144883353812424e-06, "loss": 0.6491, "step": 2558 }, { "epoch": 6.256723716381418, "grad_norm": 0.21359266073758404, "learning_rate": 3.14364640178757e-06, "loss": 0.6015, "step": 2559 }, { "epoch": 6.25916870415648, "grad_norm": 0.26574542446944377, "learning_rate": 3.1424092809838952e-06, "loss": 0.4261, "step": 2560 }, { "epoch": 6.26161369193154, "grad_norm": 0.15649352546963427, "learning_rate": 3.141171991725801e-06, "loss": 0.5906, "step": 2561 }, { "epoch": 6.264058679706602, "grad_norm": 0.22242197179468262, "learning_rate": 3.1399345343377335e-06, "loss": 0.5517, "step": 2562 }, { "epoch": 6.266503667481663, "grad_norm": 0.22402933204778624, "learning_rate": 3.1386969091441818e-06, "loss": 0.5985, "step": 2563 }, { "epoch": 6.2689486552567235, "grad_norm": 0.12688411178481487, "learning_rate": 3.1374591164696797e-06, "loss": 0.2545, "step": 2564 }, { "epoch": 6.271393643031785, "grad_norm": 0.26120048639574395, "learning_rate": 3.1362211566388062e-06, "loss": 0.5335, "step": 2565 }, { "epoch": 6.273838630806846, "grad_norm": 0.2763981462653861, "learning_rate": 3.1349830299761807e-06, "loss": 0.6066, "step": 2566 }, { "epoch": 6.276283618581907, "grad_norm": 0.35923355819891295, "learning_rate": 3.1337447368064688e-06, "loss": 0.3717, "step": 2567 }, { "epoch": 6.278728606356968, "grad_norm": 0.29447716705757104, "learning_rate": 3.1325062774543797e-06, "loss": 0.7071, "step": 2568 }, { "epoch": 6.28117359413203, "grad_norm": 0.22013361056074715, "learning_rate": 3.1312676522446663e-06, "loss": 0.4954, "step": 2569 }, { "epoch": 6.28361858190709, "grad_norm": 0.17737047831735606, "learning_rate": 3.1300288615021233e-06, "loss": 0.4877, "step": 2570 }, { "epoch": 6.286063569682152, "grad_norm": 0.23894168058087356, "learning_rate": 3.1287899055515914e-06, "loss": 0.6056, "step": 2571 }, { "epoch": 6.288508557457213, "grad_norm": 0.2558137720883508, "learning_rate": 3.1275507847179507e-06, "loss": 0.7645, "step": 2572 }, { "epoch": 6.290953545232274, "grad_norm": 0.16682292042884223, "learning_rate": 3.12631149932613e-06, "loss": 0.4147, "step": 2573 }, { "epoch": 6.293398533007335, "grad_norm": 0.22948517031074533, "learning_rate": 3.125072049701097e-06, "loss": 0.5882, "step": 2574 }, { "epoch": 6.295843520782396, "grad_norm": 0.2005988469138682, "learning_rate": 3.123832436167864e-06, "loss": 0.3994, "step": 2575 }, { "epoch": 6.298288508557457, "grad_norm": 0.23239011232948692, "learning_rate": 3.1225926590514853e-06, "loss": 0.4304, "step": 2576 }, { "epoch": 6.300733496332518, "grad_norm": 0.2552286923519756, "learning_rate": 3.12135271867706e-06, "loss": 0.7305, "step": 2577 }, { "epoch": 6.30317848410758, "grad_norm": 0.2030008074198635, "learning_rate": 3.1201126153697275e-06, "loss": 0.5212, "step": 2578 }, { "epoch": 6.30562347188264, "grad_norm": 0.24745038198436645, "learning_rate": 3.1188723494546723e-06, "loss": 1.2205, "step": 2579 }, { "epoch": 6.308068459657702, "grad_norm": 0.15748317085593955, "learning_rate": 3.117631921257121e-06, "loss": 0.5496, "step": 2580 }, { "epoch": 6.310513447432763, "grad_norm": 0.12852504029890663, "learning_rate": 3.116391331102341e-06, "loss": 0.4139, "step": 2581 }, { "epoch": 6.312958435207824, "grad_norm": 0.22214521695113232, "learning_rate": 3.1151505793156433e-06, "loss": 0.5323, "step": 2582 }, { "epoch": 6.315403422982885, "grad_norm": 0.2233074061278456, "learning_rate": 3.1139096662223822e-06, "loss": 0.5108, "step": 2583 }, { "epoch": 6.3178484107579465, "grad_norm": 0.29352310305741575, "learning_rate": 3.1126685921479545e-06, "loss": 0.4481, "step": 2584 }, { "epoch": 6.320293398533007, "grad_norm": 0.2514815794016727, "learning_rate": 3.111427357417796e-06, "loss": 0.5029, "step": 2585 }, { "epoch": 6.322738386308068, "grad_norm": 0.19136425463549375, "learning_rate": 3.1101859623573886e-06, "loss": 0.7785, "step": 2586 }, { "epoch": 6.32518337408313, "grad_norm": 0.16416622497684552, "learning_rate": 3.1089444072922537e-06, "loss": 0.4277, "step": 2587 }, { "epoch": 6.32762836185819, "grad_norm": 0.20160817436562767, "learning_rate": 3.1077026925479553e-06, "loss": 0.9888, "step": 2588 }, { "epoch": 6.330073349633252, "grad_norm": 0.1981144325444725, "learning_rate": 3.1064608184501e-06, "loss": 0.7836, "step": 2589 }, { "epoch": 6.332518337408313, "grad_norm": 0.202957447876274, "learning_rate": 3.105218785324336e-06, "loss": 0.4073, "step": 2590 }, { "epoch": 6.334963325183374, "grad_norm": 0.3023134816194395, "learning_rate": 3.1039765934963515e-06, "loss": 0.5137, "step": 2591 }, { "epoch": 6.337408312958435, "grad_norm": 0.22151347639408875, "learning_rate": 3.1027342432918784e-06, "loss": 0.5849, "step": 2592 }, { "epoch": 6.3398533007334965, "grad_norm": 0.26270029158589314, "learning_rate": 3.1014917350366893e-06, "loss": 0.6923, "step": 2593 }, { "epoch": 6.342298288508557, "grad_norm": 0.29995123460797396, "learning_rate": 3.1002490690565983e-06, "loss": 0.4826, "step": 2594 }, { "epoch": 6.3447432762836184, "grad_norm": 0.21975879882477406, "learning_rate": 3.099006245677461e-06, "loss": 0.6614, "step": 2595 }, { "epoch": 6.34718826405868, "grad_norm": 0.24819246308470977, "learning_rate": 3.0977632652251742e-06, "loss": 0.8443, "step": 2596 }, { "epoch": 6.34963325183374, "grad_norm": 0.2783422709337147, "learning_rate": 3.0965201280256753e-06, "loss": 0.584, "step": 2597 }, { "epoch": 6.352078239608802, "grad_norm": 0.19643894490004454, "learning_rate": 3.095276834404944e-06, "loss": 0.518, "step": 2598 }, { "epoch": 6.354523227383863, "grad_norm": 0.23907666706004613, "learning_rate": 3.094033384688999e-06, "loss": 0.859, "step": 2599 }, { "epoch": 6.356968215158924, "grad_norm": 0.20978596056628784, "learning_rate": 3.0927897792039034e-06, "loss": 0.4227, "step": 2600 }, { "epoch": 6.359413202933985, "grad_norm": 0.32676125304010617, "learning_rate": 3.0915460182757586e-06, "loss": 0.4254, "step": 2601 }, { "epoch": 6.361858190709047, "grad_norm": 0.1976576170849751, "learning_rate": 3.090302102230705e-06, "loss": 0.5292, "step": 2602 }, { "epoch": 6.364303178484108, "grad_norm": 0.19039374279580804, "learning_rate": 3.089058031394928e-06, "loss": 0.4586, "step": 2603 }, { "epoch": 6.3667481662591685, "grad_norm": 0.20728358996109225, "learning_rate": 3.0878138060946506e-06, "loss": 0.5421, "step": 2604 }, { "epoch": 6.36919315403423, "grad_norm": 0.25487566401951084, "learning_rate": 3.0865694266561376e-06, "loss": 0.7118, "step": 2605 }, { "epoch": 6.371638141809291, "grad_norm": 0.25448513293231295, "learning_rate": 3.0853248934056935e-06, "loss": 0.4779, "step": 2606 }, { "epoch": 6.374083129584352, "grad_norm": 0.22245020829264056, "learning_rate": 3.084080206669663e-06, "loss": 0.4621, "step": 2607 }, { "epoch": 6.376528117359413, "grad_norm": 0.23893368835176065, "learning_rate": 3.0828353667744326e-06, "loss": 0.3396, "step": 2608 }, { "epoch": 6.378973105134475, "grad_norm": 0.19805314554787354, "learning_rate": 3.0815903740464266e-06, "loss": 0.8295, "step": 2609 }, { "epoch": 6.381418092909535, "grad_norm": 0.31429540250501625, "learning_rate": 3.0803452288121116e-06, "loss": 0.736, "step": 2610 }, { "epoch": 6.383863080684597, "grad_norm": 0.3680373955125361, "learning_rate": 3.079099931397993e-06, "loss": 0.5517, "step": 2611 }, { "epoch": 6.386308068459658, "grad_norm": 0.16574992332841876, "learning_rate": 3.077854482130615e-06, "loss": 0.5557, "step": 2612 }, { "epoch": 6.388753056234719, "grad_norm": 0.24630814219741592, "learning_rate": 3.0766088813365646e-06, "loss": 0.6251, "step": 2613 }, { "epoch": 6.39119804400978, "grad_norm": 0.2517810928564044, "learning_rate": 3.0753631293424662e-06, "loss": 0.5654, "step": 2614 }, { "epoch": 6.393643031784841, "grad_norm": 0.24910906025312432, "learning_rate": 3.0741172264749843e-06, "loss": 0.3776, "step": 2615 }, { "epoch": 6.396088019559902, "grad_norm": 0.27301101124350163, "learning_rate": 3.072871173060823e-06, "loss": 0.4949, "step": 2616 }, { "epoch": 6.398533007334963, "grad_norm": 0.24345622147713936, "learning_rate": 3.0716249694267265e-06, "loss": 0.4231, "step": 2617 }, { "epoch": 6.400977995110025, "grad_norm": 0.38778318143143486, "learning_rate": 3.0703786158994773e-06, "loss": 0.4189, "step": 2618 }, { "epoch": 6.403422982885085, "grad_norm": 0.17208381711521498, "learning_rate": 3.0691321128058993e-06, "loss": 0.4015, "step": 2619 }, { "epoch": 6.405867970660147, "grad_norm": 0.19602949419441026, "learning_rate": 3.0678854604728524e-06, "loss": 0.8393, "step": 2620 }, { "epoch": 6.408312958435208, "grad_norm": 0.2506024382072708, "learning_rate": 3.0666386592272383e-06, "loss": 0.5956, "step": 2621 }, { "epoch": 6.410757946210269, "grad_norm": 0.3161966822281516, "learning_rate": 3.065391709395996e-06, "loss": 0.8634, "step": 2622 }, { "epoch": 6.41320293398533, "grad_norm": 0.2266830823934656, "learning_rate": 3.0641446113061056e-06, "loss": 1.142, "step": 2623 }, { "epoch": 6.4156479217603914, "grad_norm": 0.28127582778881116, "learning_rate": 3.0628973652845844e-06, "loss": 0.3812, "step": 2624 }, { "epoch": 6.418092909535452, "grad_norm": 0.15016607763061554, "learning_rate": 3.061649971658488e-06, "loss": 0.4832, "step": 2625 }, { "epoch": 6.420537897310513, "grad_norm": 0.25080755257887083, "learning_rate": 3.060402430754912e-06, "loss": 0.5978, "step": 2626 }, { "epoch": 6.422982885085575, "grad_norm": 0.19395190758050576, "learning_rate": 3.0591547429009904e-06, "loss": 0.4887, "step": 2627 }, { "epoch": 6.425427872860635, "grad_norm": 0.24081837594303346, "learning_rate": 3.0579069084238956e-06, "loss": 0.4192, "step": 2628 }, { "epoch": 6.427872860635697, "grad_norm": 0.2411738263676204, "learning_rate": 3.0566589276508378e-06, "loss": 0.5571, "step": 2629 }, { "epoch": 6.430317848410758, "grad_norm": 0.28248861885516363, "learning_rate": 3.0554108009090673e-06, "loss": 0.3714, "step": 2630 }, { "epoch": 6.432762836185819, "grad_norm": 0.2433957390618439, "learning_rate": 3.0541625285258697e-06, "loss": 0.7324, "step": 2631 }, { "epoch": 6.43520782396088, "grad_norm": 0.2592034021177652, "learning_rate": 3.0529141108285716e-06, "loss": 0.4371, "step": 2632 }, { "epoch": 6.4376528117359415, "grad_norm": 0.18264399442590148, "learning_rate": 3.051665548144537e-06, "loss": 0.7301, "step": 2633 }, { "epoch": 6.440097799511002, "grad_norm": 0.3052687934894556, "learning_rate": 3.0504168408011675e-06, "loss": 0.647, "step": 2634 }, { "epoch": 6.442542787286063, "grad_norm": 0.19463382291350106, "learning_rate": 3.0491679891259028e-06, "loss": 0.5206, "step": 2635 }, { "epoch": 6.444987775061125, "grad_norm": 0.2236750490856417, "learning_rate": 3.0479189934462193e-06, "loss": 0.427, "step": 2636 }, { "epoch": 6.447432762836186, "grad_norm": 0.18561564322400415, "learning_rate": 3.046669854089633e-06, "loss": 0.4594, "step": 2637 }, { "epoch": 6.449877750611247, "grad_norm": 0.16007845207349647, "learning_rate": 3.045420571383698e-06, "loss": 0.6762, "step": 2638 }, { "epoch": 6.452322738386308, "grad_norm": 0.23037334715367425, "learning_rate": 3.044171145656003e-06, "loss": 0.5492, "step": 2639 }, { "epoch": 6.45476772616137, "grad_norm": 0.23897160401969714, "learning_rate": 3.0429215772341774e-06, "loss": 0.6096, "step": 2640 }, { "epoch": 6.45721271393643, "grad_norm": 0.1514064211222451, "learning_rate": 3.0416718664458856e-06, "loss": 0.4867, "step": 2641 }, { "epoch": 6.459657701711492, "grad_norm": 0.20611106386340128, "learning_rate": 3.04042201361883e-06, "loss": 0.7408, "step": 2642 }, { "epoch": 6.462102689486553, "grad_norm": 0.28683155873909766, "learning_rate": 3.0391720190807526e-06, "loss": 0.6533, "step": 2643 }, { "epoch": 6.4645476772616135, "grad_norm": 0.17145523096021667, "learning_rate": 3.0379218831594294e-06, "loss": 0.5002, "step": 2644 }, { "epoch": 6.466992665036675, "grad_norm": 0.3013053151233808, "learning_rate": 3.0366716061826747e-06, "loss": 0.6106, "step": 2645 }, { "epoch": 6.469437652811736, "grad_norm": 0.27490909468164043, "learning_rate": 3.035421188478339e-06, "loss": 0.6135, "step": 2646 }, { "epoch": 6.471882640586797, "grad_norm": 0.23641831376824862, "learning_rate": 3.034170630374311e-06, "loss": 0.4138, "step": 2647 }, { "epoch": 6.474327628361858, "grad_norm": 0.21608543498617774, "learning_rate": 3.032919932198517e-06, "loss": 0.6937, "step": 2648 }, { "epoch": 6.47677261613692, "grad_norm": 0.2505049747415038, "learning_rate": 3.0316690942789166e-06, "loss": 0.5515, "step": 2649 }, { "epoch": 6.47921760391198, "grad_norm": 0.21512728303048317, "learning_rate": 3.03041811694351e-06, "loss": 0.7614, "step": 2650 }, { "epoch": 6.481662591687042, "grad_norm": 0.20998272842913945, "learning_rate": 3.02916700052033e-06, "loss": 0.6434, "step": 2651 }, { "epoch": 6.484107579462103, "grad_norm": 0.2525983228130296, "learning_rate": 3.027915745337449e-06, "loss": 0.427, "step": 2652 }, { "epoch": 6.486552567237164, "grad_norm": 0.38571700008245036, "learning_rate": 3.0266643517229754e-06, "loss": 0.8491, "step": 2653 }, { "epoch": 6.488997555012225, "grad_norm": 0.20265827562066865, "learning_rate": 3.025412820005053e-06, "loss": 0.4355, "step": 2654 }, { "epoch": 6.491442542787286, "grad_norm": 0.3051546548153466, "learning_rate": 3.0241611505118617e-06, "loss": 0.9612, "step": 2655 }, { "epoch": 6.493887530562347, "grad_norm": 0.1964812738010243, "learning_rate": 3.0229093435716177e-06, "loss": 0.7518, "step": 2656 }, { "epoch": 6.496332518337408, "grad_norm": 0.17488816197186746, "learning_rate": 3.0216573995125743e-06, "loss": 0.4341, "step": 2657 }, { "epoch": 6.49877750611247, "grad_norm": 0.19993841109438337, "learning_rate": 3.0204053186630184e-06, "loss": 0.4158, "step": 2658 }, { "epoch": 6.50122249388753, "grad_norm": 0.2579996171861024, "learning_rate": 3.0191531013512754e-06, "loss": 0.5411, "step": 2659 }, { "epoch": 6.503667481662592, "grad_norm": 0.23437155010463537, "learning_rate": 3.0179007479057054e-06, "loss": 0.5494, "step": 2660 }, { "epoch": 6.506112469437653, "grad_norm": 0.2698069491630078, "learning_rate": 3.0166482586547035e-06, "loss": 0.9092, "step": 2661 }, { "epoch": 6.508557457212714, "grad_norm": 0.27178839949031947, "learning_rate": 3.015395633926702e-06, "loss": 0.6301, "step": 2662 }, { "epoch": 6.511002444987775, "grad_norm": 0.33727667953944246, "learning_rate": 3.014142874050166e-06, "loss": 0.4578, "step": 2663 }, { "epoch": 6.513447432762836, "grad_norm": 0.2233925985354532, "learning_rate": 3.0128899793536004e-06, "loss": 0.6856, "step": 2664 }, { "epoch": 6.515892420537897, "grad_norm": 0.2893950352519993, "learning_rate": 3.011636950165541e-06, "loss": 0.5547, "step": 2665 }, { "epoch": 6.518337408312958, "grad_norm": 0.24246571529305033, "learning_rate": 3.01038378681456e-06, "loss": 0.3582, "step": 2666 }, { "epoch": 6.52078239608802, "grad_norm": 0.22241750449147513, "learning_rate": 3.0091304896292685e-06, "loss": 1.056, "step": 2667 }, { "epoch": 6.52322738386308, "grad_norm": 0.20996244740252007, "learning_rate": 3.0078770589383067e-06, "loss": 0.605, "step": 2668 }, { "epoch": 6.525672371638142, "grad_norm": 0.23586272198530978, "learning_rate": 3.006623495070355e-06, "loss": 0.8006, "step": 2669 }, { "epoch": 6.528117359413203, "grad_norm": 0.22481452191041768, "learning_rate": 3.0053697983541254e-06, "loss": 0.6126, "step": 2670 }, { "epoch": 6.530562347188264, "grad_norm": 0.2904064181232295, "learning_rate": 3.004115969118365e-06, "loss": 0.4795, "step": 2671 }, { "epoch": 6.533007334963325, "grad_norm": 0.25087349970587886, "learning_rate": 3.0028620076918586e-06, "loss": 0.8884, "step": 2672 }, { "epoch": 6.5354523227383865, "grad_norm": 0.27735085746983523, "learning_rate": 3.0016079144034226e-06, "loss": 0.5938, "step": 2673 }, { "epoch": 6.537897310513447, "grad_norm": 0.1753704005772011, "learning_rate": 3.0003536895819084e-06, "loss": 0.3975, "step": 2674 }, { "epoch": 6.540342298288508, "grad_norm": 0.1557245782344478, "learning_rate": 2.9990993335562036e-06, "loss": 0.4411, "step": 2675 }, { "epoch": 6.54278728606357, "grad_norm": 0.26516177430613525, "learning_rate": 2.997844846655227e-06, "loss": 1.0124, "step": 2676 }, { "epoch": 6.54523227383863, "grad_norm": 0.18843645144325918, "learning_rate": 2.9965902292079375e-06, "loss": 0.5905, "step": 2677 }, { "epoch": 6.547677261613692, "grad_norm": 0.2047826022467107, "learning_rate": 2.995335481543321e-06, "loss": 0.4208, "step": 2678 }, { "epoch": 6.550122249388753, "grad_norm": 0.19923272382698506, "learning_rate": 2.9940806039904025e-06, "loss": 0.4718, "step": 2679 }, { "epoch": 6.552567237163814, "grad_norm": 0.2550964154261786, "learning_rate": 2.9928255968782394e-06, "loss": 0.5048, "step": 2680 }, { "epoch": 6.555012224938875, "grad_norm": 0.22377564850326986, "learning_rate": 2.991570460535923e-06, "loss": 0.5758, "step": 2681 }, { "epoch": 6.557457212713937, "grad_norm": 0.2311359389950381, "learning_rate": 2.99031519529258e-06, "loss": 0.7457, "step": 2682 }, { "epoch": 6.559902200488997, "grad_norm": 0.17256223584885527, "learning_rate": 2.989059801477369e-06, "loss": 0.4341, "step": 2683 }, { "epoch": 6.5623471882640585, "grad_norm": 0.14548126215494764, "learning_rate": 2.9878042794194816e-06, "loss": 0.4921, "step": 2684 }, { "epoch": 6.56479217603912, "grad_norm": 0.24706978448397027, "learning_rate": 2.986548629448146e-06, "loss": 0.3941, "step": 2685 }, { "epoch": 6.56723716381418, "grad_norm": 0.22201557159348928, "learning_rate": 2.985292851892622e-06, "loss": 0.8555, "step": 2686 }, { "epoch": 6.569682151589242, "grad_norm": 0.2688686015455099, "learning_rate": 2.9840369470822033e-06, "loss": 0.951, "step": 2687 }, { "epoch": 6.572127139364303, "grad_norm": 0.25251141761663465, "learning_rate": 2.9827809153462175e-06, "loss": 0.7463, "step": 2688 }, { "epoch": 6.574572127139365, "grad_norm": 0.17171794332440718, "learning_rate": 2.981524757014024e-06, "loss": 0.7028, "step": 2689 }, { "epoch": 6.577017114914425, "grad_norm": 0.3269105034715147, "learning_rate": 2.9802684724150153e-06, "loss": 0.5727, "step": 2690 }, { "epoch": 6.579462102689487, "grad_norm": 0.2591005183102207, "learning_rate": 2.9790120618786194e-06, "loss": 0.8143, "step": 2691 }, { "epoch": 6.581907090464548, "grad_norm": 0.20389425061484726, "learning_rate": 2.9777555257342967e-06, "loss": 0.7032, "step": 2692 }, { "epoch": 6.5843520782396086, "grad_norm": 0.24826773708105962, "learning_rate": 2.976498864311538e-06, "loss": 0.4452, "step": 2693 }, { "epoch": 6.58679706601467, "grad_norm": 0.24916078402016142, "learning_rate": 2.97524207793987e-06, "loss": 0.4132, "step": 2694 }, { "epoch": 6.589242053789731, "grad_norm": 0.3055012626057449, "learning_rate": 2.973985166948849e-06, "loss": 0.5207, "step": 2695 }, { "epoch": 6.591687041564792, "grad_norm": 0.20532255796825322, "learning_rate": 2.972728131668068e-06, "loss": 0.5616, "step": 2696 }, { "epoch": 6.594132029339853, "grad_norm": 0.16931948927088916, "learning_rate": 2.971470972427149e-06, "loss": 0.4439, "step": 2697 }, { "epoch": 6.596577017114915, "grad_norm": 0.224220677110465, "learning_rate": 2.970213689555749e-06, "loss": 0.4907, "step": 2698 }, { "epoch": 6.599022004889975, "grad_norm": 0.23675718780016247, "learning_rate": 2.9689562833835554e-06, "loss": 0.4266, "step": 2699 }, { "epoch": 6.601466992665037, "grad_norm": 0.2193748048340769, "learning_rate": 2.967698754240289e-06, "loss": 0.6999, "step": 2700 }, { "epoch": 6.603911980440098, "grad_norm": 0.2623786958337752, "learning_rate": 2.9664411024557028e-06, "loss": 0.4871, "step": 2701 }, { "epoch": 6.606356968215159, "grad_norm": 0.24164537885918347, "learning_rate": 2.9651833283595825e-06, "loss": 0.4935, "step": 2702 }, { "epoch": 6.60880195599022, "grad_norm": 0.15721826415161416, "learning_rate": 2.9639254322817452e-06, "loss": 0.2632, "step": 2703 }, { "epoch": 6.611246943765281, "grad_norm": 0.28990486519600583, "learning_rate": 2.9626674145520395e-06, "loss": 0.4998, "step": 2704 }, { "epoch": 6.613691931540342, "grad_norm": 0.14627604416645817, "learning_rate": 2.9614092755003458e-06, "loss": 0.425, "step": 2705 }, { "epoch": 6.616136919315403, "grad_norm": 0.21084199247646593, "learning_rate": 2.960151015456578e-06, "loss": 0.679, "step": 2706 }, { "epoch": 6.618581907090465, "grad_norm": 0.23032614314046862, "learning_rate": 2.958892634750681e-06, "loss": 0.5144, "step": 2707 }, { "epoch": 6.621026894865525, "grad_norm": 0.20662132597692687, "learning_rate": 2.9576341337126307e-06, "loss": 0.4933, "step": 2708 }, { "epoch": 6.623471882640587, "grad_norm": 0.18756664168992057, "learning_rate": 2.9563755126724348e-06, "loss": 0.4756, "step": 2709 }, { "epoch": 6.625916870415648, "grad_norm": 0.28438284606941416, "learning_rate": 2.9551167719601316e-06, "loss": 0.7255, "step": 2710 }, { "epoch": 6.628361858190709, "grad_norm": 0.19967237692839773, "learning_rate": 2.9538579119057926e-06, "loss": 0.6223, "step": 2711 }, { "epoch": 6.63080684596577, "grad_norm": 0.31763163723840154, "learning_rate": 2.95259893283952e-06, "loss": 0.3949, "step": 2712 }, { "epoch": 6.6332518337408315, "grad_norm": 0.2155008705970982, "learning_rate": 2.9513398350914467e-06, "loss": 0.4276, "step": 2713 }, { "epoch": 6.635696821515893, "grad_norm": 0.3625415938722578, "learning_rate": 2.950080618991737e-06, "loss": 0.4067, "step": 2714 }, { "epoch": 6.638141809290953, "grad_norm": 0.2895888761224593, "learning_rate": 2.9488212848705853e-06, "loss": 0.6283, "step": 2715 }, { "epoch": 6.640586797066015, "grad_norm": 0.26193784786323143, "learning_rate": 2.947561833058219e-06, "loss": 0.7917, "step": 2716 }, { "epoch": 6.643031784841076, "grad_norm": 0.2492031825452439, "learning_rate": 2.946302263884895e-06, "loss": 0.461, "step": 2717 }, { "epoch": 6.645476772616137, "grad_norm": 0.25541623107479616, "learning_rate": 2.9450425776809005e-06, "loss": 1.3095, "step": 2718 }, { "epoch": 6.647921760391198, "grad_norm": 0.38052565682056416, "learning_rate": 2.943782774776555e-06, "loss": 0.456, "step": 2719 }, { "epoch": 6.65036674816626, "grad_norm": 0.18461887494915855, "learning_rate": 2.942522855502206e-06, "loss": 0.5895, "step": 2720 }, { "epoch": 6.65281173594132, "grad_norm": 0.2992363306762597, "learning_rate": 2.941262820188236e-06, "loss": 0.8049, "step": 2721 }, { "epoch": 6.6552567237163816, "grad_norm": 0.30724322442935603, "learning_rate": 2.9400026691650516e-06, "loss": 0.6763, "step": 2722 }, { "epoch": 6.657701711491443, "grad_norm": 0.21154868438804283, "learning_rate": 2.938742402763097e-06, "loss": 0.3707, "step": 2723 }, { "epoch": 6.6601466992665035, "grad_norm": 0.19229092939292675, "learning_rate": 2.9374820213128403e-06, "loss": 0.5442, "step": 2724 }, { "epoch": 6.662591687041565, "grad_norm": 0.2742948361715638, "learning_rate": 2.9362215251447825e-06, "loss": 0.4725, "step": 2725 }, { "epoch": 6.665036674816626, "grad_norm": 0.3666432675646573, "learning_rate": 2.9349609145894555e-06, "loss": 0.6125, "step": 2726 }, { "epoch": 6.667481662591687, "grad_norm": 0.27497653231406094, "learning_rate": 2.9337001899774194e-06, "loss": 0.5669, "step": 2727 }, { "epoch": 6.669926650366748, "grad_norm": 0.2258595381231396, "learning_rate": 2.932439351639266e-06, "loss": 0.8942, "step": 2728 }, { "epoch": 6.67237163814181, "grad_norm": 1.0147960895206174, "learning_rate": 2.9311783999056155e-06, "loss": 0.3725, "step": 2729 }, { "epoch": 6.67481662591687, "grad_norm": 0.25218958077865056, "learning_rate": 2.929917335107118e-06, "loss": 0.5526, "step": 2730 }, { "epoch": 6.677261613691932, "grad_norm": 0.16792298084885113, "learning_rate": 2.9286561575744545e-06, "loss": 0.5027, "step": 2731 }, { "epoch": 6.679706601466993, "grad_norm": 0.19157012028497086, "learning_rate": 2.9273948676383336e-06, "loss": 0.7786, "step": 2732 }, { "epoch": 6.6821515892420535, "grad_norm": 0.20042667939340975, "learning_rate": 2.9261334656294937e-06, "loss": 0.4554, "step": 2733 }, { "epoch": 6.684596577017115, "grad_norm": 0.24652258489059, "learning_rate": 2.9248719518787054e-06, "loss": 0.7173, "step": 2734 }, { "epoch": 6.687041564792176, "grad_norm": 0.28255033205159097, "learning_rate": 2.9236103267167655e-06, "loss": 0.5164, "step": 2735 }, { "epoch": 6.689486552567237, "grad_norm": 0.2649261641607428, "learning_rate": 2.922348590474501e-06, "loss": 0.3937, "step": 2736 }, { "epoch": 6.691931540342298, "grad_norm": 0.31930246367682547, "learning_rate": 2.921086743482767e-06, "loss": 0.4017, "step": 2737 }, { "epoch": 6.69437652811736, "grad_norm": 0.2812279678445192, "learning_rate": 2.9198247860724504e-06, "loss": 0.8392, "step": 2738 }, { "epoch": 6.69682151589242, "grad_norm": 0.18960881525836548, "learning_rate": 2.918562718574464e-06, "loss": 0.4895, "step": 2739 }, { "epoch": 6.699266503667482, "grad_norm": 0.19709767991086352, "learning_rate": 2.917300541319752e-06, "loss": 0.4432, "step": 2740 }, { "epoch": 6.701711491442543, "grad_norm": 0.19926816703391803, "learning_rate": 2.916038254639285e-06, "loss": 0.8889, "step": 2741 }, { "epoch": 6.704156479217604, "grad_norm": 0.23519372874048605, "learning_rate": 2.914775858864064e-06, "loss": 0.3823, "step": 2742 }, { "epoch": 6.706601466992665, "grad_norm": 0.16451325553688165, "learning_rate": 2.913513354325119e-06, "loss": 0.5356, "step": 2743 }, { "epoch": 6.709046454767726, "grad_norm": 0.1727199848078932, "learning_rate": 2.912250741353505e-06, "loss": 0.589, "step": 2744 }, { "epoch": 6.711491442542787, "grad_norm": 0.2781649841616548, "learning_rate": 2.91098802028031e-06, "loss": 0.4254, "step": 2745 }, { "epoch": 6.713936430317848, "grad_norm": 0.2744545367492934, "learning_rate": 2.9097251914366487e-06, "loss": 0.8466, "step": 2746 }, { "epoch": 6.71638141809291, "grad_norm": 0.21735279559568932, "learning_rate": 2.908462255153663e-06, "loss": 0.7528, "step": 2747 }, { "epoch": 6.71882640586797, "grad_norm": 0.18385081154694746, "learning_rate": 2.9071992117625237e-06, "loss": 0.6961, "step": 2748 }, { "epoch": 6.721271393643032, "grad_norm": 0.23424880872975262, "learning_rate": 2.905936061594429e-06, "loss": 0.8013, "step": 2749 }, { "epoch": 6.723716381418093, "grad_norm": 0.370127315376921, "learning_rate": 2.9046728049806066e-06, "loss": 0.446, "step": 2750 }, { "epoch": 6.726161369193154, "grad_norm": 0.19917553839152738, "learning_rate": 2.903409442252312e-06, "loss": 0.6816, "step": 2751 }, { "epoch": 6.728606356968215, "grad_norm": 0.24988300606577307, "learning_rate": 2.9021459737408264e-06, "loss": 0.4542, "step": 2752 }, { "epoch": 6.7310513447432765, "grad_norm": 0.198693501637806, "learning_rate": 2.9008823997774617e-06, "loss": 0.4508, "step": 2753 }, { "epoch": 6.733496332518337, "grad_norm": 0.2987059254981822, "learning_rate": 2.899618720693554e-06, "loss": 0.7597, "step": 2754 }, { "epoch": 6.735941320293398, "grad_norm": 0.16310346787961943, "learning_rate": 2.89835493682047e-06, "loss": 0.513, "step": 2755 }, { "epoch": 6.73838630806846, "grad_norm": 0.2059609108865939, "learning_rate": 2.897091048489603e-06, "loss": 0.4991, "step": 2756 }, { "epoch": 6.74083129584352, "grad_norm": 0.18402849371140545, "learning_rate": 2.895827056032373e-06, "loss": 0.3627, "step": 2757 }, { "epoch": 6.743276283618582, "grad_norm": 0.24634234265554922, "learning_rate": 2.894562959780228e-06, "loss": 0.3881, "step": 2758 }, { "epoch": 6.745721271393643, "grad_norm": 0.2990195768822749, "learning_rate": 2.8932987600646426e-06, "loss": 0.6247, "step": 2759 }, { "epoch": 6.748166259168704, "grad_norm": 0.22955393709081237, "learning_rate": 2.8920344572171195e-06, "loss": 0.7099, "step": 2760 }, { "epoch": 6.750611246943765, "grad_norm": 0.24466288292807217, "learning_rate": 2.8907700515691877e-06, "loss": 1.1852, "step": 2761 }, { "epoch": 6.7530562347188265, "grad_norm": 0.30472795388821744, "learning_rate": 2.8895055434524028e-06, "loss": 0.4398, "step": 2762 }, { "epoch": 6.755501222493887, "grad_norm": 0.20664302174119287, "learning_rate": 2.8882409331983487e-06, "loss": 0.8953, "step": 2763 }, { "epoch": 6.7579462102689485, "grad_norm": 0.2328990996188546, "learning_rate": 2.8869762211386335e-06, "loss": 0.3735, "step": 2764 }, { "epoch": 6.76039119804401, "grad_norm": 0.2732119444567348, "learning_rate": 2.8857114076048948e-06, "loss": 0.4616, "step": 2765 }, { "epoch": 6.76283618581907, "grad_norm": 0.20187718409519675, "learning_rate": 2.884446492928796e-06, "loss": 0.5036, "step": 2766 }, { "epoch": 6.765281173594132, "grad_norm": 0.2564015954416233, "learning_rate": 2.8831814774420265e-06, "loss": 0.6892, "step": 2767 }, { "epoch": 6.767726161369193, "grad_norm": 0.36373327251497733, "learning_rate": 2.8819163614763026e-06, "loss": 0.4086, "step": 2768 }, { "epoch": 6.770171149144255, "grad_norm": 0.33241102667606953, "learning_rate": 2.8806511453633644e-06, "loss": 0.5613, "step": 2769 }, { "epoch": 6.772616136919315, "grad_norm": 0.21117446139519003, "learning_rate": 2.879385829434983e-06, "loss": 0.6965, "step": 2770 }, { "epoch": 6.775061124694377, "grad_norm": 0.19735893864714288, "learning_rate": 2.878120414022953e-06, "loss": 0.4242, "step": 2771 }, { "epoch": 6.777506112469438, "grad_norm": 0.2629117529727301, "learning_rate": 2.8768548994590943e-06, "loss": 0.6384, "step": 2772 }, { "epoch": 6.7799511002444985, "grad_norm": 0.2816011177731885, "learning_rate": 2.875589286075255e-06, "loss": 0.4578, "step": 2773 }, { "epoch": 6.78239608801956, "grad_norm": 0.3300256350434343, "learning_rate": 2.874323574203306e-06, "loss": 0.4178, "step": 2774 }, { "epoch": 6.784841075794621, "grad_norm": 0.20584583164061476, "learning_rate": 2.873057764175148e-06, "loss": 0.9352, "step": 2775 }, { "epoch": 6.787286063569682, "grad_norm": 0.4074261674837491, "learning_rate": 2.8717918563227033e-06, "loss": 0.3981, "step": 2776 }, { "epoch": 6.789731051344743, "grad_norm": 0.25685936415912963, "learning_rate": 2.870525850977924e-06, "loss": 0.7321, "step": 2777 }, { "epoch": 6.792176039119805, "grad_norm": 0.24280540137406426, "learning_rate": 2.8692597484727845e-06, "loss": 0.4047, "step": 2778 }, { "epoch": 6.794621026894865, "grad_norm": 0.29043875898454935, "learning_rate": 2.8679935491392858e-06, "loss": 0.4859, "step": 2779 }, { "epoch": 6.797066014669927, "grad_norm": 0.164345140183799, "learning_rate": 2.8667272533094547e-06, "loss": 0.8862, "step": 2780 }, { "epoch": 6.799511002444988, "grad_norm": 0.21405966188164813, "learning_rate": 2.865460861315343e-06, "loss": 0.6696, "step": 2781 }, { "epoch": 6.801955990220049, "grad_norm": 0.29081377205013303, "learning_rate": 2.8641943734890275e-06, "loss": 0.4452, "step": 2782 }, { "epoch": 6.80440097799511, "grad_norm": 0.24536436426114933, "learning_rate": 2.862927790162611e-06, "loss": 0.4858, "step": 2783 }, { "epoch": 6.806845965770171, "grad_norm": 0.3157539307182182, "learning_rate": 2.8616611116682193e-06, "loss": 0.4516, "step": 2784 }, { "epoch": 6.809290953545232, "grad_norm": 0.2031114985673134, "learning_rate": 2.8603943383380057e-06, "loss": 0.8251, "step": 2785 }, { "epoch": 6.811735941320293, "grad_norm": 0.279626524171972, "learning_rate": 2.8591274705041453e-06, "loss": 1.0458, "step": 2786 }, { "epoch": 6.814180929095355, "grad_norm": 0.3242376603844505, "learning_rate": 2.857860508498843e-06, "loss": 0.7359, "step": 2787 }, { "epoch": 6.816625916870415, "grad_norm": 0.20884423801849686, "learning_rate": 2.8565934526543226e-06, "loss": 0.68, "step": 2788 }, { "epoch": 6.819070904645477, "grad_norm": 0.16276661716179386, "learning_rate": 2.855326303302836e-06, "loss": 0.5919, "step": 2789 }, { "epoch": 6.821515892420538, "grad_norm": 0.24677520783481247, "learning_rate": 2.8540590607766594e-06, "loss": 0.5566, "step": 2790 }, { "epoch": 6.823960880195599, "grad_norm": 0.16615301463507698, "learning_rate": 2.8527917254080916e-06, "loss": 0.4758, "step": 2791 }, { "epoch": 6.82640586797066, "grad_norm": 0.2095359279365771, "learning_rate": 2.851524297529457e-06, "loss": 0.3986, "step": 2792 }, { "epoch": 6.8288508557457215, "grad_norm": 0.27380082906367376, "learning_rate": 2.850256777473106e-06, "loss": 0.2664, "step": 2793 }, { "epoch": 6.831295843520783, "grad_norm": 0.252485887655719, "learning_rate": 2.848989165571409e-06, "loss": 0.3373, "step": 2794 }, { "epoch": 6.833740831295843, "grad_norm": 0.2752948505743244, "learning_rate": 2.8477214621567646e-06, "loss": 1.5643, "step": 2795 }, { "epoch": 6.836185819070905, "grad_norm": 0.32543176822072833, "learning_rate": 2.846453667561593e-06, "loss": 0.5343, "step": 2796 }, { "epoch": 6.838630806845966, "grad_norm": 0.3075030356717426, "learning_rate": 2.8451857821183383e-06, "loss": 0.9246, "step": 2797 }, { "epoch": 6.841075794621027, "grad_norm": 0.21719772656087993, "learning_rate": 2.84391780615947e-06, "loss": 0.4434, "step": 2798 }, { "epoch": 6.843520782396088, "grad_norm": 0.23325966516724952, "learning_rate": 2.8426497400174803e-06, "loss": 0.7113, "step": 2799 }, { "epoch": 6.84596577017115, "grad_norm": 0.21647800007884022, "learning_rate": 2.8413815840248855e-06, "loss": 0.7238, "step": 2800 }, { "epoch": 6.84841075794621, "grad_norm": 0.27048519433463364, "learning_rate": 2.8401133385142244e-06, "loss": 0.4382, "step": 2801 }, { "epoch": 6.8508557457212715, "grad_norm": 0.18601097103292863, "learning_rate": 2.8388450038180603e-06, "loss": 0.2983, "step": 2802 }, { "epoch": 6.853300733496333, "grad_norm": 0.21018733614206458, "learning_rate": 2.837576580268979e-06, "loss": 0.5112, "step": 2803 }, { "epoch": 6.8557457212713935, "grad_norm": 0.2577631653028338, "learning_rate": 2.836308068199591e-06, "loss": 0.5394, "step": 2804 }, { "epoch": 6.858190709046455, "grad_norm": 0.2606307959223212, "learning_rate": 2.8350394679425292e-06, "loss": 0.4073, "step": 2805 }, { "epoch": 6.860635696821516, "grad_norm": 0.21119054002772333, "learning_rate": 2.8337707798304496e-06, "loss": 0.6045, "step": 2806 }, { "epoch": 6.863080684596577, "grad_norm": 0.18527861172121762, "learning_rate": 2.832502004196031e-06, "loss": 0.6509, "step": 2807 }, { "epoch": 6.865525672371638, "grad_norm": 0.4518854001438533, "learning_rate": 2.8312331413719745e-06, "loss": 0.3333, "step": 2808 }, { "epoch": 6.8679706601467, "grad_norm": 0.2742515820423122, "learning_rate": 2.829964191691007e-06, "loss": 0.427, "step": 2809 }, { "epoch": 6.87041564792176, "grad_norm": 0.27879854249861935, "learning_rate": 2.828695155485875e-06, "loss": 0.5961, "step": 2810 }, { "epoch": 6.872860635696822, "grad_norm": 0.19661120593022052, "learning_rate": 2.827426033089349e-06, "loss": 0.3912, "step": 2811 }, { "epoch": 6.875305623471883, "grad_norm": 0.20068993321178927, "learning_rate": 2.826156824834222e-06, "loss": 0.5391, "step": 2812 }, { "epoch": 6.8777506112469435, "grad_norm": 0.18969871197521632, "learning_rate": 2.824887531053308e-06, "loss": 0.4676, "step": 2813 }, { "epoch": 6.880195599022005, "grad_norm": 0.20889441044113433, "learning_rate": 2.8236181520794476e-06, "loss": 0.5877, "step": 2814 }, { "epoch": 6.882640586797066, "grad_norm": 0.2364876938075912, "learning_rate": 2.8223486882455e-06, "loss": 0.6308, "step": 2815 }, { "epoch": 6.885085574572127, "grad_norm": 0.25701925051803837, "learning_rate": 2.8210791398843462e-06, "loss": 0.4281, "step": 2816 }, { "epoch": 6.887530562347188, "grad_norm": 0.19670853692453427, "learning_rate": 2.8198095073288927e-06, "loss": 0.3746, "step": 2817 }, { "epoch": 6.88997555012225, "grad_norm": 0.18468337424437048, "learning_rate": 2.8185397909120647e-06, "loss": 0.4062, "step": 2818 }, { "epoch": 6.89242053789731, "grad_norm": 0.26482491036256023, "learning_rate": 2.817269990966812e-06, "loss": 0.8197, "step": 2819 }, { "epoch": 6.894865525672372, "grad_norm": 0.296123128878938, "learning_rate": 2.8160001078261056e-06, "loss": 0.663, "step": 2820 }, { "epoch": 6.897310513447433, "grad_norm": 0.323392070416714, "learning_rate": 2.8147301418229373e-06, "loss": 0.3726, "step": 2821 }, { "epoch": 6.899755501222494, "grad_norm": 0.2458417002010351, "learning_rate": 2.813460093290321e-06, "loss": 0.9851, "step": 2822 }, { "epoch": 6.902200488997555, "grad_norm": 0.21911391718046638, "learning_rate": 2.812189962561292e-06, "loss": 0.4563, "step": 2823 }, { "epoch": 6.904645476772616, "grad_norm": 0.2704856297046878, "learning_rate": 2.8109197499689083e-06, "loss": 0.5959, "step": 2824 }, { "epoch": 6.907090464547677, "grad_norm": 0.2520313035666214, "learning_rate": 2.80964945584625e-06, "loss": 0.5115, "step": 2825 }, { "epoch": 6.909535452322738, "grad_norm": 0.25291875628267846, "learning_rate": 2.8083790805264153e-06, "loss": 0.4304, "step": 2826 }, { "epoch": 6.9119804400978, "grad_norm": 0.16666761708475025, "learning_rate": 2.807108624342527e-06, "loss": 0.5669, "step": 2827 }, { "epoch": 6.91442542787286, "grad_norm": 0.17386870018651085, "learning_rate": 2.8058380876277265e-06, "loss": 0.5577, "step": 2828 }, { "epoch": 6.916870415647922, "grad_norm": 0.17696185295288228, "learning_rate": 2.804567470715178e-06, "loss": 0.5349, "step": 2829 }, { "epoch": 6.919315403422983, "grad_norm": 0.20857481435771388, "learning_rate": 2.8032967739380678e-06, "loss": 0.8764, "step": 2830 }, { "epoch": 6.921760391198044, "grad_norm": 0.279417430575681, "learning_rate": 2.8020259976296006e-06, "loss": 0.7464, "step": 2831 }, { "epoch": 6.924205378973105, "grad_norm": 0.3099109123344021, "learning_rate": 2.800755142123003e-06, "loss": 0.6357, "step": 2832 }, { "epoch": 6.9266503667481665, "grad_norm": 0.16070560251874458, "learning_rate": 2.7994842077515216e-06, "loss": 0.5408, "step": 2833 }, { "epoch": 6.929095354523227, "grad_norm": 0.23790132839621222, "learning_rate": 2.798213194848426e-06, "loss": 0.3719, "step": 2834 }, { "epoch": 6.931540342298288, "grad_norm": 0.1416649488690476, "learning_rate": 2.7969421037470037e-06, "loss": 0.4203, "step": 2835 }, { "epoch": 6.93398533007335, "grad_norm": 0.16597424079124218, "learning_rate": 2.795670934780565e-06, "loss": 0.2303, "step": 2836 }, { "epoch": 6.93643031784841, "grad_norm": 0.17889120704636863, "learning_rate": 2.794399688282438e-06, "loss": 0.9721, "step": 2837 }, { "epoch": 6.938875305623472, "grad_norm": 0.2186810405583919, "learning_rate": 2.793128364585974e-06, "loss": 0.5193, "step": 2838 }, { "epoch": 6.941320293398533, "grad_norm": 0.3560748118375034, "learning_rate": 2.7918569640245437e-06, "loss": 0.5849, "step": 2839 }, { "epoch": 6.943765281173594, "grad_norm": 0.2502336209852303, "learning_rate": 2.7905854869315347e-06, "loss": 0.6671, "step": 2840 }, { "epoch": 6.946210268948655, "grad_norm": 0.3532463416050091, "learning_rate": 2.78931393364036e-06, "loss": 0.5823, "step": 2841 }, { "epoch": 6.9486552567237165, "grad_norm": 0.18908993704629637, "learning_rate": 2.788042304484449e-06, "loss": 0.6043, "step": 2842 }, { "epoch": 6.951100244498777, "grad_norm": 0.16310714500755577, "learning_rate": 2.7867705997972523e-06, "loss": 0.6818, "step": 2843 }, { "epoch": 6.9535452322738385, "grad_norm": 0.24819913839430116, "learning_rate": 2.78549881991224e-06, "loss": 0.9136, "step": 2844 }, { "epoch": 6.9559902200489, "grad_norm": 0.1912671282555043, "learning_rate": 2.7842269651629008e-06, "loss": 0.3358, "step": 2845 }, { "epoch": 6.95843520782396, "grad_norm": 0.23391566127328417, "learning_rate": 2.7829550358827455e-06, "loss": 0.7382, "step": 2846 }, { "epoch": 6.960880195599022, "grad_norm": 0.26613179607738646, "learning_rate": 2.7816830324053026e-06, "loss": 0.4464, "step": 2847 }, { "epoch": 6.963325183374083, "grad_norm": 0.2181186003855521, "learning_rate": 2.7804109550641205e-06, "loss": 0.4594, "step": 2848 }, { "epoch": 6.965770171149144, "grad_norm": 0.18790626434449093, "learning_rate": 2.779138804192767e-06, "loss": 0.877, "step": 2849 }, { "epoch": 6.968215158924205, "grad_norm": 0.24991768155867214, "learning_rate": 2.7778665801248294e-06, "loss": 0.4746, "step": 2850 }, { "epoch": 6.970660146699267, "grad_norm": 0.1995437717850146, "learning_rate": 2.776594283193913e-06, "loss": 0.6947, "step": 2851 }, { "epoch": 6.973105134474328, "grad_norm": 0.21548287625896684, "learning_rate": 2.775321913733644e-06, "loss": 0.7359, "step": 2852 }, { "epoch": 6.9755501222493885, "grad_norm": 0.2600861711723258, "learning_rate": 2.7740494720776673e-06, "loss": 0.4591, "step": 2853 }, { "epoch": 6.97799511002445, "grad_norm": 0.19812222263502807, "learning_rate": 2.772776958559646e-06, "loss": 0.7522, "step": 2854 }, { "epoch": 6.980440097799511, "grad_norm": 0.2274583461427906, "learning_rate": 2.7715043735132615e-06, "loss": 0.4138, "step": 2855 }, { "epoch": 6.982885085574572, "grad_norm": 0.19302459766376526, "learning_rate": 2.7702317172722143e-06, "loss": 0.4676, "step": 2856 }, { "epoch": 6.985330073349633, "grad_norm": 0.2437648146181491, "learning_rate": 2.768958990170225e-06, "loss": 0.4823, "step": 2857 }, { "epoch": 6.987775061124695, "grad_norm": 0.2844858934455108, "learning_rate": 2.7676861925410326e-06, "loss": 0.3401, "step": 2858 }, { "epoch": 6.990220048899755, "grad_norm": 0.2947085837465676, "learning_rate": 2.7664133247183927e-06, "loss": 0.6423, "step": 2859 }, { "epoch": 6.992665036674817, "grad_norm": 0.28026310584723535, "learning_rate": 2.7651403870360803e-06, "loss": 0.6117, "step": 2860 }, { "epoch": 6.995110024449878, "grad_norm": 0.16974081277840003, "learning_rate": 2.7638673798278894e-06, "loss": 0.8068, "step": 2861 }, { "epoch": 6.997555012224939, "grad_norm": 0.25464037365789094, "learning_rate": 2.76259430342763e-06, "loss": 0.5885, "step": 2862 }, { "epoch": 7.0, "grad_norm": 0.16322625467003435, "learning_rate": 2.761321158169134e-06, "loss": 0.418, "step": 2863 }, { "epoch": 7.002444987775061, "grad_norm": 0.2803745876184196, "learning_rate": 2.7600479443862484e-06, "loss": 0.5407, "step": 2864 }, { "epoch": 7.004889975550122, "grad_norm": 0.23088286437399547, "learning_rate": 2.7587746624128387e-06, "loss": 0.7657, "step": 2865 }, { "epoch": 7.007334963325183, "grad_norm": 0.211075860092187, "learning_rate": 2.7575013125827887e-06, "loss": 0.6514, "step": 2866 }, { "epoch": 7.009779951100245, "grad_norm": 0.22747524363771285, "learning_rate": 2.75622789523e-06, "loss": 0.6311, "step": 2867 }, { "epoch": 7.012224938875305, "grad_norm": 0.26568755421704543, "learning_rate": 2.754954410688392e-06, "loss": 0.9956, "step": 2868 }, { "epoch": 7.014669926650367, "grad_norm": 0.26519804678688574, "learning_rate": 2.7536808592919012e-06, "loss": 0.5011, "step": 2869 }, { "epoch": 7.017114914425428, "grad_norm": 0.2130430904612387, "learning_rate": 2.7524072413744817e-06, "loss": 1.2497, "step": 2870 }, { "epoch": 7.019559902200489, "grad_norm": 0.22564891452350325, "learning_rate": 2.751133557270106e-06, "loss": 0.8184, "step": 2871 }, { "epoch": 7.02200488997555, "grad_norm": 0.3505898093982352, "learning_rate": 2.749859807312762e-06, "loss": 0.7188, "step": 2872 }, { "epoch": 7.0244498777506115, "grad_norm": 0.15052571682476995, "learning_rate": 2.7485859918364577e-06, "loss": 0.3633, "step": 2873 }, { "epoch": 7.026894865525672, "grad_norm": 0.24242577175350447, "learning_rate": 2.7473121111752155e-06, "loss": 0.4543, "step": 2874 }, { "epoch": 7.029339853300733, "grad_norm": 0.2626842315138586, "learning_rate": 2.7460381656630768e-06, "loss": 0.7194, "step": 2875 }, { "epoch": 7.031784841075795, "grad_norm": 0.2732009367635997, "learning_rate": 2.7447641556340987e-06, "loss": 0.446, "step": 2876 }, { "epoch": 7.034229828850855, "grad_norm": 0.18694373464323058, "learning_rate": 2.743490081422356e-06, "loss": 0.5515, "step": 2877 }, { "epoch": 7.036674816625917, "grad_norm": 0.307523530491107, "learning_rate": 2.74221594336194e-06, "loss": 0.7568, "step": 2878 }, { "epoch": 7.039119804400978, "grad_norm": 0.18285192211744286, "learning_rate": 2.740941741786959e-06, "loss": 0.5321, "step": 2879 }, { "epoch": 7.041564792176039, "grad_norm": 0.2691301659888327, "learning_rate": 2.739667477031538e-06, "loss": 0.9386, "step": 2880 }, { "epoch": 7.0440097799511, "grad_norm": 0.2584152581453055, "learning_rate": 2.738393149429818e-06, "loss": 0.5119, "step": 2881 }, { "epoch": 7.0464547677261615, "grad_norm": 0.19308514640073143, "learning_rate": 2.7371187593159577e-06, "loss": 0.3915, "step": 2882 }, { "epoch": 7.048899755501223, "grad_norm": 0.27246911867563556, "learning_rate": 2.73584430702413e-06, "loss": 0.4662, "step": 2883 }, { "epoch": 7.0513447432762835, "grad_norm": 0.18812136140108413, "learning_rate": 2.7345697928885263e-06, "loss": 0.4338, "step": 2884 }, { "epoch": 7.053789731051345, "grad_norm": 0.22176889392026958, "learning_rate": 2.7332952172433535e-06, "loss": 0.6034, "step": 2885 }, { "epoch": 7.056234718826406, "grad_norm": 0.4688783208501427, "learning_rate": 2.7320205804228337e-06, "loss": 0.4336, "step": 2886 }, { "epoch": 7.058679706601467, "grad_norm": 0.16761452844997968, "learning_rate": 2.7307458827612075e-06, "loss": 0.625, "step": 2887 }, { "epoch": 7.061124694376528, "grad_norm": 0.33137347259019007, "learning_rate": 2.729471124592728e-06, "loss": 0.5385, "step": 2888 }, { "epoch": 7.06356968215159, "grad_norm": 0.2702831926650596, "learning_rate": 2.728196306251667e-06, "loss": 1.0968, "step": 2889 }, { "epoch": 7.06601466992665, "grad_norm": 0.2289703240948399, "learning_rate": 2.726921428072311e-06, "loss": 0.5042, "step": 2890 }, { "epoch": 7.068459657701712, "grad_norm": 0.15061584641231482, "learning_rate": 2.7256464903889614e-06, "loss": 0.6853, "step": 2891 }, { "epoch": 7.070904645476773, "grad_norm": 0.25300586316964646, "learning_rate": 2.724371493535937e-06, "loss": 0.4191, "step": 2892 }, { "epoch": 7.0733496332518335, "grad_norm": 0.21906541187270398, "learning_rate": 2.723096437847571e-06, "loss": 0.8312, "step": 2893 }, { "epoch": 7.075794621026895, "grad_norm": 0.19925159206363205, "learning_rate": 2.7218213236582114e-06, "loss": 0.7383, "step": 2894 }, { "epoch": 7.078239608801956, "grad_norm": 0.31152798439838525, "learning_rate": 2.7205461513022237e-06, "loss": 0.4381, "step": 2895 }, { "epoch": 7.080684596577017, "grad_norm": 0.3486666521629648, "learning_rate": 2.7192709211139855e-06, "loss": 0.3554, "step": 2896 }, { "epoch": 7.083129584352078, "grad_norm": 0.26358769216485556, "learning_rate": 2.7179956334278933e-06, "loss": 0.6358, "step": 2897 }, { "epoch": 7.08557457212714, "grad_norm": 0.27558719539655036, "learning_rate": 2.7167202885783563e-06, "loss": 0.4475, "step": 2898 }, { "epoch": 7.0880195599022, "grad_norm": 0.13034945863927172, "learning_rate": 2.7154448868997974e-06, "loss": 0.3353, "step": 2899 }, { "epoch": 7.090464547677262, "grad_norm": 0.21670235433208634, "learning_rate": 2.714169428726658e-06, "loss": 0.4382, "step": 2900 }, { "epoch": 7.092909535452323, "grad_norm": 0.16897747468006827, "learning_rate": 2.7128939143933914e-06, "loss": 0.5487, "step": 2901 }, { "epoch": 7.095354523227384, "grad_norm": 0.26073954062784044, "learning_rate": 2.7116183442344672e-06, "loss": 0.9542, "step": 2902 }, { "epoch": 7.097799511002445, "grad_norm": 0.3093657401425847, "learning_rate": 2.7103427185843687e-06, "loss": 0.4449, "step": 2903 }, { "epoch": 7.100244498777506, "grad_norm": 0.2155042553841293, "learning_rate": 2.7090670377775947e-06, "loss": 0.6862, "step": 2904 }, { "epoch": 7.102689486552567, "grad_norm": 0.19141219379610852, "learning_rate": 2.7077913021486574e-06, "loss": 0.5114, "step": 2905 }, { "epoch": 7.105134474327628, "grad_norm": 0.23082767452471636, "learning_rate": 2.706515512032084e-06, "loss": 0.5656, "step": 2906 }, { "epoch": 7.10757946210269, "grad_norm": 0.25779437907799657, "learning_rate": 2.705239667762416e-06, "loss": 0.5559, "step": 2907 }, { "epoch": 7.11002444987775, "grad_norm": 0.2695268250447425, "learning_rate": 2.7039637696742095e-06, "loss": 0.4092, "step": 2908 }, { "epoch": 7.112469437652812, "grad_norm": 0.23132305638522085, "learning_rate": 2.7026878181020328e-06, "loss": 0.4901, "step": 2909 }, { "epoch": 7.114914425427873, "grad_norm": 0.2665686056791714, "learning_rate": 2.70141181338047e-06, "loss": 0.3599, "step": 2910 }, { "epoch": 7.117359413202934, "grad_norm": 0.1679841045410817, "learning_rate": 2.700135755844119e-06, "loss": 0.584, "step": 2911 }, { "epoch": 7.119804400977995, "grad_norm": 0.2558683853890477, "learning_rate": 2.698859645827593e-06, "loss": 0.4002, "step": 2912 }, { "epoch": 7.1222493887530565, "grad_norm": 0.30949856081948834, "learning_rate": 2.6975834836655148e-06, "loss": 0.4174, "step": 2913 }, { "epoch": 7.124694376528117, "grad_norm": 0.3680640146746572, "learning_rate": 2.6963072696925247e-06, "loss": 0.3545, "step": 2914 }, { "epoch": 7.127139364303178, "grad_norm": 0.2821535466417029, "learning_rate": 2.6950310042432736e-06, "loss": 0.8195, "step": 2915 }, { "epoch": 7.12958435207824, "grad_norm": 0.32110940189330334, "learning_rate": 2.693754687652429e-06, "loss": 0.495, "step": 2916 }, { "epoch": 7.1320293398533, "grad_norm": 0.27139841801002657, "learning_rate": 2.692478320254671e-06, "loss": 0.7873, "step": 2917 }, { "epoch": 7.134474327628362, "grad_norm": 0.3627069503060499, "learning_rate": 2.6912019023846907e-06, "loss": 0.8211, "step": 2918 }, { "epoch": 7.136919315403423, "grad_norm": 0.2770772859573193, "learning_rate": 2.689925434377195e-06, "loss": 0.6837, "step": 2919 }, { "epoch": 7.139364303178484, "grad_norm": 0.22039436161782777, "learning_rate": 2.688648916566902e-06, "loss": 0.8853, "step": 2920 }, { "epoch": 7.141809290953545, "grad_norm": 0.2875204119327131, "learning_rate": 2.6873723492885446e-06, "loss": 0.5072, "step": 2921 }, { "epoch": 7.1442542787286065, "grad_norm": 0.2028580979142081, "learning_rate": 2.686095732876869e-06, "loss": 0.511, "step": 2922 }, { "epoch": 7.146699266503667, "grad_norm": 0.22505244739592664, "learning_rate": 2.684819067666632e-06, "loss": 0.6521, "step": 2923 }, { "epoch": 7.1491442542787285, "grad_norm": 0.2213350866866343, "learning_rate": 2.683542353992604e-06, "loss": 0.5769, "step": 2924 }, { "epoch": 7.15158924205379, "grad_norm": 0.33583246074464174, "learning_rate": 2.6822655921895696e-06, "loss": 0.6502, "step": 2925 }, { "epoch": 7.15403422982885, "grad_norm": 0.30042754931647303, "learning_rate": 2.6809887825923246e-06, "loss": 0.5514, "step": 2926 }, { "epoch": 7.156479217603912, "grad_norm": 0.31121705307781394, "learning_rate": 2.679711925535678e-06, "loss": 0.3817, "step": 2927 }, { "epoch": 7.158924205378973, "grad_norm": 0.24563668641683023, "learning_rate": 2.6784350213544506e-06, "loss": 0.5166, "step": 2928 }, { "epoch": 7.161369193154035, "grad_norm": 0.2259532147608978, "learning_rate": 2.6771580703834765e-06, "loss": 0.7099, "step": 2929 }, { "epoch": 7.163814180929095, "grad_norm": 0.22245470278937962, "learning_rate": 2.6758810729575996e-06, "loss": 0.4901, "step": 2930 }, { "epoch": 7.166259168704157, "grad_norm": 0.24832582897620054, "learning_rate": 2.67460402941168e-06, "loss": 0.9396, "step": 2931 }, { "epoch": 7.168704156479218, "grad_norm": 0.19387990831694119, "learning_rate": 2.6733269400805873e-06, "loss": 0.4139, "step": 2932 }, { "epoch": 7.1711491442542785, "grad_norm": 0.20145170067938317, "learning_rate": 2.6720498052992027e-06, "loss": 0.682, "step": 2933 }, { "epoch": 7.17359413202934, "grad_norm": 0.22821181254687828, "learning_rate": 2.670772625402421e-06, "loss": 0.6827, "step": 2934 }, { "epoch": 7.176039119804401, "grad_norm": 0.2549198815844666, "learning_rate": 2.6694954007251473e-06, "loss": 0.7026, "step": 2935 }, { "epoch": 7.178484107579462, "grad_norm": 0.27325397442169697, "learning_rate": 2.6682181316022997e-06, "loss": 0.4772, "step": 2936 }, { "epoch": 7.180929095354523, "grad_norm": 0.26054492466893603, "learning_rate": 2.6669408183688073e-06, "loss": 0.5789, "step": 2937 }, { "epoch": 7.183374083129585, "grad_norm": 0.313714585961581, "learning_rate": 2.665663461359611e-06, "loss": 0.6954, "step": 2938 }, { "epoch": 7.185819070904645, "grad_norm": 0.20054052361492203, "learning_rate": 2.6643860609096627e-06, "loss": 0.6214, "step": 2939 }, { "epoch": 7.188264058679707, "grad_norm": 0.22081203247082012, "learning_rate": 2.663108617353926e-06, "loss": 0.575, "step": 2940 }, { "epoch": 7.190709046454768, "grad_norm": 0.16713815639057567, "learning_rate": 2.661831131027377e-06, "loss": 0.8715, "step": 2941 }, { "epoch": 7.193154034229829, "grad_norm": 0.24276798644266592, "learning_rate": 2.660553602265e-06, "loss": 0.7361, "step": 2942 }, { "epoch": 7.19559902200489, "grad_norm": 0.2605901492973475, "learning_rate": 2.6592760314017927e-06, "loss": 0.4941, "step": 2943 }, { "epoch": 7.198044009779951, "grad_norm": 0.1786826033971783, "learning_rate": 2.6579984187727653e-06, "loss": 0.5443, "step": 2944 }, { "epoch": 7.200488997555012, "grad_norm": 0.2606748743012102, "learning_rate": 2.6567207647129343e-06, "loss": 0.3753, "step": 2945 }, { "epoch": 7.202933985330073, "grad_norm": 0.19414882758983018, "learning_rate": 2.6554430695573323e-06, "loss": 0.3664, "step": 2946 }, { "epoch": 7.205378973105135, "grad_norm": 0.2264733794119016, "learning_rate": 2.654165333640999e-06, "loss": 0.6219, "step": 2947 }, { "epoch": 7.207823960880195, "grad_norm": 0.19034874069577673, "learning_rate": 2.652887557298986e-06, "loss": 0.6806, "step": 2948 }, { "epoch": 7.210268948655257, "grad_norm": 0.2752382280083933, "learning_rate": 2.6516097408663554e-06, "loss": 0.6122, "step": 2949 }, { "epoch": 7.212713936430318, "grad_norm": 0.3066266089459573, "learning_rate": 2.6503318846781806e-06, "loss": 0.7968, "step": 2950 }, { "epoch": 7.215158924205379, "grad_norm": 0.16419497836946748, "learning_rate": 2.6490539890695444e-06, "loss": 0.3694, "step": 2951 }, { "epoch": 7.21760391198044, "grad_norm": 0.17491037978819649, "learning_rate": 2.6477760543755397e-06, "loss": 0.4807, "step": 2952 }, { "epoch": 7.2200488997555015, "grad_norm": 0.24596675408316465, "learning_rate": 2.6464980809312714e-06, "loss": 0.4241, "step": 2953 }, { "epoch": 7.222493887530562, "grad_norm": 0.21827950017245704, "learning_rate": 2.645220069071852e-06, "loss": 0.7709, "step": 2954 }, { "epoch": 7.224938875305623, "grad_norm": 0.2578390717253169, "learning_rate": 2.6439420191324068e-06, "loss": 0.42, "step": 2955 }, { "epoch": 7.227383863080685, "grad_norm": 0.2694821835003641, "learning_rate": 2.642663931448069e-06, "loss": 0.5034, "step": 2956 }, { "epoch": 7.229828850855745, "grad_norm": 0.20612655245266992, "learning_rate": 2.6413858063539825e-06, "loss": 1.0793, "step": 2957 }, { "epoch": 7.232273838630807, "grad_norm": 0.2426643406424326, "learning_rate": 2.640107644185301e-06, "loss": 0.5392, "step": 2958 }, { "epoch": 7.234718826405868, "grad_norm": 0.16444199449606253, "learning_rate": 2.6388294452771877e-06, "loss": 0.4558, "step": 2959 }, { "epoch": 7.237163814180929, "grad_norm": 0.2619059134833961, "learning_rate": 2.637551209964816e-06, "loss": 0.552, "step": 2960 }, { "epoch": 7.23960880195599, "grad_norm": 0.18439667078066344, "learning_rate": 2.636272938583369e-06, "loss": 0.7145, "step": 2961 }, { "epoch": 7.2420537897310515, "grad_norm": 0.23863734257390962, "learning_rate": 2.6349946314680377e-06, "loss": 0.4409, "step": 2962 }, { "epoch": 7.244498777506112, "grad_norm": 0.2198552145322951, "learning_rate": 2.633716288954023e-06, "loss": 0.6169, "step": 2963 }, { "epoch": 7.2469437652811735, "grad_norm": 0.22326119636185135, "learning_rate": 2.6324379113765375e-06, "loss": 0.4382, "step": 2964 }, { "epoch": 7.249388753056235, "grad_norm": 0.27014790237821074, "learning_rate": 2.631159499070799e-06, "loss": 0.4749, "step": 2965 }, { "epoch": 7.251833740831296, "grad_norm": 0.24081356311916513, "learning_rate": 2.629881052372038e-06, "loss": 0.6134, "step": 2966 }, { "epoch": 7.254278728606357, "grad_norm": 0.17161117902722794, "learning_rate": 2.628602571615492e-06, "loss": 0.7874, "step": 2967 }, { "epoch": 7.256723716381418, "grad_norm": 0.2398725251477415, "learning_rate": 2.6273240571364085e-06, "loss": 0.5432, "step": 2968 }, { "epoch": 7.25916870415648, "grad_norm": 0.19982384076847412, "learning_rate": 2.626045509270041e-06, "loss": 0.5497, "step": 2969 }, { "epoch": 7.26161369193154, "grad_norm": 0.1894593131207446, "learning_rate": 2.6247669283516556e-06, "loss": 0.3953, "step": 2970 }, { "epoch": 7.264058679706602, "grad_norm": 0.1687955973744008, "learning_rate": 2.623488314716527e-06, "loss": 0.6704, "step": 2971 }, { "epoch": 7.266503667481663, "grad_norm": 0.27730200852178416, "learning_rate": 2.622209668699935e-06, "loss": 0.8758, "step": 2972 }, { "epoch": 7.2689486552567235, "grad_norm": 0.17811972796571376, "learning_rate": 2.62093099063717e-06, "loss": 0.8755, "step": 2973 }, { "epoch": 7.271393643031785, "grad_norm": 0.21224666534498302, "learning_rate": 2.6196522808635312e-06, "loss": 0.9706, "step": 2974 }, { "epoch": 7.273838630806846, "grad_norm": 0.32741921302892496, "learning_rate": 2.6183735397143245e-06, "loss": 0.5269, "step": 2975 }, { "epoch": 7.276283618581907, "grad_norm": 0.274112751574388, "learning_rate": 2.617094767524867e-06, "loss": 0.5993, "step": 2976 }, { "epoch": 7.278728606356968, "grad_norm": 0.3280601352893266, "learning_rate": 2.615815964630481e-06, "loss": 0.4208, "step": 2977 }, { "epoch": 7.28117359413203, "grad_norm": 0.310599068435982, "learning_rate": 2.6145371313664974e-06, "loss": 0.5437, "step": 2978 }, { "epoch": 7.28361858190709, "grad_norm": 0.2557657092366433, "learning_rate": 2.6132582680682555e-06, "loss": 0.5243, "step": 2979 }, { "epoch": 7.286063569682152, "grad_norm": 0.2930215435854808, "learning_rate": 2.6119793750711037e-06, "loss": 0.4333, "step": 2980 }, { "epoch": 7.288508557457213, "grad_norm": 0.17914355849993288, "learning_rate": 2.610700452710397e-06, "loss": 0.6089, "step": 2981 }, { "epoch": 7.290953545232274, "grad_norm": 0.1947525712560195, "learning_rate": 2.6094215013214968e-06, "loss": 0.5047, "step": 2982 }, { "epoch": 7.293398533007335, "grad_norm": 0.27300799867964864, "learning_rate": 2.6081425212397743e-06, "loss": 0.5266, "step": 2983 }, { "epoch": 7.295843520782396, "grad_norm": 0.253010527314013, "learning_rate": 2.606863512800607e-06, "loss": 0.6578, "step": 2984 }, { "epoch": 7.298288508557457, "grad_norm": 0.15130497976652685, "learning_rate": 2.605584476339381e-06, "loss": 0.3647, "step": 2985 }, { "epoch": 7.300733496332518, "grad_norm": 0.2721948823865963, "learning_rate": 2.6043054121914874e-06, "loss": 0.6244, "step": 2986 }, { "epoch": 7.30317848410758, "grad_norm": 0.26666221734950074, "learning_rate": 2.603026320692328e-06, "loss": 0.4325, "step": 2987 }, { "epoch": 7.30562347188264, "grad_norm": 0.18867001521198126, "learning_rate": 2.601747202177308e-06, "loss": 0.5654, "step": 2988 }, { "epoch": 7.308068459657702, "grad_norm": 0.16113026859571675, "learning_rate": 2.600468056981843e-06, "loss": 0.467, "step": 2989 }, { "epoch": 7.310513447432763, "grad_norm": 0.23391896775037566, "learning_rate": 2.599188885441353e-06, "loss": 0.495, "step": 2990 }, { "epoch": 7.312958435207824, "grad_norm": 0.24518063952663302, "learning_rate": 2.5979096878912674e-06, "loss": 0.642, "step": 2991 }, { "epoch": 7.315403422982885, "grad_norm": 0.24886738902605468, "learning_rate": 2.5966304646670205e-06, "loss": 0.6152, "step": 2992 }, { "epoch": 7.3178484107579465, "grad_norm": 0.38492841103188963, "learning_rate": 2.595351216104054e-06, "loss": 0.4035, "step": 2993 }, { "epoch": 7.320293398533007, "grad_norm": 0.23033118490359544, "learning_rate": 2.5940719425378146e-06, "loss": 0.5255, "step": 2994 }, { "epoch": 7.322738386308068, "grad_norm": 0.18682112409446436, "learning_rate": 2.592792644303759e-06, "loss": 0.3984, "step": 2995 }, { "epoch": 7.32518337408313, "grad_norm": 0.2569320141099276, "learning_rate": 2.5915133217373486e-06, "loss": 0.4159, "step": 2996 }, { "epoch": 7.32762836185819, "grad_norm": 0.20084271876316015, "learning_rate": 2.5902339751740495e-06, "loss": 0.4319, "step": 2997 }, { "epoch": 7.330073349633252, "grad_norm": 0.21885052279069006, "learning_rate": 2.588954604949337e-06, "loss": 0.4847, "step": 2998 }, { "epoch": 7.332518337408313, "grad_norm": 0.2607211894337378, "learning_rate": 2.58767521139869e-06, "loss": 0.6601, "step": 2999 }, { "epoch": 7.334963325183374, "grad_norm": 0.18573235146562372, "learning_rate": 2.586395794857597e-06, "loss": 0.7328, "step": 3000 }, { "epoch": 7.337408312958435, "grad_norm": 0.19264480822199387, "learning_rate": 2.5851163556615477e-06, "loss": 0.3931, "step": 3001 }, { "epoch": 7.3398533007334965, "grad_norm": 0.15248358598988995, "learning_rate": 2.5838368941460424e-06, "loss": 0.401, "step": 3002 }, { "epoch": 7.342298288508557, "grad_norm": 0.2635920588085101, "learning_rate": 2.582557410646584e-06, "loss": 0.4985, "step": 3003 }, { "epoch": 7.3447432762836184, "grad_norm": 0.24724241576032033, "learning_rate": 2.581277905498682e-06, "loss": 0.5974, "step": 3004 }, { "epoch": 7.34718826405868, "grad_norm": 0.3105746730882682, "learning_rate": 2.579998379037853e-06, "loss": 0.6453, "step": 3005 }, { "epoch": 7.34963325183374, "grad_norm": 0.2837998957809835, "learning_rate": 2.5787188315996174e-06, "loss": 0.979, "step": 3006 }, { "epoch": 7.352078239608802, "grad_norm": 0.17875702232896482, "learning_rate": 2.5774392635195034e-06, "loss": 0.3907, "step": 3007 }, { "epoch": 7.354523227383863, "grad_norm": 0.22796479694519697, "learning_rate": 2.576159675133041e-06, "loss": 0.3553, "step": 3008 }, { "epoch": 7.356968215158924, "grad_norm": 0.21862653271171265, "learning_rate": 2.574880066775768e-06, "loss": 0.8928, "step": 3009 }, { "epoch": 7.359413202933985, "grad_norm": 0.22612851527264033, "learning_rate": 2.573600438783228e-06, "loss": 1.0224, "step": 3010 }, { "epoch": 7.361858190709047, "grad_norm": 0.31028712900532934, "learning_rate": 2.5723207914909683e-06, "loss": 0.4225, "step": 3011 }, { "epoch": 7.364303178484108, "grad_norm": 0.18582083063007593, "learning_rate": 2.571041125234541e-06, "loss": 0.6432, "step": 3012 }, { "epoch": 7.3667481662591685, "grad_norm": 0.39592482881641045, "learning_rate": 2.5697614403495047e-06, "loss": 0.4614, "step": 3013 }, { "epoch": 7.36919315403423, "grad_norm": 0.32490787442781377, "learning_rate": 2.5684817371714215e-06, "loss": 0.7597, "step": 3014 }, { "epoch": 7.371638141809291, "grad_norm": 0.24597339568802967, "learning_rate": 2.5672020160358594e-06, "loss": 0.6171, "step": 3015 }, { "epoch": 7.374083129584352, "grad_norm": 0.13920933361593787, "learning_rate": 2.56592227727839e-06, "loss": 0.4047, "step": 3016 }, { "epoch": 7.376528117359413, "grad_norm": 0.3537658207729003, "learning_rate": 2.564642521234591e-06, "loss": 0.3657, "step": 3017 }, { "epoch": 7.378973105134475, "grad_norm": 0.2527338755418982, "learning_rate": 2.563362748240043e-06, "loss": 0.4189, "step": 3018 }, { "epoch": 7.381418092909535, "grad_norm": 0.28029910000597685, "learning_rate": 2.562082958630332e-06, "loss": 0.7656, "step": 3019 }, { "epoch": 7.383863080684597, "grad_norm": 0.2868369160858378, "learning_rate": 2.560803152741048e-06, "loss": 0.3976, "step": 3020 }, { "epoch": 7.386308068459658, "grad_norm": 0.33983470571230534, "learning_rate": 2.559523330907786e-06, "loss": 0.4035, "step": 3021 }, { "epoch": 7.388753056234719, "grad_norm": 0.20548505186591842, "learning_rate": 2.558243493466143e-06, "loss": 0.505, "step": 3022 }, { "epoch": 7.39119804400978, "grad_norm": 0.22560695512715223, "learning_rate": 2.5569636407517236e-06, "loss": 0.5681, "step": 3023 }, { "epoch": 7.393643031784841, "grad_norm": 0.30957612893826314, "learning_rate": 2.5556837731001327e-06, "loss": 0.4167, "step": 3024 }, { "epoch": 7.396088019559902, "grad_norm": 0.25213066813495016, "learning_rate": 2.554403890846983e-06, "loss": 0.4206, "step": 3025 }, { "epoch": 7.398533007334963, "grad_norm": 0.29377561120107787, "learning_rate": 2.5531239943278875e-06, "loss": 0.4275, "step": 3026 }, { "epoch": 7.400977995110025, "grad_norm": 0.24972624162853088, "learning_rate": 2.551844083878464e-06, "loss": 0.6882, "step": 3027 }, { "epoch": 7.403422982885085, "grad_norm": 0.1800902171482418, "learning_rate": 2.550564159834335e-06, "loss": 0.387, "step": 3028 }, { "epoch": 7.405867970660147, "grad_norm": 0.29228083711624503, "learning_rate": 2.549284222531126e-06, "loss": 0.4235, "step": 3029 }, { "epoch": 7.408312958435208, "grad_norm": 0.2377301917286694, "learning_rate": 2.548004272304466e-06, "loss": 0.5987, "step": 3030 }, { "epoch": 7.410757946210269, "grad_norm": 0.25098074720481267, "learning_rate": 2.546724309489986e-06, "loss": 0.7531, "step": 3031 }, { "epoch": 7.41320293398533, "grad_norm": 0.1585232265398779, "learning_rate": 2.5454443344233237e-06, "loss": 0.4742, "step": 3032 }, { "epoch": 7.4156479217603914, "grad_norm": 0.28072409603685156, "learning_rate": 2.5441643474401154e-06, "loss": 0.6785, "step": 3033 }, { "epoch": 7.418092909535452, "grad_norm": 0.2604901554050247, "learning_rate": 2.5428843488760047e-06, "loss": 0.5039, "step": 3034 }, { "epoch": 7.420537897310513, "grad_norm": 0.27961694355764205, "learning_rate": 2.5416043390666366e-06, "loss": 0.8087, "step": 3035 }, { "epoch": 7.422982885085575, "grad_norm": 0.22774640603189428, "learning_rate": 2.5403243183476583e-06, "loss": 0.3912, "step": 3036 }, { "epoch": 7.425427872860635, "grad_norm": 0.2501693371116074, "learning_rate": 2.5390442870547204e-06, "loss": 0.897, "step": 3037 }, { "epoch": 7.427872860635697, "grad_norm": 0.16049190845485004, "learning_rate": 2.5377642455234764e-06, "loss": 0.641, "step": 3038 }, { "epoch": 7.430317848410758, "grad_norm": 0.2490419741601378, "learning_rate": 2.5364841940895824e-06, "loss": 0.9267, "step": 3039 }, { "epoch": 7.432762836185819, "grad_norm": 0.39942779544388296, "learning_rate": 2.5352041330886983e-06, "loss": 0.7282, "step": 3040 }, { "epoch": 7.43520782396088, "grad_norm": 0.33902692858378725, "learning_rate": 2.533924062856485e-06, "loss": 0.4725, "step": 3041 }, { "epoch": 7.4376528117359415, "grad_norm": 0.16852391063269725, "learning_rate": 2.5326439837286056e-06, "loss": 0.4283, "step": 3042 }, { "epoch": 7.440097799511002, "grad_norm": 0.27402820569518577, "learning_rate": 2.5313638960407257e-06, "loss": 0.4424, "step": 3043 }, { "epoch": 7.442542787286063, "grad_norm": 0.1691815766016492, "learning_rate": 2.5300838001285138e-06, "loss": 0.4659, "step": 3044 }, { "epoch": 7.444987775061125, "grad_norm": 0.21312398871392432, "learning_rate": 2.5288036963276418e-06, "loss": 0.6387, "step": 3045 }, { "epoch": 7.447432762836186, "grad_norm": 0.2580434180449843, "learning_rate": 2.5275235849737816e-06, "loss": 0.3497, "step": 3046 }, { "epoch": 7.449877750611247, "grad_norm": 0.22142529211194736, "learning_rate": 2.526243466402607e-06, "loss": 0.5879, "step": 3047 }, { "epoch": 7.452322738386308, "grad_norm": 0.28453282966537363, "learning_rate": 2.524963340949795e-06, "loss": 0.6065, "step": 3048 }, { "epoch": 7.45476772616137, "grad_norm": 0.25583762718860864, "learning_rate": 2.523683208951023e-06, "loss": 0.5536, "step": 3049 }, { "epoch": 7.45721271393643, "grad_norm": 0.2513097814568369, "learning_rate": 2.522403070741971e-06, "loss": 0.5697, "step": 3050 }, { "epoch": 7.459657701711492, "grad_norm": 0.24574963961874632, "learning_rate": 2.5211229266583227e-06, "loss": 0.9711, "step": 3051 }, { "epoch": 7.462102689486553, "grad_norm": 0.17398616180687465, "learning_rate": 2.519842777035759e-06, "loss": 0.8388, "step": 3052 }, { "epoch": 7.4645476772616135, "grad_norm": 0.2483442806086505, "learning_rate": 2.5185626222099634e-06, "loss": 0.9939, "step": 3053 }, { "epoch": 7.466992665036675, "grad_norm": 0.332713206201648, "learning_rate": 2.5172824625166246e-06, "loss": 0.6806, "step": 3054 }, { "epoch": 7.469437652811736, "grad_norm": 0.1952017106663902, "learning_rate": 2.5160022982914283e-06, "loss": 0.4928, "step": 3055 }, { "epoch": 7.471882640586797, "grad_norm": 0.23318188727798092, "learning_rate": 2.514722129870063e-06, "loss": 0.9781, "step": 3056 }, { "epoch": 7.474327628361858, "grad_norm": 0.2955992672319146, "learning_rate": 2.5134419575882186e-06, "loss": 0.4968, "step": 3057 }, { "epoch": 7.47677261613692, "grad_norm": 0.19652001076603975, "learning_rate": 2.512161781781585e-06, "loss": 0.6841, "step": 3058 }, { "epoch": 7.47921760391198, "grad_norm": 0.27635151418185444, "learning_rate": 2.5108816027858534e-06, "loss": 0.4746, "step": 3059 }, { "epoch": 7.481662591687042, "grad_norm": 0.20029156472941967, "learning_rate": 2.509601420936717e-06, "loss": 0.4187, "step": 3060 }, { "epoch": 7.484107579462103, "grad_norm": 0.1940639684185768, "learning_rate": 2.508321236569868e-06, "loss": 0.4481, "step": 3061 }, { "epoch": 7.486552567237164, "grad_norm": 0.3534214377274368, "learning_rate": 2.507041050021001e-06, "loss": 0.8606, "step": 3062 }, { "epoch": 7.488997555012225, "grad_norm": 0.24147646593089964, "learning_rate": 2.505760861625808e-06, "loss": 0.389, "step": 3063 }, { "epoch": 7.491442542787286, "grad_norm": 0.2827565574547815, "learning_rate": 2.5044806717199867e-06, "loss": 0.3694, "step": 3064 }, { "epoch": 7.493887530562347, "grad_norm": 0.19286131188641506, "learning_rate": 2.5032004806392296e-06, "loss": 0.6074, "step": 3065 }, { "epoch": 7.496332518337408, "grad_norm": 0.22638559719212145, "learning_rate": 2.5019202887192335e-06, "loss": 0.4574, "step": 3066 }, { "epoch": 7.49877750611247, "grad_norm": 0.2197834027513471, "learning_rate": 2.5006400962956946e-06, "loss": 0.3832, "step": 3067 }, { "epoch": 7.50122249388753, "grad_norm": 0.299025486887785, "learning_rate": 2.4993599037043063e-06, "loss": 0.6896, "step": 3068 }, { "epoch": 7.503667481662592, "grad_norm": 0.2095614338150072, "learning_rate": 2.498079711280767e-06, "loss": 0.6673, "step": 3069 }, { "epoch": 7.506112469437653, "grad_norm": 0.2958210297138419, "learning_rate": 2.4967995193607713e-06, "loss": 0.442, "step": 3070 }, { "epoch": 7.508557457212714, "grad_norm": 0.20643013406650393, "learning_rate": 2.495519328280015e-06, "loss": 0.8165, "step": 3071 }, { "epoch": 7.511002444987775, "grad_norm": 0.22687363631541038, "learning_rate": 2.4942391383741925e-06, "loss": 0.6408, "step": 3072 }, { "epoch": 7.513447432762836, "grad_norm": 0.17539261811180762, "learning_rate": 2.4929589499790004e-06, "loss": 0.6777, "step": 3073 }, { "epoch": 7.515892420537897, "grad_norm": 0.25795241122804413, "learning_rate": 2.491678763430132e-06, "loss": 0.5883, "step": 3074 }, { "epoch": 7.518337408312958, "grad_norm": 0.2542978446735659, "learning_rate": 2.4903985790632835e-06, "loss": 0.7852, "step": 3075 }, { "epoch": 7.52078239608802, "grad_norm": 0.23306459767291018, "learning_rate": 2.489118397214147e-06, "loss": 0.8453, "step": 3076 }, { "epoch": 7.52322738386308, "grad_norm": 0.19316702078147083, "learning_rate": 2.487838218218416e-06, "loss": 0.5335, "step": 3077 }, { "epoch": 7.525672371638142, "grad_norm": 0.2412896765339769, "learning_rate": 2.486558042411782e-06, "loss": 0.7603, "step": 3078 }, { "epoch": 7.528117359413203, "grad_norm": 0.24760459132221221, "learning_rate": 2.4852778701299375e-06, "loss": 0.3275, "step": 3079 }, { "epoch": 7.530562347188264, "grad_norm": 0.18954719812295986, "learning_rate": 2.4839977017085726e-06, "loss": 0.6596, "step": 3080 }, { "epoch": 7.533007334963325, "grad_norm": 0.3242798383202998, "learning_rate": 2.4827175374833767e-06, "loss": 0.3886, "step": 3081 }, { "epoch": 7.5354523227383865, "grad_norm": 0.16208012684464868, "learning_rate": 2.481437377790037e-06, "loss": 0.7222, "step": 3082 }, { "epoch": 7.537897310513447, "grad_norm": 0.20860461337524006, "learning_rate": 2.480157222964243e-06, "loss": 0.4343, "step": 3083 }, { "epoch": 7.540342298288508, "grad_norm": 0.21025417567686325, "learning_rate": 2.4788770733416777e-06, "loss": 0.4839, "step": 3084 }, { "epoch": 7.54278728606357, "grad_norm": 0.31637854851569014, "learning_rate": 2.4775969292580287e-06, "loss": 0.3542, "step": 3085 }, { "epoch": 7.54523227383863, "grad_norm": 0.27428815704601006, "learning_rate": 2.4763167910489775e-06, "loss": 0.6499, "step": 3086 }, { "epoch": 7.547677261613692, "grad_norm": 0.1851423567150205, "learning_rate": 2.4750366590502063e-06, "loss": 0.5171, "step": 3087 }, { "epoch": 7.550122249388753, "grad_norm": 0.3918767615715888, "learning_rate": 2.4737565335973934e-06, "loss": 0.584, "step": 3088 }, { "epoch": 7.552567237163814, "grad_norm": 0.136371493376507, "learning_rate": 2.4724764150262188e-06, "loss": 0.4393, "step": 3089 }, { "epoch": 7.555012224938875, "grad_norm": 0.21472356174138402, "learning_rate": 2.4711963036723586e-06, "loss": 0.4288, "step": 3090 }, { "epoch": 7.557457212713937, "grad_norm": 1.1333014224540965, "learning_rate": 2.4699161998714866e-06, "loss": 0.6945, "step": 3091 }, { "epoch": 7.559902200488997, "grad_norm": 0.21766224747912963, "learning_rate": 2.4686361039592756e-06, "loss": 0.4151, "step": 3092 }, { "epoch": 7.5623471882640585, "grad_norm": 0.24123092111061495, "learning_rate": 2.467356016271396e-06, "loss": 0.3888, "step": 3093 }, { "epoch": 7.56479217603912, "grad_norm": 0.2530918459185112, "learning_rate": 2.466075937143517e-06, "loss": 0.8052, "step": 3094 }, { "epoch": 7.56723716381418, "grad_norm": 0.26064207752990876, "learning_rate": 2.4647958669113017e-06, "loss": 0.4154, "step": 3095 }, { "epoch": 7.569682151589242, "grad_norm": 0.2969673794101501, "learning_rate": 2.463515805910418e-06, "loss": 1.0655, "step": 3096 }, { "epoch": 7.572127139364303, "grad_norm": 0.29479804469298926, "learning_rate": 2.4622357544765244e-06, "loss": 0.5498, "step": 3097 }, { "epoch": 7.574572127139365, "grad_norm": 0.2570075059243293, "learning_rate": 2.46095571294528e-06, "loss": 0.3113, "step": 3098 }, { "epoch": 7.577017114914425, "grad_norm": 0.21717549246514384, "learning_rate": 2.4596756816523425e-06, "loss": 0.7294, "step": 3099 }, { "epoch": 7.579462102689487, "grad_norm": 0.20854011607966358, "learning_rate": 2.458395660933364e-06, "loss": 0.5262, "step": 3100 }, { "epoch": 7.581907090464548, "grad_norm": 0.20348584720553187, "learning_rate": 2.4571156511239957e-06, "loss": 0.7973, "step": 3101 }, { "epoch": 7.5843520782396086, "grad_norm": 0.20996518095982614, "learning_rate": 2.455835652559885e-06, "loss": 0.5112, "step": 3102 }, { "epoch": 7.58679706601467, "grad_norm": 0.2798613367066497, "learning_rate": 2.4545556655766776e-06, "loss": 0.4554, "step": 3103 }, { "epoch": 7.589242053789731, "grad_norm": 0.21628406734140962, "learning_rate": 2.4532756905100146e-06, "loss": 0.8099, "step": 3104 }, { "epoch": 7.591687041564792, "grad_norm": 0.24931404934942428, "learning_rate": 2.4519957276955354e-06, "loss": 1.1828, "step": 3105 }, { "epoch": 7.594132029339853, "grad_norm": 0.2970211474157629, "learning_rate": 2.4507157774688743e-06, "loss": 0.6834, "step": 3106 }, { "epoch": 7.596577017114915, "grad_norm": 0.18084727239322743, "learning_rate": 2.4494358401656654e-06, "loss": 0.3969, "step": 3107 }, { "epoch": 7.599022004889975, "grad_norm": 0.25043634630076206, "learning_rate": 2.4481559161215364e-06, "loss": 0.3496, "step": 3108 }, { "epoch": 7.601466992665037, "grad_norm": 0.23266827103149407, "learning_rate": 2.4468760056721134e-06, "loss": 0.6346, "step": 3109 }, { "epoch": 7.603911980440098, "grad_norm": 0.2709130000379762, "learning_rate": 2.445596109153018e-06, "loss": 1.2344, "step": 3110 }, { "epoch": 7.606356968215159, "grad_norm": 0.22920038629441936, "learning_rate": 2.4443162268998677e-06, "loss": 0.3802, "step": 3111 }, { "epoch": 7.60880195599022, "grad_norm": 0.3238605862359861, "learning_rate": 2.4430363592482772e-06, "loss": 0.5317, "step": 3112 }, { "epoch": 7.611246943765281, "grad_norm": 0.18694501470826708, "learning_rate": 2.441756506533858e-06, "loss": 0.6354, "step": 3113 }, { "epoch": 7.613691931540342, "grad_norm": 0.12149681604714048, "learning_rate": 2.4404766690922154e-06, "loss": 0.2385, "step": 3114 }, { "epoch": 7.616136919315403, "grad_norm": 0.20149419441038058, "learning_rate": 2.439196847258953e-06, "loss": 0.4682, "step": 3115 }, { "epoch": 7.618581907090465, "grad_norm": 0.18316129613057858, "learning_rate": 2.4379170413696683e-06, "loss": 0.6557, "step": 3116 }, { "epoch": 7.621026894865525, "grad_norm": 0.24878583243713595, "learning_rate": 2.4366372517599573e-06, "loss": 0.3668, "step": 3117 }, { "epoch": 7.623471882640587, "grad_norm": 0.19572351963178702, "learning_rate": 2.435357478765409e-06, "loss": 0.5577, "step": 3118 }, { "epoch": 7.625916870415648, "grad_norm": 0.24939790969271688, "learning_rate": 2.4340777227216102e-06, "loss": 0.7628, "step": 3119 }, { "epoch": 7.628361858190709, "grad_norm": 0.29580147799702244, "learning_rate": 2.4327979839641414e-06, "loss": 1.1039, "step": 3120 }, { "epoch": 7.63080684596577, "grad_norm": 0.24793205193479567, "learning_rate": 2.4315182628285794e-06, "loss": 1.0408, "step": 3121 }, { "epoch": 7.6332518337408315, "grad_norm": 0.23509392737206194, "learning_rate": 2.430238559650496e-06, "loss": 0.4099, "step": 3122 }, { "epoch": 7.635696821515893, "grad_norm": 0.24806324467933052, "learning_rate": 2.42895887476546e-06, "loss": 0.369, "step": 3123 }, { "epoch": 7.638141809290953, "grad_norm": 0.2037916016765037, "learning_rate": 2.427679208509033e-06, "loss": 0.6029, "step": 3124 }, { "epoch": 7.640586797066015, "grad_norm": 0.28029880706299704, "learning_rate": 2.4263995612167733e-06, "loss": 1.2196, "step": 3125 }, { "epoch": 7.643031784841076, "grad_norm": 0.26595026334896166, "learning_rate": 2.4251199332242326e-06, "loss": 1.1411, "step": 3126 }, { "epoch": 7.645476772616137, "grad_norm": 0.23682166046414943, "learning_rate": 2.42384032486696e-06, "loss": 0.4842, "step": 3127 }, { "epoch": 7.647921760391198, "grad_norm": 0.25539084379869303, "learning_rate": 2.422560736480497e-06, "loss": 0.3856, "step": 3128 }, { "epoch": 7.65036674816626, "grad_norm": 0.26142277643795847, "learning_rate": 2.4212811684003826e-06, "loss": 0.3374, "step": 3129 }, { "epoch": 7.65281173594132, "grad_norm": 0.2257427591586434, "learning_rate": 2.4200016209621473e-06, "loss": 0.3485, "step": 3130 }, { "epoch": 7.6552567237163816, "grad_norm": 0.33243368849347354, "learning_rate": 2.4187220945013187e-06, "loss": 0.4164, "step": 3131 }, { "epoch": 7.657701711491443, "grad_norm": 0.18678351332570467, "learning_rate": 2.417442589353417e-06, "loss": 0.3696, "step": 3132 }, { "epoch": 7.6601466992665035, "grad_norm": 0.18927821584272086, "learning_rate": 2.4161631058539584e-06, "loss": 0.6585, "step": 3133 }, { "epoch": 7.662591687041565, "grad_norm": 0.26544126333496704, "learning_rate": 2.414883644338453e-06, "loss": 0.6283, "step": 3134 }, { "epoch": 7.665036674816626, "grad_norm": 0.3545344959476537, "learning_rate": 2.4136042051424044e-06, "loss": 0.7948, "step": 3135 }, { "epoch": 7.667481662591687, "grad_norm": 0.2381216169463349, "learning_rate": 2.4123247886013103e-06, "loss": 0.3896, "step": 3136 }, { "epoch": 7.669926650366748, "grad_norm": 0.19688833541210066, "learning_rate": 2.411045395050664e-06, "loss": 0.4053, "step": 3137 }, { "epoch": 7.67237163814181, "grad_norm": 0.2632038454551941, "learning_rate": 2.4097660248259505e-06, "loss": 0.7826, "step": 3138 }, { "epoch": 7.67481662591687, "grad_norm": 0.24262971968419264, "learning_rate": 2.408486678262652e-06, "loss": 0.7517, "step": 3139 }, { "epoch": 7.677261613691932, "grad_norm": 0.19645329809199671, "learning_rate": 2.4072073556962413e-06, "loss": 0.4941, "step": 3140 }, { "epoch": 7.679706601466993, "grad_norm": 0.3042721898639073, "learning_rate": 2.405928057462186e-06, "loss": 0.4733, "step": 3141 }, { "epoch": 7.6821515892420535, "grad_norm": 0.15354177989267473, "learning_rate": 2.404648783895947e-06, "loss": 0.6951, "step": 3142 }, { "epoch": 7.684596577017115, "grad_norm": 0.24551650479184892, "learning_rate": 2.40336953533298e-06, "loss": 0.4915, "step": 3143 }, { "epoch": 7.687041564792176, "grad_norm": 0.20497693746367474, "learning_rate": 2.402090312108733e-06, "loss": 0.5237, "step": 3144 }, { "epoch": 7.689486552567237, "grad_norm": 0.23141152778065371, "learning_rate": 2.4008111145586473e-06, "loss": 0.3957, "step": 3145 }, { "epoch": 7.691931540342298, "grad_norm": 0.24998408901827956, "learning_rate": 2.399531943018158e-06, "loss": 0.7646, "step": 3146 }, { "epoch": 7.69437652811736, "grad_norm": 0.1812420897044591, "learning_rate": 2.3982527978226928e-06, "loss": 0.4292, "step": 3147 }, { "epoch": 7.69682151589242, "grad_norm": 0.24762521513003954, "learning_rate": 2.3969736793076724e-06, "loss": 0.6067, "step": 3148 }, { "epoch": 7.699266503667482, "grad_norm": 0.28776322788633224, "learning_rate": 2.3956945878085126e-06, "loss": 0.4628, "step": 3149 }, { "epoch": 7.701711491442543, "grad_norm": 0.19873109913489917, "learning_rate": 2.3944155236606196e-06, "loss": 0.4561, "step": 3150 }, { "epoch": 7.704156479217604, "grad_norm": 0.2718957934517454, "learning_rate": 2.3931364871993936e-06, "loss": 0.711, "step": 3151 }, { "epoch": 7.706601466992665, "grad_norm": 0.3742228184716042, "learning_rate": 2.391857478760226e-06, "loss": 0.6987, "step": 3152 }, { "epoch": 7.709046454767726, "grad_norm": 0.22814577541179415, "learning_rate": 2.3905784986785036e-06, "loss": 0.4985, "step": 3153 }, { "epoch": 7.711491442542787, "grad_norm": 0.29769822406075214, "learning_rate": 2.3892995472896043e-06, "loss": 0.5882, "step": 3154 }, { "epoch": 7.713936430317848, "grad_norm": 0.26045635976734643, "learning_rate": 2.388020624928897e-06, "loss": 0.3918, "step": 3155 }, { "epoch": 7.71638141809291, "grad_norm": 0.2316618379215414, "learning_rate": 2.386741731931745e-06, "loss": 0.6423, "step": 3156 }, { "epoch": 7.71882640586797, "grad_norm": 0.17477930855745918, "learning_rate": 2.385462868633504e-06, "loss": 0.4273, "step": 3157 }, { "epoch": 7.721271393643032, "grad_norm": 0.30484145992022255, "learning_rate": 2.3841840353695204e-06, "loss": 0.6654, "step": 3158 }, { "epoch": 7.723716381418093, "grad_norm": 0.19539622918562286, "learning_rate": 2.3829052324751334e-06, "loss": 0.8415, "step": 3159 }, { "epoch": 7.726161369193154, "grad_norm": 0.38854475430340346, "learning_rate": 2.3816264602856755e-06, "loss": 0.9295, "step": 3160 }, { "epoch": 7.728606356968215, "grad_norm": 0.2634508293024988, "learning_rate": 2.38034771913647e-06, "loss": 0.8087, "step": 3161 }, { "epoch": 7.7310513447432765, "grad_norm": 0.28004455048678034, "learning_rate": 2.3790690093628303e-06, "loss": 0.5717, "step": 3162 }, { "epoch": 7.733496332518337, "grad_norm": 0.2457687233142582, "learning_rate": 2.3777903313000658e-06, "loss": 0.4699, "step": 3163 }, { "epoch": 7.735941320293398, "grad_norm": 0.26529724540965743, "learning_rate": 2.376511685283474e-06, "loss": 0.6874, "step": 3164 }, { "epoch": 7.73838630806846, "grad_norm": 0.22724169588157916, "learning_rate": 2.375233071648345e-06, "loss": 0.4439, "step": 3165 }, { "epoch": 7.74083129584352, "grad_norm": 0.2543940101695516, "learning_rate": 2.37395449072996e-06, "loss": 0.635, "step": 3166 }, { "epoch": 7.743276283618582, "grad_norm": 0.23534333702766563, "learning_rate": 2.3726759428635932e-06, "loss": 0.5333, "step": 3167 }, { "epoch": 7.745721271393643, "grad_norm": 0.29225670957661554, "learning_rate": 2.371397428384509e-06, "loss": 0.3734, "step": 3168 }, { "epoch": 7.748166259168704, "grad_norm": 0.2908015095675076, "learning_rate": 2.370118947627963e-06, "loss": 0.5655, "step": 3169 }, { "epoch": 7.750611246943765, "grad_norm": 0.3161393612001849, "learning_rate": 2.368840500929201e-06, "loss": 0.5279, "step": 3170 }, { "epoch": 7.7530562347188265, "grad_norm": 0.23357802992854354, "learning_rate": 2.3675620886234634e-06, "loss": 0.6345, "step": 3171 }, { "epoch": 7.755501222493887, "grad_norm": 0.3005708796550451, "learning_rate": 2.366283711045977e-06, "loss": 0.2983, "step": 3172 }, { "epoch": 7.7579462102689485, "grad_norm": 0.24185348150688785, "learning_rate": 2.365005368531963e-06, "loss": 0.6028, "step": 3173 }, { "epoch": 7.76039119804401, "grad_norm": 0.26559729372548946, "learning_rate": 2.363727061416632e-06, "loss": 0.3782, "step": 3174 }, { "epoch": 7.76283618581907, "grad_norm": 0.2391491338159959, "learning_rate": 2.3624487900351843e-06, "loss": 0.7549, "step": 3175 }, { "epoch": 7.765281173594132, "grad_norm": 0.2831720594614054, "learning_rate": 2.3611705547228127e-06, "loss": 0.6142, "step": 3176 }, { "epoch": 7.767726161369193, "grad_norm": 0.19778844692055547, "learning_rate": 2.3598923558147e-06, "loss": 0.6207, "step": 3177 }, { "epoch": 7.770171149144255, "grad_norm": 0.21700360897866425, "learning_rate": 2.3586141936460188e-06, "loss": 0.8188, "step": 3178 }, { "epoch": 7.772616136919315, "grad_norm": 0.20446390751614257, "learning_rate": 2.3573360685519324e-06, "loss": 0.419, "step": 3179 }, { "epoch": 7.775061124694377, "grad_norm": 0.2146638171160273, "learning_rate": 2.3560579808675945e-06, "loss": 0.4322, "step": 3180 }, { "epoch": 7.777506112469438, "grad_norm": 0.2395970243405885, "learning_rate": 2.3547799309281483e-06, "loss": 0.9569, "step": 3181 }, { "epoch": 7.7799511002444985, "grad_norm": 0.21724586072791693, "learning_rate": 2.353501919068729e-06, "loss": 0.3865, "step": 3182 }, { "epoch": 7.78239608801956, "grad_norm": 0.2821998542033634, "learning_rate": 2.3522239456244607e-06, "loss": 0.8293, "step": 3183 }, { "epoch": 7.784841075794621, "grad_norm": 0.3151105134185895, "learning_rate": 2.3509460109304564e-06, "loss": 0.4647, "step": 3184 }, { "epoch": 7.787286063569682, "grad_norm": 0.2243361820710655, "learning_rate": 2.34966811532182e-06, "loss": 0.705, "step": 3185 }, { "epoch": 7.789731051344743, "grad_norm": 0.3260290177692361, "learning_rate": 2.348390259133645e-06, "loss": 0.7008, "step": 3186 }, { "epoch": 7.792176039119805, "grad_norm": 0.21681754867041578, "learning_rate": 2.3471124427010146e-06, "loss": 0.4151, "step": 3187 }, { "epoch": 7.794621026894865, "grad_norm": 0.30992780644441953, "learning_rate": 2.345834666359002e-06, "loss": 0.5602, "step": 3188 }, { "epoch": 7.797066014669927, "grad_norm": 0.2811711957974211, "learning_rate": 2.344556930442669e-06, "loss": 0.4655, "step": 3189 }, { "epoch": 7.799511002444988, "grad_norm": 0.3994448716397458, "learning_rate": 2.343279235287066e-06, "loss": 1.112, "step": 3190 }, { "epoch": 7.801955990220049, "grad_norm": 0.18318610060823692, "learning_rate": 2.3420015812272355e-06, "loss": 0.6131, "step": 3191 }, { "epoch": 7.80440097799511, "grad_norm": 0.2694047533989525, "learning_rate": 2.340723968598207e-06, "loss": 0.38, "step": 3192 }, { "epoch": 7.806845965770171, "grad_norm": 0.34058683775136644, "learning_rate": 2.3394463977350006e-06, "loss": 0.4159, "step": 3193 }, { "epoch": 7.809290953545232, "grad_norm": 0.20797645109804502, "learning_rate": 2.338168868972624e-06, "loss": 0.3349, "step": 3194 }, { "epoch": 7.811735941320293, "grad_norm": 0.2280936349902658, "learning_rate": 2.3368913826460747e-06, "loss": 0.3447, "step": 3195 }, { "epoch": 7.814180929095355, "grad_norm": 0.3069921428014094, "learning_rate": 2.3356139390903377e-06, "loss": 0.794, "step": 3196 }, { "epoch": 7.816625916870415, "grad_norm": 0.31277784198727426, "learning_rate": 2.3343365386403895e-06, "loss": 0.5676, "step": 3197 }, { "epoch": 7.819070904645477, "grad_norm": 0.27399383642020914, "learning_rate": 2.3330591816311935e-06, "loss": 0.4982, "step": 3198 }, { "epoch": 7.821515892420538, "grad_norm": 0.28448812562833087, "learning_rate": 2.331781868397701e-06, "loss": 0.3521, "step": 3199 }, { "epoch": 7.823960880195599, "grad_norm": 0.3810281906669007, "learning_rate": 2.3305045992748535e-06, "loss": 0.7999, "step": 3200 }, { "epoch": 7.82640586797066, "grad_norm": 0.3215459745154409, "learning_rate": 2.32922737459758e-06, "loss": 0.5269, "step": 3201 }, { "epoch": 7.8288508557457215, "grad_norm": 0.27854362919366615, "learning_rate": 2.3279501947007973e-06, "loss": 0.6282, "step": 3202 }, { "epoch": 7.831295843520783, "grad_norm": 0.29443093097456946, "learning_rate": 2.326673059919413e-06, "loss": 0.4309, "step": 3203 }, { "epoch": 7.833740831295843, "grad_norm": 0.30518041813809504, "learning_rate": 2.3253959705883205e-06, "loss": 0.3812, "step": 3204 }, { "epoch": 7.836185819070905, "grad_norm": 0.32789582768920966, "learning_rate": 2.324118927042401e-06, "loss": 0.6941, "step": 3205 }, { "epoch": 7.838630806845966, "grad_norm": 0.3312699980873302, "learning_rate": 2.3228419296165243e-06, "loss": 0.8747, "step": 3206 }, { "epoch": 7.841075794621027, "grad_norm": 0.16034457439578667, "learning_rate": 2.32156497864555e-06, "loss": 0.487, "step": 3207 }, { "epoch": 7.843520782396088, "grad_norm": 0.17820043916342254, "learning_rate": 2.320288074464323e-06, "loss": 0.412, "step": 3208 }, { "epoch": 7.84596577017115, "grad_norm": 0.13428850030942405, "learning_rate": 2.319011217407676e-06, "loss": 0.306, "step": 3209 }, { "epoch": 7.84841075794621, "grad_norm": 0.2769796786345839, "learning_rate": 2.317734407810431e-06, "loss": 0.5125, "step": 3210 }, { "epoch": 7.8508557457212715, "grad_norm": 0.2091112752025157, "learning_rate": 2.3164576460073968e-06, "loss": 0.9005, "step": 3211 }, { "epoch": 7.853300733496333, "grad_norm": 0.19643153493471402, "learning_rate": 2.3151809323333695e-06, "loss": 0.3908, "step": 3212 }, { "epoch": 7.8557457212713935, "grad_norm": 0.15205628903794896, "learning_rate": 2.3139042671231314e-06, "loss": 0.754, "step": 3213 }, { "epoch": 7.858190709046455, "grad_norm": 0.17306439900821313, "learning_rate": 2.3126276507114554e-06, "loss": 0.4363, "step": 3214 }, { "epoch": 7.860635696821516, "grad_norm": 0.28057051863693905, "learning_rate": 2.3113510834330987e-06, "loss": 0.5, "step": 3215 }, { "epoch": 7.863080684596577, "grad_norm": 0.20826017686313453, "learning_rate": 2.3100745656228056e-06, "loss": 0.7434, "step": 3216 }, { "epoch": 7.865525672371638, "grad_norm": 0.31609110995425865, "learning_rate": 2.30879809761531e-06, "loss": 0.8658, "step": 3217 }, { "epoch": 7.8679706601467, "grad_norm": 0.3029161121405763, "learning_rate": 2.30752167974533e-06, "loss": 0.4779, "step": 3218 }, { "epoch": 7.87041564792176, "grad_norm": 0.2418983548846202, "learning_rate": 2.3062453123475716e-06, "loss": 0.9823, "step": 3219 }, { "epoch": 7.872860635696822, "grad_norm": 0.18393846840435493, "learning_rate": 2.304968995756727e-06, "loss": 0.4982, "step": 3220 }, { "epoch": 7.875305623471883, "grad_norm": 0.2974337019145741, "learning_rate": 2.303692730307477e-06, "loss": 0.5631, "step": 3221 }, { "epoch": 7.8777506112469435, "grad_norm": 0.21946059741874746, "learning_rate": 2.3024165163344865e-06, "loss": 0.3665, "step": 3222 }, { "epoch": 7.880195599022005, "grad_norm": 0.25182871704461207, "learning_rate": 2.3011403541724084e-06, "loss": 0.9837, "step": 3223 }, { "epoch": 7.882640586797066, "grad_norm": 0.255778740461956, "learning_rate": 2.299864244155881e-06, "loss": 0.6821, "step": 3224 }, { "epoch": 7.885085574572127, "grad_norm": 0.35564665562460485, "learning_rate": 2.2985881866195305e-06, "loss": 0.4816, "step": 3225 }, { "epoch": 7.887530562347188, "grad_norm": 0.21068956579030376, "learning_rate": 2.297312181897968e-06, "loss": 0.4047, "step": 3226 }, { "epoch": 7.88997555012225, "grad_norm": 0.2656364748045013, "learning_rate": 2.296036230325792e-06, "loss": 0.6599, "step": 3227 }, { "epoch": 7.89242053789731, "grad_norm": 0.22146173859832524, "learning_rate": 2.294760332237585e-06, "loss": 0.5366, "step": 3228 }, { "epoch": 7.894865525672372, "grad_norm": 0.18617491550036522, "learning_rate": 2.2934844879679167e-06, "loss": 0.5427, "step": 3229 }, { "epoch": 7.897310513447433, "grad_norm": 0.27111302158369094, "learning_rate": 2.2922086978513434e-06, "loss": 0.776, "step": 3230 }, { "epoch": 7.899755501222494, "grad_norm": 0.24665173031329884, "learning_rate": 2.290932962222406e-06, "loss": 0.4599, "step": 3231 }, { "epoch": 7.902200488997555, "grad_norm": 0.24451211112846627, "learning_rate": 2.289657281415632e-06, "loss": 0.4712, "step": 3232 }, { "epoch": 7.904645476772616, "grad_norm": 0.3948875188231338, "learning_rate": 2.288381655765534e-06, "loss": 0.5739, "step": 3233 }, { "epoch": 7.907090464547677, "grad_norm": 0.39050795022372625, "learning_rate": 2.287106085606609e-06, "loss": 0.4051, "step": 3234 }, { "epoch": 7.909535452322738, "grad_norm": 0.31680193181172295, "learning_rate": 2.285830571273343e-06, "loss": 0.563, "step": 3235 }, { "epoch": 7.9119804400978, "grad_norm": 0.33373864718743146, "learning_rate": 2.284555113100203e-06, "loss": 0.4937, "step": 3236 }, { "epoch": 7.91442542787286, "grad_norm": 0.24548489686152225, "learning_rate": 2.2832797114216445e-06, "loss": 0.7264, "step": 3237 }, { "epoch": 7.916870415647922, "grad_norm": 0.18656614994953785, "learning_rate": 2.282004366572107e-06, "loss": 0.7359, "step": 3238 }, { "epoch": 7.919315403422983, "grad_norm": 0.26130581292892213, "learning_rate": 2.280729078886015e-06, "loss": 0.5839, "step": 3239 }, { "epoch": 7.921760391198044, "grad_norm": 0.228604495419214, "learning_rate": 2.2794538486977767e-06, "loss": 0.6304, "step": 3240 }, { "epoch": 7.924205378973105, "grad_norm": 0.2797424308407201, "learning_rate": 2.2781786763417894e-06, "loss": 0.3997, "step": 3241 }, { "epoch": 7.9266503667481665, "grad_norm": 0.2383957439768026, "learning_rate": 2.27690356215243e-06, "loss": 1.1731, "step": 3242 }, { "epoch": 7.929095354523227, "grad_norm": 0.21404238275490492, "learning_rate": 2.2756285064640645e-06, "loss": 0.5782, "step": 3243 }, { "epoch": 7.931540342298288, "grad_norm": 0.15510037145429412, "learning_rate": 2.2743535096110394e-06, "loss": 0.439, "step": 3244 }, { "epoch": 7.93398533007335, "grad_norm": 0.2777662451941109, "learning_rate": 2.2730785719276893e-06, "loss": 0.6197, "step": 3245 }, { "epoch": 7.93643031784841, "grad_norm": 0.20822673822689605, "learning_rate": 2.271803693748333e-06, "loss": 0.4595, "step": 3246 }, { "epoch": 7.938875305623472, "grad_norm": 0.21044260981543203, "learning_rate": 2.2705288754072724e-06, "loss": 0.6432, "step": 3247 }, { "epoch": 7.941320293398533, "grad_norm": 0.3093710070447849, "learning_rate": 2.2692541172387933e-06, "loss": 0.4343, "step": 3248 }, { "epoch": 7.943765281173594, "grad_norm": 0.3270340408205683, "learning_rate": 2.2679794195771667e-06, "loss": 0.4525, "step": 3249 }, { "epoch": 7.946210268948655, "grad_norm": 0.26553413509787116, "learning_rate": 2.266704782756647e-06, "loss": 0.5808, "step": 3250 }, { "epoch": 7.9486552567237165, "grad_norm": 0.17919662978193723, "learning_rate": 2.2654302071114746e-06, "loss": 0.5499, "step": 3251 }, { "epoch": 7.951100244498777, "grad_norm": 0.19762948581713405, "learning_rate": 2.2641556929758714e-06, "loss": 0.5006, "step": 3252 }, { "epoch": 7.9535452322738385, "grad_norm": 0.2026054606749769, "learning_rate": 2.262881240684044e-06, "loss": 0.7326, "step": 3253 }, { "epoch": 7.9559902200489, "grad_norm": 0.20306689087156107, "learning_rate": 2.2616068505701825e-06, "loss": 0.4867, "step": 3254 }, { "epoch": 7.95843520782396, "grad_norm": 0.43497799236102697, "learning_rate": 2.260332522968463e-06, "loss": 0.5227, "step": 3255 }, { "epoch": 7.960880195599022, "grad_norm": 0.25885414864675954, "learning_rate": 2.259058258213041e-06, "loss": 0.4592, "step": 3256 }, { "epoch": 7.963325183374083, "grad_norm": 0.280833723150612, "learning_rate": 2.2577840566380606e-06, "loss": 0.3524, "step": 3257 }, { "epoch": 7.965770171149144, "grad_norm": 0.23013473488437863, "learning_rate": 2.256509918577645e-06, "loss": 0.9061, "step": 3258 }, { "epoch": 7.968215158924205, "grad_norm": 0.22034064583006813, "learning_rate": 2.255235844365902e-06, "loss": 0.5337, "step": 3259 }, { "epoch": 7.970660146699267, "grad_norm": 0.31109651275930195, "learning_rate": 2.2539618343369236e-06, "loss": 0.6184, "step": 3260 }, { "epoch": 7.973105134474328, "grad_norm": 0.2963571529769519, "learning_rate": 2.252687888824785e-06, "loss": 0.3998, "step": 3261 }, { "epoch": 7.9755501222493885, "grad_norm": 0.26493851618912834, "learning_rate": 2.251414008163543e-06, "loss": 0.4372, "step": 3262 }, { "epoch": 7.97799511002445, "grad_norm": 0.2589890421631379, "learning_rate": 2.250140192687239e-06, "loss": 0.4442, "step": 3263 }, { "epoch": 7.980440097799511, "grad_norm": 0.23826811148711516, "learning_rate": 2.2488664427298953e-06, "loss": 0.6974, "step": 3264 }, { "epoch": 7.982885085574572, "grad_norm": 0.29756901477902586, "learning_rate": 2.2475927586255195e-06, "loss": 0.4962, "step": 3265 }, { "epoch": 7.985330073349633, "grad_norm": 0.18120215433475353, "learning_rate": 2.246319140708099e-06, "loss": 0.4656, "step": 3266 }, { "epoch": 7.987775061124695, "grad_norm": 0.34522225863218786, "learning_rate": 2.2450455893116084e-06, "loss": 0.5383, "step": 3267 }, { "epoch": 7.990220048899755, "grad_norm": 0.3162570980239538, "learning_rate": 2.2437721047700004e-06, "loss": 0.7995, "step": 3268 }, { "epoch": 7.992665036674817, "grad_norm": 0.25755895302927456, "learning_rate": 2.242498687417212e-06, "loss": 0.6439, "step": 3269 }, { "epoch": 7.995110024449878, "grad_norm": 0.17985710227091536, "learning_rate": 2.241225337587162e-06, "loss": 0.497, "step": 3270 }, { "epoch": 7.997555012224939, "grad_norm": 0.2793589515870948, "learning_rate": 2.2399520556137524e-06, "loss": 0.9224, "step": 3271 }, { "epoch": 8.0, "grad_norm": 0.16483927629548378, "learning_rate": 2.238678841830867e-06, "loss": 0.4565, "step": 3272 }, { "epoch": 8.00244498777506, "grad_norm": 0.20966436356805296, "learning_rate": 2.237405696572371e-06, "loss": 0.6797, "step": 3273 }, { "epoch": 8.004889975550123, "grad_norm": 0.3033158216942618, "learning_rate": 2.236132620172112e-06, "loss": 0.4024, "step": 3274 }, { "epoch": 8.007334963325183, "grad_norm": 0.2934947694569621, "learning_rate": 2.234859612963921e-06, "loss": 0.5648, "step": 3275 }, { "epoch": 8.009779951100244, "grad_norm": 0.24567766275521513, "learning_rate": 2.2335866752816085e-06, "loss": 0.5651, "step": 3276 }, { "epoch": 8.012224938875306, "grad_norm": 0.28683418643508823, "learning_rate": 2.2323138074589673e-06, "loss": 0.3594, "step": 3277 }, { "epoch": 8.014669926650367, "grad_norm": 0.34829030597822436, "learning_rate": 2.2310410098297746e-06, "loss": 0.7458, "step": 3278 }, { "epoch": 8.017114914425427, "grad_norm": 0.22910489223440902, "learning_rate": 2.2297682827277857e-06, "loss": 0.765, "step": 3279 }, { "epoch": 8.01955990220049, "grad_norm": 0.29258807743206355, "learning_rate": 2.2284956264867394e-06, "loss": 0.5015, "step": 3280 }, { "epoch": 8.02200488997555, "grad_norm": 0.2869198065234492, "learning_rate": 2.227223041440355e-06, "loss": 0.3421, "step": 3281 }, { "epoch": 8.02444987775061, "grad_norm": 0.19175866844328074, "learning_rate": 2.2259505279223335e-06, "loss": 0.6012, "step": 3282 }, { "epoch": 8.026894865525673, "grad_norm": 0.1944543923646408, "learning_rate": 2.2246780862663567e-06, "loss": 0.3996, "step": 3283 }, { "epoch": 8.029339853300733, "grad_norm": 0.20996111324000724, "learning_rate": 2.2234057168060875e-06, "loss": 0.6123, "step": 3284 }, { "epoch": 8.031784841075794, "grad_norm": 0.2589008677014097, "learning_rate": 2.222133419875172e-06, "loss": 0.8901, "step": 3285 }, { "epoch": 8.034229828850856, "grad_norm": 0.17460551725506557, "learning_rate": 2.2208611958072344e-06, "loss": 0.5167, "step": 3286 }, { "epoch": 8.036674816625917, "grad_norm": 0.2680406664567495, "learning_rate": 2.2195890449358807e-06, "loss": 0.6727, "step": 3287 }, { "epoch": 8.039119804400977, "grad_norm": 0.1965472965844495, "learning_rate": 2.218316967594698e-06, "loss": 0.7956, "step": 3288 }, { "epoch": 8.04156479217604, "grad_norm": 0.26894507983137267, "learning_rate": 2.2170449641172545e-06, "loss": 0.4245, "step": 3289 }, { "epoch": 8.0440097799511, "grad_norm": 0.2614031105042637, "learning_rate": 2.215773034837099e-06, "loss": 0.407, "step": 3290 }, { "epoch": 8.04645476772616, "grad_norm": 0.28128198962168793, "learning_rate": 2.214501180087761e-06, "loss": 0.4564, "step": 3291 }, { "epoch": 8.048899755501223, "grad_norm": 0.37807689248766874, "learning_rate": 2.213229400202748e-06, "loss": 0.3756, "step": 3292 }, { "epoch": 8.051344743276283, "grad_norm": 0.24016845795409608, "learning_rate": 2.211957695515552e-06, "loss": 0.8035, "step": 3293 }, { "epoch": 8.053789731051344, "grad_norm": 0.23773057321647217, "learning_rate": 2.2106860663596404e-06, "loss": 0.8945, "step": 3294 }, { "epoch": 8.056234718826406, "grad_norm": 0.2945353519289615, "learning_rate": 2.2094145130684657e-06, "loss": 0.8274, "step": 3295 }, { "epoch": 8.058679706601467, "grad_norm": 0.20002552096846707, "learning_rate": 2.208143035975458e-06, "loss": 0.6665, "step": 3296 }, { "epoch": 8.061124694376527, "grad_norm": 0.2399752748552482, "learning_rate": 2.206871635414027e-06, "loss": 0.8377, "step": 3297 }, { "epoch": 8.06356968215159, "grad_norm": 0.297450056383463, "learning_rate": 2.2056003117175628e-06, "loss": 0.4946, "step": 3298 }, { "epoch": 8.06601466992665, "grad_norm": 0.24031986741383574, "learning_rate": 2.2043290652194354e-06, "loss": 0.6912, "step": 3299 }, { "epoch": 8.06845965770171, "grad_norm": 0.3109731827357909, "learning_rate": 2.2030578962529968e-06, "loss": 1.1028, "step": 3300 }, { "epoch": 8.070904645476773, "grad_norm": 0.2882215914484284, "learning_rate": 2.201786805151575e-06, "loss": 0.4467, "step": 3301 }, { "epoch": 8.073349633251834, "grad_norm": 0.18925992291693913, "learning_rate": 2.2005157922484793e-06, "loss": 0.4117, "step": 3302 }, { "epoch": 8.075794621026894, "grad_norm": 0.24852423088940928, "learning_rate": 2.199244857876998e-06, "loss": 0.4681, "step": 3303 }, { "epoch": 8.078239608801956, "grad_norm": 0.17809239993774464, "learning_rate": 2.1979740023704e-06, "loss": 0.3467, "step": 3304 }, { "epoch": 8.080684596577017, "grad_norm": 0.19610735340762786, "learning_rate": 2.1967032260619326e-06, "loss": 0.7391, "step": 3305 }, { "epoch": 8.083129584352077, "grad_norm": 0.24257174498061096, "learning_rate": 2.1954325292848224e-06, "loss": 0.4964, "step": 3306 }, { "epoch": 8.08557457212714, "grad_norm": 0.2676488731244071, "learning_rate": 2.194161912372275e-06, "loss": 0.3909, "step": 3307 }, { "epoch": 8.0880195599022, "grad_norm": 0.1676290116316388, "learning_rate": 2.1928913756574745e-06, "loss": 0.4186, "step": 3308 }, { "epoch": 8.09046454767726, "grad_norm": 0.3654213870694616, "learning_rate": 2.1916209194735847e-06, "loss": 0.2666, "step": 3309 }, { "epoch": 8.092909535452323, "grad_norm": 0.35652037365058786, "learning_rate": 2.1903505441537504e-06, "loss": 0.4306, "step": 3310 }, { "epoch": 8.095354523227384, "grad_norm": 0.21613964975179473, "learning_rate": 2.1890802500310917e-06, "loss": 0.9138, "step": 3311 }, { "epoch": 8.097799511002446, "grad_norm": 0.3772319397740884, "learning_rate": 2.1878100374387088e-06, "loss": 0.3566, "step": 3312 }, { "epoch": 8.100244498777506, "grad_norm": 0.1957001359507549, "learning_rate": 2.18653990670968e-06, "loss": 0.3775, "step": 3313 }, { "epoch": 8.102689486552567, "grad_norm": 0.1562721947646665, "learning_rate": 2.1852698581770635e-06, "loss": 0.3096, "step": 3314 }, { "epoch": 8.10513447432763, "grad_norm": 0.3130643678278964, "learning_rate": 2.1839998921738952e-06, "loss": 0.6474, "step": 3315 }, { "epoch": 8.10757946210269, "grad_norm": 0.2728111836077955, "learning_rate": 2.1827300090331882e-06, "loss": 0.4167, "step": 3316 }, { "epoch": 8.11002444987775, "grad_norm": 0.13170344815930668, "learning_rate": 2.181460209087936e-06, "loss": 0.3108, "step": 3317 }, { "epoch": 8.112469437652813, "grad_norm": 0.1990113643389976, "learning_rate": 2.1801904926711086e-06, "loss": 0.902, "step": 3318 }, { "epoch": 8.114914425427873, "grad_norm": 0.3466188007064488, "learning_rate": 2.1789208601156546e-06, "loss": 0.8578, "step": 3319 }, { "epoch": 8.117359413202934, "grad_norm": 0.26783782390502947, "learning_rate": 2.1776513117545005e-06, "loss": 0.5227, "step": 3320 }, { "epoch": 8.119804400977996, "grad_norm": 0.22102632472072922, "learning_rate": 2.1763818479205524e-06, "loss": 0.6049, "step": 3321 }, { "epoch": 8.122249388753056, "grad_norm": 0.3476238586802372, "learning_rate": 2.175112468946692e-06, "loss": 0.5577, "step": 3322 }, { "epoch": 8.124694376528117, "grad_norm": 0.32182590725407917, "learning_rate": 2.1738431751657786e-06, "loss": 0.6853, "step": 3323 }, { "epoch": 8.12713936430318, "grad_norm": 0.2402886490455442, "learning_rate": 2.1725739669106514e-06, "loss": 0.3808, "step": 3324 }, { "epoch": 8.12958435207824, "grad_norm": 0.28320159314786564, "learning_rate": 2.1713048445141258e-06, "loss": 0.676, "step": 3325 }, { "epoch": 8.1320293398533, "grad_norm": 0.3037174715692259, "learning_rate": 2.1700358083089937e-06, "loss": 0.4805, "step": 3326 }, { "epoch": 8.134474327628363, "grad_norm": 0.22168510148589918, "learning_rate": 2.1687668586280264e-06, "loss": 0.3918, "step": 3327 }, { "epoch": 8.136919315403423, "grad_norm": 0.16482402201319796, "learning_rate": 2.1674979958039702e-06, "loss": 0.4132, "step": 3328 }, { "epoch": 8.139364303178484, "grad_norm": 0.25571129611684856, "learning_rate": 2.1662292201695517e-06, "loss": 0.8492, "step": 3329 }, { "epoch": 8.141809290953546, "grad_norm": 0.2024948433133835, "learning_rate": 2.1649605320574716e-06, "loss": 0.8337, "step": 3330 }, { "epoch": 8.144254278728607, "grad_norm": 0.23716956650170803, "learning_rate": 2.1636919318004095e-06, "loss": 0.945, "step": 3331 }, { "epoch": 8.146699266503667, "grad_norm": 0.22239581704359357, "learning_rate": 2.162423419731022e-06, "loss": 0.5735, "step": 3332 }, { "epoch": 8.14914425427873, "grad_norm": 0.27260363303174445, "learning_rate": 2.1611549961819405e-06, "loss": 0.4612, "step": 3333 }, { "epoch": 8.15158924205379, "grad_norm": 0.2045198430600031, "learning_rate": 2.1598866614857764e-06, "loss": 0.9936, "step": 3334 }, { "epoch": 8.15403422982885, "grad_norm": 0.260414411549753, "learning_rate": 2.1586184159751153e-06, "loss": 0.6533, "step": 3335 }, { "epoch": 8.156479217603913, "grad_norm": 0.22760596742884684, "learning_rate": 2.15735025998252e-06, "loss": 1.1346, "step": 3336 }, { "epoch": 8.158924205378973, "grad_norm": 0.2593018291387644, "learning_rate": 2.1560821938405304e-06, "loss": 0.511, "step": 3337 }, { "epoch": 8.161369193154034, "grad_norm": 0.2185253968022397, "learning_rate": 2.1548142178816625e-06, "loss": 0.5877, "step": 3338 }, { "epoch": 8.163814180929096, "grad_norm": 0.26085755194928273, "learning_rate": 2.1535463324384083e-06, "loss": 0.4246, "step": 3339 }, { "epoch": 8.166259168704157, "grad_norm": 0.26407723762602575, "learning_rate": 2.1522785378432366e-06, "loss": 0.9678, "step": 3340 }, { "epoch": 8.168704156479217, "grad_norm": 0.23066863650752187, "learning_rate": 2.1510108344285923e-06, "loss": 0.4286, "step": 3341 }, { "epoch": 8.17114914425428, "grad_norm": 0.3533463113288656, "learning_rate": 2.149743222526895e-06, "loss": 0.449, "step": 3342 }, { "epoch": 8.17359413202934, "grad_norm": 0.32523346008625154, "learning_rate": 2.148475702470543e-06, "loss": 0.6602, "step": 3343 }, { "epoch": 8.1760391198044, "grad_norm": 0.2218149508281891, "learning_rate": 2.147208274591909e-06, "loss": 0.3142, "step": 3344 }, { "epoch": 8.178484107579463, "grad_norm": 0.27337454581191045, "learning_rate": 2.1459409392233414e-06, "loss": 0.3963, "step": 3345 }, { "epoch": 8.180929095354523, "grad_norm": 0.33572395258140497, "learning_rate": 2.1446736966971647e-06, "loss": 0.4172, "step": 3346 }, { "epoch": 8.183374083129584, "grad_norm": 0.18174562819312717, "learning_rate": 2.1434065473456778e-06, "loss": 0.5893, "step": 3347 }, { "epoch": 8.185819070904646, "grad_norm": 0.3187460703982711, "learning_rate": 2.142139491501158e-06, "loss": 0.3628, "step": 3348 }, { "epoch": 8.188264058679707, "grad_norm": 0.229579750361833, "learning_rate": 2.140872529495855e-06, "loss": 0.6271, "step": 3349 }, { "epoch": 8.190709046454767, "grad_norm": 0.21759990670649823, "learning_rate": 2.139605661661996e-06, "loss": 0.4626, "step": 3350 }, { "epoch": 8.19315403422983, "grad_norm": 0.2425253018586069, "learning_rate": 2.138338888331782e-06, "loss": 0.5089, "step": 3351 }, { "epoch": 8.19559902200489, "grad_norm": 0.2755832683214657, "learning_rate": 2.1370722098373896e-06, "loss": 0.6217, "step": 3352 }, { "epoch": 8.19804400977995, "grad_norm": 0.1948822978017345, "learning_rate": 2.1358056265109724e-06, "loss": 0.5413, "step": 3353 }, { "epoch": 8.200488997555013, "grad_norm": 0.25747125288334, "learning_rate": 2.134539138684657e-06, "loss": 0.4911, "step": 3354 }, { "epoch": 8.202933985330073, "grad_norm": 0.33537374928304037, "learning_rate": 2.1332727466905457e-06, "loss": 0.5571, "step": 3355 }, { "epoch": 8.205378973105134, "grad_norm": 0.300579616927296, "learning_rate": 2.132006450860715e-06, "loss": 0.6801, "step": 3356 }, { "epoch": 8.207823960880196, "grad_norm": 0.22554301242116664, "learning_rate": 2.1307402515272164e-06, "loss": 0.5255, "step": 3357 }, { "epoch": 8.210268948655257, "grad_norm": 0.1858291724045866, "learning_rate": 2.129474149022077e-06, "loss": 0.5286, "step": 3358 }, { "epoch": 8.212713936430317, "grad_norm": 0.18648916050311798, "learning_rate": 2.1282081436772975e-06, "loss": 0.5646, "step": 3359 }, { "epoch": 8.21515892420538, "grad_norm": 0.25587507494362327, "learning_rate": 2.1269422358248537e-06, "loss": 0.4062, "step": 3360 }, { "epoch": 8.21760391198044, "grad_norm": 0.2925418272781181, "learning_rate": 2.125676425796695e-06, "loss": 0.4998, "step": 3361 }, { "epoch": 8.2200488997555, "grad_norm": 0.2932527219095575, "learning_rate": 2.1244107139247467e-06, "loss": 0.597, "step": 3362 }, { "epoch": 8.222493887530563, "grad_norm": 0.1951304442942124, "learning_rate": 2.1231451005409057e-06, "loss": 0.4422, "step": 3363 }, { "epoch": 8.224938875305623, "grad_norm": 0.19665287770304266, "learning_rate": 2.1218795859770473e-06, "loss": 0.5439, "step": 3364 }, { "epoch": 8.227383863080684, "grad_norm": 0.44932913170944594, "learning_rate": 2.120614170565017e-06, "loss": 0.5757, "step": 3365 }, { "epoch": 8.229828850855746, "grad_norm": 0.2970313936281067, "learning_rate": 2.119348854636636e-06, "loss": 0.4703, "step": 3366 }, { "epoch": 8.232273838630807, "grad_norm": 0.4727216724489321, "learning_rate": 2.1180836385236987e-06, "loss": 0.6709, "step": 3367 }, { "epoch": 8.234718826405867, "grad_norm": 0.21503357278070417, "learning_rate": 2.116818522557974e-06, "loss": 0.6792, "step": 3368 }, { "epoch": 8.23716381418093, "grad_norm": 0.20375855782870206, "learning_rate": 2.1155535070712043e-06, "loss": 0.8595, "step": 3369 }, { "epoch": 8.23960880195599, "grad_norm": 0.22397210605989334, "learning_rate": 2.114288592395106e-06, "loss": 0.381, "step": 3370 }, { "epoch": 8.24205378973105, "grad_norm": 0.19497901458370331, "learning_rate": 2.1130237788613673e-06, "loss": 0.6178, "step": 3371 }, { "epoch": 8.244498777506113, "grad_norm": 0.30867577787826517, "learning_rate": 2.1117590668016525e-06, "loss": 0.6962, "step": 3372 }, { "epoch": 8.246943765281173, "grad_norm": 0.19803752531891974, "learning_rate": 2.110494456547598e-06, "loss": 0.4969, "step": 3373 }, { "epoch": 8.249388753056234, "grad_norm": 0.2169016215344782, "learning_rate": 2.1092299484308127e-06, "loss": 0.7604, "step": 3374 }, { "epoch": 8.251833740831296, "grad_norm": 0.33423973079752384, "learning_rate": 2.107965542782881e-06, "loss": 0.5164, "step": 3375 }, { "epoch": 8.254278728606357, "grad_norm": 0.30088352178476596, "learning_rate": 2.1067012399353578e-06, "loss": 0.7118, "step": 3376 }, { "epoch": 8.256723716381417, "grad_norm": 0.3345517177433198, "learning_rate": 2.1054370402197723e-06, "loss": 0.512, "step": 3377 }, { "epoch": 8.25916870415648, "grad_norm": 0.19397875087021696, "learning_rate": 2.1041729439676275e-06, "loss": 0.6044, "step": 3378 }, { "epoch": 8.26161369193154, "grad_norm": 0.2796766997854982, "learning_rate": 2.1029089515103977e-06, "loss": 0.9442, "step": 3379 }, { "epoch": 8.2640586797066, "grad_norm": 0.2510208419655631, "learning_rate": 2.101645063179531e-06, "loss": 0.484, "step": 3380 }, { "epoch": 8.266503667481663, "grad_norm": 0.2829611678227683, "learning_rate": 2.100381279306447e-06, "loss": 0.3747, "step": 3381 }, { "epoch": 8.268948655256724, "grad_norm": 0.1780699095012843, "learning_rate": 2.09911760022254e-06, "loss": 0.3773, "step": 3382 }, { "epoch": 8.271393643031784, "grad_norm": 0.2202648710978936, "learning_rate": 2.0978540262591745e-06, "loss": 0.8265, "step": 3383 }, { "epoch": 8.273838630806846, "grad_norm": 0.26987873896587145, "learning_rate": 2.0965905577476886e-06, "loss": 0.8246, "step": 3384 }, { "epoch": 8.276283618581907, "grad_norm": 0.2071182980440185, "learning_rate": 2.095327195019394e-06, "loss": 0.378, "step": 3385 }, { "epoch": 8.278728606356967, "grad_norm": 0.22715977351774883, "learning_rate": 2.094063938405572e-06, "loss": 0.5309, "step": 3386 }, { "epoch": 8.28117359413203, "grad_norm": 0.2091454551842603, "learning_rate": 2.0928007882374775e-06, "loss": 0.3418, "step": 3387 }, { "epoch": 8.28361858190709, "grad_norm": 0.23582710552066402, "learning_rate": 2.0915377448463383e-06, "loss": 0.6727, "step": 3388 }, { "epoch": 8.28606356968215, "grad_norm": 0.2877486045425973, "learning_rate": 2.090274808563352e-06, "loss": 0.594, "step": 3389 }, { "epoch": 8.288508557457213, "grad_norm": 0.22243680534312027, "learning_rate": 2.0890119797196906e-06, "loss": 0.5697, "step": 3390 }, { "epoch": 8.290953545232274, "grad_norm": 0.3072022173931324, "learning_rate": 2.0877492586464955e-06, "loss": 0.8175, "step": 3391 }, { "epoch": 8.293398533007334, "grad_norm": 0.31225453177874035, "learning_rate": 2.086486645674883e-06, "loss": 0.6288, "step": 3392 }, { "epoch": 8.295843520782396, "grad_norm": 0.2338806616940788, "learning_rate": 2.085224141135937e-06, "loss": 0.5839, "step": 3393 }, { "epoch": 8.298288508557457, "grad_norm": 0.3356195927529672, "learning_rate": 2.0839617453607165e-06, "loss": 0.4523, "step": 3394 }, { "epoch": 8.30073349633252, "grad_norm": 0.3090457311145743, "learning_rate": 2.0826994586802483e-06, "loss": 0.7139, "step": 3395 }, { "epoch": 8.30317848410758, "grad_norm": 0.21844868274931487, "learning_rate": 2.0814372814255363e-06, "loss": 0.3581, "step": 3396 }, { "epoch": 8.30562347188264, "grad_norm": 0.32577250562808757, "learning_rate": 2.08017521392755e-06, "loss": 0.3167, "step": 3397 }, { "epoch": 8.3080684596577, "grad_norm": 0.23337886064912508, "learning_rate": 2.0789132565172334e-06, "loss": 0.6412, "step": 3398 }, { "epoch": 8.310513447432763, "grad_norm": 0.17878435674897483, "learning_rate": 2.0776514095255004e-06, "loss": 0.6377, "step": 3399 }, { "epoch": 8.312958435207824, "grad_norm": 0.31572463304418785, "learning_rate": 2.076389673283236e-06, "loss": 0.9649, "step": 3400 }, { "epoch": 8.315403422982886, "grad_norm": 0.2675760035307563, "learning_rate": 2.075128048121295e-06, "loss": 0.8356, "step": 3401 }, { "epoch": 8.317848410757946, "grad_norm": 0.2952103395822672, "learning_rate": 2.0738665343705067e-06, "loss": 0.6255, "step": 3402 }, { "epoch": 8.320293398533007, "grad_norm": 0.16655698430568028, "learning_rate": 2.072605132361668e-06, "loss": 0.4255, "step": 3403 }, { "epoch": 8.32273838630807, "grad_norm": 0.2050784296849249, "learning_rate": 2.0713438424255472e-06, "loss": 0.4992, "step": 3404 }, { "epoch": 8.32518337408313, "grad_norm": 0.2273616927984424, "learning_rate": 2.070082664892883e-06, "loss": 0.4345, "step": 3405 }, { "epoch": 8.32762836185819, "grad_norm": 0.28025948483705754, "learning_rate": 2.068821600094385e-06, "loss": 0.4779, "step": 3406 }, { "epoch": 8.330073349633253, "grad_norm": 0.27508873012165336, "learning_rate": 2.0675606483607343e-06, "loss": 0.8361, "step": 3407 }, { "epoch": 8.332518337408313, "grad_norm": 0.1781506826075628, "learning_rate": 2.0662998100225806e-06, "loss": 0.3488, "step": 3408 }, { "epoch": 8.334963325183374, "grad_norm": 0.16844370223706442, "learning_rate": 2.0650390854105454e-06, "loss": 0.3299, "step": 3409 }, { "epoch": 8.337408312958436, "grad_norm": 0.32402860328147687, "learning_rate": 2.0637784748552183e-06, "loss": 0.445, "step": 3410 }, { "epoch": 8.339853300733497, "grad_norm": 0.2819396934826585, "learning_rate": 2.062517978687161e-06, "loss": 0.4548, "step": 3411 }, { "epoch": 8.342298288508557, "grad_norm": 0.3211030707301094, "learning_rate": 2.061257597236904e-06, "loss": 0.4618, "step": 3412 }, { "epoch": 8.34474327628362, "grad_norm": 0.19514637441586846, "learning_rate": 2.0599973308349488e-06, "loss": 0.5277, "step": 3413 }, { "epoch": 8.34718826405868, "grad_norm": 0.2529062358165802, "learning_rate": 2.0587371798117656e-06, "loss": 0.4326, "step": 3414 }, { "epoch": 8.34963325183374, "grad_norm": 0.3118818630202157, "learning_rate": 2.0574771444977944e-06, "loss": 0.6284, "step": 3415 }, { "epoch": 8.352078239608803, "grad_norm": 0.29677829176884524, "learning_rate": 2.0562172252234465e-06, "loss": 0.6335, "step": 3416 }, { "epoch": 8.354523227383863, "grad_norm": 0.19344817468517547, "learning_rate": 2.0549574223191e-06, "loss": 0.6191, "step": 3417 }, { "epoch": 8.356968215158924, "grad_norm": 0.25026792360165995, "learning_rate": 2.0536977361151056e-06, "loss": 0.6727, "step": 3418 }, { "epoch": 8.359413202933986, "grad_norm": 0.17765642151833175, "learning_rate": 2.0524381669417814e-06, "loss": 0.8783, "step": 3419 }, { "epoch": 8.361858190709047, "grad_norm": 0.20665953394130498, "learning_rate": 2.0511787151294155e-06, "loss": 1.0192, "step": 3420 }, { "epoch": 8.364303178484107, "grad_norm": 0.23322126659676687, "learning_rate": 2.049919381008264e-06, "loss": 0.4594, "step": 3421 }, { "epoch": 8.36674816625917, "grad_norm": 0.22952978689422676, "learning_rate": 2.048660164908554e-06, "loss": 0.8095, "step": 3422 }, { "epoch": 8.36919315403423, "grad_norm": 0.1647336948038653, "learning_rate": 2.047401067160481e-06, "loss": 0.4817, "step": 3423 }, { "epoch": 8.37163814180929, "grad_norm": 0.22749716011141102, "learning_rate": 2.0461420880942083e-06, "loss": 0.6213, "step": 3424 }, { "epoch": 8.374083129584353, "grad_norm": 0.3723366899042355, "learning_rate": 2.0448832280398692e-06, "loss": 0.4867, "step": 3425 }, { "epoch": 8.376528117359413, "grad_norm": 0.38826891638444855, "learning_rate": 2.0436244873275665e-06, "loss": 0.3349, "step": 3426 }, { "epoch": 8.378973105134474, "grad_norm": 0.23517561514999716, "learning_rate": 2.0423658662873692e-06, "loss": 0.4286, "step": 3427 }, { "epoch": 8.381418092909536, "grad_norm": 0.2427328008607717, "learning_rate": 2.041107365249319e-06, "loss": 0.3834, "step": 3428 }, { "epoch": 8.383863080684597, "grad_norm": 0.2808909523934966, "learning_rate": 2.0398489845434218e-06, "loss": 0.545, "step": 3429 }, { "epoch": 8.386308068459657, "grad_norm": 0.22841285788294444, "learning_rate": 2.0385907244996546e-06, "loss": 0.7469, "step": 3430 }, { "epoch": 8.38875305623472, "grad_norm": 0.22083418274828928, "learning_rate": 2.0373325854479613e-06, "loss": 0.4581, "step": 3431 }, { "epoch": 8.39119804400978, "grad_norm": 0.2040634099796583, "learning_rate": 2.036074567718255e-06, "loss": 0.3826, "step": 3432 }, { "epoch": 8.39364303178484, "grad_norm": 0.29919685022484543, "learning_rate": 2.034816671640418e-06, "loss": 0.5029, "step": 3433 }, { "epoch": 8.396088019559903, "grad_norm": 0.3630374426822063, "learning_rate": 2.0335588975442976e-06, "loss": 0.7438, "step": 3434 }, { "epoch": 8.398533007334963, "grad_norm": 0.26018754246078385, "learning_rate": 2.0323012457597117e-06, "loss": 0.4561, "step": 3435 }, { "epoch": 8.400977995110024, "grad_norm": 0.3552612284614733, "learning_rate": 2.0310437166164455e-06, "loss": 0.4431, "step": 3436 }, { "epoch": 8.403422982885086, "grad_norm": 0.21130114213829806, "learning_rate": 2.0297863104442524e-06, "loss": 0.7768, "step": 3437 }, { "epoch": 8.405867970660147, "grad_norm": 0.27095450763579304, "learning_rate": 2.028529027572851e-06, "loss": 0.7652, "step": 3438 }, { "epoch": 8.408312958435207, "grad_norm": 0.23179304885781649, "learning_rate": 2.0272718683319324e-06, "loss": 0.591, "step": 3439 }, { "epoch": 8.41075794621027, "grad_norm": 0.22587948916367723, "learning_rate": 2.0260148330511516e-06, "loss": 0.6797, "step": 3440 }, { "epoch": 8.41320293398533, "grad_norm": 0.1663087905902326, "learning_rate": 2.024757922060131e-06, "loss": 0.6021, "step": 3441 }, { "epoch": 8.41564792176039, "grad_norm": 0.23632069414957482, "learning_rate": 2.0235011356884626e-06, "loss": 0.4362, "step": 3442 }, { "epoch": 8.418092909535453, "grad_norm": 0.2929898914490159, "learning_rate": 2.022244474265704e-06, "loss": 0.4028, "step": 3443 }, { "epoch": 8.420537897310513, "grad_norm": 0.23262149448981922, "learning_rate": 2.020987938121381e-06, "loss": 0.7849, "step": 3444 }, { "epoch": 8.422982885085574, "grad_norm": 0.2327857269211227, "learning_rate": 2.019731527584985e-06, "loss": 0.4748, "step": 3445 }, { "epoch": 8.425427872860636, "grad_norm": 0.2738723621020514, "learning_rate": 2.0184752429859773e-06, "loss": 0.3634, "step": 3446 }, { "epoch": 8.427872860635697, "grad_norm": 0.21150595229738153, "learning_rate": 2.0172190846537838e-06, "loss": 0.7538, "step": 3447 }, { "epoch": 8.430317848410757, "grad_norm": 0.29809412622435655, "learning_rate": 2.0159630529177975e-06, "loss": 0.5818, "step": 3448 }, { "epoch": 8.43276283618582, "grad_norm": 0.2140657593249767, "learning_rate": 2.0147071481073784e-06, "loss": 0.9393, "step": 3449 }, { "epoch": 8.43520782396088, "grad_norm": 0.25685127894637233, "learning_rate": 2.0134513705518545e-06, "loss": 0.4599, "step": 3450 }, { "epoch": 8.43765281173594, "grad_norm": 0.3010748797119755, "learning_rate": 2.012195720580519e-06, "loss": 0.7133, "step": 3451 }, { "epoch": 8.440097799511003, "grad_norm": 0.28334849859695055, "learning_rate": 2.0109401985226325e-06, "loss": 0.5436, "step": 3452 }, { "epoch": 8.442542787286063, "grad_norm": 0.2844726500699958, "learning_rate": 2.0096848047074207e-06, "loss": 0.7464, "step": 3453 }, { "epoch": 8.444987775061124, "grad_norm": 0.2720857460371934, "learning_rate": 2.0084295394640777e-06, "loss": 0.8159, "step": 3454 }, { "epoch": 8.447432762836186, "grad_norm": 0.15513839200603916, "learning_rate": 2.0071744031217614e-06, "loss": 0.555, "step": 3455 }, { "epoch": 8.449877750611247, "grad_norm": 0.19640942693321717, "learning_rate": 2.0059193960095983e-06, "loss": 0.5562, "step": 3456 }, { "epoch": 8.452322738386307, "grad_norm": 0.26330371947760495, "learning_rate": 2.0046645184566802e-06, "loss": 0.3978, "step": 3457 }, { "epoch": 8.45476772616137, "grad_norm": 0.24033992797584902, "learning_rate": 2.003409770792064e-06, "loss": 0.6038, "step": 3458 }, { "epoch": 8.45721271393643, "grad_norm": 0.18700285439486017, "learning_rate": 2.0021551533447724e-06, "loss": 0.5415, "step": 3459 }, { "epoch": 8.45965770171149, "grad_norm": 0.22936795936443075, "learning_rate": 2.0009006664437973e-06, "loss": 1.1404, "step": 3460 }, { "epoch": 8.462102689486553, "grad_norm": 0.29091826064494175, "learning_rate": 1.999646310418092e-06, "loss": 0.8825, "step": 3461 }, { "epoch": 8.464547677261614, "grad_norm": 0.1876587646209225, "learning_rate": 1.998392085596578e-06, "loss": 0.628, "step": 3462 }, { "epoch": 8.466992665036674, "grad_norm": 0.21211714009735663, "learning_rate": 1.997137992308142e-06, "loss": 0.5015, "step": 3463 }, { "epoch": 8.469437652811736, "grad_norm": 0.2487273800376849, "learning_rate": 1.9958840308816356e-06, "loss": 0.378, "step": 3464 }, { "epoch": 8.471882640586797, "grad_norm": 0.2617102295436207, "learning_rate": 1.994630201645876e-06, "loss": 0.396, "step": 3465 }, { "epoch": 8.474327628361857, "grad_norm": 0.46322961722230716, "learning_rate": 1.993376504929646e-06, "loss": 0.7566, "step": 3466 }, { "epoch": 8.47677261613692, "grad_norm": 0.3227873329415639, "learning_rate": 1.992122941061694e-06, "loss": 0.5154, "step": 3467 }, { "epoch": 8.47921760391198, "grad_norm": 0.25595505532255103, "learning_rate": 1.9908695103707328e-06, "loss": 0.4571, "step": 3468 }, { "epoch": 8.48166259168704, "grad_norm": 0.25688017162328164, "learning_rate": 1.9896162131854403e-06, "loss": 0.8192, "step": 3469 }, { "epoch": 8.484107579462103, "grad_norm": 0.2534667793603329, "learning_rate": 1.98836304983446e-06, "loss": 0.4847, "step": 3470 }, { "epoch": 8.486552567237164, "grad_norm": 0.19904472783850125, "learning_rate": 1.9871100206464e-06, "loss": 0.6501, "step": 3471 }, { "epoch": 8.488997555012224, "grad_norm": 0.2426345197295938, "learning_rate": 1.9858571259498337e-06, "loss": 0.8233, "step": 3472 }, { "epoch": 8.491442542787286, "grad_norm": 0.3098659857495727, "learning_rate": 1.984604366073299e-06, "loss": 0.5228, "step": 3473 }, { "epoch": 8.493887530562347, "grad_norm": 0.23323184752158677, "learning_rate": 1.9833517413452973e-06, "loss": 1.165, "step": 3474 }, { "epoch": 8.496332518337407, "grad_norm": 0.22369321632061312, "learning_rate": 1.982099252094295e-06, "loss": 0.3996, "step": 3475 }, { "epoch": 8.49877750611247, "grad_norm": 0.24718967362547936, "learning_rate": 1.980846898648725e-06, "loss": 0.5318, "step": 3476 }, { "epoch": 8.50122249388753, "grad_norm": 0.17581178230781888, "learning_rate": 1.9795946813369825e-06, "loss": 0.6048, "step": 3477 }, { "epoch": 8.503667481662593, "grad_norm": 0.2326584189182196, "learning_rate": 1.9783426004874274e-06, "loss": 0.5056, "step": 3478 }, { "epoch": 8.506112469437653, "grad_norm": 0.3021945024852293, "learning_rate": 1.9770906564283828e-06, "loss": 0.8403, "step": 3479 }, { "epoch": 8.508557457212714, "grad_norm": 0.3463324406417476, "learning_rate": 1.975838849488139e-06, "loss": 0.4415, "step": 3480 }, { "epoch": 8.511002444987774, "grad_norm": 0.3697625240035511, "learning_rate": 1.974587179994947e-06, "loss": 0.8968, "step": 3481 }, { "epoch": 8.513447432762836, "grad_norm": 0.17100665827733202, "learning_rate": 1.973335648277024e-06, "loss": 0.6178, "step": 3482 }, { "epoch": 8.515892420537897, "grad_norm": 0.31159033677965736, "learning_rate": 1.972084254662551e-06, "loss": 0.9117, "step": 3483 }, { "epoch": 8.51833740831296, "grad_norm": 0.22136098447305175, "learning_rate": 1.9708329994796707e-06, "loss": 0.8505, "step": 3484 }, { "epoch": 8.52078239608802, "grad_norm": 0.26238335861967294, "learning_rate": 1.969581883056491e-06, "loss": 0.4293, "step": 3485 }, { "epoch": 8.52322738386308, "grad_norm": 0.20846764913793728, "learning_rate": 1.968330905721084e-06, "loss": 0.3413, "step": 3486 }, { "epoch": 8.525672371638143, "grad_norm": 0.2462977978996641, "learning_rate": 1.967080067801484e-06, "loss": 0.7905, "step": 3487 }, { "epoch": 8.528117359413203, "grad_norm": 0.24242753971382194, "learning_rate": 1.9658293696256893e-06, "loss": 0.4386, "step": 3488 }, { "epoch": 8.530562347188264, "grad_norm": 0.2526537077565564, "learning_rate": 1.9645788115216614e-06, "loss": 0.3959, "step": 3489 }, { "epoch": 8.533007334963326, "grad_norm": 0.2529371426102152, "learning_rate": 1.9633283938173265e-06, "loss": 0.4898, "step": 3490 }, { "epoch": 8.535452322738386, "grad_norm": 0.18851739780634966, "learning_rate": 1.962078116840572e-06, "loss": 0.9808, "step": 3491 }, { "epoch": 8.537897310513447, "grad_norm": 0.28748391227190234, "learning_rate": 1.9608279809192474e-06, "loss": 0.542, "step": 3492 }, { "epoch": 8.54034229828851, "grad_norm": 0.1649721954538384, "learning_rate": 1.9595779863811697e-06, "loss": 0.4289, "step": 3493 }, { "epoch": 8.54278728606357, "grad_norm": 0.2221071967383684, "learning_rate": 1.9583281335541152e-06, "loss": 0.645, "step": 3494 }, { "epoch": 8.54523227383863, "grad_norm": 0.25201814625354385, "learning_rate": 1.9570784227658234e-06, "loss": 0.4984, "step": 3495 }, { "epoch": 8.547677261613693, "grad_norm": 0.21172051968087555, "learning_rate": 1.9558288543439973e-06, "loss": 0.5099, "step": 3496 }, { "epoch": 8.550122249388753, "grad_norm": 0.20991644584812197, "learning_rate": 1.9545794286163027e-06, "loss": 0.5876, "step": 3497 }, { "epoch": 8.552567237163814, "grad_norm": 0.20930536048535836, "learning_rate": 1.9533301459103677e-06, "loss": 0.4725, "step": 3498 }, { "epoch": 8.555012224938876, "grad_norm": 0.21017253622942617, "learning_rate": 1.9520810065537815e-06, "loss": 0.9064, "step": 3499 }, { "epoch": 8.557457212713937, "grad_norm": 0.29022485731845893, "learning_rate": 1.9508320108740985e-06, "loss": 0.4033, "step": 3500 }, { "epoch": 8.559902200488997, "grad_norm": 0.27130152721460377, "learning_rate": 1.9495831591988333e-06, "loss": 0.4446, "step": 3501 }, { "epoch": 8.56234718826406, "grad_norm": 0.28189091625406437, "learning_rate": 1.9483344518554632e-06, "loss": 0.3453, "step": 3502 }, { "epoch": 8.56479217603912, "grad_norm": 0.20880743274939603, "learning_rate": 1.947085889171429e-06, "loss": 0.6393, "step": 3503 }, { "epoch": 8.56723716381418, "grad_norm": 0.2846776112193436, "learning_rate": 1.945837471474131e-06, "loss": 0.5495, "step": 3504 }, { "epoch": 8.569682151589243, "grad_norm": 0.2429138249945023, "learning_rate": 1.9445891990909335e-06, "loss": 0.4722, "step": 3505 }, { "epoch": 8.572127139364303, "grad_norm": 0.2495870228371255, "learning_rate": 1.9433410723491627e-06, "loss": 0.4256, "step": 3506 }, { "epoch": 8.574572127139364, "grad_norm": 0.3697444792189972, "learning_rate": 1.9420930915761053e-06, "loss": 0.5657, "step": 3507 }, { "epoch": 8.577017114914426, "grad_norm": 0.20177825430554475, "learning_rate": 1.9408452570990104e-06, "loss": 0.5508, "step": 3508 }, { "epoch": 8.579462102689487, "grad_norm": 0.25046160816116525, "learning_rate": 1.9395975692450886e-06, "loss": 0.8596, "step": 3509 }, { "epoch": 8.581907090464547, "grad_norm": 0.17699240289447116, "learning_rate": 1.938350028341513e-06, "loss": 0.4062, "step": 3510 }, { "epoch": 8.58435207823961, "grad_norm": 0.18910200825032344, "learning_rate": 1.937102634715417e-06, "loss": 1.0002, "step": 3511 }, { "epoch": 8.58679706601467, "grad_norm": 0.23239455434756917, "learning_rate": 1.9358553886938952e-06, "loss": 0.3862, "step": 3512 }, { "epoch": 8.58924205378973, "grad_norm": 0.24163750549647728, "learning_rate": 1.934608290604004e-06, "loss": 0.4689, "step": 3513 }, { "epoch": 8.591687041564793, "grad_norm": 0.16407933128204633, "learning_rate": 1.9333613407727625e-06, "loss": 0.6044, "step": 3514 }, { "epoch": 8.594132029339853, "grad_norm": 0.20372024372004213, "learning_rate": 1.932114539527148e-06, "loss": 0.3832, "step": 3515 }, { "epoch": 8.596577017114914, "grad_norm": 0.404097054579615, "learning_rate": 1.9308678871941015e-06, "loss": 0.4433, "step": 3516 }, { "epoch": 8.599022004889976, "grad_norm": 0.2042974117653784, "learning_rate": 1.929621384100523e-06, "loss": 0.5233, "step": 3517 }, { "epoch": 8.601466992665037, "grad_norm": 0.26809622469822847, "learning_rate": 1.9283750305732747e-06, "loss": 0.3648, "step": 3518 }, { "epoch": 8.603911980440097, "grad_norm": 0.35456141185891565, "learning_rate": 1.9271288269391773e-06, "loss": 0.4103, "step": 3519 }, { "epoch": 8.60635696821516, "grad_norm": 0.27894431595251695, "learning_rate": 1.925882773525017e-06, "loss": 0.3682, "step": 3520 }, { "epoch": 8.60880195599022, "grad_norm": 0.3000402634321122, "learning_rate": 1.924636870657535e-06, "loss": 0.4124, "step": 3521 }, { "epoch": 8.61124694376528, "grad_norm": 0.28394657959099134, "learning_rate": 1.9233911186634367e-06, "loss": 0.7768, "step": 3522 }, { "epoch": 8.613691931540343, "grad_norm": 0.28027335269104686, "learning_rate": 1.9221455178693858e-06, "loss": 0.8926, "step": 3523 }, { "epoch": 8.616136919315403, "grad_norm": 0.29206777414134233, "learning_rate": 1.920900068602008e-06, "loss": 0.5853, "step": 3524 }, { "epoch": 8.618581907090464, "grad_norm": 0.2876923464255937, "learning_rate": 1.9196547711878884e-06, "loss": 0.6171, "step": 3525 }, { "epoch": 8.621026894865526, "grad_norm": 0.3861950516456796, "learning_rate": 1.9184096259535734e-06, "loss": 0.7075, "step": 3526 }, { "epoch": 8.623471882640587, "grad_norm": 0.26286690408869506, "learning_rate": 1.9171646332255682e-06, "loss": 0.8028, "step": 3527 }, { "epoch": 8.625916870415647, "grad_norm": 0.27436038805453095, "learning_rate": 1.9159197933303376e-06, "loss": 0.7116, "step": 3528 }, { "epoch": 8.62836185819071, "grad_norm": 0.22404469343288536, "learning_rate": 1.914675106594307e-06, "loss": 0.9514, "step": 3529 }, { "epoch": 8.63080684596577, "grad_norm": 0.25793644445227665, "learning_rate": 1.913430573343863e-06, "loss": 0.5078, "step": 3530 }, { "epoch": 8.63325183374083, "grad_norm": 0.27709460765148414, "learning_rate": 1.9121861939053506e-06, "loss": 0.4828, "step": 3531 }, { "epoch": 8.635696821515893, "grad_norm": 0.18781442664342946, "learning_rate": 1.9109419686050733e-06, "loss": 0.6476, "step": 3532 }, { "epoch": 8.638141809290953, "grad_norm": 0.17832987303233894, "learning_rate": 1.909697897769296e-06, "loss": 0.6052, "step": 3533 }, { "epoch": 8.640586797066014, "grad_norm": 0.34380449294481163, "learning_rate": 1.908453981724243e-06, "loss": 0.5571, "step": 3534 }, { "epoch": 8.643031784841076, "grad_norm": 0.3291903673278451, "learning_rate": 1.9072102207960964e-06, "loss": 0.4599, "step": 3535 }, { "epoch": 8.645476772616137, "grad_norm": 0.2474979389749802, "learning_rate": 1.9059666153110006e-06, "loss": 0.6129, "step": 3536 }, { "epoch": 8.647921760391197, "grad_norm": 0.2733955512016475, "learning_rate": 1.9047231655950568e-06, "loss": 0.6318, "step": 3537 }, { "epoch": 8.65036674816626, "grad_norm": 0.2476370397992299, "learning_rate": 1.9034798719743253e-06, "loss": 0.413, "step": 3538 }, { "epoch": 8.65281173594132, "grad_norm": 0.21549234716228272, "learning_rate": 1.9022367347748264e-06, "loss": 0.6836, "step": 3539 }, { "epoch": 8.65525672371638, "grad_norm": 0.23096758833599285, "learning_rate": 1.9009937543225397e-06, "loss": 0.6656, "step": 3540 }, { "epoch": 8.657701711491443, "grad_norm": 0.20289537968424193, "learning_rate": 1.8997509309434023e-06, "loss": 0.4325, "step": 3541 }, { "epoch": 8.660146699266503, "grad_norm": 0.16243383038822015, "learning_rate": 1.8985082649633115e-06, "loss": 0.5114, "step": 3542 }, { "epoch": 8.662591687041564, "grad_norm": 0.3310952435561895, "learning_rate": 1.8972657567081224e-06, "loss": 0.5086, "step": 3543 }, { "epoch": 8.665036674816626, "grad_norm": 0.25989128465423195, "learning_rate": 1.8960234065036495e-06, "loss": 0.7747, "step": 3544 }, { "epoch": 8.667481662591687, "grad_norm": 0.3117353261024849, "learning_rate": 1.8947812146756645e-06, "loss": 0.4592, "step": 3545 }, { "epoch": 8.669926650366747, "grad_norm": 0.291014918859351, "learning_rate": 1.8935391815498998e-06, "loss": 0.7344, "step": 3546 }, { "epoch": 8.67237163814181, "grad_norm": 0.29487924447870956, "learning_rate": 1.8922973074520451e-06, "loss": 0.4169, "step": 3547 }, { "epoch": 8.67481662591687, "grad_norm": 0.2106458073265303, "learning_rate": 1.8910555927077471e-06, "loss": 0.4149, "step": 3548 }, { "epoch": 8.67726161369193, "grad_norm": 0.23161344642698872, "learning_rate": 1.8898140376426118e-06, "loss": 0.4984, "step": 3549 }, { "epoch": 8.679706601466993, "grad_norm": 0.2920424107188469, "learning_rate": 1.8885726425822044e-06, "loss": 0.5177, "step": 3550 }, { "epoch": 8.682151589242054, "grad_norm": 0.20094702529289873, "learning_rate": 1.8873314078520466e-06, "loss": 0.4431, "step": 3551 }, { "epoch": 8.684596577017114, "grad_norm": 0.26483461221684934, "learning_rate": 1.8860903337776182e-06, "loss": 0.831, "step": 3552 }, { "epoch": 8.687041564792176, "grad_norm": 0.30028524212987706, "learning_rate": 1.8848494206843574e-06, "loss": 0.5183, "step": 3553 }, { "epoch": 8.689486552567237, "grad_norm": 0.3340435237146265, "learning_rate": 1.8836086688976605e-06, "loss": 0.4664, "step": 3554 }, { "epoch": 8.6919315403423, "grad_norm": 0.24168574000166768, "learning_rate": 1.8823680787428805e-06, "loss": 0.6054, "step": 3555 }, { "epoch": 8.69437652811736, "grad_norm": 0.18809454416548121, "learning_rate": 1.8811276505453277e-06, "loss": 0.377, "step": 3556 }, { "epoch": 8.69682151589242, "grad_norm": 0.2791547233976903, "learning_rate": 1.879887384630273e-06, "loss": 0.5847, "step": 3557 }, { "epoch": 8.69926650366748, "grad_norm": 0.15870852643894778, "learning_rate": 1.8786472813229408e-06, "loss": 0.3086, "step": 3558 }, { "epoch": 8.701711491442543, "grad_norm": 0.2661033255739425, "learning_rate": 1.8774073409485149e-06, "loss": 0.461, "step": 3559 }, { "epoch": 8.704156479217604, "grad_norm": 0.1954390829026525, "learning_rate": 1.876167563832137e-06, "loss": 0.5994, "step": 3560 }, { "epoch": 8.706601466992666, "grad_norm": 0.2209018184894722, "learning_rate": 1.8749279502989037e-06, "loss": 0.4499, "step": 3561 }, { "epoch": 8.709046454767726, "grad_norm": 0.2579623621111341, "learning_rate": 1.8736885006738708e-06, "loss": 0.7331, "step": 3562 }, { "epoch": 8.711491442542787, "grad_norm": 0.19758607921099702, "learning_rate": 1.8724492152820495e-06, "loss": 0.4074, "step": 3563 }, { "epoch": 8.713936430317847, "grad_norm": 0.20391215032763163, "learning_rate": 1.87121009444841e-06, "loss": 1.1082, "step": 3564 }, { "epoch": 8.71638141809291, "grad_norm": 0.1999394149660434, "learning_rate": 1.8699711384978781e-06, "loss": 0.8326, "step": 3565 }, { "epoch": 8.71882640586797, "grad_norm": 0.32513091702383184, "learning_rate": 1.8687323477553354e-06, "loss": 0.6517, "step": 3566 }, { "epoch": 8.721271393643033, "grad_norm": 0.1843675802642467, "learning_rate": 1.8674937225456207e-06, "loss": 0.3082, "step": 3567 }, { "epoch": 8.723716381418093, "grad_norm": 0.19508695125818107, "learning_rate": 1.866255263193532e-06, "loss": 0.3884, "step": 3568 }, { "epoch": 8.726161369193154, "grad_norm": 0.2795456489020726, "learning_rate": 1.86501697002382e-06, "loss": 0.7367, "step": 3569 }, { "epoch": 8.728606356968216, "grad_norm": 0.2638947225523053, "learning_rate": 1.8637788433611948e-06, "loss": 0.5503, "step": 3570 }, { "epoch": 8.731051344743276, "grad_norm": 0.2984692933370658, "learning_rate": 1.8625408835303205e-06, "loss": 0.4435, "step": 3571 }, { "epoch": 8.733496332518337, "grad_norm": 0.1680395301381349, "learning_rate": 1.861303090855819e-06, "loss": 0.5352, "step": 3572 }, { "epoch": 8.7359413202934, "grad_norm": 0.18032241384480857, "learning_rate": 1.8600654656622675e-06, "loss": 0.4048, "step": 3573 }, { "epoch": 8.73838630806846, "grad_norm": 0.18451461045994727, "learning_rate": 1.8588280082741999e-06, "loss": 0.3899, "step": 3574 }, { "epoch": 8.74083129584352, "grad_norm": 0.19278557122166826, "learning_rate": 1.857590719016106e-06, "loss": 0.4494, "step": 3575 }, { "epoch": 8.743276283618583, "grad_norm": 0.2670709610212619, "learning_rate": 1.8563535982124314e-06, "loss": 0.8321, "step": 3576 }, { "epoch": 8.745721271393643, "grad_norm": 0.2369556950403544, "learning_rate": 1.8551166461875761e-06, "loss": 0.7048, "step": 3577 }, { "epoch": 8.748166259168704, "grad_norm": 0.19792029199228522, "learning_rate": 1.8538798632658994e-06, "loss": 0.3984, "step": 3578 }, { "epoch": 8.750611246943766, "grad_norm": 0.20337492509153593, "learning_rate": 1.8526432497717128e-06, "loss": 0.5945, "step": 3579 }, { "epoch": 8.753056234718827, "grad_norm": 0.27425567054014555, "learning_rate": 1.8514068060292856e-06, "loss": 0.4128, "step": 3580 }, { "epoch": 8.755501222493887, "grad_norm": 0.2374559670832558, "learning_rate": 1.8501705323628405e-06, "loss": 0.6566, "step": 3581 }, { "epoch": 8.75794621026895, "grad_norm": 0.2271241467376953, "learning_rate": 1.8489344290965578e-06, "loss": 0.7612, "step": 3582 }, { "epoch": 8.76039119804401, "grad_norm": 0.17744630326547883, "learning_rate": 1.8476984965545708e-06, "loss": 0.4367, "step": 3583 }, { "epoch": 8.76283618581907, "grad_norm": 0.29642515539022385, "learning_rate": 1.846462735060971e-06, "loss": 0.8888, "step": 3584 }, { "epoch": 8.765281173594133, "grad_norm": 0.2956561935943152, "learning_rate": 1.8452271449398018e-06, "loss": 0.6676, "step": 3585 }, { "epoch": 8.767726161369193, "grad_norm": 0.3481583250220868, "learning_rate": 1.8439917265150641e-06, "loss": 0.4657, "step": 3586 }, { "epoch": 8.770171149144254, "grad_norm": 0.17487483307159501, "learning_rate": 1.8427564801107123e-06, "loss": 0.3737, "step": 3587 }, { "epoch": 8.772616136919316, "grad_norm": 0.3204048859324367, "learning_rate": 1.8415214060506568e-06, "loss": 0.3674, "step": 3588 }, { "epoch": 8.775061124694377, "grad_norm": 0.2737594055342559, "learning_rate": 1.8402865046587618e-06, "loss": 0.7675, "step": 3589 }, { "epoch": 8.777506112469437, "grad_norm": 0.19345359628899317, "learning_rate": 1.8390517762588472e-06, "loss": 0.514, "step": 3590 }, { "epoch": 8.7799511002445, "grad_norm": 0.21344587476710078, "learning_rate": 1.8378172211746875e-06, "loss": 0.7672, "step": 3591 }, { "epoch": 8.78239608801956, "grad_norm": 0.25356720115852227, "learning_rate": 1.8365828397300103e-06, "loss": 0.6055, "step": 3592 }, { "epoch": 8.78484107579462, "grad_norm": 0.25905199134501394, "learning_rate": 1.8353486322484988e-06, "loss": 0.4837, "step": 3593 }, { "epoch": 8.787286063569683, "grad_norm": 0.17390934203312977, "learning_rate": 1.834114599053791e-06, "loss": 0.7742, "step": 3594 }, { "epoch": 8.789731051344743, "grad_norm": 0.24831720076112573, "learning_rate": 1.8328807404694785e-06, "loss": 0.4213, "step": 3595 }, { "epoch": 8.792176039119804, "grad_norm": 0.27056706558634086, "learning_rate": 1.8316470568191069e-06, "loss": 0.5978, "step": 3596 }, { "epoch": 8.794621026894866, "grad_norm": 0.23147927279437583, "learning_rate": 1.8304135484261763e-06, "loss": 0.6103, "step": 3597 }, { "epoch": 8.797066014669927, "grad_norm": 0.33076328852681885, "learning_rate": 1.8291802156141414e-06, "loss": 0.7115, "step": 3598 }, { "epoch": 8.799511002444987, "grad_norm": 0.3902915468455055, "learning_rate": 1.8279470587064089e-06, "loss": 0.7379, "step": 3599 }, { "epoch": 8.80195599022005, "grad_norm": 0.19933831404052776, "learning_rate": 1.8267140780263426e-06, "loss": 0.5269, "step": 3600 }, { "epoch": 8.80440097799511, "grad_norm": 0.25602405112150456, "learning_rate": 1.8254812738972576e-06, "loss": 0.5156, "step": 3601 }, { "epoch": 8.80684596577017, "grad_norm": 0.3238957148239051, "learning_rate": 1.8242486466424236e-06, "loss": 0.5862, "step": 3602 }, { "epoch": 8.809290953545233, "grad_norm": 0.2650617154850533, "learning_rate": 1.8230161965850626e-06, "loss": 1.0186, "step": 3603 }, { "epoch": 8.811735941320293, "grad_norm": 0.2126378721161555, "learning_rate": 1.8217839240483527e-06, "loss": 0.4014, "step": 3604 }, { "epoch": 8.814180929095354, "grad_norm": 0.32915470229914523, "learning_rate": 1.8205518293554227e-06, "loss": 0.7862, "step": 3605 }, { "epoch": 8.816625916870416, "grad_norm": 0.3288675811089684, "learning_rate": 1.819319912829357e-06, "loss": 0.4115, "step": 3606 }, { "epoch": 8.819070904645477, "grad_norm": 0.27349779861110085, "learning_rate": 1.8180881747931912e-06, "loss": 0.6449, "step": 3607 }, { "epoch": 8.821515892420537, "grad_norm": 0.3366097350630797, "learning_rate": 1.8168566155699161e-06, "loss": 0.4608, "step": 3608 }, { "epoch": 8.8239608801956, "grad_norm": 0.2398592912437631, "learning_rate": 1.8156252354824746e-06, "loss": 0.5226, "step": 3609 }, { "epoch": 8.82640586797066, "grad_norm": 0.25126285784834845, "learning_rate": 1.8143940348537619e-06, "loss": 1.1248, "step": 3610 }, { "epoch": 8.82885085574572, "grad_norm": 0.20497404551365436, "learning_rate": 1.8131630140066284e-06, "loss": 0.4013, "step": 3611 }, { "epoch": 8.831295843520783, "grad_norm": 0.17657946261528143, "learning_rate": 1.8119321732638756e-06, "loss": 0.3747, "step": 3612 }, { "epoch": 8.833740831295843, "grad_norm": 0.2112306497322087, "learning_rate": 1.8107015129482568e-06, "loss": 0.4105, "step": 3613 }, { "epoch": 8.836185819070904, "grad_norm": 0.22649199266512973, "learning_rate": 1.8094710333824812e-06, "loss": 0.8881, "step": 3614 }, { "epoch": 8.838630806845966, "grad_norm": 0.17546627290486372, "learning_rate": 1.8082407348892078e-06, "loss": 0.3832, "step": 3615 }, { "epoch": 8.841075794621027, "grad_norm": 0.2826507214981961, "learning_rate": 1.8070106177910488e-06, "loss": 1.0206, "step": 3616 }, { "epoch": 8.843520782396087, "grad_norm": 0.30125207639250673, "learning_rate": 1.8057806824105689e-06, "loss": 0.5728, "step": 3617 }, { "epoch": 8.84596577017115, "grad_norm": 0.22873835128877792, "learning_rate": 1.8045509290702862e-06, "loss": 0.34, "step": 3618 }, { "epoch": 8.84841075794621, "grad_norm": 0.4193489042586497, "learning_rate": 1.8033213580926696e-06, "loss": 0.68, "step": 3619 }, { "epoch": 8.85085574572127, "grad_norm": 0.2858781203351356, "learning_rate": 1.80209196980014e-06, "loss": 0.4504, "step": 3620 }, { "epoch": 8.853300733496333, "grad_norm": 0.2958818440981415, "learning_rate": 1.8008627645150727e-06, "loss": 0.6342, "step": 3621 }, { "epoch": 8.855745721271393, "grad_norm": 0.28012187701396, "learning_rate": 1.7996337425597932e-06, "loss": 0.8687, "step": 3622 }, { "epoch": 8.858190709046454, "grad_norm": 0.3814683350693685, "learning_rate": 1.7984049042565777e-06, "loss": 0.4411, "step": 3623 }, { "epoch": 8.860635696821516, "grad_norm": 0.2640221794278391, "learning_rate": 1.7971762499276576e-06, "loss": 0.5539, "step": 3624 }, { "epoch": 8.863080684596577, "grad_norm": 0.20738841897062282, "learning_rate": 1.7959477798952135e-06, "loss": 0.6179, "step": 3625 }, { "epoch": 8.865525672371637, "grad_norm": 0.3262876550672289, "learning_rate": 1.7947194944813779e-06, "loss": 0.4065, "step": 3626 }, { "epoch": 8.8679706601467, "grad_norm": 0.4221722697792616, "learning_rate": 1.7934913940082355e-06, "loss": 0.4429, "step": 3627 }, { "epoch": 8.87041564792176, "grad_norm": 0.2229330865453247, "learning_rate": 1.7922634787978227e-06, "loss": 0.8856, "step": 3628 }, { "epoch": 8.87286063569682, "grad_norm": 0.25642928739595605, "learning_rate": 1.791035749172127e-06, "loss": 0.8247, "step": 3629 }, { "epoch": 8.875305623471883, "grad_norm": 0.26881034316757757, "learning_rate": 1.789808205453087e-06, "loss": 0.8383, "step": 3630 }, { "epoch": 8.877750611246944, "grad_norm": 0.23335074805961314, "learning_rate": 1.788580847962592e-06, "loss": 0.5438, "step": 3631 }, { "epoch": 8.880195599022004, "grad_norm": 0.20788433365989567, "learning_rate": 1.7873536770224847e-06, "loss": 0.4696, "step": 3632 }, { "epoch": 8.882640586797066, "grad_norm": 0.29934702043428696, "learning_rate": 1.7861266929545561e-06, "loss": 0.4366, "step": 3633 }, { "epoch": 8.885085574572127, "grad_norm": 0.22062837977260127, "learning_rate": 1.7848998960805515e-06, "loss": 0.7434, "step": 3634 }, { "epoch": 8.887530562347187, "grad_norm": 0.13359902350951794, "learning_rate": 1.783673286722163e-06, "loss": 0.4307, "step": 3635 }, { "epoch": 8.88997555012225, "grad_norm": 0.35273608991692523, "learning_rate": 1.7824468652010366e-06, "loss": 0.6067, "step": 3636 }, { "epoch": 8.89242053789731, "grad_norm": 0.20696255054202498, "learning_rate": 1.7812206318387674e-06, "loss": 0.87, "step": 3637 }, { "epoch": 8.894865525672373, "grad_norm": 0.2399146374565134, "learning_rate": 1.779994586956903e-06, "loss": 0.4494, "step": 3638 }, { "epoch": 8.897310513447433, "grad_norm": 0.2326272397900849, "learning_rate": 1.7787687308769394e-06, "loss": 0.4104, "step": 3639 }, { "epoch": 8.899755501222494, "grad_norm": 0.2801005163790302, "learning_rate": 1.777543063920325e-06, "loss": 0.5831, "step": 3640 }, { "epoch": 8.902200488997554, "grad_norm": 0.25036850775354236, "learning_rate": 1.7763175864084563e-06, "loss": 0.8582, "step": 3641 }, { "epoch": 8.904645476772616, "grad_norm": 0.2770522128615707, "learning_rate": 1.7750922986626828e-06, "loss": 0.4781, "step": 3642 }, { "epoch": 8.907090464547677, "grad_norm": 0.35283891841933107, "learning_rate": 1.773867201004303e-06, "loss": 0.3935, "step": 3643 }, { "epoch": 8.90953545232274, "grad_norm": 0.2634912974010283, "learning_rate": 1.7726422937545653e-06, "loss": 0.8075, "step": 3644 }, { "epoch": 8.9119804400978, "grad_norm": 0.20034590320795434, "learning_rate": 1.7714175772346686e-06, "loss": 0.4773, "step": 3645 }, { "epoch": 8.91442542787286, "grad_norm": 0.2097835495187987, "learning_rate": 1.7701930517657615e-06, "loss": 0.3724, "step": 3646 }, { "epoch": 8.91687041564792, "grad_norm": 0.44833926481000913, "learning_rate": 1.7689687176689418e-06, "loss": 0.5903, "step": 3647 }, { "epoch": 8.919315403422983, "grad_norm": 0.21785872488112026, "learning_rate": 1.7677445752652594e-06, "loss": 0.3931, "step": 3648 }, { "epoch": 8.921760391198044, "grad_norm": 0.4219317473076237, "learning_rate": 1.7665206248757116e-06, "loss": 0.4517, "step": 3649 }, { "epoch": 8.924205378973106, "grad_norm": 0.23288670727080563, "learning_rate": 1.7652968668212462e-06, "loss": 1.0049, "step": 3650 }, { "epoch": 8.926650366748166, "grad_norm": 0.3105504328060746, "learning_rate": 1.76407330142276e-06, "loss": 0.6928, "step": 3651 }, { "epoch": 8.929095354523227, "grad_norm": 0.2448374618102944, "learning_rate": 1.7628499290011014e-06, "loss": 0.8116, "step": 3652 }, { "epoch": 8.93154034229829, "grad_norm": 0.3930291356974308, "learning_rate": 1.7616267498770644e-06, "loss": 0.9907, "step": 3653 }, { "epoch": 8.93398533007335, "grad_norm": 0.20867402496773718, "learning_rate": 1.7604037643713973e-06, "loss": 0.4163, "step": 3654 }, { "epoch": 8.93643031784841, "grad_norm": 0.3972959669402503, "learning_rate": 1.7591809728047934e-06, "loss": 0.7532, "step": 3655 }, { "epoch": 8.938875305623473, "grad_norm": 0.39822776357470846, "learning_rate": 1.7579583754978968e-06, "loss": 0.3948, "step": 3656 }, { "epoch": 8.941320293398533, "grad_norm": 0.37464443040936035, "learning_rate": 1.7567359727713e-06, "loss": 0.3813, "step": 3657 }, { "epoch": 8.943765281173594, "grad_norm": 0.2734178913534183, "learning_rate": 1.7555137649455462e-06, "loss": 0.6545, "step": 3658 }, { "epoch": 8.946210268948656, "grad_norm": 0.23281213624014122, "learning_rate": 1.7542917523411252e-06, "loss": 0.6916, "step": 3659 }, { "epoch": 8.948655256723717, "grad_norm": 0.27526596644154083, "learning_rate": 1.7530699352784772e-06, "loss": 0.7506, "step": 3660 }, { "epoch": 8.951100244498777, "grad_norm": 0.2202602641955853, "learning_rate": 1.7518483140779895e-06, "loss": 0.5424, "step": 3661 }, { "epoch": 8.95354523227384, "grad_norm": 0.3514005572126245, "learning_rate": 1.750626889060001e-06, "loss": 0.4009, "step": 3662 }, { "epoch": 8.9559902200489, "grad_norm": 0.18485340344583961, "learning_rate": 1.7494056605447956e-06, "loss": 0.4246, "step": 3663 }, { "epoch": 8.95843520782396, "grad_norm": 0.19706697309874063, "learning_rate": 1.748184628852609e-06, "loss": 0.5644, "step": 3664 }, { "epoch": 8.960880195599023, "grad_norm": 0.23839806335486152, "learning_rate": 1.7469637943036223e-06, "loss": 0.5784, "step": 3665 }, { "epoch": 8.963325183374083, "grad_norm": 0.32638494809175844, "learning_rate": 1.7457431572179672e-06, "loss": 0.6155, "step": 3666 }, { "epoch": 8.965770171149144, "grad_norm": 0.21498940232261818, "learning_rate": 1.7445227179157215e-06, "loss": 0.4677, "step": 3667 }, { "epoch": 8.968215158924206, "grad_norm": 0.3010381909948852, "learning_rate": 1.7433024767169138e-06, "loss": 0.2288, "step": 3668 }, { "epoch": 8.970660146699267, "grad_norm": 0.25056184123472935, "learning_rate": 1.7420824339415181e-06, "loss": 0.6996, "step": 3669 }, { "epoch": 8.973105134474327, "grad_norm": 0.4725189358441569, "learning_rate": 1.7408625899094587e-06, "loss": 0.447, "step": 3670 }, { "epoch": 8.97555012224939, "grad_norm": 0.26131544802175893, "learning_rate": 1.7396429449406049e-06, "loss": 0.8393, "step": 3671 }, { "epoch": 8.97799511002445, "grad_norm": 0.17502893508039313, "learning_rate": 1.7384234993547772e-06, "loss": 0.6801, "step": 3672 }, { "epoch": 8.98044009779951, "grad_norm": 0.31462350844714715, "learning_rate": 1.7372042534717414e-06, "loss": 0.4786, "step": 3673 }, { "epoch": 8.982885085574573, "grad_norm": 0.2007139309560344, "learning_rate": 1.735985207611211e-06, "loss": 0.5823, "step": 3674 }, { "epoch": 8.985330073349633, "grad_norm": 0.2681694303288693, "learning_rate": 1.7347663620928495e-06, "loss": 0.5757, "step": 3675 }, { "epoch": 8.987775061124694, "grad_norm": 0.2402669852138904, "learning_rate": 1.7335477172362648e-06, "loss": 0.3457, "step": 3676 }, { "epoch": 8.990220048899756, "grad_norm": 0.3139336397866662, "learning_rate": 1.7323292733610136e-06, "loss": 0.4574, "step": 3677 }, { "epoch": 8.992665036674817, "grad_norm": 0.31739826620724554, "learning_rate": 1.7311110307866003e-06, "loss": 0.4143, "step": 3678 }, { "epoch": 8.995110024449877, "grad_norm": 0.1857274865172723, "learning_rate": 1.7298929898324757e-06, "loss": 0.5478, "step": 3679 }, { "epoch": 8.99755501222494, "grad_norm": 0.2418060200439178, "learning_rate": 1.7286751508180377e-06, "loss": 1.0249, "step": 3680 }, { "epoch": 9.0, "grad_norm": 0.2897884308653675, "learning_rate": 1.7274575140626318e-06, "loss": 0.3965, "step": 3681 }, { "epoch": 9.00244498777506, "grad_norm": 0.3228548685613579, "learning_rate": 1.7262400798855506e-06, "loss": 0.4992, "step": 3682 }, { "epoch": 9.004889975550123, "grad_norm": 0.2840729176965229, "learning_rate": 1.7250228486060327e-06, "loss": 0.3513, "step": 3683 }, { "epoch": 9.007334963325183, "grad_norm": 0.2740017510081472, "learning_rate": 1.7238058205432645e-06, "loss": 1.3055, "step": 3684 }, { "epoch": 9.009779951100244, "grad_norm": 0.32951515299543316, "learning_rate": 1.7225889960163769e-06, "loss": 0.496, "step": 3685 }, { "epoch": 9.012224938875306, "grad_norm": 0.34827803623910913, "learning_rate": 1.7213723753444517e-06, "loss": 0.7676, "step": 3686 }, { "epoch": 9.014669926650367, "grad_norm": 0.34380633238781994, "learning_rate": 1.720155958846513e-06, "loss": 0.67, "step": 3687 }, { "epoch": 9.017114914425427, "grad_norm": 0.19034952715698975, "learning_rate": 1.7189397468415342e-06, "loss": 0.4378, "step": 3688 }, { "epoch": 9.01955990220049, "grad_norm": 0.19875623691519614, "learning_rate": 1.7177237396484333e-06, "loss": 0.4273, "step": 3689 }, { "epoch": 9.02200488997555, "grad_norm": 0.21094727794638565, "learning_rate": 1.7165079375860755e-06, "loss": 0.372, "step": 3690 }, { "epoch": 9.02444987775061, "grad_norm": 0.22221619136656934, "learning_rate": 1.7152923409732708e-06, "loss": 0.6271, "step": 3691 }, { "epoch": 9.026894865525673, "grad_norm": 0.2648316594089732, "learning_rate": 1.7140769501287779e-06, "loss": 0.6083, "step": 3692 }, { "epoch": 9.029339853300733, "grad_norm": 0.17813606891911415, "learning_rate": 1.7128617653712995e-06, "loss": 0.5181, "step": 3693 }, { "epoch": 9.031784841075794, "grad_norm": 0.25785981157895854, "learning_rate": 1.7116467870194848e-06, "loss": 0.7364, "step": 3694 }, { "epoch": 9.034229828850856, "grad_norm": 0.20646755174345485, "learning_rate": 1.7104320153919282e-06, "loss": 0.4577, "step": 3695 }, { "epoch": 9.036674816625917, "grad_norm": 0.26057312865616206, "learning_rate": 1.7092174508071723e-06, "loss": 0.3341, "step": 3696 }, { "epoch": 9.039119804400977, "grad_norm": 0.3090236312135023, "learning_rate": 1.7080030935837028e-06, "loss": 0.4987, "step": 3697 }, { "epoch": 9.04156479217604, "grad_norm": 0.19374664850984177, "learning_rate": 1.706788944039952e-06, "loss": 0.4914, "step": 3698 }, { "epoch": 9.0440097799511, "grad_norm": 0.17954118068271388, "learning_rate": 1.705575002494298e-06, "loss": 0.5855, "step": 3699 }, { "epoch": 9.04645476772616, "grad_norm": 0.3357297595797145, "learning_rate": 1.704361269265064e-06, "loss": 0.7081, "step": 3700 }, { "epoch": 9.048899755501223, "grad_norm": 0.22778642926523404, "learning_rate": 1.7031477446705177e-06, "loss": 0.4111, "step": 3701 }, { "epoch": 9.051344743276283, "grad_norm": 0.29308149555047897, "learning_rate": 1.7019344290288745e-06, "loss": 0.6798, "step": 3702 }, { "epoch": 9.053789731051344, "grad_norm": 0.2262081971621139, "learning_rate": 1.7007213226582925e-06, "loss": 1.1001, "step": 3703 }, { "epoch": 9.056234718826406, "grad_norm": 0.2887851866884744, "learning_rate": 1.6995084258768768e-06, "loss": 0.6088, "step": 3704 }, { "epoch": 9.058679706601467, "grad_norm": 0.17916203359554328, "learning_rate": 1.698295739002675e-06, "loss": 0.6397, "step": 3705 }, { "epoch": 9.061124694376527, "grad_norm": 0.2160105868811308, "learning_rate": 1.6970832623536831e-06, "loss": 0.3685, "step": 3706 }, { "epoch": 9.06356968215159, "grad_norm": 0.33142951810255217, "learning_rate": 1.6958709962478393e-06, "loss": 0.5572, "step": 3707 }, { "epoch": 9.06601466992665, "grad_norm": 0.2087830952399488, "learning_rate": 1.6946589410030287e-06, "loss": 0.647, "step": 3708 }, { "epoch": 9.06845965770171, "grad_norm": 0.20544846030078978, "learning_rate": 1.693447096937079e-06, "loss": 0.5347, "step": 3709 }, { "epoch": 9.070904645476773, "grad_norm": 0.3136290222947587, "learning_rate": 1.6922354643677635e-06, "loss": 0.6184, "step": 3710 }, { "epoch": 9.073349633251834, "grad_norm": 0.20037297567558165, "learning_rate": 1.6910240436127998e-06, "loss": 0.5755, "step": 3711 }, { "epoch": 9.075794621026894, "grad_norm": 0.28240057248266964, "learning_rate": 1.6898128349898506e-06, "loss": 0.4307, "step": 3712 }, { "epoch": 9.078239608801956, "grad_norm": 0.21289114597158304, "learning_rate": 1.6886018388165225e-06, "loss": 0.5909, "step": 3713 }, { "epoch": 9.080684596577017, "grad_norm": 0.2560310542811439, "learning_rate": 1.6873910554103665e-06, "loss": 0.5309, "step": 3714 }, { "epoch": 9.083129584352077, "grad_norm": 0.1899935635694557, "learning_rate": 1.686180485088877e-06, "loss": 0.4388, "step": 3715 }, { "epoch": 9.08557457212714, "grad_norm": 0.17742918419061687, "learning_rate": 1.6849701281694939e-06, "loss": 0.4545, "step": 3716 }, { "epoch": 9.0880195599022, "grad_norm": 0.2944865736054724, "learning_rate": 1.6837599849695997e-06, "loss": 1.0101, "step": 3717 }, { "epoch": 9.09046454767726, "grad_norm": 0.23322332682570077, "learning_rate": 1.6825500558065227e-06, "loss": 0.6963, "step": 3718 }, { "epoch": 9.092909535452323, "grad_norm": 0.20870916576435303, "learning_rate": 1.6813403409975343e-06, "loss": 0.7526, "step": 3719 }, { "epoch": 9.095354523227384, "grad_norm": 0.21417039188762763, "learning_rate": 1.6801308408598482e-06, "loss": 0.5864, "step": 3720 }, { "epoch": 9.097799511002446, "grad_norm": 0.23700313280661495, "learning_rate": 1.6789215557106235e-06, "loss": 0.508, "step": 3721 }, { "epoch": 9.100244498777506, "grad_norm": 0.17053782317739405, "learning_rate": 1.6777124858669627e-06, "loss": 0.5249, "step": 3722 }, { "epoch": 9.102689486552567, "grad_norm": 0.3076115590266008, "learning_rate": 1.6765036316459116e-06, "loss": 0.6926, "step": 3723 }, { "epoch": 9.10513447432763, "grad_norm": 0.2842074746253879, "learning_rate": 1.675294993364459e-06, "loss": 0.3693, "step": 3724 }, { "epoch": 9.10757946210269, "grad_norm": 0.35671516508547924, "learning_rate": 1.6740865713395375e-06, "loss": 0.5412, "step": 3725 }, { "epoch": 9.11002444987775, "grad_norm": 0.3030085451625398, "learning_rate": 1.6728783658880237e-06, "loss": 0.6223, "step": 3726 }, { "epoch": 9.112469437652813, "grad_norm": 0.364928485959764, "learning_rate": 1.6716703773267362e-06, "loss": 0.5489, "step": 3727 }, { "epoch": 9.114914425427873, "grad_norm": 0.36402170204914, "learning_rate": 1.6704626059724365e-06, "loss": 0.4529, "step": 3728 }, { "epoch": 9.117359413202934, "grad_norm": 0.32530071850513187, "learning_rate": 1.6692550521418315e-06, "loss": 0.795, "step": 3729 }, { "epoch": 9.119804400977996, "grad_norm": 0.24370227121304872, "learning_rate": 1.668047716151569e-06, "loss": 0.6319, "step": 3730 }, { "epoch": 9.122249388753056, "grad_norm": 0.23501026701979078, "learning_rate": 1.6668405983182396e-06, "loss": 0.8241, "step": 3731 }, { "epoch": 9.124694376528117, "grad_norm": 0.21172114971586284, "learning_rate": 1.6656336989583774e-06, "loss": 0.837, "step": 3732 }, { "epoch": 9.12713936430318, "grad_norm": 0.3014212597291553, "learning_rate": 1.6644270183884591e-06, "loss": 0.7785, "step": 3733 }, { "epoch": 9.12958435207824, "grad_norm": 0.23603569185838502, "learning_rate": 1.6632205569249043e-06, "loss": 0.458, "step": 3734 }, { "epoch": 9.1320293398533, "grad_norm": 0.19829630956475972, "learning_rate": 1.6620143148840743e-06, "loss": 0.6728, "step": 3735 }, { "epoch": 9.134474327628363, "grad_norm": 0.30043491651235604, "learning_rate": 1.6608082925822736e-06, "loss": 0.9533, "step": 3736 }, { "epoch": 9.136919315403423, "grad_norm": 0.2716964335105591, "learning_rate": 1.659602490335749e-06, "loss": 0.458, "step": 3737 }, { "epoch": 9.139364303178484, "grad_norm": 0.2368201972049681, "learning_rate": 1.6583969084606882e-06, "loss": 0.7162, "step": 3738 }, { "epoch": 9.141809290953546, "grad_norm": 0.22048558433269078, "learning_rate": 1.6571915472732244e-06, "loss": 0.6522, "step": 3739 }, { "epoch": 9.144254278728607, "grad_norm": 0.1781830068774942, "learning_rate": 1.6559864070894296e-06, "loss": 0.4682, "step": 3740 }, { "epoch": 9.146699266503667, "grad_norm": 0.20422760914985105, "learning_rate": 1.6547814882253199e-06, "loss": 0.3882, "step": 3741 }, { "epoch": 9.14914425427873, "grad_norm": 0.24461098580943325, "learning_rate": 1.6535767909968523e-06, "loss": 0.8867, "step": 3742 }, { "epoch": 9.15158924205379, "grad_norm": 0.26469061251149767, "learning_rate": 1.6523723157199256e-06, "loss": 0.8181, "step": 3743 }, { "epoch": 9.15403422982885, "grad_norm": 0.16941790869842885, "learning_rate": 1.6511680627103815e-06, "loss": 0.744, "step": 3744 }, { "epoch": 9.156479217603913, "grad_norm": 0.2265218410292478, "learning_rate": 1.6499640322840015e-06, "loss": 0.4978, "step": 3745 }, { "epoch": 9.158924205378973, "grad_norm": 0.1738015802257966, "learning_rate": 1.6487602247565113e-06, "loss": 0.512, "step": 3746 }, { "epoch": 9.161369193154034, "grad_norm": 0.17724390341372845, "learning_rate": 1.647556640443576e-06, "loss": 0.9639, "step": 3747 }, { "epoch": 9.163814180929096, "grad_norm": 0.19530979704047577, "learning_rate": 1.6463532796608033e-06, "loss": 0.4142, "step": 3748 }, { "epoch": 9.166259168704157, "grad_norm": 0.36818484818698294, "learning_rate": 1.6451501427237409e-06, "loss": 0.5216, "step": 3749 }, { "epoch": 9.168704156479217, "grad_norm": 0.27025054447724406, "learning_rate": 1.6439472299478804e-06, "loss": 0.8026, "step": 3750 }, { "epoch": 9.17114914425428, "grad_norm": 0.2535552631533771, "learning_rate": 1.6427445416486526e-06, "loss": 0.8205, "step": 3751 }, { "epoch": 9.17359413202934, "grad_norm": 0.31793232299305474, "learning_rate": 1.6415420781414298e-06, "loss": 0.6869, "step": 3752 }, { "epoch": 9.1760391198044, "grad_norm": 0.2601201240994496, "learning_rate": 1.6403398397415257e-06, "loss": 0.797, "step": 3753 }, { "epoch": 9.178484107579463, "grad_norm": 0.2747501640079801, "learning_rate": 1.639137826764194e-06, "loss": 0.4444, "step": 3754 }, { "epoch": 9.180929095354523, "grad_norm": 0.299737542945522, "learning_rate": 1.6379360395246302e-06, "loss": 0.4071, "step": 3755 }, { "epoch": 9.183374083129584, "grad_norm": 0.22879057273475878, "learning_rate": 1.6367344783379713e-06, "loss": 0.5856, "step": 3756 }, { "epoch": 9.185819070904646, "grad_norm": 0.23508058587943995, "learning_rate": 1.6355331435192931e-06, "loss": 0.7579, "step": 3757 }, { "epoch": 9.188264058679707, "grad_norm": 0.2544698070829883, "learning_rate": 1.6343320353836136e-06, "loss": 0.5149, "step": 3758 }, { "epoch": 9.190709046454767, "grad_norm": 0.26067841063301544, "learning_rate": 1.6331311542458905e-06, "loss": 0.6982, "step": 3759 }, { "epoch": 9.19315403422983, "grad_norm": 0.36520995289866987, "learning_rate": 1.6319305004210218e-06, "loss": 0.3495, "step": 3760 }, { "epoch": 9.19559902200489, "grad_norm": 0.1892094000401061, "learning_rate": 1.6307300742238483e-06, "loss": 0.5657, "step": 3761 }, { "epoch": 9.19804400977995, "grad_norm": 0.2474017554639616, "learning_rate": 1.629529875969148e-06, "loss": 0.3428, "step": 3762 }, { "epoch": 9.200488997555013, "grad_norm": 0.25912866439841126, "learning_rate": 1.6283299059716407e-06, "loss": 0.9238, "step": 3763 }, { "epoch": 9.202933985330073, "grad_norm": 0.2650997651041663, "learning_rate": 1.6271301645459858e-06, "loss": 0.5867, "step": 3764 }, { "epoch": 9.205378973105134, "grad_norm": 0.32228420738556607, "learning_rate": 1.6259306520067822e-06, "loss": 0.4173, "step": 3765 }, { "epoch": 9.207823960880196, "grad_norm": 0.506277531583115, "learning_rate": 1.6247313686685707e-06, "loss": 0.7579, "step": 3766 }, { "epoch": 9.210268948655257, "grad_norm": 0.1915610389020002, "learning_rate": 1.6235323148458307e-06, "loss": 0.3681, "step": 3767 }, { "epoch": 9.212713936430317, "grad_norm": 0.1671853483903209, "learning_rate": 1.622333490852981e-06, "loss": 0.7176, "step": 3768 }, { "epoch": 9.21515892420538, "grad_norm": 0.23659874535830658, "learning_rate": 1.6211348970043807e-06, "loss": 0.8612, "step": 3769 }, { "epoch": 9.21760391198044, "grad_norm": 0.3086686194070926, "learning_rate": 1.6199365336143297e-06, "loss": 0.5305, "step": 3770 }, { "epoch": 9.2200488997555, "grad_norm": 0.23544422666156783, "learning_rate": 1.6187384009970639e-06, "loss": 1.073, "step": 3771 }, { "epoch": 9.222493887530563, "grad_norm": 0.31378595219232436, "learning_rate": 1.6175404994667637e-06, "loss": 0.5475, "step": 3772 }, { "epoch": 9.224938875305623, "grad_norm": 0.2913281005225974, "learning_rate": 1.6163428293375455e-06, "loss": 0.4706, "step": 3773 }, { "epoch": 9.227383863080684, "grad_norm": 0.2124143285837525, "learning_rate": 1.6151453909234649e-06, "loss": 0.6314, "step": 3774 }, { "epoch": 9.229828850855746, "grad_norm": 0.17088870560903544, "learning_rate": 1.6139481845385186e-06, "loss": 0.6387, "step": 3775 }, { "epoch": 9.232273838630807, "grad_norm": 0.22286428608958236, "learning_rate": 1.6127512104966413e-06, "loss": 0.437, "step": 3776 }, { "epoch": 9.234718826405867, "grad_norm": 0.2856527412349572, "learning_rate": 1.6115544691117074e-06, "loss": 0.7422, "step": 3777 }, { "epoch": 9.23716381418093, "grad_norm": 0.29922213258273284, "learning_rate": 1.6103579606975291e-06, "loss": 0.3501, "step": 3778 }, { "epoch": 9.23960880195599, "grad_norm": 0.256104060447046, "learning_rate": 1.6091616855678583e-06, "loss": 0.4883, "step": 3779 }, { "epoch": 9.24205378973105, "grad_norm": 0.3106199968381704, "learning_rate": 1.6079656440363864e-06, "loss": 0.4477, "step": 3780 }, { "epoch": 9.244498777506113, "grad_norm": 0.24593255016935583, "learning_rate": 1.6067698364167418e-06, "loss": 0.3889, "step": 3781 }, { "epoch": 9.246943765281173, "grad_norm": 0.13855123344890902, "learning_rate": 1.6055742630224939e-06, "loss": 0.3928, "step": 3782 }, { "epoch": 9.249388753056234, "grad_norm": 0.20610135502255728, "learning_rate": 1.6043789241671489e-06, "loss": 0.8226, "step": 3783 }, { "epoch": 9.251833740831296, "grad_norm": 0.24231371367318064, "learning_rate": 1.6031838201641517e-06, "loss": 0.4779, "step": 3784 }, { "epoch": 9.254278728606357, "grad_norm": 0.3166356702014923, "learning_rate": 1.6019889513268867e-06, "loss": 0.5674, "step": 3785 }, { "epoch": 9.256723716381417, "grad_norm": 0.35368932502481937, "learning_rate": 1.6007943179686752e-06, "loss": 0.475, "step": 3786 }, { "epoch": 9.25916870415648, "grad_norm": 0.429863933558201, "learning_rate": 1.599599920402778e-06, "loss": 0.3623, "step": 3787 }, { "epoch": 9.26161369193154, "grad_norm": 0.24244222291930287, "learning_rate": 1.5984057589423923e-06, "loss": 0.6596, "step": 3788 }, { "epoch": 9.2640586797066, "grad_norm": 0.4125940137785228, "learning_rate": 1.5972118339006554e-06, "loss": 0.3446, "step": 3789 }, { "epoch": 9.266503667481663, "grad_norm": 0.30449985060980256, "learning_rate": 1.5960181455906421e-06, "loss": 0.8954, "step": 3790 }, { "epoch": 9.268948655256724, "grad_norm": 0.23175075675837326, "learning_rate": 1.5948246943253646e-06, "loss": 0.6334, "step": 3791 }, { "epoch": 9.271393643031784, "grad_norm": 0.20916099129299942, "learning_rate": 1.5936314804177716e-06, "loss": 0.694, "step": 3792 }, { "epoch": 9.273838630806846, "grad_norm": 0.28891342757622707, "learning_rate": 1.5924385041807536e-06, "loss": 0.4279, "step": 3793 }, { "epoch": 9.276283618581907, "grad_norm": 0.2404900443998669, "learning_rate": 1.5912457659271347e-06, "loss": 0.7967, "step": 3794 }, { "epoch": 9.278728606356967, "grad_norm": 0.2750882550804834, "learning_rate": 1.5900532659696789e-06, "loss": 0.7529, "step": 3795 }, { "epoch": 9.28117359413203, "grad_norm": 0.30912787736361075, "learning_rate": 1.5888610046210866e-06, "loss": 0.4239, "step": 3796 }, { "epoch": 9.28361858190709, "grad_norm": 0.2508255942827207, "learning_rate": 1.5876689821939966e-06, "loss": 0.4374, "step": 3797 }, { "epoch": 9.28606356968215, "grad_norm": 0.23100608921988958, "learning_rate": 1.5864771990009831e-06, "loss": 0.4373, "step": 3798 }, { "epoch": 9.288508557457213, "grad_norm": 0.28046214821294163, "learning_rate": 1.5852856553545604e-06, "loss": 0.3865, "step": 3799 }, { "epoch": 9.290953545232274, "grad_norm": 0.390727020032442, "learning_rate": 1.584094351567178e-06, "loss": 0.3697, "step": 3800 }, { "epoch": 9.293398533007334, "grad_norm": 0.2185051166879761, "learning_rate": 1.582903287951223e-06, "loss": 0.3471, "step": 3801 }, { "epoch": 9.295843520782396, "grad_norm": 0.22014036037899523, "learning_rate": 1.5817124648190192e-06, "loss": 0.718, "step": 3802 }, { "epoch": 9.298288508557457, "grad_norm": 0.27577483119513485, "learning_rate": 1.5805218824828273e-06, "loss": 0.3761, "step": 3803 }, { "epoch": 9.30073349633252, "grad_norm": 0.2625618545995872, "learning_rate": 1.579331541254846e-06, "loss": 0.5664, "step": 3804 }, { "epoch": 9.30317848410758, "grad_norm": 0.36177283522201026, "learning_rate": 1.5781414414472108e-06, "loss": 0.4816, "step": 3805 }, { "epoch": 9.30562347188264, "grad_norm": 0.31061749211889167, "learning_rate": 1.5769515833719922e-06, "loss": 0.6579, "step": 3806 }, { "epoch": 9.3080684596577, "grad_norm": 0.2036908061958182, "learning_rate": 1.5757619673411984e-06, "loss": 0.8445, "step": 3807 }, { "epoch": 9.310513447432763, "grad_norm": 0.1590616427309036, "learning_rate": 1.5745725936667728e-06, "loss": 0.5905, "step": 3808 }, { "epoch": 9.312958435207824, "grad_norm": 0.327381254181539, "learning_rate": 1.5733834626605984e-06, "loss": 0.655, "step": 3809 }, { "epoch": 9.315403422982886, "grad_norm": 0.26672461852296214, "learning_rate": 1.5721945746344916e-06, "loss": 0.4225, "step": 3810 }, { "epoch": 9.317848410757946, "grad_norm": 0.24269502735188786, "learning_rate": 1.5710059299002062e-06, "loss": 0.4836, "step": 3811 }, { "epoch": 9.320293398533007, "grad_norm": 0.340286871284897, "learning_rate": 1.5698175287694314e-06, "loss": 0.7579, "step": 3812 }, { "epoch": 9.32273838630807, "grad_norm": 0.2805701708215805, "learning_rate": 1.5686293715537931e-06, "loss": 0.4126, "step": 3813 }, { "epoch": 9.32518337408313, "grad_norm": 0.25583317919518245, "learning_rate": 1.5674414585648546e-06, "loss": 0.5164, "step": 3814 }, { "epoch": 9.32762836185819, "grad_norm": 0.21931110676576776, "learning_rate": 1.5662537901141134e-06, "loss": 0.3632, "step": 3815 }, { "epoch": 9.330073349633253, "grad_norm": 0.2300101633693936, "learning_rate": 1.565066366513003e-06, "loss": 0.5609, "step": 3816 }, { "epoch": 9.332518337408313, "grad_norm": 0.2004595426909433, "learning_rate": 1.5638791880728934e-06, "loss": 0.4431, "step": 3817 }, { "epoch": 9.334963325183374, "grad_norm": 0.3680087940803977, "learning_rate": 1.5626922551050893e-06, "loss": 0.6512, "step": 3818 }, { "epoch": 9.337408312958436, "grad_norm": 0.33855738738370084, "learning_rate": 1.5615055679208328e-06, "loss": 1.0626, "step": 3819 }, { "epoch": 9.339853300733497, "grad_norm": 0.17288329992854268, "learning_rate": 1.5603191268312996e-06, "loss": 0.4605, "step": 3820 }, { "epoch": 9.342298288508557, "grad_norm": 0.25364415036630994, "learning_rate": 1.5591329321476021e-06, "loss": 0.5384, "step": 3821 }, { "epoch": 9.34474327628362, "grad_norm": 0.3390621030101342, "learning_rate": 1.5579469841807868e-06, "loss": 0.4693, "step": 3822 }, { "epoch": 9.34718826405868, "grad_norm": 0.3291154609181924, "learning_rate": 1.5567612832418378e-06, "loss": 0.8564, "step": 3823 }, { "epoch": 9.34963325183374, "grad_norm": 0.25787013124519487, "learning_rate": 1.5555758296416712e-06, "loss": 0.7168, "step": 3824 }, { "epoch": 9.352078239608803, "grad_norm": 0.24642638379136683, "learning_rate": 1.5543906236911423e-06, "loss": 0.8619, "step": 3825 }, { "epoch": 9.354523227383863, "grad_norm": 0.2607012277639971, "learning_rate": 1.5532056657010383e-06, "loss": 0.4176, "step": 3826 }, { "epoch": 9.356968215158924, "grad_norm": 0.21880186727465328, "learning_rate": 1.552020955982082e-06, "loss": 0.5805, "step": 3827 }, { "epoch": 9.359413202933986, "grad_norm": 0.24452764654808748, "learning_rate": 1.5508364948449311e-06, "loss": 0.6891, "step": 3828 }, { "epoch": 9.361858190709047, "grad_norm": 0.261778046619333, "learning_rate": 1.5496522826001795e-06, "loss": 0.4183, "step": 3829 }, { "epoch": 9.364303178484107, "grad_norm": 0.24457021633396422, "learning_rate": 1.5484683195583542e-06, "loss": 0.6153, "step": 3830 }, { "epoch": 9.36674816625917, "grad_norm": 0.2788064621134573, "learning_rate": 1.5472846060299175e-06, "loss": 0.7051, "step": 3831 }, { "epoch": 9.36919315403423, "grad_norm": 0.18139567183952385, "learning_rate": 1.5461011423252651e-06, "loss": 0.3695, "step": 3832 }, { "epoch": 9.37163814180929, "grad_norm": 0.2343131057488101, "learning_rate": 1.54491792875473e-06, "loss": 0.4752, "step": 3833 }, { "epoch": 9.374083129584353, "grad_norm": 0.21484676209749104, "learning_rate": 1.5437349656285773e-06, "loss": 1.0612, "step": 3834 }, { "epoch": 9.376528117359413, "grad_norm": 0.25035329740727935, "learning_rate": 1.5425522532570052e-06, "loss": 0.4031, "step": 3835 }, { "epoch": 9.378973105134474, "grad_norm": 0.26642890222957083, "learning_rate": 1.5413697919501511e-06, "loss": 0.8965, "step": 3836 }, { "epoch": 9.381418092909536, "grad_norm": 0.2719855763447448, "learning_rate": 1.5401875820180818e-06, "loss": 0.681, "step": 3837 }, { "epoch": 9.383863080684597, "grad_norm": 0.23550996583705056, "learning_rate": 1.5390056237707995e-06, "loss": 0.5173, "step": 3838 }, { "epoch": 9.386308068459657, "grad_norm": 0.2269722291731216, "learning_rate": 1.5378239175182414e-06, "loss": 0.4988, "step": 3839 }, { "epoch": 9.38875305623472, "grad_norm": 0.2748394123876589, "learning_rate": 1.5366424635702777e-06, "loss": 0.3344, "step": 3840 }, { "epoch": 9.39119804400978, "grad_norm": 0.34625633943457906, "learning_rate": 1.5354612622367127e-06, "loss": 0.4166, "step": 3841 }, { "epoch": 9.39364303178484, "grad_norm": 0.41023735900581076, "learning_rate": 1.5342803138272839e-06, "loss": 0.3633, "step": 3842 }, { "epoch": 9.396088019559903, "grad_norm": 0.27251988760932444, "learning_rate": 1.5330996186516638e-06, "loss": 0.4134, "step": 3843 }, { "epoch": 9.398533007334963, "grad_norm": 0.26046606941966505, "learning_rate": 1.5319191770194575e-06, "loss": 0.427, "step": 3844 }, { "epoch": 9.400977995110024, "grad_norm": 0.18528006823525117, "learning_rate": 1.5307389892402036e-06, "loss": 0.4514, "step": 3845 }, { "epoch": 9.403422982885086, "grad_norm": 0.42183434125366914, "learning_rate": 1.5295590556233736e-06, "loss": 0.6703, "step": 3846 }, { "epoch": 9.405867970660147, "grad_norm": 0.3267238734885524, "learning_rate": 1.528379376478375e-06, "loss": 0.7672, "step": 3847 }, { "epoch": 9.408312958435207, "grad_norm": 0.2593569367199505, "learning_rate": 1.5271999521145447e-06, "loss": 0.437, "step": 3848 }, { "epoch": 9.41075794621027, "grad_norm": 0.3508763176999022, "learning_rate": 1.5260207828411565e-06, "loss": 0.7, "step": 3849 }, { "epoch": 9.41320293398533, "grad_norm": 0.3925283838131019, "learning_rate": 1.5248418689674146e-06, "loss": 0.3971, "step": 3850 }, { "epoch": 9.41564792176039, "grad_norm": 0.282276042080178, "learning_rate": 1.5236632108024573e-06, "loss": 0.5469, "step": 3851 }, { "epoch": 9.418092909535453, "grad_norm": 0.2483119944427397, "learning_rate": 1.5224848086553557e-06, "loss": 0.4509, "step": 3852 }, { "epoch": 9.420537897310513, "grad_norm": 0.21266501317302733, "learning_rate": 1.521306662835114e-06, "loss": 0.6013, "step": 3853 }, { "epoch": 9.422982885085574, "grad_norm": 0.18382479990475614, "learning_rate": 1.5201287736506693e-06, "loss": 0.3147, "step": 3854 }, { "epoch": 9.425427872860636, "grad_norm": 0.1716753842204437, "learning_rate": 1.5189511414108905e-06, "loss": 0.519, "step": 3855 }, { "epoch": 9.427872860635697, "grad_norm": 0.27258692756150965, "learning_rate": 1.5177737664245789e-06, "loss": 0.3697, "step": 3856 }, { "epoch": 9.430317848410757, "grad_norm": 0.38881125335082445, "learning_rate": 1.5165966490004714e-06, "loss": 0.5634, "step": 3857 }, { "epoch": 9.43276283618582, "grad_norm": 0.19525853814771407, "learning_rate": 1.5154197894472334e-06, "loss": 0.4505, "step": 3858 }, { "epoch": 9.43520782396088, "grad_norm": 0.2580931913442357, "learning_rate": 1.5142431880734653e-06, "loss": 0.3955, "step": 3859 }, { "epoch": 9.43765281173594, "grad_norm": 0.37381110033515363, "learning_rate": 1.513066845187699e-06, "loss": 0.5465, "step": 3860 }, { "epoch": 9.440097799511003, "grad_norm": 0.2513291669597155, "learning_rate": 1.511890761098398e-06, "loss": 0.6013, "step": 3861 }, { "epoch": 9.442542787286063, "grad_norm": 0.2966032921538477, "learning_rate": 1.5107149361139583e-06, "loss": 0.4522, "step": 3862 }, { "epoch": 9.444987775061124, "grad_norm": 0.26345497385855476, "learning_rate": 1.5095393705427086e-06, "loss": 0.4001, "step": 3863 }, { "epoch": 9.447432762836186, "grad_norm": 0.20996728950229512, "learning_rate": 1.508364064692909e-06, "loss": 0.7981, "step": 3864 }, { "epoch": 9.449877750611247, "grad_norm": 0.2173885801355247, "learning_rate": 1.5071890188727517e-06, "loss": 0.523, "step": 3865 }, { "epoch": 9.452322738386307, "grad_norm": 0.28345858612891267, "learning_rate": 1.50601423339036e-06, "loss": 0.3318, "step": 3866 }, { "epoch": 9.45476772616137, "grad_norm": 0.34560425686590196, "learning_rate": 1.5048397085537897e-06, "loss": 0.4381, "step": 3867 }, { "epoch": 9.45721271393643, "grad_norm": 0.19511512400930675, "learning_rate": 1.5036654446710285e-06, "loss": 0.5559, "step": 3868 }, { "epoch": 9.45965770171149, "grad_norm": 0.28881050558040083, "learning_rate": 1.5024914420499958e-06, "loss": 0.6507, "step": 3869 }, { "epoch": 9.462102689486553, "grad_norm": 0.22786352830792714, "learning_rate": 1.5013177009985413e-06, "loss": 0.398, "step": 3870 }, { "epoch": 9.464547677261614, "grad_norm": 0.2496813565935898, "learning_rate": 1.5001442218244467e-06, "loss": 0.4082, "step": 3871 }, { "epoch": 9.466992665036674, "grad_norm": 0.15235222252555736, "learning_rate": 1.4989710048354248e-06, "loss": 0.4066, "step": 3872 }, { "epoch": 9.469437652811736, "grad_norm": 0.15865061826644536, "learning_rate": 1.4977980503391207e-06, "loss": 0.3741, "step": 3873 }, { "epoch": 9.471882640586797, "grad_norm": 0.24645026010520765, "learning_rate": 1.4966253586431096e-06, "loss": 0.7443, "step": 3874 }, { "epoch": 9.474327628361857, "grad_norm": 0.26070164880621505, "learning_rate": 1.4954529300548982e-06, "loss": 0.6742, "step": 3875 }, { "epoch": 9.47677261613692, "grad_norm": 0.398964430535094, "learning_rate": 1.4942807648819234e-06, "loss": 0.4894, "step": 3876 }, { "epoch": 9.47921760391198, "grad_norm": 0.21327871634526044, "learning_rate": 1.4931088634315548e-06, "loss": 0.7871, "step": 3877 }, { "epoch": 9.48166259168704, "grad_norm": 0.2598754737216216, "learning_rate": 1.4919372260110906e-06, "loss": 0.42, "step": 3878 }, { "epoch": 9.484107579462103, "grad_norm": 0.26201329993546274, "learning_rate": 1.4907658529277624e-06, "loss": 0.3751, "step": 3879 }, { "epoch": 9.486552567237164, "grad_norm": 0.20246205288724684, "learning_rate": 1.4895947444887305e-06, "loss": 0.3743, "step": 3880 }, { "epoch": 9.488997555012224, "grad_norm": 0.1893651462240564, "learning_rate": 1.488423901001086e-06, "loss": 0.3432, "step": 3881 }, { "epoch": 9.491442542787286, "grad_norm": 0.22569244998743596, "learning_rate": 1.4872533227718505e-06, "loss": 0.4795, "step": 3882 }, { "epoch": 9.493887530562347, "grad_norm": 0.36169043847349547, "learning_rate": 1.4860830101079777e-06, "loss": 0.6255, "step": 3883 }, { "epoch": 9.496332518337407, "grad_norm": 0.32117371882228074, "learning_rate": 1.4849129633163495e-06, "loss": 0.4783, "step": 3884 }, { "epoch": 9.49877750611247, "grad_norm": 0.2728824745205743, "learning_rate": 1.4837431827037787e-06, "loss": 0.8893, "step": 3885 }, { "epoch": 9.50122249388753, "grad_norm": 0.18820591643508766, "learning_rate": 1.4825736685770088e-06, "loss": 0.8901, "step": 3886 }, { "epoch": 9.503667481662593, "grad_norm": 0.32183934404663295, "learning_rate": 1.4814044212427137e-06, "loss": 0.4946, "step": 3887 }, { "epoch": 9.506112469437653, "grad_norm": 0.29151050706296716, "learning_rate": 1.4802354410074962e-06, "loss": 0.4417, "step": 3888 }, { "epoch": 9.508557457212714, "grad_norm": 0.23114284631504753, "learning_rate": 1.479066728177889e-06, "loss": 0.4494, "step": 3889 }, { "epoch": 9.511002444987774, "grad_norm": 0.21831575849741888, "learning_rate": 1.4778982830603571e-06, "loss": 0.859, "step": 3890 }, { "epoch": 9.513447432762836, "grad_norm": 0.20068416926450727, "learning_rate": 1.4767301059612928e-06, "loss": 0.5435, "step": 3891 }, { "epoch": 9.515892420537897, "grad_norm": 0.28036754127399266, "learning_rate": 1.4755621971870182e-06, "loss": 0.8721, "step": 3892 }, { "epoch": 9.51833740831296, "grad_norm": 0.388235189390517, "learning_rate": 1.474394557043787e-06, "loss": 0.3815, "step": 3893 }, { "epoch": 9.52078239608802, "grad_norm": 0.26923631005010584, "learning_rate": 1.4732271858377804e-06, "loss": 0.3218, "step": 3894 }, { "epoch": 9.52322738386308, "grad_norm": 0.24796812926152387, "learning_rate": 1.4720600838751098e-06, "loss": 0.8166, "step": 3895 }, { "epoch": 9.525672371638143, "grad_norm": 0.2741246544608282, "learning_rate": 1.4708932514618156e-06, "loss": 0.7367, "step": 3896 }, { "epoch": 9.528117359413203, "grad_norm": 0.32784599263797454, "learning_rate": 1.4697266889038692e-06, "loss": 0.7006, "step": 3897 }, { "epoch": 9.530562347188264, "grad_norm": 0.2348504195082149, "learning_rate": 1.4685603965071694e-06, "loss": 0.4815, "step": 3898 }, { "epoch": 9.533007334963326, "grad_norm": 0.4112360504452804, "learning_rate": 1.4673943745775431e-06, "loss": 0.4292, "step": 3899 }, { "epoch": 9.535452322738386, "grad_norm": 0.2882630232102204, "learning_rate": 1.4662286234207511e-06, "loss": 0.4322, "step": 3900 }, { "epoch": 9.537897310513447, "grad_norm": 0.20217662879806014, "learning_rate": 1.4650631433424783e-06, "loss": 0.4758, "step": 3901 }, { "epoch": 9.54034229828851, "grad_norm": 0.2501333131137029, "learning_rate": 1.46389793464834e-06, "loss": 0.4085, "step": 3902 }, { "epoch": 9.54278728606357, "grad_norm": 0.2618879827763672, "learning_rate": 1.4627329976438813e-06, "loss": 0.425, "step": 3903 }, { "epoch": 9.54523227383863, "grad_norm": 0.3232067974029118, "learning_rate": 1.4615683326345755e-06, "loss": 0.3979, "step": 3904 }, { "epoch": 9.547677261613693, "grad_norm": 0.2781067773006322, "learning_rate": 1.460403939925824e-06, "loss": 0.6261, "step": 3905 }, { "epoch": 9.550122249388753, "grad_norm": 0.2536830901072346, "learning_rate": 1.4592398198229567e-06, "loss": 1.1174, "step": 3906 }, { "epoch": 9.552567237163814, "grad_norm": 0.2828240899196044, "learning_rate": 1.4580759726312332e-06, "loss": 1.2102, "step": 3907 }, { "epoch": 9.555012224938876, "grad_norm": 0.2836336620117512, "learning_rate": 1.4569123986558403e-06, "loss": 0.5564, "step": 3908 }, { "epoch": 9.557457212713937, "grad_norm": 0.2186569259068507, "learning_rate": 1.455749098201895e-06, "loss": 0.432, "step": 3909 }, { "epoch": 9.559902200488997, "grad_norm": 0.27671182214937934, "learning_rate": 1.454586071574439e-06, "loss": 0.4719, "step": 3910 }, { "epoch": 9.56234718826406, "grad_norm": 0.29097522782909824, "learning_rate": 1.453423319078448e-06, "loss": 0.959, "step": 3911 }, { "epoch": 9.56479217603912, "grad_norm": 0.2516092262333198, "learning_rate": 1.4522608410188187e-06, "loss": 1.0286, "step": 3912 }, { "epoch": 9.56723716381418, "grad_norm": 0.3167854996866788, "learning_rate": 1.4510986377003827e-06, "loss": 0.4634, "step": 3913 }, { "epoch": 9.569682151589243, "grad_norm": 0.2956333783163217, "learning_rate": 1.4499367094278933e-06, "loss": 1.2497, "step": 3914 }, { "epoch": 9.572127139364303, "grad_norm": 0.12452372093432812, "learning_rate": 1.4487750565060363e-06, "loss": 0.3454, "step": 3915 }, { "epoch": 9.574572127139364, "grad_norm": 0.25758294161491413, "learning_rate": 1.4476136792394243e-06, "loss": 0.9257, "step": 3916 }, { "epoch": 9.577017114914426, "grad_norm": 0.34424993896770056, "learning_rate": 1.4464525779325952e-06, "loss": 0.464, "step": 3917 }, { "epoch": 9.579462102689487, "grad_norm": 0.19024653218826845, "learning_rate": 1.4452917528900174e-06, "loss": 0.3934, "step": 3918 }, { "epoch": 9.581907090464547, "grad_norm": 0.3765922551182458, "learning_rate": 1.4441312044160865e-06, "loss": 1.132, "step": 3919 }, { "epoch": 9.58435207823961, "grad_norm": 0.26661955382907154, "learning_rate": 1.442970932815123e-06, "loss": 0.8372, "step": 3920 }, { "epoch": 9.58679706601467, "grad_norm": 0.2263815769966482, "learning_rate": 1.4418109383913781e-06, "loss": 0.7024, "step": 3921 }, { "epoch": 9.58924205378973, "grad_norm": 0.2409953320817237, "learning_rate": 1.4406512214490287e-06, "loss": 0.7336, "step": 3922 }, { "epoch": 9.591687041564793, "grad_norm": 0.1855472481546824, "learning_rate": 1.4394917822921796e-06, "loss": 0.5094, "step": 3923 }, { "epoch": 9.594132029339853, "grad_norm": 0.2428079246696036, "learning_rate": 1.4383326212248607e-06, "loss": 0.4238, "step": 3924 }, { "epoch": 9.596577017114914, "grad_norm": 0.26581152342214764, "learning_rate": 1.4371737385510316e-06, "loss": 0.4901, "step": 3925 }, { "epoch": 9.599022004889976, "grad_norm": 0.17244434653124355, "learning_rate": 1.4360151345745788e-06, "loss": 0.6601, "step": 3926 }, { "epoch": 9.601466992665037, "grad_norm": 0.2793973728887399, "learning_rate": 1.4348568095993128e-06, "loss": 0.5162, "step": 3927 }, { "epoch": 9.603911980440097, "grad_norm": 0.27592847910560553, "learning_rate": 1.433698763928974e-06, "loss": 0.4485, "step": 3928 }, { "epoch": 9.60635696821516, "grad_norm": 0.3052432638800618, "learning_rate": 1.4325409978672292e-06, "loss": 0.5086, "step": 3929 }, { "epoch": 9.60880195599022, "grad_norm": 0.2136516872521334, "learning_rate": 1.4313835117176694e-06, "loss": 0.8255, "step": 3930 }, { "epoch": 9.61124694376528, "grad_norm": 0.2528188484313757, "learning_rate": 1.430226305783815e-06, "loss": 0.5319, "step": 3931 }, { "epoch": 9.613691931540343, "grad_norm": 0.16787268151409984, "learning_rate": 1.4290693803691119e-06, "loss": 0.3191, "step": 3932 }, { "epoch": 9.616136919315403, "grad_norm": 0.2627155902395158, "learning_rate": 1.427912735776933e-06, "loss": 0.467, "step": 3933 }, { "epoch": 9.618581907090464, "grad_norm": 0.2182565257468506, "learning_rate": 1.4267563723105754e-06, "loss": 0.4459, "step": 3934 }, { "epoch": 9.621026894865526, "grad_norm": 0.24942486480444884, "learning_rate": 1.4256002902732653e-06, "loss": 0.5695, "step": 3935 }, { "epoch": 9.623471882640587, "grad_norm": 0.34547539878290207, "learning_rate": 1.4244444899681542e-06, "loss": 0.6361, "step": 3936 }, { "epoch": 9.625916870415647, "grad_norm": 0.2265519687684426, "learning_rate": 1.4232889716983178e-06, "loss": 0.4542, "step": 3937 }, { "epoch": 9.62836185819071, "grad_norm": 0.21419583897858963, "learning_rate": 1.4221337357667609e-06, "loss": 0.4851, "step": 3938 }, { "epoch": 9.63080684596577, "grad_norm": 0.20182049869359583, "learning_rate": 1.4209787824764126e-06, "loss": 0.6504, "step": 3939 }, { "epoch": 9.63325183374083, "grad_norm": 0.26157328597494645, "learning_rate": 1.4198241121301273e-06, "loss": 0.5001, "step": 3940 }, { "epoch": 9.635696821515893, "grad_norm": 0.31257514794025604, "learning_rate": 1.4186697250306874e-06, "loss": 0.6232, "step": 3941 }, { "epoch": 9.638141809290953, "grad_norm": 0.17477144754131954, "learning_rate": 1.4175156214807967e-06, "loss": 0.3635, "step": 3942 }, { "epoch": 9.640586797066014, "grad_norm": 0.26209661829220765, "learning_rate": 1.4163618017830916e-06, "loss": 1.2046, "step": 3943 }, { "epoch": 9.643031784841076, "grad_norm": 0.22135618638862203, "learning_rate": 1.415208266240127e-06, "loss": 0.5394, "step": 3944 }, { "epoch": 9.645476772616137, "grad_norm": 0.18221718699621778, "learning_rate": 1.4140550151543876e-06, "loss": 0.6964, "step": 3945 }, { "epoch": 9.647921760391197, "grad_norm": 0.30832798801944217, "learning_rate": 1.4129020488282825e-06, "loss": 0.3581, "step": 3946 }, { "epoch": 9.65036674816626, "grad_norm": 0.21838010494159443, "learning_rate": 1.4117493675641447e-06, "loss": 0.4118, "step": 3947 }, { "epoch": 9.65281173594132, "grad_norm": 0.17583045315217985, "learning_rate": 1.4105969716642344e-06, "loss": 0.2516, "step": 3948 }, { "epoch": 9.65525672371638, "grad_norm": 0.2810287788402748, "learning_rate": 1.4094448614307366e-06, "loss": 0.6859, "step": 3949 }, { "epoch": 9.657701711491443, "grad_norm": 0.32960458590376607, "learning_rate": 1.4082930371657597e-06, "loss": 0.6625, "step": 3950 }, { "epoch": 9.660146699266503, "grad_norm": 0.3286936825929337, "learning_rate": 1.40714149917134e-06, "loss": 0.5802, "step": 3951 }, { "epoch": 9.662591687041564, "grad_norm": 0.18696778489621488, "learning_rate": 1.4059902477494353e-06, "loss": 0.3198, "step": 3952 }, { "epoch": 9.665036674816626, "grad_norm": 0.27252071291482494, "learning_rate": 1.4048392832019311e-06, "loss": 0.9785, "step": 3953 }, { "epoch": 9.667481662591687, "grad_norm": 0.2170101571231686, "learning_rate": 1.4036886058306366e-06, "loss": 0.5676, "step": 3954 }, { "epoch": 9.669926650366747, "grad_norm": 0.24876790622604092, "learning_rate": 1.4025382159372857e-06, "loss": 0.6646, "step": 3955 }, { "epoch": 9.67237163814181, "grad_norm": 0.30067965876125013, "learning_rate": 1.401388113823538e-06, "loss": 0.3403, "step": 3956 }, { "epoch": 9.67481662591687, "grad_norm": 0.27085007712143144, "learning_rate": 1.4002382997909745e-06, "loss": 0.378, "step": 3957 }, { "epoch": 9.67726161369193, "grad_norm": 0.3428239285869144, "learning_rate": 1.3990887741411036e-06, "loss": 0.7776, "step": 3958 }, { "epoch": 9.679706601466993, "grad_norm": 0.34318448564404536, "learning_rate": 1.397939537175359e-06, "loss": 0.4245, "step": 3959 }, { "epoch": 9.682151589242054, "grad_norm": 0.18813678324033897, "learning_rate": 1.3967905891950939e-06, "loss": 0.4382, "step": 3960 }, { "epoch": 9.684596577017114, "grad_norm": 0.2481956458728824, "learning_rate": 1.3956419305015916e-06, "loss": 0.541, "step": 3961 }, { "epoch": 9.687041564792176, "grad_norm": 0.2759394854994073, "learning_rate": 1.3944935613960544e-06, "loss": 0.6456, "step": 3962 }, { "epoch": 9.689486552567237, "grad_norm": 0.23648839449579018, "learning_rate": 1.393345482179612e-06, "loss": 0.8678, "step": 3963 }, { "epoch": 9.6919315403423, "grad_norm": 0.2775187916056447, "learning_rate": 1.3921976931533168e-06, "loss": 0.4647, "step": 3964 }, { "epoch": 9.69437652811736, "grad_norm": 0.212561005425861, "learning_rate": 1.3910501946181457e-06, "loss": 0.6494, "step": 3965 }, { "epoch": 9.69682151589242, "grad_norm": 0.2393154283233034, "learning_rate": 1.3899029868749997e-06, "loss": 0.5593, "step": 3966 }, { "epoch": 9.69926650366748, "grad_norm": 0.21942466467555843, "learning_rate": 1.3887560702247016e-06, "loss": 0.6501, "step": 3967 }, { "epoch": 9.701711491442543, "grad_norm": 0.1543197263424504, "learning_rate": 1.3876094449679994e-06, "loss": 0.4184, "step": 3968 }, { "epoch": 9.704156479217604, "grad_norm": 0.23833409175962472, "learning_rate": 1.386463111405566e-06, "loss": 0.4069, "step": 3969 }, { "epoch": 9.706601466992666, "grad_norm": 0.2700969070574167, "learning_rate": 1.3853170698379941e-06, "loss": 0.4828, "step": 3970 }, { "epoch": 9.709046454767726, "grad_norm": 0.18209115182032812, "learning_rate": 1.3841713205658044e-06, "loss": 0.4128, "step": 3971 }, { "epoch": 9.711491442542787, "grad_norm": 0.2657672354640742, "learning_rate": 1.383025863889436e-06, "loss": 0.4539, "step": 3972 }, { "epoch": 9.713936430317847, "grad_norm": 0.18597257338833875, "learning_rate": 1.3818807001092555e-06, "loss": 0.4192, "step": 3973 }, { "epoch": 9.71638141809291, "grad_norm": 0.2849990895018088, "learning_rate": 1.3807358295255505e-06, "loss": 0.5531, "step": 3974 }, { "epoch": 9.71882640586797, "grad_norm": 0.2018641079331819, "learning_rate": 1.3795912524385324e-06, "loss": 0.7408, "step": 3975 }, { "epoch": 9.721271393643033, "grad_norm": 0.18256542688006575, "learning_rate": 1.3784469691483365e-06, "loss": 0.5101, "step": 3976 }, { "epoch": 9.723716381418093, "grad_norm": 0.4149618047659524, "learning_rate": 1.3773029799550187e-06, "loss": 0.3591, "step": 3977 }, { "epoch": 9.726161369193154, "grad_norm": 0.35049615587324584, "learning_rate": 1.376159285158559e-06, "loss": 0.2927, "step": 3978 }, { "epoch": 9.728606356968216, "grad_norm": 0.3019717508892536, "learning_rate": 1.3750158850588621e-06, "loss": 0.6626, "step": 3979 }, { "epoch": 9.731051344743276, "grad_norm": 0.3290939071048803, "learning_rate": 1.3738727799557517e-06, "loss": 0.665, "step": 3980 }, { "epoch": 9.733496332518337, "grad_norm": 0.2797276764815221, "learning_rate": 1.3727299701489781e-06, "loss": 0.971, "step": 3981 }, { "epoch": 9.7359413202934, "grad_norm": 0.19758045797299298, "learning_rate": 1.3715874559382097e-06, "loss": 0.4284, "step": 3982 }, { "epoch": 9.73838630806846, "grad_norm": 0.3391785181832523, "learning_rate": 1.3704452376230418e-06, "loss": 0.4663, "step": 3983 }, { "epoch": 9.74083129584352, "grad_norm": 0.28474792783621183, "learning_rate": 1.3693033155029902e-06, "loss": 0.3899, "step": 3984 }, { "epoch": 9.743276283618583, "grad_norm": 0.2151835339931385, "learning_rate": 1.3681616898774907e-06, "loss": 0.4189, "step": 3985 }, { "epoch": 9.745721271393643, "grad_norm": 0.20035558222650202, "learning_rate": 1.3670203610459073e-06, "loss": 0.4837, "step": 3986 }, { "epoch": 9.748166259168704, "grad_norm": 0.3322480947920425, "learning_rate": 1.3658793293075198e-06, "loss": 1.0254, "step": 3987 }, { "epoch": 9.750611246943766, "grad_norm": 0.31196751197145167, "learning_rate": 1.3647385949615339e-06, "loss": 0.4131, "step": 3988 }, { "epoch": 9.753056234718827, "grad_norm": 0.34403764525947955, "learning_rate": 1.3635981583070768e-06, "loss": 0.533, "step": 3989 }, { "epoch": 9.755501222493887, "grad_norm": 0.24855990948278087, "learning_rate": 1.3624580196431953e-06, "loss": 0.6523, "step": 3990 }, { "epoch": 9.75794621026895, "grad_norm": 0.2448847344317707, "learning_rate": 1.3613181792688624e-06, "loss": 0.8425, "step": 3991 }, { "epoch": 9.76039119804401, "grad_norm": 0.3983280525019243, "learning_rate": 1.3601786374829672e-06, "loss": 0.7042, "step": 3992 }, { "epoch": 9.76283618581907, "grad_norm": 0.22289976454558866, "learning_rate": 1.3590393945843259e-06, "loss": 0.7046, "step": 3993 }, { "epoch": 9.765281173594133, "grad_norm": 0.2806807111791484, "learning_rate": 1.3579004508716743e-06, "loss": 1.2044, "step": 3994 }, { "epoch": 9.767726161369193, "grad_norm": 0.356165075161078, "learning_rate": 1.3567618066436678e-06, "loss": 0.3877, "step": 3995 }, { "epoch": 9.770171149144254, "grad_norm": 0.2732755945353237, "learning_rate": 1.3556234621988856e-06, "loss": 0.8471, "step": 3996 }, { "epoch": 9.772616136919316, "grad_norm": 0.3104435388910349, "learning_rate": 1.354485417835828e-06, "loss": 0.5921, "step": 3997 }, { "epoch": 9.775061124694377, "grad_norm": 0.33305744081239336, "learning_rate": 1.3533476738529167e-06, "loss": 0.4819, "step": 3998 }, { "epoch": 9.777506112469437, "grad_norm": 0.2609144342432876, "learning_rate": 1.352210230548494e-06, "loss": 0.7856, "step": 3999 }, { "epoch": 9.7799511002445, "grad_norm": 0.2346505471202344, "learning_rate": 1.3510730882208228e-06, "loss": 0.3456, "step": 4000 }, { "epoch": 9.78239608801956, "grad_norm": 0.3187090038197892, "learning_rate": 1.3499362471680893e-06, "loss": 0.4256, "step": 4001 }, { "epoch": 9.78484107579462, "grad_norm": 0.4188158182956479, "learning_rate": 1.3487997076883974e-06, "loss": 0.4658, "step": 4002 }, { "epoch": 9.787286063569683, "grad_norm": 0.30772753842958833, "learning_rate": 1.3476634700797747e-06, "loss": 0.9264, "step": 4003 }, { "epoch": 9.789731051344743, "grad_norm": 0.29770352658904536, "learning_rate": 1.3465275346401696e-06, "loss": 0.434, "step": 4004 }, { "epoch": 9.792176039119804, "grad_norm": 0.2148016282006695, "learning_rate": 1.3453919016674486e-06, "loss": 0.4837, "step": 4005 }, { "epoch": 9.794621026894866, "grad_norm": 0.2511468203283835, "learning_rate": 1.3442565714594013e-06, "loss": 0.8167, "step": 4006 }, { "epoch": 9.797066014669927, "grad_norm": 0.18083728849801017, "learning_rate": 1.343121544313738e-06, "loss": 0.4708, "step": 4007 }, { "epoch": 9.799511002444987, "grad_norm": 0.4324889402612805, "learning_rate": 1.3419868205280878e-06, "loss": 0.4683, "step": 4008 }, { "epoch": 9.80195599022005, "grad_norm": 0.21189145286353303, "learning_rate": 1.340852400400003e-06, "loss": 0.4278, "step": 4009 }, { "epoch": 9.80440097799511, "grad_norm": 0.24058925943763806, "learning_rate": 1.339718284226952e-06, "loss": 0.3901, "step": 4010 }, { "epoch": 9.80684596577017, "grad_norm": 0.34018629658408606, "learning_rate": 1.3385844723063284e-06, "loss": 0.3652, "step": 4011 }, { "epoch": 9.809290953545233, "grad_norm": 0.20768885335650156, "learning_rate": 1.337450964935442e-06, "loss": 0.4695, "step": 4012 }, { "epoch": 9.811735941320293, "grad_norm": 0.27313987056560646, "learning_rate": 1.3363177624115245e-06, "loss": 0.4484, "step": 4013 }, { "epoch": 9.814180929095354, "grad_norm": 0.22675647169767135, "learning_rate": 1.335184865031729e-06, "loss": 0.5852, "step": 4014 }, { "epoch": 9.816625916870416, "grad_norm": 0.23221625648773955, "learning_rate": 1.3340522730931254e-06, "loss": 0.5577, "step": 4015 }, { "epoch": 9.819070904645477, "grad_norm": 0.19306287736802782, "learning_rate": 1.3329199868927057e-06, "loss": 0.5059, "step": 4016 }, { "epoch": 9.821515892420537, "grad_norm": 0.24331144070605967, "learning_rate": 1.3317880067273813e-06, "loss": 0.4951, "step": 4017 }, { "epoch": 9.8239608801956, "grad_norm": 0.3584028422872811, "learning_rate": 1.330656332893984e-06, "loss": 0.8374, "step": 4018 }, { "epoch": 9.82640586797066, "grad_norm": 0.24255779782869188, "learning_rate": 1.3295249656892648e-06, "loss": 1.1743, "step": 4019 }, { "epoch": 9.82885085574572, "grad_norm": 0.28504017675566035, "learning_rate": 1.328393905409892e-06, "loss": 0.7278, "step": 4020 }, { "epoch": 9.831295843520783, "grad_norm": 0.23545291624481923, "learning_rate": 1.3272631523524588e-06, "loss": 0.5571, "step": 4021 }, { "epoch": 9.833740831295843, "grad_norm": 0.27442221546789286, "learning_rate": 1.3261327068134707e-06, "loss": 0.4213, "step": 4022 }, { "epoch": 9.836185819070904, "grad_norm": 0.26127903365357846, "learning_rate": 1.3250025690893587e-06, "loss": 1.0642, "step": 4023 }, { "epoch": 9.838630806845966, "grad_norm": 0.27470608252322676, "learning_rate": 1.3238727394764713e-06, "loss": 0.6067, "step": 4024 }, { "epoch": 9.841075794621027, "grad_norm": 0.2573437619683948, "learning_rate": 1.322743218271074e-06, "loss": 0.4248, "step": 4025 }, { "epoch": 9.843520782396087, "grad_norm": 0.21265396359805472, "learning_rate": 1.3216140057693535e-06, "loss": 0.5927, "step": 4026 }, { "epoch": 9.84596577017115, "grad_norm": 0.27504638164778344, "learning_rate": 1.320485102267417e-06, "loss": 0.4278, "step": 4027 }, { "epoch": 9.84841075794621, "grad_norm": 0.24169818341469193, "learning_rate": 1.319356508061285e-06, "loss": 0.5972, "step": 4028 }, { "epoch": 9.85085574572127, "grad_norm": 0.20905556195545266, "learning_rate": 1.318228223446905e-06, "loss": 0.4348, "step": 4029 }, { "epoch": 9.853300733496333, "grad_norm": 0.18281378945650573, "learning_rate": 1.3171002487201367e-06, "loss": 0.4958, "step": 4030 }, { "epoch": 9.855745721271393, "grad_norm": 0.3652458517947851, "learning_rate": 1.315972584176762e-06, "loss": 1.224, "step": 4031 }, { "epoch": 9.858190709046454, "grad_norm": 0.2405041016435086, "learning_rate": 1.314845230112479e-06, "loss": 0.4903, "step": 4032 }, { "epoch": 9.860635696821516, "grad_norm": 0.30627644931498266, "learning_rate": 1.3137181868229062e-06, "loss": 0.5234, "step": 4033 }, { "epoch": 9.863080684596577, "grad_norm": 0.3803625027981747, "learning_rate": 1.3125914546035818e-06, "loss": 0.4284, "step": 4034 }, { "epoch": 9.865525672371637, "grad_norm": 0.24174079896188697, "learning_rate": 1.3114650337499578e-06, "loss": 0.6802, "step": 4035 }, { "epoch": 9.8679706601467, "grad_norm": 0.24785162980587833, "learning_rate": 1.3103389245574099e-06, "loss": 0.4202, "step": 4036 }, { "epoch": 9.87041564792176, "grad_norm": 0.2300014848512258, "learning_rate": 1.3092131273212294e-06, "loss": 0.4173, "step": 4037 }, { "epoch": 9.87286063569682, "grad_norm": 0.31246322713184427, "learning_rate": 1.3080876423366247e-06, "loss": 0.3977, "step": 4038 }, { "epoch": 9.875305623471883, "grad_norm": 0.24769167867706052, "learning_rate": 1.3069624698987249e-06, "loss": 0.8479, "step": 4039 }, { "epoch": 9.877750611246944, "grad_norm": 0.22506985121027237, "learning_rate": 1.3058376103025757e-06, "loss": 0.699, "step": 4040 }, { "epoch": 9.880195599022004, "grad_norm": 0.30200295002971855, "learning_rate": 1.3047130638431421e-06, "loss": 0.512, "step": 4041 }, { "epoch": 9.882640586797066, "grad_norm": 0.2136967120677066, "learning_rate": 1.3035888308153038e-06, "loss": 0.4234, "step": 4042 }, { "epoch": 9.885085574572127, "grad_norm": 0.22256333342879878, "learning_rate": 1.3024649115138615e-06, "loss": 0.4512, "step": 4043 }, { "epoch": 9.887530562347187, "grad_norm": 0.36412595704100104, "learning_rate": 1.301341306233534e-06, "loss": 0.7507, "step": 4044 }, { "epoch": 9.88997555012225, "grad_norm": 0.283661783817512, "learning_rate": 1.3002180152689536e-06, "loss": 0.3735, "step": 4045 }, { "epoch": 9.89242053789731, "grad_norm": 0.16748915632260827, "learning_rate": 1.2990950389146742e-06, "loss": 0.4292, "step": 4046 }, { "epoch": 9.894865525672373, "grad_norm": 0.2947914039367817, "learning_rate": 1.297972377465167e-06, "loss": 0.472, "step": 4047 }, { "epoch": 9.897310513447433, "grad_norm": 0.21103275150947443, "learning_rate": 1.2968500312148174e-06, "loss": 0.8662, "step": 4048 }, { "epoch": 9.899755501222494, "grad_norm": 0.31252306666777663, "learning_rate": 1.2957280004579317e-06, "loss": 0.7353, "step": 4049 }, { "epoch": 9.902200488997554, "grad_norm": 0.2923266986094965, "learning_rate": 1.2946062854887315e-06, "loss": 0.666, "step": 4050 }, { "epoch": 9.904645476772616, "grad_norm": 0.27575854504308867, "learning_rate": 1.2934848866013574e-06, "loss": 0.5038, "step": 4051 }, { "epoch": 9.907090464547677, "grad_norm": 0.26167143181598956, "learning_rate": 1.292363804089864e-06, "loss": 0.5366, "step": 4052 }, { "epoch": 9.90953545232274, "grad_norm": 0.21611890609422782, "learning_rate": 1.2912430382482252e-06, "loss": 0.6822, "step": 4053 }, { "epoch": 9.9119804400978, "grad_norm": 0.22680242164140751, "learning_rate": 1.2901225893703328e-06, "loss": 0.5049, "step": 4054 }, { "epoch": 9.91442542787286, "grad_norm": 0.25993264570727953, "learning_rate": 1.2890024577499926e-06, "loss": 0.617, "step": 4055 }, { "epoch": 9.91687041564792, "grad_norm": 0.16596350901869003, "learning_rate": 1.2878826436809294e-06, "loss": 0.5196, "step": 4056 }, { "epoch": 9.919315403422983, "grad_norm": 0.3077170796631459, "learning_rate": 1.286763147456785e-06, "loss": 0.6832, "step": 4057 }, { "epoch": 9.921760391198044, "grad_norm": 0.289863672865671, "learning_rate": 1.2856439693711147e-06, "loss": 1.0394, "step": 4058 }, { "epoch": 9.924205378973106, "grad_norm": 0.22318121335478194, "learning_rate": 1.284525109717395e-06, "loss": 0.3355, "step": 4059 }, { "epoch": 9.926650366748166, "grad_norm": 0.27983770280342873, "learning_rate": 1.2834065687890137e-06, "loss": 0.6487, "step": 4060 }, { "epoch": 9.929095354523227, "grad_norm": 0.2516985367876321, "learning_rate": 1.2822883468792813e-06, "loss": 0.716, "step": 4061 }, { "epoch": 9.93154034229829, "grad_norm": 0.16509823194721587, "learning_rate": 1.2811704442814185e-06, "loss": 0.3455, "step": 4062 }, { "epoch": 9.93398533007335, "grad_norm": 0.19651460729146775, "learning_rate": 1.280052861288566e-06, "loss": 0.5983, "step": 4063 }, { "epoch": 9.93643031784841, "grad_norm": 0.19802406598885053, "learning_rate": 1.27893559819378e-06, "loss": 0.5085, "step": 4064 }, { "epoch": 9.938875305623473, "grad_norm": 0.2714731735601288, "learning_rate": 1.2778186552900318e-06, "loss": 0.381, "step": 4065 }, { "epoch": 9.941320293398533, "grad_norm": 0.2653730714754877, "learning_rate": 1.276702032870209e-06, "loss": 1.4207, "step": 4066 }, { "epoch": 9.943765281173594, "grad_norm": 0.20978572867131567, "learning_rate": 1.2755857312271174e-06, "loss": 0.4533, "step": 4067 }, { "epoch": 9.946210268948656, "grad_norm": 0.35499504720385117, "learning_rate": 1.2744697506534745e-06, "loss": 0.4246, "step": 4068 }, { "epoch": 9.948655256723717, "grad_norm": 0.217983903720869, "learning_rate": 1.2733540914419178e-06, "loss": 0.4871, "step": 4069 }, { "epoch": 9.951100244498777, "grad_norm": 0.3619942625932222, "learning_rate": 1.2722387538849973e-06, "loss": 0.5489, "step": 4070 }, { "epoch": 9.95354523227384, "grad_norm": 0.26996248526651395, "learning_rate": 1.2711237382751806e-06, "loss": 0.8182, "step": 4071 }, { "epoch": 9.9559902200489, "grad_norm": 0.2683802099011985, "learning_rate": 1.2700090449048503e-06, "loss": 0.5322, "step": 4072 }, { "epoch": 9.95843520782396, "grad_norm": 0.2791425885932258, "learning_rate": 1.2688946740663047e-06, "loss": 0.7948, "step": 4073 }, { "epoch": 9.960880195599023, "grad_norm": 0.3167766846215492, "learning_rate": 1.2677806260517581e-06, "loss": 0.4159, "step": 4074 }, { "epoch": 9.963325183374083, "grad_norm": 0.2512942284505814, "learning_rate": 1.2666669011533375e-06, "loss": 0.3536, "step": 4075 }, { "epoch": 9.965770171149144, "grad_norm": 0.1922633846371159, "learning_rate": 1.2655534996630883e-06, "loss": 0.4131, "step": 4076 }, { "epoch": 9.968215158924206, "grad_norm": 0.2829681281489595, "learning_rate": 1.2644404218729705e-06, "loss": 0.6787, "step": 4077 }, { "epoch": 9.970660146699267, "grad_norm": 0.27886341223811234, "learning_rate": 1.2633276680748568e-06, "loss": 0.4792, "step": 4078 }, { "epoch": 9.973105134474327, "grad_norm": 0.2884935247227573, "learning_rate": 1.2622152385605384e-06, "loss": 0.4922, "step": 4079 }, { "epoch": 9.97555012224939, "grad_norm": 0.2810686692717784, "learning_rate": 1.2611031336217183e-06, "loss": 0.491, "step": 4080 }, { "epoch": 9.97799511002445, "grad_norm": 0.17739240444636853, "learning_rate": 1.259991353550017e-06, "loss": 0.4317, "step": 4081 }, { "epoch": 9.98044009779951, "grad_norm": 0.14834440006009075, "learning_rate": 1.2588798986369678e-06, "loss": 0.4045, "step": 4082 }, { "epoch": 9.982885085574573, "grad_norm": 0.38700062986101735, "learning_rate": 1.2577687691740201e-06, "loss": 0.4821, "step": 4083 }, { "epoch": 9.985330073349633, "grad_norm": 0.2838176558247498, "learning_rate": 1.2566579654525384e-06, "loss": 1.0047, "step": 4084 }, { "epoch": 9.987775061124694, "grad_norm": 0.20977476004828025, "learning_rate": 1.2555474877637991e-06, "loss": 1.0607, "step": 4085 }, { "epoch": 9.990220048899756, "grad_norm": 0.2427352226732089, "learning_rate": 1.2544373363989958e-06, "loss": 0.3205, "step": 4086 }, { "epoch": 9.992665036674817, "grad_norm": 0.24564461294695167, "learning_rate": 1.2533275116492367e-06, "loss": 0.6627, "step": 4087 }, { "epoch": 9.995110024449877, "grad_norm": 0.2921670297897772, "learning_rate": 1.2522180138055407e-06, "loss": 0.4503, "step": 4088 }, { "epoch": 9.99755501222494, "grad_norm": 0.2963943919570437, "learning_rate": 1.2511088431588462e-06, "loss": 0.601, "step": 4089 }, { "epoch": 10.0, "grad_norm": 0.3269205605410965, "learning_rate": 1.2500000000000007e-06, "loss": 0.4649, "step": 4090 }, { "epoch": 10.00244498777506, "grad_norm": 0.20843576283697876, "learning_rate": 1.2488914846197694e-06, "loss": 0.4783, "step": 4091 }, { "epoch": 10.004889975550123, "grad_norm": 0.21228257893915267, "learning_rate": 1.2477832973088306e-06, "loss": 0.6509, "step": 4092 }, { "epoch": 10.007334963325183, "grad_norm": 0.2572391703630435, "learning_rate": 1.2466754383577758e-06, "loss": 0.3411, "step": 4093 }, { "epoch": 10.009779951100244, "grad_norm": 0.241545368594539, "learning_rate": 1.2455679080571125e-06, "loss": 0.7343, "step": 4094 }, { "epoch": 10.012224938875306, "grad_norm": 0.35998029760473993, "learning_rate": 1.2444607066972586e-06, "loss": 0.6575, "step": 4095 }, { "epoch": 10.014669926650367, "grad_norm": 0.2044765807201473, "learning_rate": 1.2433538345685483e-06, "loss": 0.4346, "step": 4096 }, { "epoch": 10.017114914425427, "grad_norm": 0.2504161142070067, "learning_rate": 1.24224729196123e-06, "loss": 0.4439, "step": 4097 }, { "epoch": 10.01955990220049, "grad_norm": 0.2903557236014858, "learning_rate": 1.2411410791654625e-06, "loss": 0.4689, "step": 4098 }, { "epoch": 10.02200488997555, "grad_norm": 0.2591738393184338, "learning_rate": 1.2400351964713217e-06, "loss": 0.8255, "step": 4099 }, { "epoch": 10.02444987775061, "grad_norm": 0.2329846759559528, "learning_rate": 1.238929644168794e-06, "loss": 0.9687, "step": 4100 }, { "epoch": 10.026894865525673, "grad_norm": 0.3081185957900225, "learning_rate": 1.2378244225477812e-06, "loss": 1.0416, "step": 4101 }, { "epoch": 10.029339853300733, "grad_norm": 0.286622540419071, "learning_rate": 1.2367195318980988e-06, "loss": 0.6481, "step": 4102 }, { "epoch": 10.031784841075794, "grad_norm": 0.31021730314554313, "learning_rate": 1.235614972509471e-06, "loss": 0.4727, "step": 4103 }, { "epoch": 10.034229828850856, "grad_norm": 0.162757958499169, "learning_rate": 1.2345107446715432e-06, "loss": 0.3342, "step": 4104 }, { "epoch": 10.036674816625917, "grad_norm": 0.27846512859041017, "learning_rate": 1.233406848673866e-06, "loss": 0.4379, "step": 4105 }, { "epoch": 10.039119804400977, "grad_norm": 0.204590491923056, "learning_rate": 1.2323032848059068e-06, "loss": 0.4143, "step": 4106 }, { "epoch": 10.04156479217604, "grad_norm": 0.2873974055529322, "learning_rate": 1.231200053357047e-06, "loss": 0.5169, "step": 4107 }, { "epoch": 10.0440097799511, "grad_norm": 0.20276498352866473, "learning_rate": 1.2300971546165766e-06, "loss": 0.4137, "step": 4108 }, { "epoch": 10.04645476772616, "grad_norm": 0.3146174274331853, "learning_rate": 1.228994588873703e-06, "loss": 0.4003, "step": 4109 }, { "epoch": 10.048899755501223, "grad_norm": 0.23904682396563412, "learning_rate": 1.227892356417542e-06, "loss": 0.5495, "step": 4110 }, { "epoch": 10.051344743276283, "grad_norm": 0.20316142027532735, "learning_rate": 1.2267904575371259e-06, "loss": 0.5436, "step": 4111 }, { "epoch": 10.053789731051344, "grad_norm": 0.2232828036353317, "learning_rate": 1.2256888925213982e-06, "loss": 0.8635, "step": 4112 }, { "epoch": 10.056234718826406, "grad_norm": 0.24173000139408476, "learning_rate": 1.2245876616592126e-06, "loss": 0.5745, "step": 4113 }, { "epoch": 10.058679706601467, "grad_norm": 0.2001999328932452, "learning_rate": 1.223486765239338e-06, "loss": 0.5304, "step": 4114 }, { "epoch": 10.061124694376527, "grad_norm": 0.27570629479538195, "learning_rate": 1.222386203550455e-06, "loss": 0.619, "step": 4115 }, { "epoch": 10.06356968215159, "grad_norm": 0.2941588126495584, "learning_rate": 1.2212859768811553e-06, "loss": 0.7085, "step": 4116 }, { "epoch": 10.06601466992665, "grad_norm": 0.24627626972521532, "learning_rate": 1.2201860855199453e-06, "loss": 0.695, "step": 4117 }, { "epoch": 10.06845965770171, "grad_norm": 0.31838314900410647, "learning_rate": 1.219086529755239e-06, "loss": 0.6182, "step": 4118 }, { "epoch": 10.070904645476773, "grad_norm": 0.31589738946603907, "learning_rate": 1.2179873098753677e-06, "loss": 0.4961, "step": 4119 }, { "epoch": 10.073349633251834, "grad_norm": 0.22257795628191507, "learning_rate": 1.2168884261685698e-06, "loss": 0.429, "step": 4120 }, { "epoch": 10.075794621026894, "grad_norm": 0.28921420264830827, "learning_rate": 1.2157898789229989e-06, "loss": 0.7576, "step": 4121 }, { "epoch": 10.078239608801956, "grad_norm": 0.29053641275724085, "learning_rate": 1.2146916684267196e-06, "loss": 0.6919, "step": 4122 }, { "epoch": 10.080684596577017, "grad_norm": 0.2737294086045129, "learning_rate": 1.2135937949677066e-06, "loss": 0.3998, "step": 4123 }, { "epoch": 10.083129584352077, "grad_norm": 0.3496990866546356, "learning_rate": 1.2124962588338483e-06, "loss": 0.8825, "step": 4124 }, { "epoch": 10.08557457212714, "grad_norm": 0.3072496222576729, "learning_rate": 1.2113990603129432e-06, "loss": 0.7952, "step": 4125 }, { "epoch": 10.0880195599022, "grad_norm": 0.330229488909154, "learning_rate": 1.2103021996927028e-06, "loss": 0.527, "step": 4126 }, { "epoch": 10.09046454767726, "grad_norm": 0.40515907641986176, "learning_rate": 1.2092056772607493e-06, "loss": 0.406, "step": 4127 }, { "epoch": 10.092909535452323, "grad_norm": 0.23914429622564956, "learning_rate": 1.2081094933046142e-06, "loss": 0.7231, "step": 4128 }, { "epoch": 10.095354523227384, "grad_norm": 0.2559082460774146, "learning_rate": 1.2070136481117442e-06, "loss": 0.594, "step": 4129 }, { "epoch": 10.097799511002446, "grad_norm": 0.269734036600957, "learning_rate": 1.2059181419694932e-06, "loss": 0.5171, "step": 4130 }, { "epoch": 10.100244498777506, "grad_norm": 0.32241860877482476, "learning_rate": 1.2048229751651283e-06, "loss": 0.4156, "step": 4131 }, { "epoch": 10.102689486552567, "grad_norm": 0.22515215823769727, "learning_rate": 1.2037281479858288e-06, "loss": 0.5137, "step": 4132 }, { "epoch": 10.10513447432763, "grad_norm": 0.30491056815591205, "learning_rate": 1.2026336607186815e-06, "loss": 0.7891, "step": 4133 }, { "epoch": 10.10757946210269, "grad_norm": 0.28790410039915304, "learning_rate": 1.2015395136506866e-06, "loss": 0.6664, "step": 4134 }, { "epoch": 10.11002444987775, "grad_norm": 0.1582398271639738, "learning_rate": 1.2004457070687548e-06, "loss": 0.3299, "step": 4135 }, { "epoch": 10.112469437652813, "grad_norm": 0.19119891549146686, "learning_rate": 1.1993522412597073e-06, "loss": 0.5576, "step": 4136 }, { "epoch": 10.114914425427873, "grad_norm": 0.30228616505192535, "learning_rate": 1.1982591165102763e-06, "loss": 0.6854, "step": 4137 }, { "epoch": 10.117359413202934, "grad_norm": 0.27561708240492117, "learning_rate": 1.1971663331071032e-06, "loss": 0.5825, "step": 4138 }, { "epoch": 10.119804400977996, "grad_norm": 0.388323690395099, "learning_rate": 1.1960738913367417e-06, "loss": 0.4349, "step": 4139 }, { "epoch": 10.122249388753056, "grad_norm": 0.17641853111428465, "learning_rate": 1.194981791485654e-06, "loss": 0.3887, "step": 4140 }, { "epoch": 10.124694376528117, "grad_norm": 0.26054947973364223, "learning_rate": 1.193890033840214e-06, "loss": 0.6125, "step": 4141 }, { "epoch": 10.12713936430318, "grad_norm": 0.21564783351249864, "learning_rate": 1.1927986186867067e-06, "loss": 0.4714, "step": 4142 }, { "epoch": 10.12958435207824, "grad_norm": 0.3126582058013417, "learning_rate": 1.1917075463113242e-06, "loss": 0.3956, "step": 4143 }, { "epoch": 10.1320293398533, "grad_norm": 0.2884270653175563, "learning_rate": 1.190616817000172e-06, "loss": 0.4076, "step": 4144 }, { "epoch": 10.134474327628363, "grad_norm": 0.3081749580720585, "learning_rate": 1.1895264310392645e-06, "loss": 0.3902, "step": 4145 }, { "epoch": 10.136919315403423, "grad_norm": 0.4556862954832288, "learning_rate": 1.1884363887145237e-06, "loss": 0.752, "step": 4146 }, { "epoch": 10.139364303178484, "grad_norm": 0.19790824809069463, "learning_rate": 1.187346690311787e-06, "loss": 0.5941, "step": 4147 }, { "epoch": 10.141809290953546, "grad_norm": 0.24690976280309146, "learning_rate": 1.1862573361167957e-06, "loss": 0.3208, "step": 4148 }, { "epoch": 10.144254278728607, "grad_norm": 0.3778130988717183, "learning_rate": 1.185168326415205e-06, "loss": 0.3549, "step": 4149 }, { "epoch": 10.146699266503667, "grad_norm": 0.20341049699409436, "learning_rate": 1.1840796614925764e-06, "loss": 0.4514, "step": 4150 }, { "epoch": 10.14914425427873, "grad_norm": 0.21723705533523527, "learning_rate": 1.182991341634384e-06, "loss": 0.644, "step": 4151 }, { "epoch": 10.15158924205379, "grad_norm": 0.24145819203642227, "learning_rate": 1.181903367126011e-06, "loss": 0.674, "step": 4152 }, { "epoch": 10.15403422982885, "grad_norm": 0.2160016856989047, "learning_rate": 1.180815738252747e-06, "loss": 0.4534, "step": 4153 }, { "epoch": 10.156479217603913, "grad_norm": 0.3441669993273657, "learning_rate": 1.1797284552997945e-06, "loss": 0.4201, "step": 4154 }, { "epoch": 10.158924205378973, "grad_norm": 0.2910476570687918, "learning_rate": 1.1786415185522645e-06, "loss": 0.5914, "step": 4155 }, { "epoch": 10.161369193154034, "grad_norm": 0.20140729786102404, "learning_rate": 1.1775549282951755e-06, "loss": 0.4155, "step": 4156 }, { "epoch": 10.163814180929096, "grad_norm": 0.29284486893101835, "learning_rate": 1.1764686848134566e-06, "loss": 0.4007, "step": 4157 }, { "epoch": 10.166259168704157, "grad_norm": 0.27338124082317167, "learning_rate": 1.1753827883919461e-06, "loss": 0.5415, "step": 4158 }, { "epoch": 10.168704156479217, "grad_norm": 0.33634477557508075, "learning_rate": 1.1742972393153918e-06, "loss": 0.8392, "step": 4159 }, { "epoch": 10.17114914425428, "grad_norm": 0.35731306434055854, "learning_rate": 1.1732120378684472e-06, "loss": 0.4808, "step": 4160 }, { "epoch": 10.17359413202934, "grad_norm": 0.21615973895853108, "learning_rate": 1.1721271843356784e-06, "loss": 0.4477, "step": 4161 }, { "epoch": 10.1760391198044, "grad_norm": 0.22801464095729906, "learning_rate": 1.17104267900156e-06, "loss": 0.4508, "step": 4162 }, { "epoch": 10.178484107579463, "grad_norm": 0.3105366872180831, "learning_rate": 1.1699585221504717e-06, "loss": 0.6155, "step": 4163 }, { "epoch": 10.180929095354523, "grad_norm": 0.21894071649710106, "learning_rate": 1.168874714066705e-06, "loss": 0.8423, "step": 4164 }, { "epoch": 10.183374083129584, "grad_norm": 0.2569470966766589, "learning_rate": 1.167791255034461e-06, "loss": 0.906, "step": 4165 }, { "epoch": 10.185819070904646, "grad_norm": 0.23821097890448478, "learning_rate": 1.1667081453378448e-06, "loss": 0.5307, "step": 4166 }, { "epoch": 10.188264058679707, "grad_norm": 0.3022091400946855, "learning_rate": 1.1656253852608736e-06, "loss": 0.7315, "step": 4167 }, { "epoch": 10.190709046454767, "grad_norm": 0.15477962231889153, "learning_rate": 1.1645429750874724e-06, "loss": 0.3853, "step": 4168 }, { "epoch": 10.19315403422983, "grad_norm": 0.30326146145813354, "learning_rate": 1.1634609151014743e-06, "loss": 0.8213, "step": 4169 }, { "epoch": 10.19559902200489, "grad_norm": 0.1763170486185931, "learning_rate": 1.1623792055866185e-06, "loss": 0.3783, "step": 4170 }, { "epoch": 10.19804400977995, "grad_norm": 0.2700999073428717, "learning_rate": 1.161297846826555e-06, "loss": 0.5596, "step": 4171 }, { "epoch": 10.200488997555013, "grad_norm": 0.28348043194799327, "learning_rate": 1.1602168391048418e-06, "loss": 1.4361, "step": 4172 }, { "epoch": 10.202933985330073, "grad_norm": 0.3812604496964369, "learning_rate": 1.159136182704942e-06, "loss": 0.3998, "step": 4173 }, { "epoch": 10.205378973105134, "grad_norm": 0.3402163874437233, "learning_rate": 1.1580558779102291e-06, "loss": 0.8034, "step": 4174 }, { "epoch": 10.207823960880196, "grad_norm": 0.21006144137395466, "learning_rate": 1.156975925003985e-06, "loss": 0.9017, "step": 4175 }, { "epoch": 10.210268948655257, "grad_norm": 0.1926071896728539, "learning_rate": 1.155896324269396e-06, "loss": 0.5015, "step": 4176 }, { "epoch": 10.212713936430317, "grad_norm": 0.2224797492015203, "learning_rate": 1.1548170759895602e-06, "loss": 0.7502, "step": 4177 }, { "epoch": 10.21515892420538, "grad_norm": 0.24848998628121807, "learning_rate": 1.153738180447478e-06, "loss": 0.5773, "step": 4178 }, { "epoch": 10.21760391198044, "grad_norm": 0.21902288233749048, "learning_rate": 1.152659637926065e-06, "loss": 0.5423, "step": 4179 }, { "epoch": 10.2200488997555, "grad_norm": 0.3049237722929354, "learning_rate": 1.1515814487081362e-06, "loss": 0.9181, "step": 4180 }, { "epoch": 10.222493887530563, "grad_norm": 0.2315209866675309, "learning_rate": 1.1505036130764186e-06, "loss": 0.5385, "step": 4181 }, { "epoch": 10.224938875305623, "grad_norm": 0.2943775324155884, "learning_rate": 1.1494261313135466e-06, "loss": 0.5121, "step": 4182 }, { "epoch": 10.227383863080684, "grad_norm": 0.4174047381364314, "learning_rate": 1.1483490037020585e-06, "loss": 0.4699, "step": 4183 }, { "epoch": 10.229828850855746, "grad_norm": 0.3253795652437366, "learning_rate": 1.1472722305244024e-06, "loss": 0.4873, "step": 4184 }, { "epoch": 10.232273838630807, "grad_norm": 0.2815085250810611, "learning_rate": 1.1461958120629345e-06, "loss": 0.6014, "step": 4185 }, { "epoch": 10.234718826405867, "grad_norm": 0.2146580038711741, "learning_rate": 1.1451197485999142e-06, "loss": 0.4959, "step": 4186 }, { "epoch": 10.23716381418093, "grad_norm": 0.2732581361140207, "learning_rate": 1.144044040417512e-06, "loss": 0.7633, "step": 4187 }, { "epoch": 10.23960880195599, "grad_norm": 0.20660481128824887, "learning_rate": 1.1429686877978006e-06, "loss": 0.6877, "step": 4188 }, { "epoch": 10.24205378973105, "grad_norm": 0.16109358894117934, "learning_rate": 1.1418936910227629e-06, "loss": 0.6768, "step": 4189 }, { "epoch": 10.244498777506113, "grad_norm": 0.21063067488195503, "learning_rate": 1.1408190503742903e-06, "loss": 0.4372, "step": 4190 }, { "epoch": 10.246943765281173, "grad_norm": 0.20266345557795643, "learning_rate": 1.1397447661341751e-06, "loss": 0.3692, "step": 4191 }, { "epoch": 10.249388753056234, "grad_norm": 0.3409355044102508, "learning_rate": 1.1386708385841214e-06, "loss": 0.4144, "step": 4192 }, { "epoch": 10.251833740831296, "grad_norm": 0.2135103580797468, "learning_rate": 1.1375972680057356e-06, "loss": 0.4847, "step": 4193 }, { "epoch": 10.254278728606357, "grad_norm": 0.28838238241776076, "learning_rate": 1.1365240546805332e-06, "loss": 0.4302, "step": 4194 }, { "epoch": 10.256723716381417, "grad_norm": 0.19391426535568707, "learning_rate": 1.1354511988899366e-06, "loss": 0.6519, "step": 4195 }, { "epoch": 10.25916870415648, "grad_norm": 0.2516415184994791, "learning_rate": 1.1343787009152709e-06, "loss": 0.6085, "step": 4196 }, { "epoch": 10.26161369193154, "grad_norm": 0.4803650111397249, "learning_rate": 1.133306561037772e-06, "loss": 0.7518, "step": 4197 }, { "epoch": 10.2640586797066, "grad_norm": 0.3299564140719025, "learning_rate": 1.1322347795385772e-06, "loss": 0.8661, "step": 4198 }, { "epoch": 10.266503667481663, "grad_norm": 0.2670071811253802, "learning_rate": 1.1311633566987333e-06, "loss": 0.63, "step": 4199 }, { "epoch": 10.268948655256724, "grad_norm": 0.26863777400741456, "learning_rate": 1.1300922927991914e-06, "loss": 0.3205, "step": 4200 }, { "epoch": 10.271393643031784, "grad_norm": 0.25421582705898893, "learning_rate": 1.1290215881208096e-06, "loss": 0.6836, "step": 4201 }, { "epoch": 10.273838630806846, "grad_norm": 0.25380212123474405, "learning_rate": 1.1279512429443524e-06, "loss": 0.4074, "step": 4202 }, { "epoch": 10.276283618581907, "grad_norm": 0.24722976283467396, "learning_rate": 1.126881257550486e-06, "loss": 0.4332, "step": 4203 }, { "epoch": 10.278728606356967, "grad_norm": 0.42665278112790245, "learning_rate": 1.1258116322197861e-06, "loss": 0.6619, "step": 4204 }, { "epoch": 10.28117359413203, "grad_norm": 0.37727090572341593, "learning_rate": 1.1247423672327349e-06, "loss": 0.3989, "step": 4205 }, { "epoch": 10.28361858190709, "grad_norm": 0.29823749757789775, "learning_rate": 1.1236734628697152e-06, "loss": 0.6988, "step": 4206 }, { "epoch": 10.28606356968215, "grad_norm": 0.24808436034893372, "learning_rate": 1.1226049194110206e-06, "loss": 0.7238, "step": 4207 }, { "epoch": 10.288508557457213, "grad_norm": 0.2505010264085653, "learning_rate": 1.1215367371368455e-06, "loss": 0.4817, "step": 4208 }, { "epoch": 10.290953545232274, "grad_norm": 0.23820776669465543, "learning_rate": 1.1204689163272931e-06, "loss": 0.4623, "step": 4209 }, { "epoch": 10.293398533007334, "grad_norm": 0.2448760834274258, "learning_rate": 1.1194014572623699e-06, "loss": 0.4795, "step": 4210 }, { "epoch": 10.295843520782396, "grad_norm": 0.21855635872086634, "learning_rate": 1.1183343602219884e-06, "loss": 0.4256, "step": 4211 }, { "epoch": 10.298288508557457, "grad_norm": 0.25934787937900033, "learning_rate": 1.1172676254859672e-06, "loss": 0.3763, "step": 4212 }, { "epoch": 10.30073349633252, "grad_norm": 0.21593605947712682, "learning_rate": 1.1162012533340257e-06, "loss": 0.8065, "step": 4213 }, { "epoch": 10.30317848410758, "grad_norm": 0.333071118152502, "learning_rate": 1.115135244045793e-06, "loss": 0.8674, "step": 4214 }, { "epoch": 10.30562347188264, "grad_norm": 0.21052561234933265, "learning_rate": 1.1140695979008017e-06, "loss": 0.5846, "step": 4215 }, { "epoch": 10.3080684596577, "grad_norm": 0.21395761980933636, "learning_rate": 1.1130043151784866e-06, "loss": 0.5805, "step": 4216 }, { "epoch": 10.310513447432763, "grad_norm": 0.2610315453236665, "learning_rate": 1.1119393961581915e-06, "loss": 0.4286, "step": 4217 }, { "epoch": 10.312958435207824, "grad_norm": 0.284873353549121, "learning_rate": 1.11087484111916e-06, "loss": 0.4579, "step": 4218 }, { "epoch": 10.315403422982886, "grad_norm": 0.2660496054519845, "learning_rate": 1.1098106503405443e-06, "loss": 0.4459, "step": 4219 }, { "epoch": 10.317848410757946, "grad_norm": 0.2221084619552763, "learning_rate": 1.1087468241014005e-06, "loss": 0.5036, "step": 4220 }, { "epoch": 10.320293398533007, "grad_norm": 0.27584253905045814, "learning_rate": 1.1076833626806852e-06, "loss": 0.502, "step": 4221 }, { "epoch": 10.32273838630807, "grad_norm": 0.2846802268178864, "learning_rate": 1.106620266357266e-06, "loss": 0.4134, "step": 4222 }, { "epoch": 10.32518337408313, "grad_norm": 0.30832109832740445, "learning_rate": 1.1055575354099083e-06, "loss": 0.6541, "step": 4223 }, { "epoch": 10.32762836185819, "grad_norm": 0.3670009084712405, "learning_rate": 1.1044951701172857e-06, "loss": 0.5937, "step": 4224 }, { "epoch": 10.330073349633253, "grad_norm": 0.3260661340177253, "learning_rate": 1.1034331707579752e-06, "loss": 0.3496, "step": 4225 }, { "epoch": 10.332518337408313, "grad_norm": 0.25570231793931764, "learning_rate": 1.102371537610456e-06, "loss": 0.4066, "step": 4226 }, { "epoch": 10.334963325183374, "grad_norm": 0.2572165327921113, "learning_rate": 1.101310270953114e-06, "loss": 0.5803, "step": 4227 }, { "epoch": 10.337408312958436, "grad_norm": 0.35429454241041924, "learning_rate": 1.100249371064236e-06, "loss": 0.4324, "step": 4228 }, { "epoch": 10.339853300733497, "grad_norm": 0.3752294459315225, "learning_rate": 1.0991888382220151e-06, "loss": 0.5597, "step": 4229 }, { "epoch": 10.342298288508557, "grad_norm": 0.36426667257913814, "learning_rate": 1.0981286727045484e-06, "loss": 0.5793, "step": 4230 }, { "epoch": 10.34474327628362, "grad_norm": 0.21342809231005616, "learning_rate": 1.0970688747898335e-06, "loss": 0.4607, "step": 4231 }, { "epoch": 10.34718826405868, "grad_norm": 0.2758253952506122, "learning_rate": 1.0960094447557742e-06, "loss": 1.2919, "step": 4232 }, { "epoch": 10.34963325183374, "grad_norm": 0.3018893020271271, "learning_rate": 1.0949503828801783e-06, "loss": 0.7432, "step": 4233 }, { "epoch": 10.352078239608803, "grad_norm": 0.2589862181460002, "learning_rate": 1.0938916894407553e-06, "loss": 0.4057, "step": 4234 }, { "epoch": 10.354523227383863, "grad_norm": 0.16981586539910698, "learning_rate": 1.09283336471512e-06, "loss": 0.5141, "step": 4235 }, { "epoch": 10.356968215158924, "grad_norm": 0.26154933006397024, "learning_rate": 1.0917754089807875e-06, "loss": 0.9263, "step": 4236 }, { "epoch": 10.359413202933986, "grad_norm": 0.2663676827423522, "learning_rate": 1.0907178225151799e-06, "loss": 0.3686, "step": 4237 }, { "epoch": 10.361858190709047, "grad_norm": 0.2888314656070139, "learning_rate": 1.0896606055956186e-06, "loss": 0.5378, "step": 4238 }, { "epoch": 10.364303178484107, "grad_norm": 0.21432782182469964, "learning_rate": 1.0886037584993311e-06, "loss": 0.8412, "step": 4239 }, { "epoch": 10.36674816625917, "grad_norm": 0.15904350964729666, "learning_rate": 1.0875472815034483e-06, "loss": 0.1969, "step": 4240 }, { "epoch": 10.36919315403423, "grad_norm": 0.3847422486539824, "learning_rate": 1.086491174885e-06, "loss": 0.335, "step": 4241 }, { "epoch": 10.37163814180929, "grad_norm": 0.244886244899407, "learning_rate": 1.0854354389209226e-06, "loss": 0.6841, "step": 4242 }, { "epoch": 10.374083129584353, "grad_norm": 0.2423179358033271, "learning_rate": 1.0843800738880545e-06, "loss": 0.5508, "step": 4243 }, { "epoch": 10.376528117359413, "grad_norm": 0.24481050768794999, "learning_rate": 1.0833250800631368e-06, "loss": 0.3785, "step": 4244 }, { "epoch": 10.378973105134474, "grad_norm": 0.41548784648028797, "learning_rate": 1.0822704577228133e-06, "loss": 0.7046, "step": 4245 }, { "epoch": 10.381418092909536, "grad_norm": 0.376757479458428, "learning_rate": 1.0812162071436288e-06, "loss": 0.5152, "step": 4246 }, { "epoch": 10.383863080684597, "grad_norm": 0.22139308348613054, "learning_rate": 1.0801623286020327e-06, "loss": 0.3987, "step": 4247 }, { "epoch": 10.386308068459657, "grad_norm": 0.3161245954958729, "learning_rate": 1.0791088223743771e-06, "loss": 0.4938, "step": 4248 }, { "epoch": 10.38875305623472, "grad_norm": 0.27960886231599424, "learning_rate": 1.078055688736913e-06, "loss": 0.7321, "step": 4249 }, { "epoch": 10.39119804400978, "grad_norm": 0.45091483854375275, "learning_rate": 1.077002927965799e-06, "loss": 0.4293, "step": 4250 }, { "epoch": 10.39364303178484, "grad_norm": 0.24169982156743752, "learning_rate": 1.0759505403370906e-06, "loss": 0.3913, "step": 4251 }, { "epoch": 10.396088019559903, "grad_norm": 0.33563048052452765, "learning_rate": 1.074898526126749e-06, "loss": 0.8785, "step": 4252 }, { "epoch": 10.398533007334963, "grad_norm": 0.22860336753584973, "learning_rate": 1.0738468856106363e-06, "loss": 0.4365, "step": 4253 }, { "epoch": 10.400977995110024, "grad_norm": 0.41847078332940246, "learning_rate": 1.0727956190645166e-06, "loss": 0.4003, "step": 4254 }, { "epoch": 10.403422982885086, "grad_norm": 0.26386490365276594, "learning_rate": 1.0717447267640575e-06, "loss": 0.5579, "step": 4255 }, { "epoch": 10.405867970660147, "grad_norm": 0.3315566599584672, "learning_rate": 1.0706942089848246e-06, "loss": 0.3855, "step": 4256 }, { "epoch": 10.408312958435207, "grad_norm": 0.26863576263656813, "learning_rate": 1.0696440660022892e-06, "loss": 0.4303, "step": 4257 }, { "epoch": 10.41075794621027, "grad_norm": 0.2921422625151501, "learning_rate": 1.068594298091823e-06, "loss": 0.4861, "step": 4258 }, { "epoch": 10.41320293398533, "grad_norm": 0.26531808905998455, "learning_rate": 1.067544905528698e-06, "loss": 0.7276, "step": 4259 }, { "epoch": 10.41564792176039, "grad_norm": 0.2860387956517466, "learning_rate": 1.0664958885880903e-06, "loss": 0.6428, "step": 4260 }, { "epoch": 10.418092909535453, "grad_norm": 0.36487813397526914, "learning_rate": 1.0654472475450748e-06, "loss": 0.6094, "step": 4261 }, { "epoch": 10.420537897310513, "grad_norm": 0.19662818281492933, "learning_rate": 1.0643989826746298e-06, "loss": 0.4207, "step": 4262 }, { "epoch": 10.422982885085574, "grad_norm": 0.3788751583815013, "learning_rate": 1.0633510942516354e-06, "loss": 0.4877, "step": 4263 }, { "epoch": 10.425427872860636, "grad_norm": 0.26201849194313936, "learning_rate": 1.062303582550869e-06, "loss": 0.61, "step": 4264 }, { "epoch": 10.427872860635697, "grad_norm": 0.3081976089537839, "learning_rate": 1.0612564478470158e-06, "loss": 0.4374, "step": 4265 }, { "epoch": 10.430317848410757, "grad_norm": 0.32562078149842, "learning_rate": 1.0602096904146561e-06, "loss": 0.4765, "step": 4266 }, { "epoch": 10.43276283618582, "grad_norm": 0.31390020641837185, "learning_rate": 1.0591633105282742e-06, "loss": 0.5158, "step": 4267 }, { "epoch": 10.43520782396088, "grad_norm": 0.21320436571176027, "learning_rate": 1.058117308462256e-06, "loss": 0.6171, "step": 4268 }, { "epoch": 10.43765281173594, "grad_norm": 0.20694455970272932, "learning_rate": 1.0570716844908854e-06, "loss": 0.7641, "step": 4269 }, { "epoch": 10.440097799511003, "grad_norm": 0.433356465180769, "learning_rate": 1.0560264388883505e-06, "loss": 0.6787, "step": 4270 }, { "epoch": 10.442542787286063, "grad_norm": 0.19627752136973292, "learning_rate": 1.0549815719287367e-06, "loss": 0.5985, "step": 4271 }, { "epoch": 10.444987775061124, "grad_norm": 0.3532235142292159, "learning_rate": 1.0539370838860334e-06, "loss": 0.6049, "step": 4272 }, { "epoch": 10.447432762836186, "grad_norm": 0.18998209868258623, "learning_rate": 1.0528929750341303e-06, "loss": 0.4402, "step": 4273 }, { "epoch": 10.449877750611247, "grad_norm": 0.2167927347712745, "learning_rate": 1.051849245646814e-06, "loss": 0.3878, "step": 4274 }, { "epoch": 10.452322738386307, "grad_norm": 0.2235549317041821, "learning_rate": 1.0508058959977757e-06, "loss": 0.8425, "step": 4275 }, { "epoch": 10.45476772616137, "grad_norm": 0.2556532345971111, "learning_rate": 1.0497629263606057e-06, "loss": 0.4573, "step": 4276 }, { "epoch": 10.45721271393643, "grad_norm": 0.26422450912041473, "learning_rate": 1.0487203370087943e-06, "loss": 0.5472, "step": 4277 }, { "epoch": 10.45965770171149, "grad_norm": 0.20641761692696617, "learning_rate": 1.0476781282157332e-06, "loss": 0.452, "step": 4278 }, { "epoch": 10.462102689486553, "grad_norm": 0.2308920442199556, "learning_rate": 1.0466363002547117e-06, "loss": 0.7833, "step": 4279 }, { "epoch": 10.464547677261614, "grad_norm": 0.2818830072693171, "learning_rate": 1.0455948533989225e-06, "loss": 0.4071, "step": 4280 }, { "epoch": 10.466992665036674, "grad_norm": 0.20014125332339097, "learning_rate": 1.044553787921455e-06, "loss": 0.3706, "step": 4281 }, { "epoch": 10.469437652811736, "grad_norm": 0.29893637195780276, "learning_rate": 1.0435131040953017e-06, "loss": 0.4334, "step": 4282 }, { "epoch": 10.471882640586797, "grad_norm": 0.2184268115591645, "learning_rate": 1.0424728021933545e-06, "loss": 0.5834, "step": 4283 }, { "epoch": 10.474327628361857, "grad_norm": 0.2789675432324769, "learning_rate": 1.0414328824884021e-06, "loss": 0.7017, "step": 4284 }, { "epoch": 10.47677261613692, "grad_norm": 0.2848807353058191, "learning_rate": 1.0403933452531368e-06, "loss": 0.4981, "step": 4285 }, { "epoch": 10.47921760391198, "grad_norm": 0.30646649759686806, "learning_rate": 1.0393541907601488e-06, "loss": 0.718, "step": 4286 }, { "epoch": 10.48166259168704, "grad_norm": 0.2670537262739069, "learning_rate": 1.0383154192819277e-06, "loss": 0.6994, "step": 4287 }, { "epoch": 10.484107579462103, "grad_norm": 0.278557544019373, "learning_rate": 1.037277031090865e-06, "loss": 0.6547, "step": 4288 }, { "epoch": 10.486552567237164, "grad_norm": 0.32952705090290973, "learning_rate": 1.0362390264592475e-06, "loss": 0.3579, "step": 4289 }, { "epoch": 10.488997555012224, "grad_norm": 0.22306291470480558, "learning_rate": 1.0352014056592654e-06, "loss": 0.8264, "step": 4290 }, { "epoch": 10.491442542787286, "grad_norm": 0.2790616778153501, "learning_rate": 1.0341641689630053e-06, "loss": 0.5048, "step": 4291 }, { "epoch": 10.493887530562347, "grad_norm": 0.20380835201691247, "learning_rate": 1.0331273166424551e-06, "loss": 0.9128, "step": 4292 }, { "epoch": 10.496332518337407, "grad_norm": 0.19651500785895423, "learning_rate": 1.032090848969502e-06, "loss": 0.6676, "step": 4293 }, { "epoch": 10.49877750611247, "grad_norm": 0.26744570732099743, "learning_rate": 1.0310547662159299e-06, "loss": 0.3443, "step": 4294 }, { "epoch": 10.50122249388753, "grad_norm": 0.32483429652313195, "learning_rate": 1.030019068653425e-06, "loss": 0.4567, "step": 4295 }, { "epoch": 10.503667481662593, "grad_norm": 0.2442696538945737, "learning_rate": 1.0289837565535688e-06, "loss": 0.4345, "step": 4296 }, { "epoch": 10.506112469437653, "grad_norm": 0.195352710619373, "learning_rate": 1.0279488301878457e-06, "loss": 0.4908, "step": 4297 }, { "epoch": 10.508557457212714, "grad_norm": 0.33427791501532744, "learning_rate": 1.0269142898276377e-06, "loss": 0.4971, "step": 4298 }, { "epoch": 10.511002444987774, "grad_norm": 0.20892654797180402, "learning_rate": 1.0258801357442227e-06, "loss": 0.3987, "step": 4299 }, { "epoch": 10.513447432762836, "grad_norm": 0.2122906776839145, "learning_rate": 1.024846368208782e-06, "loss": 0.533, "step": 4300 }, { "epoch": 10.515892420537897, "grad_norm": 0.3228820210091164, "learning_rate": 1.0238129874923905e-06, "loss": 0.7088, "step": 4301 }, { "epoch": 10.51833740831296, "grad_norm": 0.3013798127110642, "learning_rate": 1.022779993866026e-06, "loss": 0.8274, "step": 4302 }, { "epoch": 10.52078239608802, "grad_norm": 0.26692594965715, "learning_rate": 1.0217473876005635e-06, "loss": 0.8578, "step": 4303 }, { "epoch": 10.52322738386308, "grad_norm": 0.24590370222879326, "learning_rate": 1.0207151689667742e-06, "loss": 0.5524, "step": 4304 }, { "epoch": 10.525672371638143, "grad_norm": 0.38640332647289244, "learning_rate": 1.0196833382353306e-06, "loss": 0.7099, "step": 4305 }, { "epoch": 10.528117359413203, "grad_norm": 0.30321522549706703, "learning_rate": 1.0186518956768026e-06, "loss": 0.5987, "step": 4306 }, { "epoch": 10.530562347188264, "grad_norm": 0.28960180745559505, "learning_rate": 1.0176208415616564e-06, "loss": 0.776, "step": 4307 }, { "epoch": 10.533007334963326, "grad_norm": 0.20923029078323754, "learning_rate": 1.0165901761602608e-06, "loss": 0.583, "step": 4308 }, { "epoch": 10.535452322738386, "grad_norm": 0.25635600129741176, "learning_rate": 1.0155598997428773e-06, "loss": 0.4153, "step": 4309 }, { "epoch": 10.537897310513447, "grad_norm": 0.2948867049894463, "learning_rate": 1.01453001257967e-06, "loss": 0.5548, "step": 4310 }, { "epoch": 10.54034229828851, "grad_norm": 0.22432599513869883, "learning_rate": 1.013500514940697e-06, "loss": 0.4419, "step": 4311 }, { "epoch": 10.54278728606357, "grad_norm": 0.2570326252242165, "learning_rate": 1.012471407095917e-06, "loss": 0.7046, "step": 4312 }, { "epoch": 10.54523227383863, "grad_norm": 0.19928909167986347, "learning_rate": 1.0114426893151864e-06, "loss": 0.5154, "step": 4313 }, { "epoch": 10.547677261613693, "grad_norm": 0.21777738208008943, "learning_rate": 1.0104143618682568e-06, "loss": 0.3773, "step": 4314 }, { "epoch": 10.550122249388753, "grad_norm": 0.1593210421520442, "learning_rate": 1.0093864250247804e-06, "loss": 0.3999, "step": 4315 }, { "epoch": 10.552567237163814, "grad_norm": 0.3586817935567945, "learning_rate": 1.0083588790543062e-06, "loss": 0.5295, "step": 4316 }, { "epoch": 10.555012224938876, "grad_norm": 0.29243419840995577, "learning_rate": 1.0073317242262787e-06, "loss": 0.4194, "step": 4317 }, { "epoch": 10.557457212713937, "grad_norm": 0.26275739205581367, "learning_rate": 1.0063049608100426e-06, "loss": 0.6337, "step": 4318 }, { "epoch": 10.559902200488997, "grad_norm": 0.2756060460916107, "learning_rate": 1.0052785890748384e-06, "loss": 0.9403, "step": 4319 }, { "epoch": 10.56234718826406, "grad_norm": 0.20829443309216508, "learning_rate": 1.004252609289805e-06, "loss": 0.9553, "step": 4320 }, { "epoch": 10.56479217603912, "grad_norm": 0.27059316195803773, "learning_rate": 1.0032270217239765e-06, "loss": 0.8934, "step": 4321 }, { "epoch": 10.56723716381418, "grad_norm": 0.3626143064646002, "learning_rate": 1.0022018266462861e-06, "loss": 0.4049, "step": 4322 }, { "epoch": 10.569682151589243, "grad_norm": 0.1927649523245912, "learning_rate": 1.0011770243255642e-06, "loss": 0.4423, "step": 4323 }, { "epoch": 10.572127139364303, "grad_norm": 0.21369774068766934, "learning_rate": 1.0001526150305357e-06, "loss": 0.7241, "step": 4324 }, { "epoch": 10.574572127139364, "grad_norm": 0.3759232350000817, "learning_rate": 9.991285990298253e-07, "loss": 0.435, "step": 4325 }, { "epoch": 10.577017114914426, "grad_norm": 0.17815386883869375, "learning_rate": 9.98104976591954e-07, "loss": 0.6977, "step": 4326 }, { "epoch": 10.579462102689487, "grad_norm": 0.17530810875953795, "learning_rate": 9.970817479853372e-07, "loss": 0.3524, "step": 4327 }, { "epoch": 10.581907090464547, "grad_norm": 0.23459776835546303, "learning_rate": 9.960589134782903e-07, "loss": 0.966, "step": 4328 }, { "epoch": 10.58435207823961, "grad_norm": 0.18033747891273672, "learning_rate": 9.950364733390233e-07, "loss": 0.398, "step": 4329 }, { "epoch": 10.58679706601467, "grad_norm": 0.21279943215915337, "learning_rate": 9.940144278356444e-07, "loss": 0.5577, "step": 4330 }, { "epoch": 10.58924205378973, "grad_norm": 0.23986507967631324, "learning_rate": 9.929927772361562e-07, "loss": 0.4536, "step": 4331 }, { "epoch": 10.591687041564793, "grad_norm": 0.23190694368363074, "learning_rate": 9.919715218084588e-07, "loss": 0.5099, "step": 4332 }, { "epoch": 10.594132029339853, "grad_norm": 0.2181617359119374, "learning_rate": 9.909506618203501e-07, "loss": 0.4519, "step": 4333 }, { "epoch": 10.596577017114914, "grad_norm": 0.2620375780093581, "learning_rate": 9.899301975395214e-07, "loss": 0.6629, "step": 4334 }, { "epoch": 10.599022004889976, "grad_norm": 0.2544811026050055, "learning_rate": 9.889101292335625e-07, "loss": 0.5131, "step": 4335 }, { "epoch": 10.601466992665037, "grad_norm": 0.24143413021218876, "learning_rate": 9.878904571699596e-07, "loss": 0.8713, "step": 4336 }, { "epoch": 10.603911980440097, "grad_norm": 0.25564385214669744, "learning_rate": 9.86871181616092e-07, "loss": 0.5029, "step": 4337 }, { "epoch": 10.60635696821516, "grad_norm": 0.38465103858679484, "learning_rate": 9.85852302839239e-07, "loss": 0.5773, "step": 4338 }, { "epoch": 10.60880195599022, "grad_norm": 0.22444021703376024, "learning_rate": 9.848338211065714e-07, "loss": 0.8214, "step": 4339 }, { "epoch": 10.61124694376528, "grad_norm": 0.2792213262790953, "learning_rate": 9.83815736685162e-07, "loss": 0.6426, "step": 4340 }, { "epoch": 10.613691931540343, "grad_norm": 0.28763017776667527, "learning_rate": 9.82798049841973e-07, "loss": 0.4784, "step": 4341 }, { "epoch": 10.616136919315403, "grad_norm": 0.29670939092781656, "learning_rate": 9.817807608438663e-07, "loss": 0.4279, "step": 4342 }, { "epoch": 10.618581907090464, "grad_norm": 0.330271466928609, "learning_rate": 9.807638699575988e-07, "loss": 0.5066, "step": 4343 }, { "epoch": 10.621026894865526, "grad_norm": 0.23499697602317826, "learning_rate": 9.797473774498212e-07, "loss": 0.8341, "step": 4344 }, { "epoch": 10.623471882640587, "grad_norm": 0.24688716661319512, "learning_rate": 9.78731283587082e-07, "loss": 0.4716, "step": 4345 }, { "epoch": 10.625916870415647, "grad_norm": 0.20704999293221601, "learning_rate": 9.77715588635825e-07, "loss": 0.4567, "step": 4346 }, { "epoch": 10.62836185819071, "grad_norm": 0.20415077914130392, "learning_rate": 9.76700292862387e-07, "loss": 0.4772, "step": 4347 }, { "epoch": 10.63080684596577, "grad_norm": 0.33140905873083754, "learning_rate": 9.756853965330035e-07, "loss": 0.6902, "step": 4348 }, { "epoch": 10.63325183374083, "grad_norm": 0.1818198379603264, "learning_rate": 9.74670899913802e-07, "loss": 0.3833, "step": 4349 }, { "epoch": 10.635696821515893, "grad_norm": 0.3250422855406044, "learning_rate": 9.736568032708068e-07, "loss": 0.4677, "step": 4350 }, { "epoch": 10.638141809290953, "grad_norm": 0.21018869986651117, "learning_rate": 9.726431068699385e-07, "loss": 0.3753, "step": 4351 }, { "epoch": 10.640586797066014, "grad_norm": 0.325376689959012, "learning_rate": 9.716298109770107e-07, "loss": 0.5235, "step": 4352 }, { "epoch": 10.643031784841076, "grad_norm": 0.3099870718506511, "learning_rate": 9.706169158577336e-07, "loss": 0.7004, "step": 4353 }, { "epoch": 10.645476772616137, "grad_norm": 0.18464272261388961, "learning_rate": 9.696044217777101e-07, "loss": 0.6622, "step": 4354 }, { "epoch": 10.647921760391197, "grad_norm": 0.26758599303735314, "learning_rate": 9.685923290024398e-07, "loss": 0.5943, "step": 4355 }, { "epoch": 10.65036674816626, "grad_norm": 0.22330967176009905, "learning_rate": 9.675806377973183e-07, "loss": 0.3678, "step": 4356 }, { "epoch": 10.65281173594132, "grad_norm": 0.2241055381192565, "learning_rate": 9.665693484276315e-07, "loss": 0.9309, "step": 4357 }, { "epoch": 10.65525672371638, "grad_norm": 0.20088571868784003, "learning_rate": 9.655584611585647e-07, "loss": 0.8415, "step": 4358 }, { "epoch": 10.657701711491443, "grad_norm": 0.2737798661349978, "learning_rate": 9.64547976255194e-07, "loss": 0.4643, "step": 4359 }, { "epoch": 10.660146699266503, "grad_norm": 0.2092178363833883, "learning_rate": 9.635378939824925e-07, "loss": 0.7475, "step": 4360 }, { "epoch": 10.662591687041564, "grad_norm": 0.24399175982209115, "learning_rate": 9.62528214605327e-07, "loss": 0.8379, "step": 4361 }, { "epoch": 10.665036674816626, "grad_norm": 0.19122089848163124, "learning_rate": 9.615189383884585e-07, "loss": 0.4814, "step": 4362 }, { "epoch": 10.667481662591687, "grad_norm": 0.19134590066254722, "learning_rate": 9.60510065596543e-07, "loss": 0.4581, "step": 4363 }, { "epoch": 10.669926650366747, "grad_norm": 0.25128884113975064, "learning_rate": 9.595015964941287e-07, "loss": 1.0461, "step": 4364 }, { "epoch": 10.67237163814181, "grad_norm": 0.3547101290265935, "learning_rate": 9.584935313456597e-07, "loss": 0.3851, "step": 4365 }, { "epoch": 10.67481662591687, "grad_norm": 0.3610904754610384, "learning_rate": 9.574858704154749e-07, "loss": 0.6249, "step": 4366 }, { "epoch": 10.67726161369193, "grad_norm": 0.3503368125276897, "learning_rate": 9.564786139678043e-07, "loss": 0.632, "step": 4367 }, { "epoch": 10.679706601466993, "grad_norm": 0.26365466941111376, "learning_rate": 9.554717622667752e-07, "loss": 0.7467, "step": 4368 }, { "epoch": 10.682151589242054, "grad_norm": 0.3095614530979863, "learning_rate": 9.54465315576406e-07, "loss": 0.5197, "step": 4369 }, { "epoch": 10.684596577017114, "grad_norm": 0.2847381720183174, "learning_rate": 9.534592741606097e-07, "loss": 1.3271, "step": 4370 }, { "epoch": 10.687041564792176, "grad_norm": 0.25389374066124937, "learning_rate": 9.524536382831947e-07, "loss": 0.6757, "step": 4371 }, { "epoch": 10.689486552567237, "grad_norm": 0.21911826613625454, "learning_rate": 9.514484082078609e-07, "loss": 0.4052, "step": 4372 }, { "epoch": 10.6919315403423, "grad_norm": 0.1764902828990768, "learning_rate": 9.504435841982041e-07, "loss": 0.6313, "step": 4373 }, { "epoch": 10.69437652811736, "grad_norm": 0.331708896072561, "learning_rate": 9.494391665177097e-07, "loss": 0.3464, "step": 4374 }, { "epoch": 10.69682151589242, "grad_norm": 0.19659752072918085, "learning_rate": 9.484351554297605e-07, "loss": 0.5552, "step": 4375 }, { "epoch": 10.69926650366748, "grad_norm": 0.22942094415746575, "learning_rate": 9.474315511976321e-07, "loss": 0.5713, "step": 4376 }, { "epoch": 10.701711491442543, "grad_norm": 0.2687075149302389, "learning_rate": 9.464283540844902e-07, "loss": 0.5856, "step": 4377 }, { "epoch": 10.704156479217604, "grad_norm": 0.1921975744919534, "learning_rate": 9.454255643533983e-07, "loss": 0.484, "step": 4378 }, { "epoch": 10.706601466992666, "grad_norm": 0.2869895532461907, "learning_rate": 9.444231822673086e-07, "loss": 0.4356, "step": 4379 }, { "epoch": 10.709046454767726, "grad_norm": 0.15579817052682807, "learning_rate": 9.434212080890701e-07, "loss": 0.4563, "step": 4380 }, { "epoch": 10.711491442542787, "grad_norm": 0.32465613513550284, "learning_rate": 9.424196420814239e-07, "loss": 0.3911, "step": 4381 }, { "epoch": 10.713936430317847, "grad_norm": 0.31794639758301396, "learning_rate": 9.414184845070007e-07, "loss": 0.681, "step": 4382 }, { "epoch": 10.71638141809291, "grad_norm": 0.2972437232345382, "learning_rate": 9.404177356283309e-07, "loss": 0.3986, "step": 4383 }, { "epoch": 10.71882640586797, "grad_norm": 0.3199035978556867, "learning_rate": 9.394173957078309e-07, "loss": 0.3613, "step": 4384 }, { "epoch": 10.721271393643033, "grad_norm": 0.3287698075179986, "learning_rate": 9.384174650078135e-07, "loss": 0.5073, "step": 4385 }, { "epoch": 10.723716381418093, "grad_norm": 0.1995913938730153, "learning_rate": 9.374179437904846e-07, "loss": 0.6236, "step": 4386 }, { "epoch": 10.726161369193154, "grad_norm": 0.2637770153765454, "learning_rate": 9.364188323179394e-07, "loss": 0.7355, "step": 4387 }, { "epoch": 10.728606356968216, "grad_norm": 0.2292466546258323, "learning_rate": 9.354201308521698e-07, "loss": 0.7841, "step": 4388 }, { "epoch": 10.731051344743276, "grad_norm": 0.2826226884825545, "learning_rate": 9.344218396550563e-07, "loss": 0.4379, "step": 4389 }, { "epoch": 10.733496332518337, "grad_norm": 0.2230679421584428, "learning_rate": 9.334239589883748e-07, "loss": 0.3922, "step": 4390 }, { "epoch": 10.7359413202934, "grad_norm": 0.3298159686786096, "learning_rate": 9.324264891137933e-07, "loss": 0.7186, "step": 4391 }, { "epoch": 10.73838630806846, "grad_norm": 0.22234638548153243, "learning_rate": 9.314294302928692e-07, "loss": 0.8327, "step": 4392 }, { "epoch": 10.74083129584352, "grad_norm": 0.38196027450070896, "learning_rate": 9.304327827870552e-07, "loss": 0.7486, "step": 4393 }, { "epoch": 10.743276283618583, "grad_norm": 0.2634863222326317, "learning_rate": 9.294365468576951e-07, "loss": 0.3903, "step": 4394 }, { "epoch": 10.745721271393643, "grad_norm": 0.2991866041113629, "learning_rate": 9.284407227660249e-07, "loss": 0.548, "step": 4395 }, { "epoch": 10.748166259168704, "grad_norm": 0.3673166958037705, "learning_rate": 9.274453107731732e-07, "loss": 0.7222, "step": 4396 }, { "epoch": 10.750611246943766, "grad_norm": 0.1851462104213581, "learning_rate": 9.264503111401579e-07, "loss": 0.6613, "step": 4397 }, { "epoch": 10.753056234718827, "grad_norm": 0.39527664798921636, "learning_rate": 9.254557241278927e-07, "loss": 0.3825, "step": 4398 }, { "epoch": 10.755501222493887, "grad_norm": 0.20968926398614893, "learning_rate": 9.24461549997179e-07, "loss": 0.4655, "step": 4399 }, { "epoch": 10.75794621026895, "grad_norm": 0.2664069933205161, "learning_rate": 9.234677890087135e-07, "loss": 0.4979, "step": 4400 }, { "epoch": 10.76039119804401, "grad_norm": 0.20330502942507978, "learning_rate": 9.224744414230832e-07, "loss": 0.3705, "step": 4401 }, { "epoch": 10.76283618581907, "grad_norm": 0.34511026553869817, "learning_rate": 9.214815075007655e-07, "loss": 0.4957, "step": 4402 }, { "epoch": 10.765281173594133, "grad_norm": 0.2422112152014097, "learning_rate": 9.204889875021309e-07, "loss": 0.7323, "step": 4403 }, { "epoch": 10.767726161369193, "grad_norm": 0.25647415461081396, "learning_rate": 9.194968816874411e-07, "loss": 0.7358, "step": 4404 }, { "epoch": 10.770171149144254, "grad_norm": 0.3964545172489735, "learning_rate": 9.185051903168485e-07, "loss": 0.5988, "step": 4405 }, { "epoch": 10.772616136919316, "grad_norm": 0.27539749723367507, "learning_rate": 9.175139136503986e-07, "loss": 0.544, "step": 4406 }, { "epoch": 10.775061124694377, "grad_norm": 0.2820327413058313, "learning_rate": 9.165230519480248e-07, "loss": 0.5782, "step": 4407 }, { "epoch": 10.777506112469437, "grad_norm": 0.3138077772670629, "learning_rate": 9.155326054695556e-07, "loss": 0.8416, "step": 4408 }, { "epoch": 10.7799511002445, "grad_norm": 0.2269564467882219, "learning_rate": 9.145425744747072e-07, "loss": 0.4036, "step": 4409 }, { "epoch": 10.78239608801956, "grad_norm": 0.24756087356975026, "learning_rate": 9.135529592230891e-07, "loss": 0.5553, "step": 4410 }, { "epoch": 10.78484107579462, "grad_norm": 0.3236770829473707, "learning_rate": 9.125637599742018e-07, "loss": 0.4281, "step": 4411 }, { "epoch": 10.787286063569683, "grad_norm": 0.3872565800486209, "learning_rate": 9.115749769874343e-07, "loss": 0.587, "step": 4412 }, { "epoch": 10.789731051344743, "grad_norm": 0.2324783179340739, "learning_rate": 9.10586610522069e-07, "loss": 0.7913, "step": 4413 }, { "epoch": 10.792176039119804, "grad_norm": 0.28654113015923355, "learning_rate": 9.095986608372784e-07, "loss": 0.7215, "step": 4414 }, { "epoch": 10.794621026894866, "grad_norm": 0.2993010275184039, "learning_rate": 9.086111281921253e-07, "loss": 0.6842, "step": 4415 }, { "epoch": 10.797066014669927, "grad_norm": 0.34360704609554243, "learning_rate": 9.076240128455644e-07, "loss": 0.5557, "step": 4416 }, { "epoch": 10.799511002444987, "grad_norm": 0.2192671455149053, "learning_rate": 9.066373150564381e-07, "loss": 0.3772, "step": 4417 }, { "epoch": 10.80195599022005, "grad_norm": 0.20585252900141074, "learning_rate": 9.056510350834829e-07, "loss": 0.491, "step": 4418 }, { "epoch": 10.80440097799511, "grad_norm": 0.2699705529613797, "learning_rate": 9.046651731853223e-07, "loss": 0.3854, "step": 4419 }, { "epoch": 10.80684596577017, "grad_norm": 0.16863328327896884, "learning_rate": 9.036797296204725e-07, "loss": 0.3168, "step": 4420 }, { "epoch": 10.809290953545233, "grad_norm": 0.21309466421484127, "learning_rate": 9.026947046473406e-07, "loss": 0.6027, "step": 4421 }, { "epoch": 10.811735941320293, "grad_norm": 0.23946532233739712, "learning_rate": 9.017100985242208e-07, "loss": 0.998, "step": 4422 }, { "epoch": 10.814180929095354, "grad_norm": 0.21603804495006582, "learning_rate": 9.007259115093e-07, "loss": 0.538, "step": 4423 }, { "epoch": 10.816625916870416, "grad_norm": 0.29546548245082865, "learning_rate": 8.997421438606557e-07, "loss": 0.5188, "step": 4424 }, { "epoch": 10.819070904645477, "grad_norm": 0.24162164792096535, "learning_rate": 8.987587958362517e-07, "loss": 0.4876, "step": 4425 }, { "epoch": 10.821515892420537, "grad_norm": 0.25630302694170654, "learning_rate": 8.97775867693948e-07, "loss": 0.442, "step": 4426 }, { "epoch": 10.8239608801956, "grad_norm": 0.19120876699089995, "learning_rate": 8.967933596914877e-07, "loss": 0.4678, "step": 4427 }, { "epoch": 10.82640586797066, "grad_norm": 0.3636273796371703, "learning_rate": 8.958112720865089e-07, "loss": 0.4169, "step": 4428 }, { "epoch": 10.82885085574572, "grad_norm": 0.1749989420365941, "learning_rate": 8.94829605136536e-07, "loss": 0.3598, "step": 4429 }, { "epoch": 10.831295843520783, "grad_norm": 0.2470284400608675, "learning_rate": 8.938483590989852e-07, "loss": 1.2318, "step": 4430 }, { "epoch": 10.833740831295843, "grad_norm": 0.29891761018647545, "learning_rate": 8.928675342311624e-07, "loss": 0.5585, "step": 4431 }, { "epoch": 10.836185819070904, "grad_norm": 0.280447994022004, "learning_rate": 8.918871307902608e-07, "loss": 0.5534, "step": 4432 }, { "epoch": 10.838630806845966, "grad_norm": 0.21079968755899087, "learning_rate": 8.909071490333652e-07, "loss": 0.636, "step": 4433 }, { "epoch": 10.841075794621027, "grad_norm": 0.3117880234247859, "learning_rate": 8.899275892174506e-07, "loss": 0.3534, "step": 4434 }, { "epoch": 10.843520782396087, "grad_norm": 0.22603199067615662, "learning_rate": 8.889484515993776e-07, "loss": 0.6899, "step": 4435 }, { "epoch": 10.84596577017115, "grad_norm": 0.2302250341189511, "learning_rate": 8.879697364359002e-07, "loss": 0.7505, "step": 4436 }, { "epoch": 10.84841075794621, "grad_norm": 0.3216128410875158, "learning_rate": 8.869914439836594e-07, "loss": 0.4419, "step": 4437 }, { "epoch": 10.85085574572127, "grad_norm": 0.22257863716300735, "learning_rate": 8.860135744991866e-07, "loss": 0.5931, "step": 4438 }, { "epoch": 10.853300733496333, "grad_norm": 0.18158574391719975, "learning_rate": 8.850361282389003e-07, "loss": 0.4167, "step": 4439 }, { "epoch": 10.855745721271393, "grad_norm": 0.32205543383096275, "learning_rate": 8.840591054591097e-07, "loss": 0.7763, "step": 4440 }, { "epoch": 10.858190709046454, "grad_norm": 0.29840217919148176, "learning_rate": 8.830825064160137e-07, "loss": 0.5101, "step": 4441 }, { "epoch": 10.860635696821516, "grad_norm": 0.2756364356076906, "learning_rate": 8.821063313656971e-07, "loss": 0.492, "step": 4442 }, { "epoch": 10.863080684596577, "grad_norm": 0.30828226502492734, "learning_rate": 8.811305805641362e-07, "loss": 0.4277, "step": 4443 }, { "epoch": 10.865525672371637, "grad_norm": 0.3405026748914933, "learning_rate": 8.80155254267196e-07, "loss": 0.669, "step": 4444 }, { "epoch": 10.8679706601467, "grad_norm": 0.2880078286270374, "learning_rate": 8.791803527306278e-07, "loss": 0.6152, "step": 4445 }, { "epoch": 10.87041564792176, "grad_norm": 0.2572118825561372, "learning_rate": 8.78205876210074e-07, "loss": 0.4644, "step": 4446 }, { "epoch": 10.87286063569682, "grad_norm": 0.38648626433353456, "learning_rate": 8.772318249610645e-07, "loss": 0.3167, "step": 4447 }, { "epoch": 10.875305623471883, "grad_norm": 0.21856404549859335, "learning_rate": 8.762581992390187e-07, "loss": 0.8888, "step": 4448 }, { "epoch": 10.877750611246944, "grad_norm": 0.2625702537584893, "learning_rate": 8.752849992992418e-07, "loss": 0.9432, "step": 4449 }, { "epoch": 10.880195599022004, "grad_norm": 0.28104352573269825, "learning_rate": 8.743122253969302e-07, "loss": 0.5355, "step": 4450 }, { "epoch": 10.882640586797066, "grad_norm": 0.2559239747828476, "learning_rate": 8.733398777871679e-07, "loss": 1.0392, "step": 4451 }, { "epoch": 10.885085574572127, "grad_norm": 0.22730916278844424, "learning_rate": 8.723679567249254e-07, "loss": 0.373, "step": 4452 }, { "epoch": 10.887530562347187, "grad_norm": 0.20767087684718552, "learning_rate": 8.713964624650633e-07, "loss": 0.7799, "step": 4453 }, { "epoch": 10.88997555012225, "grad_norm": 0.20657583949687602, "learning_rate": 8.704253952623309e-07, "loss": 0.6948, "step": 4454 }, { "epoch": 10.89242053789731, "grad_norm": 0.30545276658983533, "learning_rate": 8.69454755371362e-07, "loss": 0.4878, "step": 4455 }, { "epoch": 10.894865525672373, "grad_norm": 0.15837107102478948, "learning_rate": 8.684845430466824e-07, "loss": 0.4047, "step": 4456 }, { "epoch": 10.897310513447433, "grad_norm": 0.2648991868962773, "learning_rate": 8.675147585427015e-07, "loss": 0.4929, "step": 4457 }, { "epoch": 10.899755501222494, "grad_norm": 0.20973692875051925, "learning_rate": 8.665454021137226e-07, "loss": 0.3618, "step": 4458 }, { "epoch": 10.902200488997554, "grad_norm": 0.26653300908609395, "learning_rate": 8.655764740139302e-07, "loss": 0.6734, "step": 4459 }, { "epoch": 10.904645476772616, "grad_norm": 0.22430595797584799, "learning_rate": 8.646079744974007e-07, "loss": 0.4768, "step": 4460 }, { "epoch": 10.907090464547677, "grad_norm": 0.2262324485591408, "learning_rate": 8.636399038180975e-07, "loss": 1.1212, "step": 4461 }, { "epoch": 10.90953545232274, "grad_norm": 0.23896216109058896, "learning_rate": 8.626722622298692e-07, "loss": 0.9531, "step": 4462 }, { "epoch": 10.9119804400978, "grad_norm": 0.28232467773131853, "learning_rate": 8.617050499864543e-07, "loss": 0.7083, "step": 4463 }, { "epoch": 10.91442542787286, "grad_norm": 0.12834222325869268, "learning_rate": 8.607382673414788e-07, "loss": 0.2334, "step": 4464 }, { "epoch": 10.91687041564792, "grad_norm": 0.3180468946084402, "learning_rate": 8.59771914548454e-07, "loss": 0.4638, "step": 4465 }, { "epoch": 10.919315403422983, "grad_norm": 0.2450976404145126, "learning_rate": 8.588059918607811e-07, "loss": 0.448, "step": 4466 }, { "epoch": 10.921760391198044, "grad_norm": 0.17675474003336317, "learning_rate": 8.578404995317452e-07, "loss": 0.7308, "step": 4467 }, { "epoch": 10.924205378973106, "grad_norm": 0.32324907896928085, "learning_rate": 8.56875437814522e-07, "loss": 0.8168, "step": 4468 }, { "epoch": 10.926650366748166, "grad_norm": 0.2890846951076616, "learning_rate": 8.559108069621719e-07, "loss": 0.8603, "step": 4469 }, { "epoch": 10.929095354523227, "grad_norm": 0.28180787028995546, "learning_rate": 8.549466072276441e-07, "loss": 0.3787, "step": 4470 }, { "epoch": 10.93154034229829, "grad_norm": 0.2699585089295434, "learning_rate": 8.539828388637739e-07, "loss": 0.2415, "step": 4471 }, { "epoch": 10.93398533007335, "grad_norm": 0.2061375638823393, "learning_rate": 8.530195021232823e-07, "loss": 0.5574, "step": 4472 }, { "epoch": 10.93643031784841, "grad_norm": 0.22817389604323257, "learning_rate": 8.520565972587788e-07, "loss": 0.3478, "step": 4473 }, { "epoch": 10.938875305623473, "grad_norm": 0.21006231657407667, "learning_rate": 8.5109412452276e-07, "loss": 0.544, "step": 4474 }, { "epoch": 10.941320293398533, "grad_norm": 0.21186009207457654, "learning_rate": 8.501320841676067e-07, "loss": 0.4297, "step": 4475 }, { "epoch": 10.943765281173594, "grad_norm": 0.268039698428206, "learning_rate": 8.491704764455894e-07, "loss": 0.7772, "step": 4476 }, { "epoch": 10.946210268948656, "grad_norm": 0.3150596163643315, "learning_rate": 8.482093016088622e-07, "loss": 0.8611, "step": 4477 }, { "epoch": 10.948655256723717, "grad_norm": 0.2261951522324787, "learning_rate": 8.472485599094679e-07, "loss": 0.6551, "step": 4478 }, { "epoch": 10.951100244498777, "grad_norm": 0.26293016278688186, "learning_rate": 8.462882515993348e-07, "loss": 0.5416, "step": 4479 }, { "epoch": 10.95354523227384, "grad_norm": 0.2504390444504082, "learning_rate": 8.453283769302778e-07, "loss": 0.7449, "step": 4480 }, { "epoch": 10.9559902200489, "grad_norm": 0.3306988849884077, "learning_rate": 8.443689361539989e-07, "loss": 0.4165, "step": 4481 }, { "epoch": 10.95843520782396, "grad_norm": 0.22036648806258632, "learning_rate": 8.434099295220841e-07, "loss": 0.699, "step": 4482 }, { "epoch": 10.960880195599023, "grad_norm": 0.27981473203785723, "learning_rate": 8.424513572860068e-07, "loss": 0.5788, "step": 4483 }, { "epoch": 10.963325183374083, "grad_norm": 0.25730042240918105, "learning_rate": 8.414932196971284e-07, "loss": 0.7325, "step": 4484 }, { "epoch": 10.965770171149144, "grad_norm": 0.2463058403824498, "learning_rate": 8.405355170066925e-07, "loss": 0.7869, "step": 4485 }, { "epoch": 10.968215158924206, "grad_norm": 0.25286765477335593, "learning_rate": 8.395782494658322e-07, "loss": 0.5196, "step": 4486 }, { "epoch": 10.970660146699267, "grad_norm": 0.31734230187658513, "learning_rate": 8.386214173255633e-07, "loss": 0.7287, "step": 4487 }, { "epoch": 10.973105134474327, "grad_norm": 0.17246194453919222, "learning_rate": 8.376650208367901e-07, "loss": 0.8018, "step": 4488 }, { "epoch": 10.97555012224939, "grad_norm": 0.2407156080368671, "learning_rate": 8.367090602503016e-07, "loss": 0.5047, "step": 4489 }, { "epoch": 10.97799511002445, "grad_norm": 0.2922035398633921, "learning_rate": 8.357535358167723e-07, "loss": 0.4366, "step": 4490 }, { "epoch": 10.98044009779951, "grad_norm": 0.2437845624651476, "learning_rate": 8.347984477867638e-07, "loss": 0.726, "step": 4491 }, { "epoch": 10.982885085574573, "grad_norm": 0.3905562708560972, "learning_rate": 8.338437964107201e-07, "loss": 0.4928, "step": 4492 }, { "epoch": 10.985330073349633, "grad_norm": 0.310411980402308, "learning_rate": 8.328895819389737e-07, "loss": 0.627, "step": 4493 }, { "epoch": 10.987775061124694, "grad_norm": 0.27887099268850135, "learning_rate": 8.319358046217421e-07, "loss": 0.8326, "step": 4494 }, { "epoch": 10.990220048899756, "grad_norm": 0.35830875990279687, "learning_rate": 8.309824647091261e-07, "loss": 0.4339, "step": 4495 }, { "epoch": 10.992665036674817, "grad_norm": 0.34317206265544303, "learning_rate": 8.300295624511146e-07, "loss": 0.8138, "step": 4496 }, { "epoch": 10.995110024449877, "grad_norm": 0.24199401731588846, "learning_rate": 8.29077098097579e-07, "loss": 0.7366, "step": 4497 }, { "epoch": 10.99755501222494, "grad_norm": 0.3603846066956584, "learning_rate": 8.281250718982778e-07, "loss": 0.5646, "step": 4498 }, { "epoch": 11.0, "grad_norm": 0.19075556512986527, "learning_rate": 8.271734841028553e-07, "loss": 0.4423, "step": 4499 }, { "epoch": 11.00244498777506, "grad_norm": 0.1831902131501335, "learning_rate": 8.262223349608367e-07, "loss": 0.871, "step": 4500 }, { "epoch": 11.004889975550123, "grad_norm": 0.22648632453837284, "learning_rate": 8.252716247216389e-07, "loss": 0.4024, "step": 4501 }, { "epoch": 11.007334963325183, "grad_norm": 0.2518849476096333, "learning_rate": 8.243213536345568e-07, "loss": 0.49, "step": 4502 }, { "epoch": 11.009779951100244, "grad_norm": 0.27280179848624325, "learning_rate": 8.233715219487745e-07, "loss": 0.7326, "step": 4503 }, { "epoch": 11.012224938875306, "grad_norm": 0.22351049791665253, "learning_rate": 8.2242212991336e-07, "loss": 0.4238, "step": 4504 }, { "epoch": 11.014669926650367, "grad_norm": 0.31340815027961766, "learning_rate": 8.214731777772647e-07, "loss": 0.4889, "step": 4505 }, { "epoch": 11.017114914425427, "grad_norm": 0.26237060656233774, "learning_rate": 8.205246657893265e-07, "loss": 0.368, "step": 4506 }, { "epoch": 11.01955990220049, "grad_norm": 0.3124421927923915, "learning_rate": 8.195765941982656e-07, "loss": 0.4683, "step": 4507 }, { "epoch": 11.02200488997555, "grad_norm": 0.3269462688986921, "learning_rate": 8.186289632526889e-07, "loss": 0.611, "step": 4508 }, { "epoch": 11.02444987775061, "grad_norm": 0.19219501516140977, "learning_rate": 8.17681773201088e-07, "loss": 0.5032, "step": 4509 }, { "epoch": 11.026894865525673, "grad_norm": 0.3722903018568003, "learning_rate": 8.167350242918356e-07, "loss": 0.6307, "step": 4510 }, { "epoch": 11.029339853300733, "grad_norm": 0.24390355541875292, "learning_rate": 8.157887167731923e-07, "loss": 0.5244, "step": 4511 }, { "epoch": 11.031784841075794, "grad_norm": 0.21709289368777013, "learning_rate": 8.148428508933012e-07, "loss": 0.4133, "step": 4512 }, { "epoch": 11.034229828850856, "grad_norm": 0.25212275623418967, "learning_rate": 8.138974269001901e-07, "loss": 0.3299, "step": 4513 }, { "epoch": 11.036674816625917, "grad_norm": 0.3200662022443012, "learning_rate": 8.129524450417719e-07, "loss": 0.4433, "step": 4514 }, { "epoch": 11.039119804400977, "grad_norm": 0.26501731998419176, "learning_rate": 8.120079055658403e-07, "loss": 0.7023, "step": 4515 }, { "epoch": 11.04156479217604, "grad_norm": 0.27023465279150916, "learning_rate": 8.110638087200773e-07, "loss": 0.5309, "step": 4516 }, { "epoch": 11.0440097799511, "grad_norm": 0.27427600503907607, "learning_rate": 8.101201547520451e-07, "loss": 0.4105, "step": 4517 }, { "epoch": 11.04645476772616, "grad_norm": 0.4488457267980421, "learning_rate": 8.091769439091917e-07, "loss": 0.6137, "step": 4518 }, { "epoch": 11.048899755501223, "grad_norm": 0.29174034757312417, "learning_rate": 8.082341764388499e-07, "loss": 0.6363, "step": 4519 }, { "epoch": 11.051344743276283, "grad_norm": 0.32632110113196966, "learning_rate": 8.072918525882328e-07, "loss": 0.5493, "step": 4520 }, { "epoch": 11.053789731051344, "grad_norm": 0.2498051767537811, "learning_rate": 8.063499726044405e-07, "loss": 0.5249, "step": 4521 }, { "epoch": 11.056234718826406, "grad_norm": 0.21472642349457857, "learning_rate": 8.054085367344553e-07, "loss": 0.7333, "step": 4522 }, { "epoch": 11.058679706601467, "grad_norm": 0.2802799571310624, "learning_rate": 8.044675452251433e-07, "loss": 0.4129, "step": 4523 }, { "epoch": 11.061124694376527, "grad_norm": 0.20538076074997694, "learning_rate": 8.035269983232547e-07, "loss": 0.6974, "step": 4524 }, { "epoch": 11.06356968215159, "grad_norm": 0.23606579013577947, "learning_rate": 8.02586896275421e-07, "loss": 0.6632, "step": 4525 }, { "epoch": 11.06601466992665, "grad_norm": 0.23537633148482667, "learning_rate": 8.016472393281602e-07, "loss": 0.7241, "step": 4526 }, { "epoch": 11.06845965770171, "grad_norm": 0.24498023749857345, "learning_rate": 8.007080277278697e-07, "loss": 0.4311, "step": 4527 }, { "epoch": 11.070904645476773, "grad_norm": 0.19327076895010875, "learning_rate": 7.997692617208339e-07, "loss": 0.4271, "step": 4528 }, { "epoch": 11.073349633251834, "grad_norm": 0.32176051550500834, "learning_rate": 7.988309415532186e-07, "loss": 0.3714, "step": 4529 }, { "epoch": 11.075794621026894, "grad_norm": 0.2934864243299337, "learning_rate": 7.978930674710719e-07, "loss": 0.7543, "step": 4530 }, { "epoch": 11.078239608801956, "grad_norm": 0.39792587179515393, "learning_rate": 7.969556397203265e-07, "loss": 0.3317, "step": 4531 }, { "epoch": 11.080684596577017, "grad_norm": 0.2581968232529534, "learning_rate": 7.960186585467975e-07, "loss": 0.4091, "step": 4532 }, { "epoch": 11.083129584352077, "grad_norm": 0.27645371154528486, "learning_rate": 7.950821241961825e-07, "loss": 0.4343, "step": 4533 }, { "epoch": 11.08557457212714, "grad_norm": 0.2635637708911598, "learning_rate": 7.941460369140633e-07, "loss": 0.5715, "step": 4534 }, { "epoch": 11.0880195599022, "grad_norm": 0.1580199215370195, "learning_rate": 7.932103969459018e-07, "loss": 0.3829, "step": 4535 }, { "epoch": 11.09046454767726, "grad_norm": 0.3242491876617342, "learning_rate": 7.922752045370457e-07, "loss": 0.4256, "step": 4536 }, { "epoch": 11.092909535452323, "grad_norm": 0.311899012830726, "learning_rate": 7.913404599327224e-07, "loss": 0.4203, "step": 4537 }, { "epoch": 11.095354523227384, "grad_norm": 0.28490314705409886, "learning_rate": 7.904061633780438e-07, "loss": 0.3744, "step": 4538 }, { "epoch": 11.097799511002446, "grad_norm": 0.2663349059105018, "learning_rate": 7.894723151180053e-07, "loss": 0.6135, "step": 4539 }, { "epoch": 11.100244498777506, "grad_norm": 0.189176489976191, "learning_rate": 7.88538915397481e-07, "loss": 0.5604, "step": 4540 }, { "epoch": 11.102689486552567, "grad_norm": 0.4524506732992623, "learning_rate": 7.876059644612307e-07, "loss": 0.3515, "step": 4541 }, { "epoch": 11.10513447432763, "grad_norm": 0.1709543263532299, "learning_rate": 7.866734625538966e-07, "loss": 0.6789, "step": 4542 }, { "epoch": 11.10757946210269, "grad_norm": 0.18420078170250767, "learning_rate": 7.857414099199995e-07, "loss": 0.5494, "step": 4543 }, { "epoch": 11.11002444987775, "grad_norm": 0.2640621849292725, "learning_rate": 7.848098068039478e-07, "loss": 0.5343, "step": 4544 }, { "epoch": 11.112469437652813, "grad_norm": 0.2338697513582221, "learning_rate": 7.83878653450027e-07, "loss": 0.5428, "step": 4545 }, { "epoch": 11.114914425427873, "grad_norm": 0.2335547888613717, "learning_rate": 7.829479501024088e-07, "loss": 0.8721, "step": 4546 }, { "epoch": 11.117359413202934, "grad_norm": 0.1675602926231691, "learning_rate": 7.820176970051424e-07, "loss": 0.5443, "step": 4547 }, { "epoch": 11.119804400977996, "grad_norm": 0.23251292590474068, "learning_rate": 7.810878944021629e-07, "loss": 0.3919, "step": 4548 }, { "epoch": 11.122249388753056, "grad_norm": 0.27223448323065885, "learning_rate": 7.801585425372868e-07, "loss": 0.4393, "step": 4549 }, { "epoch": 11.124694376528117, "grad_norm": 0.27283738069206576, "learning_rate": 7.792296416542094e-07, "loss": 0.5953, "step": 4550 }, { "epoch": 11.12713936430318, "grad_norm": 0.20335304915968613, "learning_rate": 7.783011919965109e-07, "loss": 0.4991, "step": 4551 }, { "epoch": 11.12958435207824, "grad_norm": 0.4183843136386102, "learning_rate": 7.773731938076526e-07, "loss": 0.4945, "step": 4552 }, { "epoch": 11.1320293398533, "grad_norm": 0.2660458440355159, "learning_rate": 7.764456473309745e-07, "loss": 0.6157, "step": 4553 }, { "epoch": 11.134474327628363, "grad_norm": 0.3907633924130284, "learning_rate": 7.75518552809704e-07, "loss": 0.3833, "step": 4554 }, { "epoch": 11.136919315403423, "grad_norm": 0.19566180334377198, "learning_rate": 7.745919104869437e-07, "loss": 0.7381, "step": 4555 }, { "epoch": 11.139364303178484, "grad_norm": 0.22549769504219344, "learning_rate": 7.736657206056822e-07, "loss": 0.6883, "step": 4556 }, { "epoch": 11.141809290953546, "grad_norm": 0.2620485949303981, "learning_rate": 7.727399834087862e-07, "loss": 0.6712, "step": 4557 }, { "epoch": 11.144254278728607, "grad_norm": 0.3061620301986605, "learning_rate": 7.718146991390061e-07, "loss": 0.4059, "step": 4558 }, { "epoch": 11.146699266503667, "grad_norm": 0.2795302981015312, "learning_rate": 7.708898680389728e-07, "loss": 0.4525, "step": 4559 }, { "epoch": 11.14914425427873, "grad_norm": 0.3161021901526408, "learning_rate": 7.699654903511971e-07, "loss": 0.4968, "step": 4560 }, { "epoch": 11.15158924205379, "grad_norm": 0.2590750742302967, "learning_rate": 7.690415663180728e-07, "loss": 0.3749, "step": 4561 }, { "epoch": 11.15403422982885, "grad_norm": 0.3045676287388493, "learning_rate": 7.681180961818746e-07, "loss": 0.6718, "step": 4562 }, { "epoch": 11.156479217603913, "grad_norm": 0.18804819154730004, "learning_rate": 7.671950801847558e-07, "loss": 0.4959, "step": 4563 }, { "epoch": 11.158924205378973, "grad_norm": 0.41019432574228276, "learning_rate": 7.662725185687536e-07, "loss": 0.8298, "step": 4564 }, { "epoch": 11.161369193154034, "grad_norm": 0.1746370030024984, "learning_rate": 7.65350411575784e-07, "loss": 0.3971, "step": 4565 }, { "epoch": 11.163814180929096, "grad_norm": 0.2810986144694296, "learning_rate": 7.64428759447646e-07, "loss": 0.5553, "step": 4566 }, { "epoch": 11.166259168704157, "grad_norm": 0.19058700712436713, "learning_rate": 7.63507562426016e-07, "loss": 0.8493, "step": 4567 }, { "epoch": 11.168704156479217, "grad_norm": 0.21990978648688894, "learning_rate": 7.625868207524539e-07, "loss": 0.8012, "step": 4568 }, { "epoch": 11.17114914425428, "grad_norm": 0.3712391035753783, "learning_rate": 7.616665346683999e-07, "loss": 0.4756, "step": 4569 }, { "epoch": 11.17359413202934, "grad_norm": 0.21110902137737841, "learning_rate": 7.607467044151726e-07, "loss": 0.5216, "step": 4570 }, { "epoch": 11.1760391198044, "grad_norm": 0.2148089662719569, "learning_rate": 7.598273302339731e-07, "loss": 0.45, "step": 4571 }, { "epoch": 11.178484107579463, "grad_norm": 0.18759823932888134, "learning_rate": 7.589084123658836e-07, "loss": 0.4838, "step": 4572 }, { "epoch": 11.180929095354523, "grad_norm": 0.23198150183659141, "learning_rate": 7.579899510518635e-07, "loss": 0.4661, "step": 4573 }, { "epoch": 11.183374083129584, "grad_norm": 0.3157374918821066, "learning_rate": 7.570719465327559e-07, "loss": 0.4142, "step": 4574 }, { "epoch": 11.185819070904646, "grad_norm": 0.287021901746644, "learning_rate": 7.561543990492803e-07, "loss": 0.36, "step": 4575 }, { "epoch": 11.188264058679707, "grad_norm": 0.3097922937251926, "learning_rate": 7.55237308842042e-07, "loss": 0.5866, "step": 4576 }, { "epoch": 11.190709046454767, "grad_norm": 0.27634681812380657, "learning_rate": 7.543206761515204e-07, "loss": 0.4172, "step": 4577 }, { "epoch": 11.19315403422983, "grad_norm": 0.22754321989399273, "learning_rate": 7.534045012180785e-07, "loss": 0.3399, "step": 4578 }, { "epoch": 11.19559902200489, "grad_norm": 0.36061130376252126, "learning_rate": 7.524887842819592e-07, "loss": 0.6007, "step": 4579 }, { "epoch": 11.19804400977995, "grad_norm": 0.25900358161474, "learning_rate": 7.515735255832829e-07, "loss": 0.5833, "step": 4580 }, { "epoch": 11.200488997555013, "grad_norm": 0.2408380938861335, "learning_rate": 7.50658725362052e-07, "loss": 0.3851, "step": 4581 }, { "epoch": 11.202933985330073, "grad_norm": 0.1824988661988685, "learning_rate": 7.497443838581489e-07, "loss": 0.4622, "step": 4582 }, { "epoch": 11.205378973105134, "grad_norm": 0.22641269223992344, "learning_rate": 7.488305013113331e-07, "loss": 0.4177, "step": 4583 }, { "epoch": 11.207823960880196, "grad_norm": 0.366831924966499, "learning_rate": 7.479170779612477e-07, "loss": 0.5357, "step": 4584 }, { "epoch": 11.210268948655257, "grad_norm": 0.23934598266249307, "learning_rate": 7.47004114047411e-07, "loss": 0.451, "step": 4585 }, { "epoch": 11.212713936430317, "grad_norm": 0.3614203778800607, "learning_rate": 7.460916098092241e-07, "loss": 0.6694, "step": 4586 }, { "epoch": 11.21515892420538, "grad_norm": 0.23226587465727236, "learning_rate": 7.451795654859664e-07, "loss": 0.6513, "step": 4587 }, { "epoch": 11.21760391198044, "grad_norm": 0.4845579284545479, "learning_rate": 7.442679813167969e-07, "loss": 0.5452, "step": 4588 }, { "epoch": 11.2200488997555, "grad_norm": 0.4496858535445958, "learning_rate": 7.433568575407546e-07, "loss": 0.8544, "step": 4589 }, { "epoch": 11.222493887530563, "grad_norm": 0.2972638589664434, "learning_rate": 7.424461943967557e-07, "loss": 0.7148, "step": 4590 }, { "epoch": 11.224938875305623, "grad_norm": 0.3927954481194539, "learning_rate": 7.415359921235971e-07, "loss": 0.5038, "step": 4591 }, { "epoch": 11.227383863080684, "grad_norm": 0.37161589837202824, "learning_rate": 7.40626250959956e-07, "loss": 0.5381, "step": 4592 }, { "epoch": 11.229828850855746, "grad_norm": 0.22740660239650776, "learning_rate": 7.397169711443853e-07, "loss": 0.3818, "step": 4593 }, { "epoch": 11.232273838630807, "grad_norm": 0.2667679299181016, "learning_rate": 7.38808152915321e-07, "loss": 0.8382, "step": 4594 }, { "epoch": 11.234718826405867, "grad_norm": 0.16583011643121837, "learning_rate": 7.378997965110743e-07, "loss": 0.6774, "step": 4595 }, { "epoch": 11.23716381418093, "grad_norm": 0.3230631245366838, "learning_rate": 7.36991902169838e-07, "loss": 0.71, "step": 4596 }, { "epoch": 11.23960880195599, "grad_norm": 0.2062874812711294, "learning_rate": 7.360844701296824e-07, "loss": 0.8562, "step": 4597 }, { "epoch": 11.24205378973105, "grad_norm": 0.23546310601048928, "learning_rate": 7.351775006285572e-07, "loss": 0.5319, "step": 4598 }, { "epoch": 11.244498777506113, "grad_norm": 0.2568574051148602, "learning_rate": 7.342709939042913e-07, "loss": 0.4844, "step": 4599 }, { "epoch": 11.246943765281173, "grad_norm": 0.2543677213404928, "learning_rate": 7.333649501945902e-07, "loss": 1.1877, "step": 4600 }, { "epoch": 11.249388753056234, "grad_norm": 0.22554058228087212, "learning_rate": 7.324593697370397e-07, "loss": 0.4854, "step": 4601 }, { "epoch": 11.251833740831296, "grad_norm": 0.19813204554163338, "learning_rate": 7.315542527691047e-07, "loss": 0.6132, "step": 4602 }, { "epoch": 11.254278728606357, "grad_norm": 0.24707319146079448, "learning_rate": 7.306495995281256e-07, "loss": 0.5403, "step": 4603 }, { "epoch": 11.256723716381417, "grad_norm": 0.2126566510122896, "learning_rate": 7.297454102513257e-07, "loss": 0.6464, "step": 4604 }, { "epoch": 11.25916870415648, "grad_norm": 0.23557256285424386, "learning_rate": 7.288416851758018e-07, "loss": 0.4614, "step": 4605 }, { "epoch": 11.26161369193154, "grad_norm": 0.25545575532357595, "learning_rate": 7.279384245385321e-07, "loss": 1.0356, "step": 4606 }, { "epoch": 11.2640586797066, "grad_norm": 0.2558023841087448, "learning_rate": 7.270356285763724e-07, "loss": 0.6308, "step": 4607 }, { "epoch": 11.266503667481663, "grad_norm": 0.3241763710407807, "learning_rate": 7.261332975260568e-07, "loss": 0.7364, "step": 4608 }, { "epoch": 11.268948655256724, "grad_norm": 0.3152978795533111, "learning_rate": 7.252314316241974e-07, "loss": 0.4471, "step": 4609 }, { "epoch": 11.271393643031784, "grad_norm": 0.34502368141596973, "learning_rate": 7.243300311072826e-07, "loss": 0.6018, "step": 4610 }, { "epoch": 11.273838630806846, "grad_norm": 0.27893116550623065, "learning_rate": 7.234290962116814e-07, "loss": 0.3969, "step": 4611 }, { "epoch": 11.276283618581907, "grad_norm": 0.22216832200851147, "learning_rate": 7.225286271736398e-07, "loss": 0.4138, "step": 4612 }, { "epoch": 11.278728606356967, "grad_norm": 0.33614711576825074, "learning_rate": 7.216286242292806e-07, "loss": 0.4885, "step": 4613 }, { "epoch": 11.28117359413203, "grad_norm": 0.40238486488548003, "learning_rate": 7.207290876146058e-07, "loss": 0.5064, "step": 4614 }, { "epoch": 11.28361858190709, "grad_norm": 0.28448362392298293, "learning_rate": 7.198300175654935e-07, "loss": 0.9145, "step": 4615 }, { "epoch": 11.28606356968215, "grad_norm": 0.18925727140961157, "learning_rate": 7.189314143177012e-07, "loss": 0.5401, "step": 4616 }, { "epoch": 11.288508557457213, "grad_norm": 0.2836482339860002, "learning_rate": 7.180332781068639e-07, "loss": 1.3362, "step": 4617 }, { "epoch": 11.290953545232274, "grad_norm": 0.2548240971359618, "learning_rate": 7.171356091684911e-07, "loss": 0.3998, "step": 4618 }, { "epoch": 11.293398533007334, "grad_norm": 0.21889221573432732, "learning_rate": 7.162384077379756e-07, "loss": 0.8033, "step": 4619 }, { "epoch": 11.295843520782396, "grad_norm": 0.23813587884859602, "learning_rate": 7.153416740505815e-07, "loss": 0.4348, "step": 4620 }, { "epoch": 11.298288508557457, "grad_norm": 0.20288273627423753, "learning_rate": 7.144454083414534e-07, "loss": 0.4553, "step": 4621 }, { "epoch": 11.30073349633252, "grad_norm": 0.16320536097192173, "learning_rate": 7.135496108456142e-07, "loss": 0.6769, "step": 4622 }, { "epoch": 11.30317848410758, "grad_norm": 0.2541503652407237, "learning_rate": 7.126542817979601e-07, "loss": 0.403, "step": 4623 }, { "epoch": 11.30562347188264, "grad_norm": 0.337534172544725, "learning_rate": 7.117594214332693e-07, "loss": 0.3935, "step": 4624 }, { "epoch": 11.3080684596577, "grad_norm": 0.3431093853038867, "learning_rate": 7.108650299861924e-07, "loss": 0.3915, "step": 4625 }, { "epoch": 11.310513447432763, "grad_norm": 0.22907634597480162, "learning_rate": 7.099711076912605e-07, "loss": 0.4055, "step": 4626 }, { "epoch": 11.312958435207824, "grad_norm": 0.29437520128522804, "learning_rate": 7.090776547828812e-07, "loss": 0.4124, "step": 4627 }, { "epoch": 11.315403422982886, "grad_norm": 0.23962019086723044, "learning_rate": 7.081846714953366e-07, "loss": 0.9317, "step": 4628 }, { "epoch": 11.317848410757946, "grad_norm": 0.2455589771839322, "learning_rate": 7.072921580627884e-07, "loss": 0.519, "step": 4629 }, { "epoch": 11.320293398533007, "grad_norm": 0.2324623865475341, "learning_rate": 7.064001147192737e-07, "loss": 0.4346, "step": 4630 }, { "epoch": 11.32273838630807, "grad_norm": 0.21736370872577693, "learning_rate": 7.05508541698707e-07, "loss": 0.9281, "step": 4631 }, { "epoch": 11.32518337408313, "grad_norm": 0.21464955540580807, "learning_rate": 7.046174392348798e-07, "loss": 0.4252, "step": 4632 }, { "epoch": 11.32762836185819, "grad_norm": 0.4279262118760031, "learning_rate": 7.037268075614579e-07, "loss": 0.3621, "step": 4633 }, { "epoch": 11.330073349633253, "grad_norm": 0.21827093917612006, "learning_rate": 7.028366469119872e-07, "loss": 0.5207, "step": 4634 }, { "epoch": 11.332518337408313, "grad_norm": 0.2302413733245278, "learning_rate": 7.019469575198861e-07, "loss": 0.4252, "step": 4635 }, { "epoch": 11.334963325183374, "grad_norm": 0.22742970545840108, "learning_rate": 7.010577396184529e-07, "loss": 0.4324, "step": 4636 }, { "epoch": 11.337408312958436, "grad_norm": 0.22898035712822712, "learning_rate": 7.001689934408612e-07, "loss": 0.7259, "step": 4637 }, { "epoch": 11.339853300733497, "grad_norm": 0.3442097042119856, "learning_rate": 6.992807192201595e-07, "loss": 0.4214, "step": 4638 }, { "epoch": 11.342298288508557, "grad_norm": 0.24909817448397711, "learning_rate": 6.98392917189274e-07, "loss": 0.7714, "step": 4639 }, { "epoch": 11.34474327628362, "grad_norm": 0.3062824989924266, "learning_rate": 6.975055875810066e-07, "loss": 0.3721, "step": 4640 }, { "epoch": 11.34718826405868, "grad_norm": 0.296249605204892, "learning_rate": 6.966187306280362e-07, "loss": 1.3165, "step": 4641 }, { "epoch": 11.34963325183374, "grad_norm": 0.2004623556421978, "learning_rate": 6.95732346562917e-07, "loss": 0.511, "step": 4642 }, { "epoch": 11.352078239608803, "grad_norm": 0.24809297969950717, "learning_rate": 6.948464356180778e-07, "loss": 0.445, "step": 4643 }, { "epoch": 11.354523227383863, "grad_norm": 0.2776184966045638, "learning_rate": 6.939609980258266e-07, "loss": 0.9459, "step": 4644 }, { "epoch": 11.356968215158924, "grad_norm": 0.2947671130421909, "learning_rate": 6.930760340183434e-07, "loss": 0.407, "step": 4645 }, { "epoch": 11.359413202933986, "grad_norm": 0.19982728558159243, "learning_rate": 6.921915438276874e-07, "loss": 0.4048, "step": 4646 }, { "epoch": 11.361858190709047, "grad_norm": 0.36442212008201647, "learning_rate": 6.913075276857922e-07, "loss": 0.438, "step": 4647 }, { "epoch": 11.364303178484107, "grad_norm": 0.21591582410523855, "learning_rate": 6.904239858244663e-07, "loss": 0.9395, "step": 4648 }, { "epoch": 11.36674816625917, "grad_norm": 0.27053122103327965, "learning_rate": 6.895409184753948e-07, "loss": 0.3216, "step": 4649 }, { "epoch": 11.36919315403423, "grad_norm": 0.24705880093966814, "learning_rate": 6.886583258701382e-07, "loss": 0.4687, "step": 4650 }, { "epoch": 11.37163814180929, "grad_norm": 0.24431842255738964, "learning_rate": 6.877762082401329e-07, "loss": 0.5262, "step": 4651 }, { "epoch": 11.374083129584353, "grad_norm": 0.16442323277314677, "learning_rate": 6.868945658166907e-07, "loss": 0.2299, "step": 4652 }, { "epoch": 11.376528117359413, "grad_norm": 0.2774048376497834, "learning_rate": 6.860133988309972e-07, "loss": 0.3688, "step": 4653 }, { "epoch": 11.378973105134474, "grad_norm": 0.3216572885423358, "learning_rate": 6.851327075141156e-07, "loss": 0.3753, "step": 4654 }, { "epoch": 11.381418092909536, "grad_norm": 0.3619002507760992, "learning_rate": 6.842524920969823e-07, "loss": 0.3418, "step": 4655 }, { "epoch": 11.383863080684597, "grad_norm": 0.2097951453983877, "learning_rate": 6.833727528104106e-07, "loss": 0.3837, "step": 4656 }, { "epoch": 11.386308068459657, "grad_norm": 0.30306011045090875, "learning_rate": 6.824934898850888e-07, "loss": 0.3994, "step": 4657 }, { "epoch": 11.38875305623472, "grad_norm": 0.2832519211499832, "learning_rate": 6.816147035515786e-07, "loss": 0.7829, "step": 4658 }, { "epoch": 11.39119804400978, "grad_norm": 0.3410385734261241, "learning_rate": 6.807363940403183e-07, "loss": 1.1079, "step": 4659 }, { "epoch": 11.39364303178484, "grad_norm": 0.2533995062679686, "learning_rate": 6.798585615816217e-07, "loss": 0.5504, "step": 4660 }, { "epoch": 11.396088019559903, "grad_norm": 0.24589591089168408, "learning_rate": 6.789812064056745e-07, "loss": 0.7333, "step": 4661 }, { "epoch": 11.398533007334963, "grad_norm": 0.30631469737648204, "learning_rate": 6.781043287425417e-07, "loss": 0.4507, "step": 4662 }, { "epoch": 11.400977995110024, "grad_norm": 0.35631632759489473, "learning_rate": 6.772279288221589e-07, "loss": 0.5311, "step": 4663 }, { "epoch": 11.403422982885086, "grad_norm": 0.20801036673709464, "learning_rate": 6.763520068743399e-07, "loss": 0.3681, "step": 4664 }, { "epoch": 11.405867970660147, "grad_norm": 0.43311386587404976, "learning_rate": 6.754765631287696e-07, "loss": 0.6555, "step": 4665 }, { "epoch": 11.408312958435207, "grad_norm": 0.31063428953038436, "learning_rate": 6.746015978150103e-07, "loss": 0.7819, "step": 4666 }, { "epoch": 11.41075794621027, "grad_norm": 0.26905037438559565, "learning_rate": 6.737271111624988e-07, "loss": 0.7104, "step": 4667 }, { "epoch": 11.41320293398533, "grad_norm": 0.20222588616903778, "learning_rate": 6.728531034005436e-07, "loss": 0.6638, "step": 4668 }, { "epoch": 11.41564792176039, "grad_norm": 0.35169031899850667, "learning_rate": 6.719795747583307e-07, "loss": 0.6985, "step": 4669 }, { "epoch": 11.418092909535453, "grad_norm": 0.30234314040246196, "learning_rate": 6.711065254649202e-07, "loss": 0.4908, "step": 4670 }, { "epoch": 11.420537897310513, "grad_norm": 0.3508136750093273, "learning_rate": 6.702339557492427e-07, "loss": 0.6284, "step": 4671 }, { "epoch": 11.422982885085574, "grad_norm": 0.22186617079664653, "learning_rate": 6.693618658401097e-07, "loss": 0.4307, "step": 4672 }, { "epoch": 11.425427872860636, "grad_norm": 0.1945600963069221, "learning_rate": 6.684902559662002e-07, "loss": 0.4137, "step": 4673 }, { "epoch": 11.427872860635697, "grad_norm": 0.35684660982244654, "learning_rate": 6.676191263560722e-07, "loss": 0.8241, "step": 4674 }, { "epoch": 11.430317848410757, "grad_norm": 0.21395545755212267, "learning_rate": 6.667484772381544e-07, "loss": 0.8457, "step": 4675 }, { "epoch": 11.43276283618582, "grad_norm": 0.3403706276315063, "learning_rate": 6.658783088407511e-07, "loss": 0.5911, "step": 4676 }, { "epoch": 11.43520782396088, "grad_norm": 0.255455419115446, "learning_rate": 6.65008621392042e-07, "loss": 0.5204, "step": 4677 }, { "epoch": 11.43765281173594, "grad_norm": 0.1782976606358615, "learning_rate": 6.641394151200767e-07, "loss": 0.4158, "step": 4678 }, { "epoch": 11.440097799511003, "grad_norm": 0.3358488155050002, "learning_rate": 6.632706902527824e-07, "loss": 0.4971, "step": 4679 }, { "epoch": 11.442542787286063, "grad_norm": 0.36234215206696213, "learning_rate": 6.624024470179591e-07, "loss": 0.5798, "step": 4680 }, { "epoch": 11.444987775061124, "grad_norm": 0.21868132752420108, "learning_rate": 6.615346856432783e-07, "loss": 0.5659, "step": 4681 }, { "epoch": 11.447432762836186, "grad_norm": 0.28577584058753497, "learning_rate": 6.606674063562882e-07, "loss": 0.4, "step": 4682 }, { "epoch": 11.449877750611247, "grad_norm": 0.32667491174364294, "learning_rate": 6.598006093844086e-07, "loss": 0.5238, "step": 4683 }, { "epoch": 11.452322738386307, "grad_norm": 0.2761867026202508, "learning_rate": 6.589342949549343e-07, "loss": 0.3843, "step": 4684 }, { "epoch": 11.45476772616137, "grad_norm": 0.35971044627052867, "learning_rate": 6.58068463295033e-07, "loss": 0.5836, "step": 4685 }, { "epoch": 11.45721271393643, "grad_norm": 0.2441430549653803, "learning_rate": 6.572031146317443e-07, "loss": 0.4677, "step": 4686 }, { "epoch": 11.45965770171149, "grad_norm": 0.2723809117908462, "learning_rate": 6.563382491919837e-07, "loss": 0.8788, "step": 4687 }, { "epoch": 11.462102689486553, "grad_norm": 0.24840652022406123, "learning_rate": 6.554738672025374e-07, "loss": 0.379, "step": 4688 }, { "epoch": 11.464547677261614, "grad_norm": 0.2684965237373586, "learning_rate": 6.546099688900668e-07, "loss": 0.4071, "step": 4689 }, { "epoch": 11.466992665036674, "grad_norm": 0.2924512040150079, "learning_rate": 6.537465544811064e-07, "loss": 0.7088, "step": 4690 }, { "epoch": 11.469437652811736, "grad_norm": 0.2759319722698551, "learning_rate": 6.528836242020623e-07, "loss": 0.4256, "step": 4691 }, { "epoch": 11.471882640586797, "grad_norm": 0.22524713931410068, "learning_rate": 6.520211782792154e-07, "loss": 0.4561, "step": 4692 }, { "epoch": 11.474327628361857, "grad_norm": 0.33794632080087766, "learning_rate": 6.51159216938717e-07, "loss": 0.6215, "step": 4693 }, { "epoch": 11.47677261613692, "grad_norm": 0.24424887888280425, "learning_rate": 6.502977404065952e-07, "loss": 0.5103, "step": 4694 }, { "epoch": 11.47921760391198, "grad_norm": 0.26569260831678365, "learning_rate": 6.494367489087489e-07, "loss": 0.7476, "step": 4695 }, { "epoch": 11.48166259168704, "grad_norm": 0.2080644996543301, "learning_rate": 6.48576242670948e-07, "loss": 0.7307, "step": 4696 }, { "epoch": 11.484107579462103, "grad_norm": 0.25665579223775065, "learning_rate": 6.477162219188385e-07, "loss": 0.6652, "step": 4697 }, { "epoch": 11.486552567237164, "grad_norm": 0.23934934641447048, "learning_rate": 6.468566868779366e-07, "loss": 0.3041, "step": 4698 }, { "epoch": 11.488997555012224, "grad_norm": 0.30332472227421503, "learning_rate": 6.459976377736321e-07, "loss": 0.3868, "step": 4699 }, { "epoch": 11.491442542787286, "grad_norm": 0.19257547416021592, "learning_rate": 6.451390748311884e-07, "loss": 0.4148, "step": 4700 }, { "epoch": 11.493887530562347, "grad_norm": 0.33737341908192064, "learning_rate": 6.442809982757389e-07, "loss": 0.4345, "step": 4701 }, { "epoch": 11.496332518337407, "grad_norm": 0.19277165576829927, "learning_rate": 6.434234083322924e-07, "loss": 0.5699, "step": 4702 }, { "epoch": 11.49877750611247, "grad_norm": 0.343635346378364, "learning_rate": 6.425663052257269e-07, "loss": 0.7501, "step": 4703 }, { "epoch": 11.50122249388753, "grad_norm": 0.2581899058828507, "learning_rate": 6.417096891807948e-07, "loss": 0.703, "step": 4704 }, { "epoch": 11.503667481662593, "grad_norm": 0.20641806504531549, "learning_rate": 6.408535604221222e-07, "loss": 0.6638, "step": 4705 }, { "epoch": 11.506112469437653, "grad_norm": 0.24382034768639388, "learning_rate": 6.39997919174204e-07, "loss": 0.4442, "step": 4706 }, { "epoch": 11.508557457212714, "grad_norm": 0.3992288951871595, "learning_rate": 6.391427656614099e-07, "loss": 0.4068, "step": 4707 }, { "epoch": 11.511002444987774, "grad_norm": 0.34332395298503443, "learning_rate": 6.382881001079797e-07, "loss": 0.3232, "step": 4708 }, { "epoch": 11.513447432762836, "grad_norm": 0.30555615997850294, "learning_rate": 6.374339227380266e-07, "loss": 1.1958, "step": 4709 }, { "epoch": 11.515892420537897, "grad_norm": 0.2774514108268633, "learning_rate": 6.365802337755364e-07, "loss": 0.3904, "step": 4710 }, { "epoch": 11.51833740831296, "grad_norm": 0.2361460885277631, "learning_rate": 6.357270334443646e-07, "loss": 1.021, "step": 4711 }, { "epoch": 11.52078239608802, "grad_norm": 0.2137875360675618, "learning_rate": 6.34874321968241e-07, "loss": 0.6281, "step": 4712 }, { "epoch": 11.52322738386308, "grad_norm": 0.20344562746111486, "learning_rate": 6.340220995707647e-07, "loss": 0.4476, "step": 4713 }, { "epoch": 11.525672371638143, "grad_norm": 0.2024586152036081, "learning_rate": 6.331703664754079e-07, "loss": 0.3174, "step": 4714 }, { "epoch": 11.528117359413203, "grad_norm": 0.30410610014554074, "learning_rate": 6.323191229055167e-07, "loss": 0.5601, "step": 4715 }, { "epoch": 11.530562347188264, "grad_norm": 0.3075859124888534, "learning_rate": 6.314683690843049e-07, "loss": 0.472, "step": 4716 }, { "epoch": 11.533007334963326, "grad_norm": 0.29315517799447915, "learning_rate": 6.306181052348603e-07, "loss": 0.6899, "step": 4717 }, { "epoch": 11.535452322738386, "grad_norm": 0.24745814130760596, "learning_rate": 6.297683315801406e-07, "loss": 0.8657, "step": 4718 }, { "epoch": 11.537897310513447, "grad_norm": 0.3148533671283106, "learning_rate": 6.289190483429763e-07, "loss": 1.0306, "step": 4719 }, { "epoch": 11.54034229828851, "grad_norm": 0.24841765765600288, "learning_rate": 6.280702557460702e-07, "loss": 0.428, "step": 4720 }, { "epoch": 11.54278728606357, "grad_norm": 0.2936211577818367, "learning_rate": 6.272219540119934e-07, "loss": 0.8074, "step": 4721 }, { "epoch": 11.54523227383863, "grad_norm": 0.3959761879911899, "learning_rate": 6.263741433631912e-07, "loss": 0.3624, "step": 4722 }, { "epoch": 11.547677261613693, "grad_norm": 0.22297246706541055, "learning_rate": 6.255268240219777e-07, "loss": 0.3807, "step": 4723 }, { "epoch": 11.550122249388753, "grad_norm": 0.20407099823515282, "learning_rate": 6.246799962105402e-07, "loss": 0.6452, "step": 4724 }, { "epoch": 11.552567237163814, "grad_norm": 0.3351603541579998, "learning_rate": 6.238336601509365e-07, "loss": 1.0374, "step": 4725 }, { "epoch": 11.555012224938876, "grad_norm": 0.2793961725256349, "learning_rate": 6.229878160650951e-07, "loss": 0.8236, "step": 4726 }, { "epoch": 11.557457212713937, "grad_norm": 0.270311613639359, "learning_rate": 6.221424641748164e-07, "loss": 0.6993, "step": 4727 }, { "epoch": 11.559902200488997, "grad_norm": 0.28054294301474886, "learning_rate": 6.212976047017693e-07, "loss": 0.6052, "step": 4728 }, { "epoch": 11.56234718826406, "grad_norm": 0.21911666361189305, "learning_rate": 6.204532378674969e-07, "loss": 0.4869, "step": 4729 }, { "epoch": 11.56479217603912, "grad_norm": 0.21644075770739324, "learning_rate": 6.196093638934111e-07, "loss": 0.4392, "step": 4730 }, { "epoch": 11.56723716381418, "grad_norm": 0.22882539990886566, "learning_rate": 6.187659830007945e-07, "loss": 0.4234, "step": 4731 }, { "epoch": 11.569682151589243, "grad_norm": 0.21863418496022324, "learning_rate": 6.179230954108015e-07, "loss": 0.6179, "step": 4732 }, { "epoch": 11.572127139364303, "grad_norm": 0.19647658553054734, "learning_rate": 6.170807013444558e-07, "loss": 0.6399, "step": 4733 }, { "epoch": 11.574572127139364, "grad_norm": 0.34102751466775194, "learning_rate": 6.162388010226525e-07, "loss": 0.7511, "step": 4734 }, { "epoch": 11.577017114914426, "grad_norm": 0.2980400506514369, "learning_rate": 6.153973946661584e-07, "loss": 0.5991, "step": 4735 }, { "epoch": 11.579462102689487, "grad_norm": 0.316686584047014, "learning_rate": 6.145564824956071e-07, "loss": 1.173, "step": 4736 }, { "epoch": 11.581907090464547, "grad_norm": 0.191132278577256, "learning_rate": 6.13716064731508e-07, "loss": 0.3644, "step": 4737 }, { "epoch": 11.58435207823961, "grad_norm": 0.25096145767759465, "learning_rate": 6.128761415942353e-07, "loss": 0.6492, "step": 4738 }, { "epoch": 11.58679706601467, "grad_norm": 0.2940977956718163, "learning_rate": 6.120367133040373e-07, "loss": 0.3848, "step": 4739 }, { "epoch": 11.58924205378973, "grad_norm": 0.3474400954389469, "learning_rate": 6.111977800810318e-07, "loss": 0.3744, "step": 4740 }, { "epoch": 11.591687041564793, "grad_norm": 0.25241867220980474, "learning_rate": 6.103593421452051e-07, "loss": 0.8455, "step": 4741 }, { "epoch": 11.594132029339853, "grad_norm": 0.2587972321140265, "learning_rate": 6.095213997164151e-07, "loss": 0.7991, "step": 4742 }, { "epoch": 11.596577017114914, "grad_norm": 0.28798146351949966, "learning_rate": 6.086839530143907e-07, "loss": 0.8349, "step": 4743 }, { "epoch": 11.599022004889976, "grad_norm": 0.15825453467876194, "learning_rate": 6.078470022587277e-07, "loss": 0.4509, "step": 4744 }, { "epoch": 11.601466992665037, "grad_norm": 0.2879051956179973, "learning_rate": 6.070105476688954e-07, "loss": 0.5089, "step": 4745 }, { "epoch": 11.603911980440097, "grad_norm": 0.25053076704622995, "learning_rate": 6.061745894642298e-07, "loss": 0.5871, "step": 4746 }, { "epoch": 11.60635696821516, "grad_norm": 0.2720979781311066, "learning_rate": 6.053391278639395e-07, "loss": 0.4148, "step": 4747 }, { "epoch": 11.60880195599022, "grad_norm": 0.16004316352268858, "learning_rate": 6.04504163087101e-07, "loss": 0.3253, "step": 4748 }, { "epoch": 11.61124694376528, "grad_norm": 0.16391112069988056, "learning_rate": 6.036696953526613e-07, "loss": 0.3976, "step": 4749 }, { "epoch": 11.613691931540343, "grad_norm": 0.24942136166954784, "learning_rate": 6.028357248794381e-07, "loss": 0.6889, "step": 4750 }, { "epoch": 11.616136919315403, "grad_norm": 0.33673629652739545, "learning_rate": 6.020022518861154e-07, "loss": 0.6074, "step": 4751 }, { "epoch": 11.618581907090464, "grad_norm": 0.28953643553549907, "learning_rate": 6.011692765912503e-07, "loss": 0.4941, "step": 4752 }, { "epoch": 11.621026894865526, "grad_norm": 0.32859882309217103, "learning_rate": 6.003367992132684e-07, "loss": 0.4491, "step": 4753 }, { "epoch": 11.623471882640587, "grad_norm": 0.28923243667055515, "learning_rate": 5.995048199704629e-07, "loss": 0.6522, "step": 4754 }, { "epoch": 11.625916870415647, "grad_norm": 0.28878511620955377, "learning_rate": 5.986733390809993e-07, "loss": 0.5271, "step": 4755 }, { "epoch": 11.62836185819071, "grad_norm": 0.30909941772357113, "learning_rate": 5.978423567629094e-07, "loss": 0.4535, "step": 4756 }, { "epoch": 11.63080684596577, "grad_norm": 0.2820366569868769, "learning_rate": 5.970118732340965e-07, "loss": 0.4451, "step": 4757 }, { "epoch": 11.63325183374083, "grad_norm": 0.37431597036083847, "learning_rate": 5.961818887123324e-07, "loss": 0.3818, "step": 4758 }, { "epoch": 11.635696821515893, "grad_norm": 0.2587991279452256, "learning_rate": 5.953524034152583e-07, "loss": 0.8937, "step": 4759 }, { "epoch": 11.638141809290953, "grad_norm": 0.39980339103324153, "learning_rate": 5.945234175603845e-07, "loss": 0.542, "step": 4760 }, { "epoch": 11.640586797066014, "grad_norm": 0.29578140274044096, "learning_rate": 5.936949313650888e-07, "loss": 0.5412, "step": 4761 }, { "epoch": 11.643031784841076, "grad_norm": 0.2630602395048049, "learning_rate": 5.928669450466204e-07, "loss": 0.7017, "step": 4762 }, { "epoch": 11.645476772616137, "grad_norm": 0.25776778687360297, "learning_rate": 5.920394588220962e-07, "loss": 0.3736, "step": 4763 }, { "epoch": 11.647921760391197, "grad_norm": 0.22098979537874971, "learning_rate": 5.912124729085012e-07, "loss": 0.848, "step": 4764 }, { "epoch": 11.65036674816626, "grad_norm": 0.23471283258667477, "learning_rate": 5.903859875226914e-07, "loss": 0.4224, "step": 4765 }, { "epoch": 11.65281173594132, "grad_norm": 0.24627112474030158, "learning_rate": 5.895600028813883e-07, "loss": 0.8884, "step": 4766 }, { "epoch": 11.65525672371638, "grad_norm": 0.3231945069244839, "learning_rate": 5.887345192011854e-07, "loss": 0.9674, "step": 4767 }, { "epoch": 11.657701711491443, "grad_norm": 0.28419382779795216, "learning_rate": 5.879095366985427e-07, "loss": 0.4535, "step": 4768 }, { "epoch": 11.660146699266503, "grad_norm": 0.29642344504321216, "learning_rate": 5.870850555897901e-07, "loss": 0.5306, "step": 4769 }, { "epoch": 11.662591687041564, "grad_norm": 0.34104366660124297, "learning_rate": 5.862610760911259e-07, "loss": 0.4235, "step": 4770 }, { "epoch": 11.665036674816626, "grad_norm": 0.3018306727180531, "learning_rate": 5.854375984186144e-07, "loss": 0.5957, "step": 4771 }, { "epoch": 11.667481662591687, "grad_norm": 0.26606144339349935, "learning_rate": 5.846146227881922e-07, "loss": 0.9826, "step": 4772 }, { "epoch": 11.669926650366747, "grad_norm": 0.24320805515066607, "learning_rate": 5.837921494156621e-07, "loss": 0.3986, "step": 4773 }, { "epoch": 11.67237163814181, "grad_norm": 0.25254769082832673, "learning_rate": 5.829701785166944e-07, "loss": 0.4094, "step": 4774 }, { "epoch": 11.67481662591687, "grad_norm": 0.28233453752064297, "learning_rate": 5.821487103068299e-07, "loss": 0.6972, "step": 4775 }, { "epoch": 11.67726161369193, "grad_norm": 0.2085014090624554, "learning_rate": 5.813277450014753e-07, "loss": 0.621, "step": 4776 }, { "epoch": 11.679706601466993, "grad_norm": 0.2643448475375848, "learning_rate": 5.80507282815907e-07, "loss": 0.3342, "step": 4777 }, { "epoch": 11.682151589242054, "grad_norm": 0.25277522450542955, "learning_rate": 5.796873239652695e-07, "loss": 0.4812, "step": 4778 }, { "epoch": 11.684596577017114, "grad_norm": 0.3323776488522753, "learning_rate": 5.788678686645729e-07, "loss": 0.4583, "step": 4779 }, { "epoch": 11.687041564792176, "grad_norm": 0.22873508873163986, "learning_rate": 5.780489171286999e-07, "loss": 0.8051, "step": 4780 }, { "epoch": 11.689486552567237, "grad_norm": 0.21825315811171078, "learning_rate": 5.772304695723962e-07, "loss": 0.7503, "step": 4781 }, { "epoch": 11.6919315403423, "grad_norm": 0.26811151089103613, "learning_rate": 5.764125262102782e-07, "loss": 0.4823, "step": 4782 }, { "epoch": 11.69437652811736, "grad_norm": 0.2798585541690888, "learning_rate": 5.7559508725683e-07, "loss": 0.9482, "step": 4783 }, { "epoch": 11.69682151589242, "grad_norm": 0.29651023612598965, "learning_rate": 5.747781529264016e-07, "loss": 0.4462, "step": 4784 }, { "epoch": 11.69926650366748, "grad_norm": 0.3061182296606879, "learning_rate": 5.739617234332132e-07, "loss": 0.4017, "step": 4785 }, { "epoch": 11.701711491442543, "grad_norm": 0.23505361088022259, "learning_rate": 5.731457989913497e-07, "loss": 0.7383, "step": 4786 }, { "epoch": 11.704156479217604, "grad_norm": 0.2510445089502352, "learning_rate": 5.723303798147662e-07, "loss": 0.5046, "step": 4787 }, { "epoch": 11.706601466992666, "grad_norm": 0.26416495690564706, "learning_rate": 5.715154661172845e-07, "loss": 0.6689, "step": 4788 }, { "epoch": 11.709046454767726, "grad_norm": 0.2989582467744812, "learning_rate": 5.707010581125921e-07, "loss": 0.4811, "step": 4789 }, { "epoch": 11.711491442542787, "grad_norm": 0.24730358018933668, "learning_rate": 5.69887156014248e-07, "loss": 0.7859, "step": 4790 }, { "epoch": 11.713936430317847, "grad_norm": 0.2535765138389413, "learning_rate": 5.690737600356735e-07, "loss": 0.7102, "step": 4791 }, { "epoch": 11.71638141809291, "grad_norm": 0.20786411735110066, "learning_rate": 5.68260870390161e-07, "loss": 0.5841, "step": 4792 }, { "epoch": 11.71882640586797, "grad_norm": 0.33021637082921645, "learning_rate": 5.67448487290869e-07, "loss": 0.74, "step": 4793 }, { "epoch": 11.721271393643033, "grad_norm": 0.3230352463070714, "learning_rate": 5.666366109508217e-07, "loss": 0.944, "step": 4794 }, { "epoch": 11.723716381418093, "grad_norm": 0.32951892122308185, "learning_rate": 5.65825241582913e-07, "loss": 0.6878, "step": 4795 }, { "epoch": 11.726161369193154, "grad_norm": 0.2660983270359221, "learning_rate": 5.650143793999014e-07, "loss": 0.8063, "step": 4796 }, { "epoch": 11.728606356968216, "grad_norm": 0.21419916204887143, "learning_rate": 5.642040246144137e-07, "loss": 0.4945, "step": 4797 }, { "epoch": 11.731051344743276, "grad_norm": 0.2159669160446207, "learning_rate": 5.633941774389446e-07, "loss": 0.3959, "step": 4798 }, { "epoch": 11.733496332518337, "grad_norm": 0.308496561972242, "learning_rate": 5.625848380858528e-07, "loss": 0.7344, "step": 4799 }, { "epoch": 11.7359413202934, "grad_norm": 0.3036914977884245, "learning_rate": 5.617760067673667e-07, "loss": 0.3874, "step": 4800 }, { "epoch": 11.73838630806846, "grad_norm": 0.3093700342797868, "learning_rate": 5.609676836955802e-07, "loss": 0.577, "step": 4801 }, { "epoch": 11.74083129584352, "grad_norm": 0.29741496865470646, "learning_rate": 5.601598690824542e-07, "loss": 0.506, "step": 4802 }, { "epoch": 11.743276283618583, "grad_norm": 0.21422546636304476, "learning_rate": 5.593525631398166e-07, "loss": 0.713, "step": 4803 }, { "epoch": 11.745721271393643, "grad_norm": 0.27688218833680933, "learning_rate": 5.585457660793606e-07, "loss": 0.5534, "step": 4804 }, { "epoch": 11.748166259168704, "grad_norm": 0.25090102609950776, "learning_rate": 5.57739478112648e-07, "loss": 0.4965, "step": 4805 }, { "epoch": 11.750611246943766, "grad_norm": 0.34169634769330615, "learning_rate": 5.569336994511043e-07, "loss": 0.3478, "step": 4806 }, { "epoch": 11.753056234718827, "grad_norm": 0.2391864794593845, "learning_rate": 5.561284303060241e-07, "loss": 0.6459, "step": 4807 }, { "epoch": 11.755501222493887, "grad_norm": 0.17319199062869503, "learning_rate": 5.553236708885682e-07, "loss": 0.5093, "step": 4808 }, { "epoch": 11.75794621026895, "grad_norm": 0.22751971750924863, "learning_rate": 5.545194214097613e-07, "loss": 0.6816, "step": 4809 }, { "epoch": 11.76039119804401, "grad_norm": 0.22590029508442067, "learning_rate": 5.537156820804967e-07, "loss": 0.4387, "step": 4810 }, { "epoch": 11.76283618581907, "grad_norm": 0.3519924473774395, "learning_rate": 5.529124531115335e-07, "loss": 0.7404, "step": 4811 }, { "epoch": 11.765281173594133, "grad_norm": 0.27736442192831967, "learning_rate": 5.521097347134965e-07, "loss": 0.3935, "step": 4812 }, { "epoch": 11.767726161369193, "grad_norm": 0.49202588078919873, "learning_rate": 5.513075270968779e-07, "loss": 0.5001, "step": 4813 }, { "epoch": 11.770171149144254, "grad_norm": 0.2547540689193078, "learning_rate": 5.505058304720329e-07, "loss": 0.5766, "step": 4814 }, { "epoch": 11.772616136919316, "grad_norm": 0.4144701496806104, "learning_rate": 5.497046450491866e-07, "loss": 0.6988, "step": 4815 }, { "epoch": 11.775061124694377, "grad_norm": 0.21293997537949624, "learning_rate": 5.489039710384264e-07, "loss": 0.6032, "step": 4816 }, { "epoch": 11.777506112469437, "grad_norm": 0.2870500018532013, "learning_rate": 5.481038086497084e-07, "loss": 0.4799, "step": 4817 }, { "epoch": 11.7799511002445, "grad_norm": 0.17725150164693032, "learning_rate": 5.473041580928543e-07, "loss": 0.6955, "step": 4818 }, { "epoch": 11.78239608801956, "grad_norm": 0.2820840681507961, "learning_rate": 5.465050195775488e-07, "loss": 0.9522, "step": 4819 }, { "epoch": 11.78484107579462, "grad_norm": 0.2731751228113001, "learning_rate": 5.457063933133455e-07, "loss": 0.9021, "step": 4820 }, { "epoch": 11.787286063569683, "grad_norm": 0.3451576462579562, "learning_rate": 5.449082795096627e-07, "loss": 0.4912, "step": 4821 }, { "epoch": 11.789731051344743, "grad_norm": 0.20933647964075164, "learning_rate": 5.441106783757827e-07, "loss": 0.4948, "step": 4822 }, { "epoch": 11.792176039119804, "grad_norm": 0.1720602283934634, "learning_rate": 5.433135901208572e-07, "loss": 0.3986, "step": 4823 }, { "epoch": 11.794621026894866, "grad_norm": 0.41036927141509577, "learning_rate": 5.425170149538986e-07, "loss": 0.6739, "step": 4824 }, { "epoch": 11.797066014669927, "grad_norm": 0.232071814916651, "learning_rate": 5.417209530837892e-07, "loss": 0.5197, "step": 4825 }, { "epoch": 11.799511002444987, "grad_norm": 0.28839594399022617, "learning_rate": 5.409254047192724e-07, "loss": 0.5995, "step": 4826 }, { "epoch": 11.80195599022005, "grad_norm": 0.30384088496321743, "learning_rate": 5.401303700689608e-07, "loss": 0.6156, "step": 4827 }, { "epoch": 11.80440097799511, "grad_norm": 0.2179882472303422, "learning_rate": 5.393358493413309e-07, "loss": 0.7632, "step": 4828 }, { "epoch": 11.80684596577017, "grad_norm": 0.23018990538877515, "learning_rate": 5.385418427447229e-07, "loss": 1.0303, "step": 4829 }, { "epoch": 11.809290953545233, "grad_norm": 0.28913188972653914, "learning_rate": 5.37748350487344e-07, "loss": 0.3733, "step": 4830 }, { "epoch": 11.811735941320293, "grad_norm": 0.223322779047922, "learning_rate": 5.369553727772672e-07, "loss": 0.342, "step": 4831 }, { "epoch": 11.814180929095354, "grad_norm": 0.21348720063181859, "learning_rate": 5.361629098224269e-07, "loss": 0.6159, "step": 4832 }, { "epoch": 11.816625916870416, "grad_norm": 0.20844972795574548, "learning_rate": 5.353709618306277e-07, "loss": 0.8031, "step": 4833 }, { "epoch": 11.819070904645477, "grad_norm": 0.22148415188121776, "learning_rate": 5.345795290095352e-07, "loss": 0.4855, "step": 4834 }, { "epoch": 11.821515892420537, "grad_norm": 0.2308077788141017, "learning_rate": 5.337886115666815e-07, "loss": 0.4505, "step": 4835 }, { "epoch": 11.8239608801956, "grad_norm": 0.275696085278996, "learning_rate": 5.329982097094627e-07, "loss": 0.5977, "step": 4836 }, { "epoch": 11.82640586797066, "grad_norm": 0.17879173104254753, "learning_rate": 5.322083236451406e-07, "loss": 0.5651, "step": 4837 }, { "epoch": 11.82885085574572, "grad_norm": 0.2528832636655907, "learning_rate": 5.314189535808423e-07, "loss": 0.406, "step": 4838 }, { "epoch": 11.831295843520783, "grad_norm": 0.3128629887857883, "learning_rate": 5.306300997235569e-07, "loss": 1.089, "step": 4839 }, { "epoch": 11.833740831295843, "grad_norm": 0.26680183813521247, "learning_rate": 5.298417622801413e-07, "loss": 0.6939, "step": 4840 }, { "epoch": 11.836185819070904, "grad_norm": 0.23178691690139192, "learning_rate": 5.290539414573157e-07, "loss": 0.4246, "step": 4841 }, { "epoch": 11.838630806845966, "grad_norm": 0.24012038258847387, "learning_rate": 5.282666374616637e-07, "loss": 0.679, "step": 4842 }, { "epoch": 11.841075794621027, "grad_norm": 0.3283111358702929, "learning_rate": 5.274798504996351e-07, "loss": 0.473, "step": 4843 }, { "epoch": 11.843520782396087, "grad_norm": 0.34528407815708956, "learning_rate": 5.266935807775433e-07, "loss": 0.4242, "step": 4844 }, { "epoch": 11.84596577017115, "grad_norm": 0.2066340649402097, "learning_rate": 5.259078285015668e-07, "loss": 0.8594, "step": 4845 }, { "epoch": 11.84841075794621, "grad_norm": 0.3025201122937535, "learning_rate": 5.251225938777469e-07, "loss": 0.7965, "step": 4846 }, { "epoch": 11.85085574572127, "grad_norm": 0.2988188090146297, "learning_rate": 5.243378771119903e-07, "loss": 0.784, "step": 4847 }, { "epoch": 11.853300733496333, "grad_norm": 0.28583556733591914, "learning_rate": 5.235536784100689e-07, "loss": 0.724, "step": 4848 }, { "epoch": 11.855745721271393, "grad_norm": 0.30604816458491546, "learning_rate": 5.227699979776158e-07, "loss": 0.6654, "step": 4849 }, { "epoch": 11.858190709046454, "grad_norm": 0.41579295242521735, "learning_rate": 5.219868360201305e-07, "loss": 0.471, "step": 4850 }, { "epoch": 11.860635696821516, "grad_norm": 0.17441021745116003, "learning_rate": 5.212041927429773e-07, "loss": 0.6786, "step": 4851 }, { "epoch": 11.863080684596577, "grad_norm": 0.22141277171609103, "learning_rate": 5.204220683513811e-07, "loss": 0.4054, "step": 4852 }, { "epoch": 11.865525672371637, "grad_norm": 0.23043113893081826, "learning_rate": 5.196404630504346e-07, "loss": 0.325, "step": 4853 }, { "epoch": 11.8679706601467, "grad_norm": 0.2376559306229159, "learning_rate": 5.188593770450906e-07, "loss": 0.6998, "step": 4854 }, { "epoch": 11.87041564792176, "grad_norm": 0.23466943129903348, "learning_rate": 5.180788105401701e-07, "loss": 0.3305, "step": 4855 }, { "epoch": 11.87286063569682, "grad_norm": 0.31841479536036726, "learning_rate": 5.172987637403537e-07, "loss": 1.2162, "step": 4856 }, { "epoch": 11.875305623471883, "grad_norm": 0.20818810144016406, "learning_rate": 5.165192368501884e-07, "loss": 0.6377, "step": 4857 }, { "epoch": 11.877750611246944, "grad_norm": 0.17997492481757107, "learning_rate": 5.157402300740844e-07, "loss": 0.4752, "step": 4858 }, { "epoch": 11.880195599022004, "grad_norm": 0.2685022301990982, "learning_rate": 5.149617436163134e-07, "loss": 0.7146, "step": 4859 }, { "epoch": 11.882640586797066, "grad_norm": 0.23387838356721757, "learning_rate": 5.141837776810141e-07, "loss": 0.688, "step": 4860 }, { "epoch": 11.885085574572127, "grad_norm": 0.22647858829098802, "learning_rate": 5.134063324721866e-07, "loss": 0.6888, "step": 4861 }, { "epoch": 11.887530562347187, "grad_norm": 0.24845238553438467, "learning_rate": 5.126294081936939e-07, "loss": 0.4501, "step": 4862 }, { "epoch": 11.88997555012225, "grad_norm": 0.23427135951762612, "learning_rate": 5.118530050492648e-07, "loss": 0.4487, "step": 4863 }, { "epoch": 11.89242053789731, "grad_norm": 0.3942334515894535, "learning_rate": 5.110771232424886e-07, "loss": 0.4053, "step": 4864 }, { "epoch": 11.894865525672373, "grad_norm": 0.36981272366594314, "learning_rate": 5.103017629768197e-07, "loss": 0.4541, "step": 4865 }, { "epoch": 11.897310513447433, "grad_norm": 0.24032147882402957, "learning_rate": 5.095269244555756e-07, "loss": 0.8607, "step": 4866 }, { "epoch": 11.899755501222494, "grad_norm": 0.24857342772203783, "learning_rate": 5.087526078819363e-07, "loss": 0.5002, "step": 4867 }, { "epoch": 11.902200488997554, "grad_norm": 0.26163835010011494, "learning_rate": 5.079788134589464e-07, "loss": 0.6311, "step": 4868 }, { "epoch": 11.904645476772616, "grad_norm": 0.32514313211445794, "learning_rate": 5.07205541389511e-07, "loss": 0.3993, "step": 4869 }, { "epoch": 11.907090464547677, "grad_norm": 0.2690740141006805, "learning_rate": 5.064327918764006e-07, "loss": 0.7586, "step": 4870 }, { "epoch": 11.90953545232274, "grad_norm": 0.13632903066672414, "learning_rate": 5.056605651222482e-07, "loss": 0.3095, "step": 4871 }, { "epoch": 11.9119804400978, "grad_norm": 0.22379803708199883, "learning_rate": 5.048888613295483e-07, "loss": 0.6278, "step": 4872 }, { "epoch": 11.91442542787286, "grad_norm": 0.37995192602830485, "learning_rate": 5.041176807006601e-07, "loss": 0.3703, "step": 4873 }, { "epoch": 11.91687041564792, "grad_norm": 0.23375625996587635, "learning_rate": 5.03347023437804e-07, "loss": 0.4041, "step": 4874 }, { "epoch": 11.919315403422983, "grad_norm": 0.2795084420195988, "learning_rate": 5.025768897430644e-07, "loss": 0.5918, "step": 4875 }, { "epoch": 11.921760391198044, "grad_norm": 0.2364884976463528, "learning_rate": 5.018072798183881e-07, "loss": 0.4875, "step": 4876 }, { "epoch": 11.924205378973106, "grad_norm": 0.19458714156769039, "learning_rate": 5.010381938655845e-07, "loss": 0.738, "step": 4877 }, { "epoch": 11.926650366748166, "grad_norm": 0.24615643969780854, "learning_rate": 5.00269632086326e-07, "loss": 0.4108, "step": 4878 }, { "epoch": 11.929095354523227, "grad_norm": 0.2660258900138615, "learning_rate": 4.995015946821458e-07, "loss": 0.5286, "step": 4879 }, { "epoch": 11.93154034229829, "grad_norm": 0.5732059739086813, "learning_rate": 4.987340818544417e-07, "loss": 0.6345, "step": 4880 }, { "epoch": 11.93398533007335, "grad_norm": 0.1964365808213967, "learning_rate": 4.979670938044737e-07, "loss": 0.6577, "step": 4881 }, { "epoch": 11.93643031784841, "grad_norm": 0.3024584516938132, "learning_rate": 4.972006307333624e-07, "loss": 0.4503, "step": 4882 }, { "epoch": 11.938875305623473, "grad_norm": 0.24161055695260578, "learning_rate": 4.964346928420932e-07, "loss": 0.5782, "step": 4883 }, { "epoch": 11.941320293398533, "grad_norm": 0.27759733566002004, "learning_rate": 4.95669280331511e-07, "loss": 0.6111, "step": 4884 }, { "epoch": 11.943765281173594, "grad_norm": 0.39818775129009115, "learning_rate": 4.949043934023257e-07, "loss": 0.4773, "step": 4885 }, { "epoch": 11.946210268948656, "grad_norm": 0.269868387692756, "learning_rate": 4.941400322551074e-07, "loss": 0.468, "step": 4886 }, { "epoch": 11.948655256723717, "grad_norm": 0.2692333049835374, "learning_rate": 4.933761970902898e-07, "loss": 0.7696, "step": 4887 }, { "epoch": 11.951100244498777, "grad_norm": 0.30646921726311277, "learning_rate": 4.926128881081684e-07, "loss": 0.542, "step": 4888 }, { "epoch": 11.95354523227384, "grad_norm": 0.1910270625763932, "learning_rate": 4.918501055088986e-07, "loss": 0.3754, "step": 4889 }, { "epoch": 11.9559902200489, "grad_norm": 0.2634752764386183, "learning_rate": 4.910878494925009e-07, "loss": 0.7436, "step": 4890 }, { "epoch": 11.95843520782396, "grad_norm": 0.2691613176999981, "learning_rate": 4.903261202588561e-07, "loss": 0.6693, "step": 4891 }, { "epoch": 11.960880195599023, "grad_norm": 0.25267810163609183, "learning_rate": 4.895649180077061e-07, "loss": 1.3046, "step": 4892 }, { "epoch": 11.963325183374083, "grad_norm": 0.23324289861531824, "learning_rate": 4.888042429386572e-07, "loss": 0.7026, "step": 4893 }, { "epoch": 11.965770171149144, "grad_norm": 0.29108133442388834, "learning_rate": 4.88044095251174e-07, "loss": 0.4995, "step": 4894 }, { "epoch": 11.968215158924206, "grad_norm": 0.3410673393714446, "learning_rate": 4.872844751445851e-07, "loss": 0.5615, "step": 4895 }, { "epoch": 11.970660146699267, "grad_norm": 0.24941392874561313, "learning_rate": 4.865253828180816e-07, "loss": 0.9909, "step": 4896 }, { "epoch": 11.973105134474327, "grad_norm": 0.25637840700090175, "learning_rate": 4.857668184707126e-07, "loss": 0.6614, "step": 4897 }, { "epoch": 11.97555012224939, "grad_norm": 0.28223123795727095, "learning_rate": 4.850087823013935e-07, "loss": 0.811, "step": 4898 }, { "epoch": 11.97799511002445, "grad_norm": 0.2518362812370028, "learning_rate": 4.842512745088973e-07, "loss": 0.4578, "step": 4899 }, { "epoch": 11.98044009779951, "grad_norm": 0.2937450438739105, "learning_rate": 4.834942952918598e-07, "loss": 0.4915, "step": 4900 }, { "epoch": 11.982885085574573, "grad_norm": 0.2965872293139258, "learning_rate": 4.827378448487793e-07, "loss": 0.5215, "step": 4901 }, { "epoch": 11.985330073349633, "grad_norm": 0.29556829670653845, "learning_rate": 4.819819233780135e-07, "loss": 0.4673, "step": 4902 }, { "epoch": 11.987775061124694, "grad_norm": 0.21410853275804778, "learning_rate": 4.812265310777828e-07, "loss": 0.5493, "step": 4903 }, { "epoch": 11.990220048899756, "grad_norm": 0.2993202197685635, "learning_rate": 4.804716681461677e-07, "loss": 0.6656, "step": 4904 }, { "epoch": 11.992665036674817, "grad_norm": 0.39750040402899234, "learning_rate": 4.797173347811109e-07, "loss": 0.3661, "step": 4905 }, { "epoch": 11.995110024449877, "grad_norm": 0.26321518557157253, "learning_rate": 4.789635311804167e-07, "loss": 0.4174, "step": 4906 }, { "epoch": 11.99755501222494, "grad_norm": 0.3585518388642963, "learning_rate": 4.782102575417474e-07, "loss": 0.4389, "step": 4907 }, { "epoch": 12.0, "grad_norm": 0.20999734913997922, "learning_rate": 4.774575140626317e-07, "loss": 0.9346, "step": 4908 }, { "epoch": 12.00244498777506, "grad_norm": 0.14974658037788552, "learning_rate": 4.7670530094045373e-07, "loss": 0.3464, "step": 4909 }, { "epoch": 12.004889975550123, "grad_norm": 0.2973883503464538, "learning_rate": 4.7595361837246195e-07, "loss": 0.8563, "step": 4910 }, { "epoch": 12.007334963325183, "grad_norm": 0.4240554816834499, "learning_rate": 4.75202466555765e-07, "loss": 0.3456, "step": 4911 }, { "epoch": 12.009779951100244, "grad_norm": 0.22702086800957727, "learning_rate": 4.744518456873312e-07, "loss": 0.4787, "step": 4912 }, { "epoch": 12.012224938875306, "grad_norm": 0.2052304755907729, "learning_rate": 4.7370175596399197e-07, "loss": 0.6494, "step": 4913 }, { "epoch": 12.014669926650367, "grad_norm": 0.3580177799502247, "learning_rate": 4.729521975824361e-07, "loss": 0.8716, "step": 4914 }, { "epoch": 12.017114914425427, "grad_norm": 0.4893259899611762, "learning_rate": 4.7220317073921605e-07, "loss": 0.8095, "step": 4915 }, { "epoch": 12.01955990220049, "grad_norm": 0.3881443725575305, "learning_rate": 4.714546756307442e-07, "loss": 0.49, "step": 4916 }, { "epoch": 12.02200488997555, "grad_norm": 0.2728848214819317, "learning_rate": 4.7070671245329185e-07, "loss": 0.4269, "step": 4917 }, { "epoch": 12.02444987775061, "grad_norm": 0.28924819765520177, "learning_rate": 4.6995928140299296e-07, "loss": 0.5143, "step": 4918 }, { "epoch": 12.026894865525673, "grad_norm": 0.3424151853744166, "learning_rate": 4.6921238267584104e-07, "loss": 0.4133, "step": 4919 }, { "epoch": 12.029339853300733, "grad_norm": 0.24767702876112604, "learning_rate": 4.684660164676896e-07, "loss": 0.5702, "step": 4920 }, { "epoch": 12.031784841075794, "grad_norm": 0.23999566822622512, "learning_rate": 4.6772018297425394e-07, "loss": 0.6688, "step": 4921 }, { "epoch": 12.034229828850856, "grad_norm": 0.22883631460327247, "learning_rate": 4.6697488239110744e-07, "loss": 0.3411, "step": 4922 }, { "epoch": 12.036674816625917, "grad_norm": 0.45427945552661875, "learning_rate": 4.662301149136861e-07, "loss": 0.7177, "step": 4923 }, { "epoch": 12.039119804400977, "grad_norm": 0.2528791992440227, "learning_rate": 4.6548588073728363e-07, "loss": 0.6214, "step": 4924 }, { "epoch": 12.04156479217604, "grad_norm": 0.30876164010648816, "learning_rate": 4.647421800570562e-07, "loss": 0.7487, "step": 4925 }, { "epoch": 12.0440097799511, "grad_norm": 0.4290919427922011, "learning_rate": 4.639990130680197e-07, "loss": 0.5046, "step": 4926 }, { "epoch": 12.04645476772616, "grad_norm": 0.24842678931905046, "learning_rate": 4.63256379965048e-07, "loss": 0.4291, "step": 4927 }, { "epoch": 12.048899755501223, "grad_norm": 0.26996187161922786, "learning_rate": 4.625142809428773e-07, "loss": 0.7915, "step": 4928 }, { "epoch": 12.051344743276283, "grad_norm": 0.3780335014734925, "learning_rate": 4.6177271619610305e-07, "loss": 0.3806, "step": 4929 }, { "epoch": 12.053789731051344, "grad_norm": 0.3572646532543196, "learning_rate": 4.610316859191805e-07, "loss": 0.4199, "step": 4930 }, { "epoch": 12.056234718826406, "grad_norm": 0.2725853249922248, "learning_rate": 4.60291190306425e-07, "loss": 0.695, "step": 4931 }, { "epoch": 12.058679706601467, "grad_norm": 0.24430421053166984, "learning_rate": 4.5955122955201076e-07, "loss": 0.3543, "step": 4932 }, { "epoch": 12.061124694376527, "grad_norm": 0.26127812398454425, "learning_rate": 4.5881180384997336e-07, "loss": 0.6491, "step": 4933 }, { "epoch": 12.06356968215159, "grad_norm": 0.2148919858733965, "learning_rate": 4.5807291339420607e-07, "loss": 0.3953, "step": 4934 }, { "epoch": 12.06601466992665, "grad_norm": 0.42044662904601915, "learning_rate": 4.573345583784633e-07, "loss": 0.4603, "step": 4935 }, { "epoch": 12.06845965770171, "grad_norm": 0.28203693741480496, "learning_rate": 4.565967389963594e-07, "loss": 0.6747, "step": 4936 }, { "epoch": 12.070904645476773, "grad_norm": 0.18577593335260578, "learning_rate": 4.5585945544136615e-07, "loss": 0.3818, "step": 4937 }, { "epoch": 12.073349633251834, "grad_norm": 0.26607732766170883, "learning_rate": 4.5512270790681684e-07, "loss": 0.5725, "step": 4938 }, { "epoch": 12.075794621026894, "grad_norm": 0.2823956957916674, "learning_rate": 4.5438649658590405e-07, "loss": 0.9092, "step": 4939 }, { "epoch": 12.078239608801956, "grad_norm": 0.29256557996793753, "learning_rate": 4.5365082167167786e-07, "loss": 0.5123, "step": 4940 }, { "epoch": 12.080684596577017, "grad_norm": 0.21145596410616554, "learning_rate": 4.529156833570508e-07, "loss": 0.402, "step": 4941 }, { "epoch": 12.083129584352077, "grad_norm": 0.2647221025831137, "learning_rate": 4.521810818347916e-07, "loss": 0.5432, "step": 4942 }, { "epoch": 12.08557457212714, "grad_norm": 0.267127167065398, "learning_rate": 4.5144701729753103e-07, "loss": 0.3557, "step": 4943 }, { "epoch": 12.0880195599022, "grad_norm": 0.3764503047214616, "learning_rate": 4.507134899377558e-07, "loss": 0.5194, "step": 4944 }, { "epoch": 12.09046454767726, "grad_norm": 0.19465521547983414, "learning_rate": 4.499804999478144e-07, "loss": 0.4199, "step": 4945 }, { "epoch": 12.092909535452323, "grad_norm": 0.2828227441637348, "learning_rate": 4.492480475199146e-07, "loss": 1.2859, "step": 4946 }, { "epoch": 12.095354523227384, "grad_norm": 0.24407840282206383, "learning_rate": 4.485161328461204e-07, "loss": 0.895, "step": 4947 }, { "epoch": 12.097799511002446, "grad_norm": 0.31045936327212736, "learning_rate": 4.4778475611835766e-07, "loss": 0.431, "step": 4948 }, { "epoch": 12.100244498777506, "grad_norm": 0.2098592875810141, "learning_rate": 4.4705391752841027e-07, "loss": 0.3738, "step": 4949 }, { "epoch": 12.102689486552567, "grad_norm": 0.24051221187610344, "learning_rate": 4.4632361726791915e-07, "loss": 1.037, "step": 4950 }, { "epoch": 12.10513447432763, "grad_norm": 0.254559882601482, "learning_rate": 4.4559385552838843e-07, "loss": 0.4883, "step": 4951 }, { "epoch": 12.10757946210269, "grad_norm": 0.3127148214794706, "learning_rate": 4.4486463250117626e-07, "loss": 0.4505, "step": 4952 }, { "epoch": 12.11002444987775, "grad_norm": 0.23420130253635035, "learning_rate": 4.441359483775032e-07, "loss": 0.3181, "step": 4953 }, { "epoch": 12.112469437652813, "grad_norm": 0.2736165661244627, "learning_rate": 4.4340780334844516e-07, "loss": 0.5839, "step": 4954 }, { "epoch": 12.114914425427873, "grad_norm": 0.2753630071857551, "learning_rate": 4.4268019760493914e-07, "loss": 0.5044, "step": 4955 }, { "epoch": 12.117359413202934, "grad_norm": 0.24233135615121223, "learning_rate": 4.419531313377809e-07, "loss": 0.5101, "step": 4956 }, { "epoch": 12.119804400977996, "grad_norm": 0.2231071059003483, "learning_rate": 4.4122660473762266e-07, "loss": 0.4862, "step": 4957 }, { "epoch": 12.122249388753056, "grad_norm": 0.1892774326366706, "learning_rate": 4.405006179949767e-07, "loss": 0.4213, "step": 4958 }, { "epoch": 12.124694376528117, "grad_norm": 0.2279861699639557, "learning_rate": 4.397751713002141e-07, "loss": 0.4399, "step": 4959 }, { "epoch": 12.12713936430318, "grad_norm": 0.316489340569337, "learning_rate": 4.390502648435624e-07, "loss": 0.4053, "step": 4960 }, { "epoch": 12.12958435207824, "grad_norm": 0.2143646131539836, "learning_rate": 4.38325898815109e-07, "loss": 0.4888, "step": 4961 }, { "epoch": 12.1320293398533, "grad_norm": 0.24333042106489217, "learning_rate": 4.376020734047995e-07, "loss": 0.4852, "step": 4962 }, { "epoch": 12.134474327628363, "grad_norm": 0.26319118922886414, "learning_rate": 4.368787888024384e-07, "loss": 0.6872, "step": 4963 }, { "epoch": 12.136919315403423, "grad_norm": 0.3009017983293089, "learning_rate": 4.361560451976857e-07, "loss": 0.4274, "step": 4964 }, { "epoch": 12.139364303178484, "grad_norm": 0.19805722915349858, "learning_rate": 4.3543384278006196e-07, "loss": 0.5214, "step": 4965 }, { "epoch": 12.141809290953546, "grad_norm": 0.2775371159143536, "learning_rate": 4.347121817389463e-07, "loss": 0.4986, "step": 4966 }, { "epoch": 12.144254278728607, "grad_norm": 0.29440794914667917, "learning_rate": 4.3399106226357317e-07, "loss": 0.6688, "step": 4967 }, { "epoch": 12.146699266503667, "grad_norm": 0.25714384591712175, "learning_rate": 4.332704845430372e-07, "loss": 0.9102, "step": 4968 }, { "epoch": 12.14914425427873, "grad_norm": 0.26350374191122117, "learning_rate": 4.32550448766291e-07, "loss": 0.6438, "step": 4969 }, { "epoch": 12.15158924205379, "grad_norm": 0.22984760882582447, "learning_rate": 4.3183095512214336e-07, "loss": 0.661, "step": 4970 }, { "epoch": 12.15403422982885, "grad_norm": 0.23714512029480686, "learning_rate": 4.3111200379926306e-07, "loss": 0.8397, "step": 4971 }, { "epoch": 12.156479217603913, "grad_norm": 0.28635306255729304, "learning_rate": 4.30393594986174e-07, "loss": 0.616, "step": 4972 }, { "epoch": 12.158924205378973, "grad_norm": 0.2813402344600723, "learning_rate": 4.296757288712616e-07, "loss": 0.546, "step": 4973 }, { "epoch": 12.161369193154034, "grad_norm": 0.26632731963454787, "learning_rate": 4.289584056427648e-07, "loss": 0.541, "step": 4974 }, { "epoch": 12.163814180929096, "grad_norm": 0.19002905742819837, "learning_rate": 4.282416254887833e-07, "loss": 0.8745, "step": 4975 }, { "epoch": 12.166259168704157, "grad_norm": 0.24860555945941368, "learning_rate": 4.2752538859727355e-07, "loss": 0.4566, "step": 4976 }, { "epoch": 12.168704156479217, "grad_norm": 0.23039239212238033, "learning_rate": 4.26809695156048e-07, "loss": 0.6884, "step": 4977 }, { "epoch": 12.17114914425428, "grad_norm": 0.18796083473835296, "learning_rate": 4.260945453527784e-07, "loss": 0.4938, "step": 4978 }, { "epoch": 12.17359413202934, "grad_norm": 0.31057752004240696, "learning_rate": 4.253799393749944e-07, "loss": 0.5371, "step": 4979 }, { "epoch": 12.1760391198044, "grad_norm": 0.226832703208997, "learning_rate": 4.2466587741008034e-07, "loss": 0.6169, "step": 4980 }, { "epoch": 12.178484107579463, "grad_norm": 0.3004969549899206, "learning_rate": 4.239523596452813e-07, "loss": 0.7041, "step": 4981 }, { "epoch": 12.180929095354523, "grad_norm": 0.2580912284484904, "learning_rate": 4.2323938626769647e-07, "loss": 0.422, "step": 4982 }, { "epoch": 12.183374083129584, "grad_norm": 0.3738895179260783, "learning_rate": 4.225269574642843e-07, "loss": 0.687, "step": 4983 }, { "epoch": 12.185819070904646, "grad_norm": 0.3097757571593831, "learning_rate": 4.218150734218604e-07, "loss": 0.4423, "step": 4984 }, { "epoch": 12.188264058679707, "grad_norm": 0.22358035129012138, "learning_rate": 4.211037343270963e-07, "loss": 0.6642, "step": 4985 }, { "epoch": 12.190709046454767, "grad_norm": 0.36784878029512597, "learning_rate": 4.2039294036652274e-07, "loss": 0.564, "step": 4986 }, { "epoch": 12.19315403422983, "grad_norm": 0.33766412668757234, "learning_rate": 4.1968269172652454e-07, "loss": 0.6897, "step": 4987 }, { "epoch": 12.19559902200489, "grad_norm": 0.24811916424188032, "learning_rate": 4.189729885933461e-07, "loss": 0.722, "step": 4988 }, { "epoch": 12.19804400977995, "grad_norm": 0.26817919242988325, "learning_rate": 4.182638311530879e-07, "loss": 0.7672, "step": 4989 }, { "epoch": 12.200488997555013, "grad_norm": 0.31152358691259346, "learning_rate": 4.175552195917065e-07, "loss": 0.4398, "step": 4990 }, { "epoch": 12.202933985330073, "grad_norm": 0.18347197414672434, "learning_rate": 4.168471540950173e-07, "loss": 0.8236, "step": 4991 }, { "epoch": 12.205378973105134, "grad_norm": 0.3233397036690673, "learning_rate": 4.161396348486896e-07, "loss": 0.4922, "step": 4992 }, { "epoch": 12.207823960880196, "grad_norm": 0.2261237803673091, "learning_rate": 4.1543266203825204e-07, "loss": 0.7712, "step": 4993 }, { "epoch": 12.210268948655257, "grad_norm": 0.2247211866231037, "learning_rate": 4.1472623584908945e-07, "loss": 0.46, "step": 4994 }, { "epoch": 12.212713936430317, "grad_norm": 0.2848126151299553, "learning_rate": 4.1402035646644216e-07, "loss": 0.8908, "step": 4995 }, { "epoch": 12.21515892420538, "grad_norm": 0.19839595979620234, "learning_rate": 4.133150240754094e-07, "loss": 0.6037, "step": 4996 }, { "epoch": 12.21760391198044, "grad_norm": 0.2876660121754633, "learning_rate": 4.1261023886094345e-07, "loss": 0.5639, "step": 4997 }, { "epoch": 12.2200488997555, "grad_norm": 0.276443185488912, "learning_rate": 4.1190600100785627e-07, "loss": 0.4118, "step": 4998 }, { "epoch": 12.222493887530563, "grad_norm": 0.18855166535058215, "learning_rate": 4.1120231070081536e-07, "loss": 0.4222, "step": 4999 }, { "epoch": 12.224938875305623, "grad_norm": 0.15577853072870454, "learning_rate": 4.1049916812434335e-07, "loss": 0.4215, "step": 5000 }, { "epoch": 12.227383863080684, "grad_norm": 0.2208354021479963, "learning_rate": 4.0979657346282175e-07, "loss": 0.5355, "step": 5001 }, { "epoch": 12.229828850855746, "grad_norm": 0.19732094560503924, "learning_rate": 4.090945269004856e-07, "loss": 0.4084, "step": 5002 }, { "epoch": 12.232273838630807, "grad_norm": 0.2865513880507124, "learning_rate": 4.0839302862142825e-07, "loss": 0.4367, "step": 5003 }, { "epoch": 12.234718826405867, "grad_norm": 0.23650126395243395, "learning_rate": 4.0769207880959837e-07, "loss": 0.6435, "step": 5004 }, { "epoch": 12.23716381418093, "grad_norm": 0.2837359383769163, "learning_rate": 4.069916776488014e-07, "loss": 0.6187, "step": 5005 }, { "epoch": 12.23960880195599, "grad_norm": 0.30455379321359616, "learning_rate": 4.062918253226988e-07, "loss": 0.7203, "step": 5006 }, { "epoch": 12.24205378973105, "grad_norm": 0.37021966736430867, "learning_rate": 4.0559252201480715e-07, "loss": 0.3214, "step": 5007 }, { "epoch": 12.244498777506113, "grad_norm": 0.3171274598994411, "learning_rate": 4.0489376790849967e-07, "loss": 0.7396, "step": 5008 }, { "epoch": 12.246943765281173, "grad_norm": 0.2852895153459756, "learning_rate": 4.04195563187007e-07, "loss": 0.7068, "step": 5009 }, { "epoch": 12.249388753056234, "grad_norm": 0.28382637494332086, "learning_rate": 4.034979080334128e-07, "loss": 0.8304, "step": 5010 }, { "epoch": 12.251833740831296, "grad_norm": 0.42378223522649555, "learning_rate": 4.028008026306593e-07, "loss": 0.4572, "step": 5011 }, { "epoch": 12.254278728606357, "grad_norm": 0.19430629363737095, "learning_rate": 4.021042471615427e-07, "loss": 0.6891, "step": 5012 }, { "epoch": 12.256723716381417, "grad_norm": 0.39115272511553995, "learning_rate": 4.0140824180871616e-07, "loss": 0.4623, "step": 5013 }, { "epoch": 12.25916870415648, "grad_norm": 0.2603520388334594, "learning_rate": 4.0071278675468893e-07, "loss": 0.4589, "step": 5014 }, { "epoch": 12.26161369193154, "grad_norm": 0.19412919349615174, "learning_rate": 4.0001788218182345e-07, "loss": 0.6218, "step": 5015 }, { "epoch": 12.2640586797066, "grad_norm": 0.24701566988964282, "learning_rate": 3.993235282723418e-07, "loss": 0.7555, "step": 5016 }, { "epoch": 12.266503667481663, "grad_norm": 0.37896591986455336, "learning_rate": 3.9862972520831767e-07, "loss": 0.6261, "step": 5017 }, { "epoch": 12.268948655256724, "grad_norm": 0.28678038675444295, "learning_rate": 3.979364731716828e-07, "loss": 0.6443, "step": 5018 }, { "epoch": 12.271393643031784, "grad_norm": 0.2124750689291723, "learning_rate": 3.972437723442246e-07, "loss": 0.7061, "step": 5019 }, { "epoch": 12.273838630806846, "grad_norm": 0.2171672028171427, "learning_rate": 3.9655162290758357e-07, "loss": 0.6889, "step": 5020 }, { "epoch": 12.276283618581907, "grad_norm": 0.2606095098441035, "learning_rate": 3.9586002504325843e-07, "loss": 0.6538, "step": 5021 }, { "epoch": 12.278728606356967, "grad_norm": 0.1885873915613051, "learning_rate": 3.9516897893260075e-07, "loss": 0.4518, "step": 5022 }, { "epoch": 12.28117359413203, "grad_norm": 0.31624293739335085, "learning_rate": 3.944784847568192e-07, "loss": 0.8837, "step": 5023 }, { "epoch": 12.28361858190709, "grad_norm": 0.22936654256411337, "learning_rate": 3.9378854269697787e-07, "loss": 0.5566, "step": 5024 }, { "epoch": 12.28606356968215, "grad_norm": 0.33992117882109785, "learning_rate": 3.930991529339936e-07, "loss": 0.5945, "step": 5025 }, { "epoch": 12.288508557457213, "grad_norm": 0.2455890228239286, "learning_rate": 3.9241031564864225e-07, "loss": 0.7229, "step": 5026 }, { "epoch": 12.290953545232274, "grad_norm": 0.2926308229577295, "learning_rate": 3.9172203102155135e-07, "loss": 0.465, "step": 5027 }, { "epoch": 12.293398533007334, "grad_norm": 0.18619148394811125, "learning_rate": 3.91034299233205e-07, "loss": 0.3258, "step": 5028 }, { "epoch": 12.295843520782396, "grad_norm": 0.33312540021219555, "learning_rate": 3.903471204639434e-07, "loss": 0.44, "step": 5029 }, { "epoch": 12.298288508557457, "grad_norm": 0.19905025703442614, "learning_rate": 3.896604948939589e-07, "loss": 0.5781, "step": 5030 }, { "epoch": 12.30073349633252, "grad_norm": 0.2597890286787564, "learning_rate": 3.889744227033018e-07, "loss": 0.4012, "step": 5031 }, { "epoch": 12.30317848410758, "grad_norm": 0.26191417147920854, "learning_rate": 3.8828890407187486e-07, "loss": 0.4122, "step": 5032 }, { "epoch": 12.30562347188264, "grad_norm": 0.2799213915498337, "learning_rate": 3.8760393917943726e-07, "loss": 0.512, "step": 5033 }, { "epoch": 12.3080684596577, "grad_norm": 0.30207005310324064, "learning_rate": 3.869195282056029e-07, "loss": 0.6469, "step": 5034 }, { "epoch": 12.310513447432763, "grad_norm": 0.722534654454409, "learning_rate": 3.862356713298387e-07, "loss": 0.4412, "step": 5035 }, { "epoch": 12.312958435207824, "grad_norm": 0.23657852516259256, "learning_rate": 3.855523687314688e-07, "loss": 0.55, "step": 5036 }, { "epoch": 12.315403422982886, "grad_norm": 0.2690300598345504, "learning_rate": 3.8486962058967043e-07, "loss": 0.5133, "step": 5037 }, { "epoch": 12.317848410757946, "grad_norm": 0.23956015325747637, "learning_rate": 3.8418742708347556e-07, "loss": 0.5295, "step": 5038 }, { "epoch": 12.320293398533007, "grad_norm": 0.2660879263387045, "learning_rate": 3.83505788391772e-07, "loss": 0.3062, "step": 5039 }, { "epoch": 12.32273838630807, "grad_norm": 0.33229498982937933, "learning_rate": 3.828247046932992e-07, "loss": 0.5089, "step": 5040 }, { "epoch": 12.32518337408313, "grad_norm": 0.18521643108560396, "learning_rate": 3.821441761666547e-07, "loss": 0.8748, "step": 5041 }, { "epoch": 12.32762836185819, "grad_norm": 0.38734838474459843, "learning_rate": 3.8146420299028737e-07, "loss": 0.4731, "step": 5042 }, { "epoch": 12.330073349633253, "grad_norm": 0.29821763679064184, "learning_rate": 3.80784785342502e-07, "loss": 0.4644, "step": 5043 }, { "epoch": 12.332518337408313, "grad_norm": 0.21019307637097884, "learning_rate": 3.801059234014581e-07, "loss": 0.4445, "step": 5044 }, { "epoch": 12.334963325183374, "grad_norm": 0.20709640539114063, "learning_rate": 3.7942761734516803e-07, "loss": 0.5038, "step": 5045 }, { "epoch": 12.337408312958436, "grad_norm": 0.20676618286550388, "learning_rate": 3.787498673514994e-07, "loss": 0.4157, "step": 5046 }, { "epoch": 12.339853300733497, "grad_norm": 0.1960462683718853, "learning_rate": 3.7807267359817377e-07, "loss": 0.2934, "step": 5047 }, { "epoch": 12.342298288508557, "grad_norm": 0.3804553137532328, "learning_rate": 3.773960362627671e-07, "loss": 0.5244, "step": 5048 }, { "epoch": 12.34474327628362, "grad_norm": 0.2984267497164799, "learning_rate": 3.7671995552270943e-07, "loss": 0.6049, "step": 5049 }, { "epoch": 12.34718826405868, "grad_norm": 0.3254535330350803, "learning_rate": 3.760444315552839e-07, "loss": 0.6505, "step": 5050 }, { "epoch": 12.34963325183374, "grad_norm": 0.31588070257782663, "learning_rate": 3.753694645376291e-07, "loss": 0.4292, "step": 5051 }, { "epoch": 12.352078239608803, "grad_norm": 0.2450116203833375, "learning_rate": 3.746950546467357e-07, "loss": 0.4816, "step": 5052 }, { "epoch": 12.354523227383863, "grad_norm": 0.2990960459893679, "learning_rate": 3.7402120205945025e-07, "loss": 0.799, "step": 5053 }, { "epoch": 12.356968215158924, "grad_norm": 0.28017150965072696, "learning_rate": 3.733479069524726e-07, "loss": 0.8368, "step": 5054 }, { "epoch": 12.359413202933986, "grad_norm": 0.3536025582885215, "learning_rate": 3.726751695023553e-07, "loss": 0.596, "step": 5055 }, { "epoch": 12.361858190709047, "grad_norm": 0.16164615708845306, "learning_rate": 3.720029898855057e-07, "loss": 0.3734, "step": 5056 }, { "epoch": 12.364303178484107, "grad_norm": 0.18055497733668746, "learning_rate": 3.7133136827818575e-07, "loss": 0.3936, "step": 5057 }, { "epoch": 12.36674816625917, "grad_norm": 0.26896393565677973, "learning_rate": 3.70660304856508e-07, "loss": 0.7527, "step": 5058 }, { "epoch": 12.36919315403423, "grad_norm": 0.2098587620020675, "learning_rate": 3.69989799796443e-07, "loss": 0.8906, "step": 5059 }, { "epoch": 12.37163814180929, "grad_norm": 0.3212849828689608, "learning_rate": 3.693198532738104e-07, "loss": 0.6486, "step": 5060 }, { "epoch": 12.374083129584353, "grad_norm": 0.32110300602575653, "learning_rate": 3.686504654642875e-07, "loss": 0.8406, "step": 5061 }, { "epoch": 12.376528117359413, "grad_norm": 0.3150234896490052, "learning_rate": 3.6798163654340107e-07, "loss": 0.4141, "step": 5062 }, { "epoch": 12.378973105134474, "grad_norm": 0.26603118025847994, "learning_rate": 3.673133666865347e-07, "loss": 0.8156, "step": 5063 }, { "epoch": 12.381418092909536, "grad_norm": 0.21901387549515933, "learning_rate": 3.666456560689241e-07, "loss": 0.6184, "step": 5064 }, { "epoch": 12.383863080684597, "grad_norm": 0.2419440298392444, "learning_rate": 3.659785048656572e-07, "loss": 0.4941, "step": 5065 }, { "epoch": 12.386308068459657, "grad_norm": 0.2408103361825238, "learning_rate": 3.653119132516772e-07, "loss": 1.0229, "step": 5066 }, { "epoch": 12.38875305623472, "grad_norm": 0.41925097861802874, "learning_rate": 3.6464588140178003e-07, "loss": 0.6872, "step": 5067 }, { "epoch": 12.39119804400978, "grad_norm": 0.19283537359871333, "learning_rate": 3.639804094906127e-07, "loss": 0.4835, "step": 5068 }, { "epoch": 12.39364303178484, "grad_norm": 0.3331807967759304, "learning_rate": 3.6331549769267975e-07, "loss": 0.5127, "step": 5069 }, { "epoch": 12.396088019559903, "grad_norm": 0.2591735292648509, "learning_rate": 3.62651146182334e-07, "loss": 0.7028, "step": 5070 }, { "epoch": 12.398533007334963, "grad_norm": 0.30381603877381846, "learning_rate": 3.619873551337855e-07, "loss": 0.3901, "step": 5071 }, { "epoch": 12.400977995110024, "grad_norm": 0.23270857688987157, "learning_rate": 3.613241247210936e-07, "loss": 0.4022, "step": 5072 }, { "epoch": 12.403422982885086, "grad_norm": 0.25348171035688427, "learning_rate": 3.6066145511817363e-07, "loss": 0.9455, "step": 5073 }, { "epoch": 12.405867970660147, "grad_norm": 0.22923200733163546, "learning_rate": 3.5999934649879304e-07, "loss": 0.8498, "step": 5074 }, { "epoch": 12.408312958435207, "grad_norm": 0.35376038981238406, "learning_rate": 3.593377990365707e-07, "loss": 0.9471, "step": 5075 }, { "epoch": 12.41075794621027, "grad_norm": 0.2550170360615306, "learning_rate": 3.586768129049803e-07, "loss": 0.4988, "step": 5076 }, { "epoch": 12.41320293398533, "grad_norm": 0.364700913099084, "learning_rate": 3.580163882773477e-07, "loss": 0.435, "step": 5077 }, { "epoch": 12.41564792176039, "grad_norm": 0.2027159631723815, "learning_rate": 3.573565253268507e-07, "loss": 0.9242, "step": 5078 }, { "epoch": 12.418092909535453, "grad_norm": 0.2663820848105799, "learning_rate": 3.566972242265207e-07, "loss": 0.9055, "step": 5079 }, { "epoch": 12.420537897310513, "grad_norm": 0.2643749357542909, "learning_rate": 3.5603848514924155e-07, "loss": 0.7919, "step": 5080 }, { "epoch": 12.422982885085574, "grad_norm": 0.2611858885328245, "learning_rate": 3.5538030826775066e-07, "loss": 0.9995, "step": 5081 }, { "epoch": 12.425427872860636, "grad_norm": 0.26408003760376697, "learning_rate": 3.547226937546353e-07, "loss": 0.4547, "step": 5082 }, { "epoch": 12.427872860635697, "grad_norm": 0.23373015151998647, "learning_rate": 3.540656417823382e-07, "loss": 0.43, "step": 5083 }, { "epoch": 12.430317848410757, "grad_norm": 0.24577071343016127, "learning_rate": 3.534091525231537e-07, "loss": 0.3435, "step": 5084 }, { "epoch": 12.43276283618582, "grad_norm": 0.19613120809263992, "learning_rate": 3.5275322614922723e-07, "loss": 0.8463, "step": 5085 }, { "epoch": 12.43520782396088, "grad_norm": 0.22819306523183447, "learning_rate": 3.52097862832558e-07, "loss": 0.6876, "step": 5086 }, { "epoch": 12.43765281173594, "grad_norm": 0.20694783884664794, "learning_rate": 3.514430627449986e-07, "loss": 0.5156, "step": 5087 }, { "epoch": 12.440097799511003, "grad_norm": 0.41915172349507196, "learning_rate": 3.507888260582507e-07, "loss": 0.8944, "step": 5088 }, { "epoch": 12.442542787286063, "grad_norm": 0.49521445385576895, "learning_rate": 3.501351529438715e-07, "loss": 0.8466, "step": 5089 }, { "epoch": 12.444987775061124, "grad_norm": 0.20868128342074022, "learning_rate": 3.4948204357326764e-07, "loss": 0.7251, "step": 5090 }, { "epoch": 12.447432762836186, "grad_norm": 0.2233425243957032, "learning_rate": 3.488294981177012e-07, "loss": 0.5372, "step": 5091 }, { "epoch": 12.449877750611247, "grad_norm": 0.30600792833956014, "learning_rate": 3.4817751674828317e-07, "loss": 0.4954, "step": 5092 }, { "epoch": 12.452322738386307, "grad_norm": 0.317578132034094, "learning_rate": 3.4752609963597836e-07, "loss": 0.6282, "step": 5093 }, { "epoch": 12.45476772616137, "grad_norm": 0.22783869795385514, "learning_rate": 3.468752469516037e-07, "loss": 0.6363, "step": 5094 }, { "epoch": 12.45721271393643, "grad_norm": 0.2829340669269438, "learning_rate": 3.4622495886582704e-07, "loss": 0.4799, "step": 5095 }, { "epoch": 12.45965770171149, "grad_norm": 0.198926052813237, "learning_rate": 3.455752355491687e-07, "loss": 0.5273, "step": 5096 }, { "epoch": 12.462102689486553, "grad_norm": 0.24207039400289684, "learning_rate": 3.44926077172002e-07, "loss": 0.828, "step": 5097 }, { "epoch": 12.464547677261614, "grad_norm": 0.17038547861866635, "learning_rate": 3.4427748390454986e-07, "loss": 0.6051, "step": 5098 }, { "epoch": 12.466992665036674, "grad_norm": 0.20057716884790783, "learning_rate": 3.4362945591688955e-07, "loss": 0.5921, "step": 5099 }, { "epoch": 12.469437652811736, "grad_norm": 0.32037509799051955, "learning_rate": 3.4298199337894687e-07, "loss": 0.4386, "step": 5100 }, { "epoch": 12.471882640586797, "grad_norm": 0.3088605730807667, "learning_rate": 3.423350964605038e-07, "loss": 0.3344, "step": 5101 }, { "epoch": 12.474327628361857, "grad_norm": 0.3510904693480172, "learning_rate": 3.4168876533118983e-07, "loss": 0.8158, "step": 5102 }, { "epoch": 12.47677261613692, "grad_norm": 0.2535272273176726, "learning_rate": 3.4104300016048836e-07, "loss": 0.4056, "step": 5103 }, { "epoch": 12.47921760391198, "grad_norm": 0.31680296525846063, "learning_rate": 3.4039780111773416e-07, "loss": 0.4245, "step": 5104 }, { "epoch": 12.48166259168704, "grad_norm": 0.2626988341174822, "learning_rate": 3.3975316837211223e-07, "loss": 0.528, "step": 5105 }, { "epoch": 12.484107579462103, "grad_norm": 0.2548743953785793, "learning_rate": 3.391091020926604e-07, "loss": 1.2104, "step": 5106 }, { "epoch": 12.486552567237164, "grad_norm": 0.27267075450532235, "learning_rate": 3.3846560244826854e-07, "loss": 0.6169, "step": 5107 }, { "epoch": 12.488997555012224, "grad_norm": 0.2593378595025868, "learning_rate": 3.3782266960767556e-07, "loss": 1.075, "step": 5108 }, { "epoch": 12.491442542787286, "grad_norm": 0.30975157856150604, "learning_rate": 3.371803037394744e-07, "loss": 0.8853, "step": 5109 }, { "epoch": 12.493887530562347, "grad_norm": 0.3462420209111442, "learning_rate": 3.36538505012107e-07, "loss": 0.3526, "step": 5110 }, { "epoch": 12.496332518337407, "grad_norm": 0.2823304314844471, "learning_rate": 3.3589727359386787e-07, "loss": 0.4869, "step": 5111 }, { "epoch": 12.49877750611247, "grad_norm": 0.3963325732300786, "learning_rate": 3.352566096529031e-07, "loss": 0.7501, "step": 5112 }, { "epoch": 12.50122249388753, "grad_norm": 0.3266685945116019, "learning_rate": 3.346165133572093e-07, "loss": 0.388, "step": 5113 }, { "epoch": 12.503667481662593, "grad_norm": 0.20047925116983273, "learning_rate": 3.339769848746346e-07, "loss": 0.3427, "step": 5114 }, { "epoch": 12.506112469437653, "grad_norm": 0.31561336095688586, "learning_rate": 3.333380243728773e-07, "loss": 0.6593, "step": 5115 }, { "epoch": 12.508557457212714, "grad_norm": 0.21036725355614108, "learning_rate": 3.326996320194878e-07, "loss": 0.7012, "step": 5116 }, { "epoch": 12.511002444987774, "grad_norm": 0.23328705171523254, "learning_rate": 3.32061807981868e-07, "loss": 0.3479, "step": 5117 }, { "epoch": 12.513447432762836, "grad_norm": 0.3292262381453281, "learning_rate": 3.314245524272683e-07, "loss": 0.4543, "step": 5118 }, { "epoch": 12.515892420537897, "grad_norm": 0.2987884360875227, "learning_rate": 3.3078786552279334e-07, "loss": 0.4598, "step": 5119 }, { "epoch": 12.51833740831296, "grad_norm": 0.2641287563723492, "learning_rate": 3.3015174743539575e-07, "loss": 0.9034, "step": 5120 }, { "epoch": 12.52078239608802, "grad_norm": 0.38346876829145204, "learning_rate": 3.2951619833188093e-07, "loss": 0.3847, "step": 5121 }, { "epoch": 12.52322738386308, "grad_norm": 0.21425120542352002, "learning_rate": 3.288812183789039e-07, "loss": 0.8103, "step": 5122 }, { "epoch": 12.525672371638143, "grad_norm": 0.3610706255456914, "learning_rate": 3.282468077429715e-07, "loss": 0.5151, "step": 5123 }, { "epoch": 12.528117359413203, "grad_norm": 0.21380993238556853, "learning_rate": 3.276129665904412e-07, "loss": 0.552, "step": 5124 }, { "epoch": 12.530562347188264, "grad_norm": 0.278544477590112, "learning_rate": 3.2697969508751924e-07, "loss": 0.5829, "step": 5125 }, { "epoch": 12.533007334963326, "grad_norm": 0.19278899466571883, "learning_rate": 3.2634699340026474e-07, "loss": 0.4843, "step": 5126 }, { "epoch": 12.535452322738386, "grad_norm": 0.23775038342976557, "learning_rate": 3.2571486169458684e-07, "loss": 0.5256, "step": 5127 }, { "epoch": 12.537897310513447, "grad_norm": 0.3682607147593088, "learning_rate": 3.2508330013624414e-07, "loss": 0.4077, "step": 5128 }, { "epoch": 12.54034229828851, "grad_norm": 0.2780970695879525, "learning_rate": 3.2445230889084745e-07, "loss": 0.4224, "step": 5129 }, { "epoch": 12.54278728606357, "grad_norm": 0.19279153416205014, "learning_rate": 3.238218881238558e-07, "loss": 0.3991, "step": 5130 }, { "epoch": 12.54523227383863, "grad_norm": 0.18314314776183935, "learning_rate": 3.2319203800058123e-07, "loss": 0.3054, "step": 5131 }, { "epoch": 12.547677261613693, "grad_norm": 0.34802404020243805, "learning_rate": 3.225627586861849e-07, "loss": 0.7133, "step": 5132 }, { "epoch": 12.550122249388753, "grad_norm": 0.28021584397509935, "learning_rate": 3.2193405034567637e-07, "loss": 0.4132, "step": 5133 }, { "epoch": 12.552567237163814, "grad_norm": 0.18543097674854112, "learning_rate": 3.2130591314392e-07, "loss": 0.594, "step": 5134 }, { "epoch": 12.555012224938876, "grad_norm": 0.3405426231227696, "learning_rate": 3.2067834724562597e-07, "loss": 0.4504, "step": 5135 }, { "epoch": 12.557457212713937, "grad_norm": 0.24331450462542334, "learning_rate": 3.2005135281535703e-07, "loss": 0.8422, "step": 5136 }, { "epoch": 12.559902200488997, "grad_norm": 0.2810885896140483, "learning_rate": 3.194249300175259e-07, "loss": 0.4946, "step": 5137 }, { "epoch": 12.56234718826406, "grad_norm": 0.21835220042502648, "learning_rate": 3.187990790163936e-07, "loss": 0.5129, "step": 5138 }, { "epoch": 12.56479217603912, "grad_norm": 0.2389457234154542, "learning_rate": 3.181737999760745e-07, "loss": 0.453, "step": 5139 }, { "epoch": 12.56723716381418, "grad_norm": 0.15684544631973282, "learning_rate": 3.1754909306052905e-07, "loss": 0.4833, "step": 5140 }, { "epoch": 12.569682151589243, "grad_norm": 0.21536582613224417, "learning_rate": 3.1692495843357096e-07, "loss": 0.6389, "step": 5141 }, { "epoch": 12.572127139364303, "grad_norm": 0.19929910359836367, "learning_rate": 3.1630139625886293e-07, "loss": 0.4154, "step": 5142 }, { "epoch": 12.574572127139364, "grad_norm": 0.27472475571797134, "learning_rate": 3.1567840669991526e-07, "loss": 0.9067, "step": 5143 }, { "epoch": 12.577017114914426, "grad_norm": 0.30476957817864464, "learning_rate": 3.150559899200928e-07, "loss": 0.3928, "step": 5144 }, { "epoch": 12.579462102689487, "grad_norm": 0.19640029645154206, "learning_rate": 3.144341460826053e-07, "loss": 0.3893, "step": 5145 }, { "epoch": 12.581907090464547, "grad_norm": 0.31917464931799105, "learning_rate": 3.138128753505157e-07, "loss": 0.3029, "step": 5146 }, { "epoch": 12.58435207823961, "grad_norm": 0.37799947166298475, "learning_rate": 3.1319217788673526e-07, "loss": 0.4446, "step": 5147 }, { "epoch": 12.58679706601467, "grad_norm": 0.37987250764314784, "learning_rate": 3.1257205385402444e-07, "loss": 0.6995, "step": 5148 }, { "epoch": 12.58924205378973, "grad_norm": 0.25492651560285007, "learning_rate": 3.119525034149945e-07, "loss": 0.3702, "step": 5149 }, { "epoch": 12.591687041564793, "grad_norm": 0.28663233699812285, "learning_rate": 3.113335267321055e-07, "loss": 0.5644, "step": 5150 }, { "epoch": 12.594132029339853, "grad_norm": 0.17811696660540566, "learning_rate": 3.10715123967667e-07, "loss": 0.3584, "step": 5151 }, { "epoch": 12.596577017114914, "grad_norm": 0.2625880400575715, "learning_rate": 3.1009729528383935e-07, "loss": 0.6949, "step": 5152 }, { "epoch": 12.599022004889976, "grad_norm": 0.19809048724931394, "learning_rate": 3.0948004084263015e-07, "loss": 0.6932, "step": 5153 }, { "epoch": 12.601466992665037, "grad_norm": 0.26855607012489074, "learning_rate": 3.088633608058986e-07, "loss": 1.1185, "step": 5154 }, { "epoch": 12.603911980440097, "grad_norm": 0.24440314110278524, "learning_rate": 3.082472553353516e-07, "loss": 0.8806, "step": 5155 }, { "epoch": 12.60635696821516, "grad_norm": 0.20145101056708536, "learning_rate": 3.076317245925464e-07, "loss": 0.6365, "step": 5156 }, { "epoch": 12.60880195599022, "grad_norm": 0.4212754967308469, "learning_rate": 3.0701676873889e-07, "loss": 0.7532, "step": 5157 }, { "epoch": 12.61124694376528, "grad_norm": 0.24099640417528337, "learning_rate": 3.0640238793563674e-07, "loss": 0.3521, "step": 5158 }, { "epoch": 12.613691931540343, "grad_norm": 0.25297117513686623, "learning_rate": 3.057885823438925e-07, "loss": 0.5874, "step": 5159 }, { "epoch": 12.616136919315403, "grad_norm": 0.14923154617562195, "learning_rate": 3.051753521246095e-07, "loss": 0.5964, "step": 5160 }, { "epoch": 12.618581907090464, "grad_norm": 0.1941740101197978, "learning_rate": 3.045626974385921e-07, "loss": 0.4657, "step": 5161 }, { "epoch": 12.621026894865526, "grad_norm": 0.2878315654533411, "learning_rate": 3.0395061844649244e-07, "loss": 0.4447, "step": 5162 }, { "epoch": 12.623471882640587, "grad_norm": 0.26458991346153937, "learning_rate": 3.0333911530881056e-07, "loss": 0.4358, "step": 5163 }, { "epoch": 12.625916870415647, "grad_norm": 0.2603447637238707, "learning_rate": 3.0272818818589705e-07, "loss": 0.8532, "step": 5164 }, { "epoch": 12.62836185819071, "grad_norm": 0.3580342947442217, "learning_rate": 3.0211783723795156e-07, "loss": 0.4814, "step": 5165 }, { "epoch": 12.63080684596577, "grad_norm": 0.2704413722845608, "learning_rate": 3.0150806262502134e-07, "loss": 0.5383, "step": 5166 }, { "epoch": 12.63325183374083, "grad_norm": 0.4006365137459578, "learning_rate": 3.0089886450700434e-07, "loss": 0.6651, "step": 5167 }, { "epoch": 12.635696821515893, "grad_norm": 0.29950743493312165, "learning_rate": 3.002902430436447e-07, "loss": 0.7257, "step": 5168 }, { "epoch": 12.638141809290953, "grad_norm": 0.2799889912585528, "learning_rate": 2.9968219839453834e-07, "loss": 0.5379, "step": 5169 }, { "epoch": 12.640586797066014, "grad_norm": 0.28243778253276686, "learning_rate": 2.990747307191275e-07, "loss": 0.6974, "step": 5170 }, { "epoch": 12.643031784841076, "grad_norm": 0.2509942892720632, "learning_rate": 2.9846784017670436e-07, "loss": 0.6443, "step": 5171 }, { "epoch": 12.645476772616137, "grad_norm": 0.26492601259450355, "learning_rate": 2.978615269264101e-07, "loss": 0.6146, "step": 5172 }, { "epoch": 12.647921760391197, "grad_norm": 0.2922459343076512, "learning_rate": 2.972557911272328e-07, "loss": 0.5994, "step": 5173 }, { "epoch": 12.65036674816626, "grad_norm": 0.3375489669821644, "learning_rate": 2.966506329380112e-07, "loss": 0.4566, "step": 5174 }, { "epoch": 12.65281173594132, "grad_norm": 0.28047494143061924, "learning_rate": 2.9604605251743134e-07, "loss": 0.4858, "step": 5175 }, { "epoch": 12.65525672371638, "grad_norm": 0.21266152121398801, "learning_rate": 2.954420500240274e-07, "loss": 0.5052, "step": 5176 }, { "epoch": 12.657701711491443, "grad_norm": 0.24093700460948447, "learning_rate": 2.9483862561618423e-07, "loss": 0.4744, "step": 5177 }, { "epoch": 12.660146699266503, "grad_norm": 0.2679128287762201, "learning_rate": 2.9423577945213195e-07, "loss": 0.4327, "step": 5178 }, { "epoch": 12.662591687041564, "grad_norm": 0.30895518904870495, "learning_rate": 2.9363351168995137e-07, "loss": 0.4151, "step": 5179 }, { "epoch": 12.665036674816626, "grad_norm": 0.18931732182159922, "learning_rate": 2.930318224875714e-07, "loss": 0.6778, "step": 5180 }, { "epoch": 12.667481662591687, "grad_norm": 0.232150789310097, "learning_rate": 2.924307120027675e-07, "loss": 0.5085, "step": 5181 }, { "epoch": 12.669926650366747, "grad_norm": 0.3806923443868992, "learning_rate": 2.918301803931656e-07, "loss": 0.494, "step": 5182 }, { "epoch": 12.67237163814181, "grad_norm": 0.32740764702296093, "learning_rate": 2.912302278162379e-07, "loss": 0.4642, "step": 5183 }, { "epoch": 12.67481662591687, "grad_norm": 0.23302610622078251, "learning_rate": 2.906308544293063e-07, "loss": 0.6467, "step": 5184 }, { "epoch": 12.67726161369193, "grad_norm": 0.22754600499941294, "learning_rate": 2.900320603895407e-07, "loss": 0.4914, "step": 5185 }, { "epoch": 12.679706601466993, "grad_norm": 0.320721160588856, "learning_rate": 2.8943384585395695e-07, "loss": 0.3906, "step": 5186 }, { "epoch": 12.682151589242054, "grad_norm": 0.24620231175735657, "learning_rate": 2.8883621097942293e-07, "loss": 0.5993, "step": 5187 }, { "epoch": 12.684596577017114, "grad_norm": 0.3011897073637522, "learning_rate": 2.882391559226502e-07, "loss": 0.3867, "step": 5188 }, { "epoch": 12.687041564792176, "grad_norm": 0.2222029109860588, "learning_rate": 2.876426808402011e-07, "loss": 0.4039, "step": 5189 }, { "epoch": 12.689486552567237, "grad_norm": 0.3329278989083132, "learning_rate": 2.8704678588848537e-07, "loss": 0.3435, "step": 5190 }, { "epoch": 12.6919315403423, "grad_norm": 0.2002252995695997, "learning_rate": 2.8645147122375955e-07, "loss": 0.4659, "step": 5191 }, { "epoch": 12.69437652811736, "grad_norm": 0.33216250466344543, "learning_rate": 2.858567370021295e-07, "loss": 0.6221, "step": 5192 }, { "epoch": 12.69682151589242, "grad_norm": 0.2409126908139336, "learning_rate": 2.8526258337954747e-07, "loss": 0.7108, "step": 5193 }, { "epoch": 12.69926650366748, "grad_norm": 0.2650223443963618, "learning_rate": 2.8466901051181413e-07, "loss": 0.4551, "step": 5194 }, { "epoch": 12.701711491442543, "grad_norm": 0.2949734096850486, "learning_rate": 2.8407601855457904e-07, "loss": 0.3564, "step": 5195 }, { "epoch": 12.704156479217604, "grad_norm": 0.2649792369120882, "learning_rate": 2.8348360766333654e-07, "loss": 0.4216, "step": 5196 }, { "epoch": 12.706601466992666, "grad_norm": 0.22943497394287962, "learning_rate": 2.828917779934315e-07, "loss": 0.4269, "step": 5197 }, { "epoch": 12.709046454767726, "grad_norm": 0.32464884793854537, "learning_rate": 2.823005297000547e-07, "loss": 0.3871, "step": 5198 }, { "epoch": 12.711491442542787, "grad_norm": 0.35896404103761553, "learning_rate": 2.817098629382453e-07, "loss": 0.3829, "step": 5199 }, { "epoch": 12.713936430317847, "grad_norm": 0.27432890475642124, "learning_rate": 2.811197778628902e-07, "loss": 0.6644, "step": 5200 }, { "epoch": 12.71638141809291, "grad_norm": 0.25066415365345374, "learning_rate": 2.8053027462872193e-07, "loss": 0.3341, "step": 5201 }, { "epoch": 12.71882640586797, "grad_norm": 0.28049607441557894, "learning_rate": 2.799413533903228e-07, "loss": 0.4482, "step": 5202 }, { "epoch": 12.721271393643033, "grad_norm": 0.2146342635734247, "learning_rate": 2.793530143021203e-07, "loss": 0.4462, "step": 5203 }, { "epoch": 12.723716381418093, "grad_norm": 0.21652588508581377, "learning_rate": 2.7876525751839136e-07, "loss": 0.4361, "step": 5204 }, { "epoch": 12.726161369193154, "grad_norm": 0.2823082625689088, "learning_rate": 2.7817808319325957e-07, "loss": 0.416, "step": 5205 }, { "epoch": 12.728606356968216, "grad_norm": 0.3781358207524691, "learning_rate": 2.7759149148069405e-07, "loss": 0.4373, "step": 5206 }, { "epoch": 12.731051344743276, "grad_norm": 0.2116006880177325, "learning_rate": 2.770054825345142e-07, "loss": 0.3855, "step": 5207 }, { "epoch": 12.733496332518337, "grad_norm": 0.3615709300162841, "learning_rate": 2.7642005650838327e-07, "loss": 0.4363, "step": 5208 }, { "epoch": 12.7359413202934, "grad_norm": 0.1852833977046477, "learning_rate": 2.7583521355581456e-07, "loss": 0.4038, "step": 5209 }, { "epoch": 12.73838630806846, "grad_norm": 0.16976770063026442, "learning_rate": 2.7525095383016756e-07, "loss": 0.3218, "step": 5210 }, { "epoch": 12.74083129584352, "grad_norm": 0.32496451054913517, "learning_rate": 2.746672774846476e-07, "loss": 0.7784, "step": 5211 }, { "epoch": 12.743276283618583, "grad_norm": 0.21009765279073134, "learning_rate": 2.740841846723091e-07, "loss": 0.595, "step": 5212 }, { "epoch": 12.745721271393643, "grad_norm": 0.25883741429036217, "learning_rate": 2.7350167554605095e-07, "loss": 0.4179, "step": 5213 }, { "epoch": 12.748166259168704, "grad_norm": 0.25094756444846045, "learning_rate": 2.729197502586209e-07, "loss": 0.8024, "step": 5214 }, { "epoch": 12.750611246943766, "grad_norm": 0.2780265350247338, "learning_rate": 2.7233840896261383e-07, "loss": 0.4074, "step": 5215 }, { "epoch": 12.753056234718827, "grad_norm": 0.2514472285340264, "learning_rate": 2.7175765181046976e-07, "loss": 0.5315, "step": 5216 }, { "epoch": 12.755501222493887, "grad_norm": 0.22773834296021908, "learning_rate": 2.711774789544774e-07, "loss": 0.4848, "step": 5217 }, { "epoch": 12.75794621026895, "grad_norm": 0.30949999053136407, "learning_rate": 2.705978905467699e-07, "loss": 0.6573, "step": 5218 }, { "epoch": 12.76039119804401, "grad_norm": 0.2688742350389799, "learning_rate": 2.7001888673932997e-07, "loss": 0.4817, "step": 5219 }, { "epoch": 12.76283618581907, "grad_norm": 0.26348197540349444, "learning_rate": 2.694404676839857e-07, "loss": 0.496, "step": 5220 }, { "epoch": 12.765281173594133, "grad_norm": 0.3018462759952364, "learning_rate": 2.688626335324107e-07, "loss": 0.6579, "step": 5221 }, { "epoch": 12.767726161369193, "grad_norm": 0.24513101512945973, "learning_rate": 2.6828538443612775e-07, "loss": 0.4866, "step": 5222 }, { "epoch": 12.770171149144254, "grad_norm": 0.3155327842275826, "learning_rate": 2.6770872054650346e-07, "loss": 0.676, "step": 5223 }, { "epoch": 12.772616136919316, "grad_norm": 0.2754944703839549, "learning_rate": 2.671326420147527e-07, "loss": 1.4302, "step": 5224 }, { "epoch": 12.775061124694377, "grad_norm": 0.2482185391421997, "learning_rate": 2.6655714899193747e-07, "loss": 0.5609, "step": 5225 }, { "epoch": 12.777506112469437, "grad_norm": 0.2767043381750146, "learning_rate": 2.659822416289634e-07, "loss": 0.5948, "step": 5226 }, { "epoch": 12.7799511002445, "grad_norm": 0.2739869438031292, "learning_rate": 2.6540792007658604e-07, "loss": 1.2608, "step": 5227 }, { "epoch": 12.78239608801956, "grad_norm": 0.24464727717917417, "learning_rate": 2.6483418448540436e-07, "loss": 0.5985, "step": 5228 }, { "epoch": 12.78484107579462, "grad_norm": 0.1362553830402673, "learning_rate": 2.6426103500586497e-07, "loss": 0.3043, "step": 5229 }, { "epoch": 12.787286063569683, "grad_norm": 0.22163382716444335, "learning_rate": 2.6368847178826206e-07, "loss": 0.5105, "step": 5230 }, { "epoch": 12.789731051344743, "grad_norm": 0.3266005319060546, "learning_rate": 2.631164949827339e-07, "loss": 0.4125, "step": 5231 }, { "epoch": 12.792176039119804, "grad_norm": 0.24664218267627727, "learning_rate": 2.625451047392666e-07, "loss": 0.4946, "step": 5232 }, { "epoch": 12.794621026894866, "grad_norm": 0.37974253690985504, "learning_rate": 2.6197430120769057e-07, "loss": 0.7891, "step": 5233 }, { "epoch": 12.797066014669927, "grad_norm": 0.24601464662923175, "learning_rate": 2.6140408453768443e-07, "loss": 0.3651, "step": 5234 }, { "epoch": 12.799511002444987, "grad_norm": 0.23614770743011776, "learning_rate": 2.608344548787722e-07, "loss": 0.7393, "step": 5235 }, { "epoch": 12.80195599022005, "grad_norm": 0.2105610735170176, "learning_rate": 2.6026541238032305e-07, "loss": 0.4872, "step": 5236 }, { "epoch": 12.80440097799511, "grad_norm": 0.3014705821190384, "learning_rate": 2.5969695719155306e-07, "loss": 0.7052, "step": 5237 }, { "epoch": 12.80684596577017, "grad_norm": 0.2993352531596114, "learning_rate": 2.591290894615253e-07, "loss": 0.6453, "step": 5238 }, { "epoch": 12.809290953545233, "grad_norm": 0.29063579728885197, "learning_rate": 2.585618093391465e-07, "loss": 0.4788, "step": 5239 }, { "epoch": 12.811735941320293, "grad_norm": 0.18878750735181637, "learning_rate": 2.5799511697317063e-07, "loss": 0.5009, "step": 5240 }, { "epoch": 12.814180929095354, "grad_norm": 0.19710273983925064, "learning_rate": 2.5742901251219755e-07, "loss": 0.6211, "step": 5241 }, { "epoch": 12.816625916870416, "grad_norm": 0.3069539859022823, "learning_rate": 2.568634961046737e-07, "loss": 0.5093, "step": 5242 }, { "epoch": 12.819070904645477, "grad_norm": 0.23243765010478865, "learning_rate": 2.5629856789888886e-07, "loss": 0.4658, "step": 5243 }, { "epoch": 12.821515892420537, "grad_norm": 0.18214907839241887, "learning_rate": 2.5573422804298116e-07, "loss": 0.5782, "step": 5244 }, { "epoch": 12.8239608801956, "grad_norm": 0.279185238796717, "learning_rate": 2.551704766849336e-07, "loss": 0.3215, "step": 5245 }, { "epoch": 12.82640586797066, "grad_norm": 0.438622213658891, "learning_rate": 2.546073139725735e-07, "loss": 0.8295, "step": 5246 }, { "epoch": 12.82885085574572, "grad_norm": 0.319508634353802, "learning_rate": 2.5404474005357616e-07, "loss": 0.7422, "step": 5247 }, { "epoch": 12.831295843520783, "grad_norm": 0.18144734016536387, "learning_rate": 2.534827550754612e-07, "loss": 0.6788, "step": 5248 }, { "epoch": 12.833740831295843, "grad_norm": 0.3149596625076904, "learning_rate": 2.5292135918559336e-07, "loss": 0.4859, "step": 5249 }, { "epoch": 12.836185819070904, "grad_norm": 0.18640102926270288, "learning_rate": 2.5236055253118427e-07, "loss": 0.4601, "step": 5250 }, { "epoch": 12.838630806845966, "grad_norm": 0.19884349237845794, "learning_rate": 2.5180033525928903e-07, "loss": 0.8453, "step": 5251 }, { "epoch": 12.841075794621027, "grad_norm": 0.30900307263420557, "learning_rate": 2.5124070751681105e-07, "loss": 0.4911, "step": 5252 }, { "epoch": 12.843520782396087, "grad_norm": 0.260310064411079, "learning_rate": 2.5068166945049626e-07, "loss": 0.4059, "step": 5253 }, { "epoch": 12.84596577017115, "grad_norm": 0.2758246031019974, "learning_rate": 2.5012322120693767e-07, "loss": 1.0052, "step": 5254 }, { "epoch": 12.84841075794621, "grad_norm": 0.30640059668481695, "learning_rate": 2.4956536293257395e-07, "loss": 0.4693, "step": 5255 }, { "epoch": 12.85085574572127, "grad_norm": 0.21853809598914828, "learning_rate": 2.49008094773687e-07, "loss": 0.5638, "step": 5256 }, { "epoch": 12.853300733496333, "grad_norm": 0.19853819714831356, "learning_rate": 2.4845141687640575e-07, "loss": 0.3486, "step": 5257 }, { "epoch": 12.855745721271393, "grad_norm": 0.4405351296806944, "learning_rate": 2.478953293867045e-07, "loss": 0.5163, "step": 5258 }, { "epoch": 12.858190709046454, "grad_norm": 0.22562241260470897, "learning_rate": 2.473398324504014e-07, "loss": 0.3981, "step": 5259 }, { "epoch": 12.860635696821516, "grad_norm": 0.22175161090902956, "learning_rate": 2.467849262131608e-07, "loss": 0.4365, "step": 5260 }, { "epoch": 12.863080684596577, "grad_norm": 0.24053610884563748, "learning_rate": 2.462306108204912e-07, "loss": 0.7823, "step": 5261 }, { "epoch": 12.865525672371637, "grad_norm": 0.30698792317480383, "learning_rate": 2.4567688641774807e-07, "loss": 0.5416, "step": 5262 }, { "epoch": 12.8679706601467, "grad_norm": 0.27238465007926205, "learning_rate": 2.4512375315012936e-07, "loss": 0.9008, "step": 5263 }, { "epoch": 12.87041564792176, "grad_norm": 0.16045767618149717, "learning_rate": 2.445712111626797e-07, "loss": 0.4154, "step": 5264 }, { "epoch": 12.87286063569682, "grad_norm": 0.23513368548845895, "learning_rate": 2.440192606002889e-07, "loss": 0.6774, "step": 5265 }, { "epoch": 12.875305623471883, "grad_norm": 0.29438278935875534, "learning_rate": 2.434679016076899e-07, "loss": 0.5803, "step": 5266 }, { "epoch": 12.877750611246944, "grad_norm": 0.346545047447368, "learning_rate": 2.429171343294623e-07, "loss": 0.4551, "step": 5267 }, { "epoch": 12.880195599022004, "grad_norm": 0.3661193023855829, "learning_rate": 2.423669589100303e-07, "loss": 0.4447, "step": 5268 }, { "epoch": 12.882640586797066, "grad_norm": 0.18746366701342593, "learning_rate": 2.4181737549366116e-07, "loss": 0.4685, "step": 5269 }, { "epoch": 12.885085574572127, "grad_norm": 0.4406280828186168, "learning_rate": 2.412683842244698e-07, "loss": 0.7355, "step": 5270 }, { "epoch": 12.887530562347187, "grad_norm": 0.22495259508090046, "learning_rate": 2.4071998524641316e-07, "loss": 0.5261, "step": 5271 }, { "epoch": 12.88997555012225, "grad_norm": 0.2764002953090943, "learning_rate": 2.401721787032943e-07, "loss": 0.4324, "step": 5272 }, { "epoch": 12.89242053789731, "grad_norm": 0.3244219316132107, "learning_rate": 2.396249647387608e-07, "loss": 0.5819, "step": 5273 }, { "epoch": 12.894865525672373, "grad_norm": 0.29823243135004235, "learning_rate": 2.390783434963045e-07, "loss": 0.3971, "step": 5274 }, { "epoch": 12.897310513447433, "grad_norm": 0.38379828998583276, "learning_rate": 2.3853231511926254e-07, "loss": 0.4241, "step": 5275 }, { "epoch": 12.899755501222494, "grad_norm": 0.2449225238525846, "learning_rate": 2.379868797508153e-07, "loss": 0.4387, "step": 5276 }, { "epoch": 12.902200488997554, "grad_norm": 0.257087109060615, "learning_rate": 2.374420375339889e-07, "loss": 0.5999, "step": 5277 }, { "epoch": 12.904645476772616, "grad_norm": 0.21015208319102757, "learning_rate": 2.368977886116536e-07, "loss": 0.8812, "step": 5278 }, { "epoch": 12.907090464547677, "grad_norm": 0.24517387618174827, "learning_rate": 2.3635413312652317e-07, "loss": 0.329, "step": 5279 }, { "epoch": 12.90953545232274, "grad_norm": 0.19854802905360885, "learning_rate": 2.3581107122115727e-07, "loss": 0.5393, "step": 5280 }, { "epoch": 12.9119804400978, "grad_norm": 0.22126762132374855, "learning_rate": 2.3526860303795862e-07, "loss": 1.0191, "step": 5281 }, { "epoch": 12.91442542787286, "grad_norm": 0.29150296524215424, "learning_rate": 2.3472672871917512e-07, "loss": 0.7109, "step": 5282 }, { "epoch": 12.91687041564792, "grad_norm": 0.1676333631320654, "learning_rate": 2.341854484068984e-07, "loss": 0.3989, "step": 5283 }, { "epoch": 12.919315403422983, "grad_norm": 0.26081201047253366, "learning_rate": 2.3364476224306487e-07, "loss": 0.3281, "step": 5284 }, { "epoch": 12.921760391198044, "grad_norm": 0.3476032310703982, "learning_rate": 2.3310467036945506e-07, "loss": 0.4529, "step": 5285 }, { "epoch": 12.924205378973106, "grad_norm": 0.3573125308026871, "learning_rate": 2.3256517292769233e-07, "loss": 0.95, "step": 5286 }, { "epoch": 12.926650366748166, "grad_norm": 0.24438454382884092, "learning_rate": 2.32026270059246e-07, "loss": 0.3649, "step": 5287 }, { "epoch": 12.929095354523227, "grad_norm": 0.2834501584720449, "learning_rate": 2.3148796190542944e-07, "loss": 0.5657, "step": 5288 }, { "epoch": 12.93154034229829, "grad_norm": 0.37949541056886554, "learning_rate": 2.3095024860739778e-07, "loss": 0.5565, "step": 5289 }, { "epoch": 12.93398533007335, "grad_norm": 0.3100830799379428, "learning_rate": 2.3041313030615326e-07, "loss": 0.4378, "step": 5290 }, { "epoch": 12.93643031784841, "grad_norm": 0.3318476031877716, "learning_rate": 2.2987660714253957e-07, "loss": 0.7293, "step": 5291 }, { "epoch": 12.938875305623473, "grad_norm": 0.2922565759173333, "learning_rate": 2.2934067925724556e-07, "loss": 0.4255, "step": 5292 }, { "epoch": 12.941320293398533, "grad_norm": 0.32286458237381965, "learning_rate": 2.288053467908044e-07, "loss": 0.8685, "step": 5293 }, { "epoch": 12.943765281173594, "grad_norm": 0.21848679340543833, "learning_rate": 2.28270609883591e-07, "loss": 0.6735, "step": 5294 }, { "epoch": 12.946210268948656, "grad_norm": 0.3104114709910301, "learning_rate": 2.2773646867582766e-07, "loss": 0.3946, "step": 5295 }, { "epoch": 12.948655256723717, "grad_norm": 0.3432079540998316, "learning_rate": 2.2720292330757732e-07, "loss": 0.4689, "step": 5296 }, { "epoch": 12.951100244498777, "grad_norm": 0.35220692407940674, "learning_rate": 2.266699739187475e-07, "loss": 0.3511, "step": 5297 }, { "epoch": 12.95354523227384, "grad_norm": 0.2123614177686676, "learning_rate": 2.2613762064909088e-07, "loss": 0.935, "step": 5298 }, { "epoch": 12.9559902200489, "grad_norm": 0.22134645456079946, "learning_rate": 2.2560586363820136e-07, "loss": 0.8635, "step": 5299 }, { "epoch": 12.95843520782396, "grad_norm": 0.26908719672239767, "learning_rate": 2.2507470302551875e-07, "loss": 0.8641, "step": 5300 }, { "epoch": 12.960880195599023, "grad_norm": 0.2664152416064593, "learning_rate": 2.2454413895032502e-07, "loss": 0.4228, "step": 5301 }, { "epoch": 12.963325183374083, "grad_norm": 0.24451647326137868, "learning_rate": 2.2401417155174603e-07, "loss": 0.8422, "step": 5302 }, { "epoch": 12.965770171149144, "grad_norm": 0.2735374076193863, "learning_rate": 2.2348480096875236e-07, "loss": 0.739, "step": 5303 }, { "epoch": 12.968215158924206, "grad_norm": 0.26235278067517126, "learning_rate": 2.2295602734015543e-07, "loss": 1.2821, "step": 5304 }, { "epoch": 12.970660146699267, "grad_norm": 0.20280595631051063, "learning_rate": 2.2242785080461348e-07, "loss": 0.4569, "step": 5305 }, { "epoch": 12.973105134474327, "grad_norm": 0.20909248654981144, "learning_rate": 2.2190027150062577e-07, "loss": 0.4857, "step": 5306 }, { "epoch": 12.97555012224939, "grad_norm": 0.16615060105985366, "learning_rate": 2.2137328956653547e-07, "loss": 0.3971, "step": 5307 }, { "epoch": 12.97799511002445, "grad_norm": 0.2905958825547304, "learning_rate": 2.2084690514052988e-07, "loss": 0.49, "step": 5308 }, { "epoch": 12.98044009779951, "grad_norm": 0.2515527370419866, "learning_rate": 2.2032111836063824e-07, "loss": 0.3889, "step": 5309 }, { "epoch": 12.982885085574573, "grad_norm": 0.29437477432934656, "learning_rate": 2.1979592936473503e-07, "loss": 1.102, "step": 5310 }, { "epoch": 12.985330073349633, "grad_norm": 0.3444615176684247, "learning_rate": 2.1927133829053539e-07, "loss": 0.8434, "step": 5311 }, { "epoch": 12.987775061124694, "grad_norm": 0.22372589132459014, "learning_rate": 2.1874734527560004e-07, "loss": 0.4278, "step": 5312 }, { "epoch": 12.990220048899756, "grad_norm": 0.31864551462775814, "learning_rate": 2.1822395045733219e-07, "loss": 0.6109, "step": 5313 }, { "epoch": 12.992665036674817, "grad_norm": 0.18661989727995962, "learning_rate": 2.1770115397297676e-07, "loss": 0.7362, "step": 5314 }, { "epoch": 12.995110024449877, "grad_norm": 0.3008802839931786, "learning_rate": 2.1717895595962407e-07, "loss": 0.3169, "step": 5315 }, { "epoch": 12.99755501222494, "grad_norm": 0.21947199878524767, "learning_rate": 2.1665735655420572e-07, "loss": 0.7924, "step": 5316 }, { "epoch": 13.0, "grad_norm": 0.19176785869123447, "learning_rate": 2.1613635589349756e-07, "loss": 0.7732, "step": 5317 }, { "epoch": 13.00244498777506, "grad_norm": 0.2507008404251588, "learning_rate": 2.156159541141181e-07, "loss": 0.4056, "step": 5318 }, { "epoch": 13.004889975550123, "grad_norm": 0.23214940599424377, "learning_rate": 2.1509615135252765e-07, "loss": 0.7119, "step": 5319 }, { "epoch": 13.007334963325183, "grad_norm": 0.3312498623122705, "learning_rate": 2.145769477450313e-07, "loss": 0.6496, "step": 5320 }, { "epoch": 13.009779951100244, "grad_norm": 0.30629841105680783, "learning_rate": 2.1405834342777514e-07, "loss": 0.4458, "step": 5321 }, { "epoch": 13.012224938875306, "grad_norm": 0.274052827589616, "learning_rate": 2.1354033853674988e-07, "loss": 0.4463, "step": 5322 }, { "epoch": 13.014669926650367, "grad_norm": 0.2445954854397924, "learning_rate": 2.1302293320778823e-07, "loss": 0.4283, "step": 5323 }, { "epoch": 13.017114914425427, "grad_norm": 0.20344198660713717, "learning_rate": 2.1250612757656524e-07, "loss": 0.7197, "step": 5324 }, { "epoch": 13.01955990220049, "grad_norm": 0.21298510893567818, "learning_rate": 2.119899217785995e-07, "loss": 0.9394, "step": 5325 }, { "epoch": 13.02200488997555, "grad_norm": 0.2399602456370534, "learning_rate": 2.114743159492519e-07, "loss": 0.4625, "step": 5326 }, { "epoch": 13.02444987775061, "grad_norm": 0.3172765932094714, "learning_rate": 2.1095931022372617e-07, "loss": 0.8134, "step": 5327 }, { "epoch": 13.026894865525673, "grad_norm": 0.45600522058629817, "learning_rate": 2.104449047370691e-07, "loss": 0.3351, "step": 5328 }, { "epoch": 13.029339853300733, "grad_norm": 0.22104308683857282, "learning_rate": 2.0993109962416825e-07, "loss": 0.6739, "step": 5329 }, { "epoch": 13.031784841075794, "grad_norm": 0.3645849599527093, "learning_rate": 2.0941789501975674e-07, "loss": 0.5293, "step": 5330 }, { "epoch": 13.034229828850856, "grad_norm": 0.2010478825834421, "learning_rate": 2.0890529105840718e-07, "loss": 0.5359, "step": 5331 }, { "epoch": 13.036674816625917, "grad_norm": 0.32287510369430916, "learning_rate": 2.083932878745365e-07, "loss": 0.4544, "step": 5332 }, { "epoch": 13.039119804400977, "grad_norm": 0.2751944598769133, "learning_rate": 2.0788188560240424e-07, "loss": 0.4229, "step": 5333 }, { "epoch": 13.04156479217604, "grad_norm": 0.3539152542983435, "learning_rate": 2.0737108437611063e-07, "loss": 0.7066, "step": 5334 }, { "epoch": 13.0440097799511, "grad_norm": 0.18101064267533062, "learning_rate": 2.0686088432960023e-07, "loss": 0.4297, "step": 5335 }, { "epoch": 13.04645476772616, "grad_norm": 0.23542239064179407, "learning_rate": 2.063512855966593e-07, "loss": 0.5085, "step": 5336 }, { "epoch": 13.048899755501223, "grad_norm": 0.28634667190884916, "learning_rate": 2.0584228831091568e-07, "loss": 1.0158, "step": 5337 }, { "epoch": 13.051344743276283, "grad_norm": 0.238126699886235, "learning_rate": 2.0533389260584125e-07, "loss": 0.6384, "step": 5338 }, { "epoch": 13.053789731051344, "grad_norm": 0.23313225719223077, "learning_rate": 2.0482609861474768e-07, "loss": 0.4673, "step": 5339 }, { "epoch": 13.056234718826406, "grad_norm": 0.2861646552161169, "learning_rate": 2.0431890647079094e-07, "loss": 0.3791, "step": 5340 }, { "epoch": 13.058679706601467, "grad_norm": 0.4015048471701256, "learning_rate": 2.0381231630696803e-07, "loss": 0.3763, "step": 5341 }, { "epoch": 13.061124694376527, "grad_norm": 0.3157175129837912, "learning_rate": 2.0330632825611847e-07, "loss": 0.4153, "step": 5342 }, { "epoch": 13.06356968215159, "grad_norm": 0.4269116108957552, "learning_rate": 2.0280094245092473e-07, "loss": 0.4999, "step": 5343 }, { "epoch": 13.06601466992665, "grad_norm": 0.43000251455104793, "learning_rate": 2.022961590239092e-07, "loss": 0.3838, "step": 5344 }, { "epoch": 13.06845965770171, "grad_norm": 0.27999802326962814, "learning_rate": 2.0179197810743844e-07, "loss": 0.4955, "step": 5345 }, { "epoch": 13.070904645476773, "grad_norm": 0.18722740493649925, "learning_rate": 2.0128839983372089e-07, "loss": 0.4067, "step": 5346 }, { "epoch": 13.073349633251834, "grad_norm": 0.3547853253482003, "learning_rate": 2.0078542433480424e-07, "loss": 0.3308, "step": 5347 }, { "epoch": 13.075794621026894, "grad_norm": 0.30296289033957463, "learning_rate": 2.0028305174258271e-07, "loss": 1.0912, "step": 5348 }, { "epoch": 13.078239608801956, "grad_norm": 0.2684021167343446, "learning_rate": 1.9978128218878818e-07, "loss": 0.3196, "step": 5349 }, { "epoch": 13.080684596577017, "grad_norm": 0.18875785564764416, "learning_rate": 1.9928011580499673e-07, "loss": 0.6256, "step": 5350 }, { "epoch": 13.083129584352077, "grad_norm": 0.27504183176476815, "learning_rate": 1.987795527226255e-07, "loss": 0.5669, "step": 5351 }, { "epoch": 13.08557457212714, "grad_norm": 0.3561250770631697, "learning_rate": 1.9827959307293338e-07, "loss": 0.3427, "step": 5352 }, { "epoch": 13.0880195599022, "grad_norm": 0.1853786784129405, "learning_rate": 1.9778023698702213e-07, "loss": 0.7762, "step": 5353 }, { "epoch": 13.09046454767726, "grad_norm": 0.2876215987968845, "learning_rate": 1.972814845958332e-07, "loss": 1.1142, "step": 5354 }, { "epoch": 13.092909535452323, "grad_norm": 0.28513341283044574, "learning_rate": 1.9678333603015133e-07, "loss": 0.702, "step": 5355 }, { "epoch": 13.095354523227384, "grad_norm": 0.42181986000768024, "learning_rate": 1.962857914206029e-07, "loss": 0.3485, "step": 5356 }, { "epoch": 13.097799511002446, "grad_norm": 0.43440265389411786, "learning_rate": 1.9578885089765465e-07, "loss": 0.9741, "step": 5357 }, { "epoch": 13.100244498777506, "grad_norm": 0.2794655809888336, "learning_rate": 1.9529251459161653e-07, "loss": 0.6586, "step": 5358 }, { "epoch": 13.102689486552567, "grad_norm": 0.17775960780612599, "learning_rate": 1.9479678263263858e-07, "loss": 0.4277, "step": 5359 }, { "epoch": 13.10513447432763, "grad_norm": 0.2680588259617518, "learning_rate": 1.943016551507143e-07, "loss": 0.4925, "step": 5360 }, { "epoch": 13.10757946210269, "grad_norm": 0.39906458816742285, "learning_rate": 1.93807132275676e-07, "loss": 0.4112, "step": 5361 }, { "epoch": 13.11002444987775, "grad_norm": 0.30471016300707354, "learning_rate": 1.933132141371996e-07, "loss": 0.4946, "step": 5362 }, { "epoch": 13.112469437652813, "grad_norm": 0.2098381639596224, "learning_rate": 1.9281990086480186e-07, "loss": 0.4547, "step": 5363 }, { "epoch": 13.114914425427873, "grad_norm": 0.2104823333765495, "learning_rate": 1.9232719258784032e-07, "loss": 0.3506, "step": 5364 }, { "epoch": 13.117359413202934, "grad_norm": 0.3558425152259091, "learning_rate": 1.9183508943551478e-07, "loss": 0.6394, "step": 5365 }, { "epoch": 13.119804400977996, "grad_norm": 0.13938805273559338, "learning_rate": 1.913435915368664e-07, "loss": 0.3192, "step": 5366 }, { "epoch": 13.122249388753056, "grad_norm": 0.26666424210265416, "learning_rate": 1.9085269902077596e-07, "loss": 0.5769, "step": 5367 }, { "epoch": 13.124694376528117, "grad_norm": 0.2485025781861825, "learning_rate": 1.9036241201596777e-07, "loss": 1.1123, "step": 5368 }, { "epoch": 13.12713936430318, "grad_norm": 0.2848209180720596, "learning_rate": 1.898727306510051e-07, "loss": 0.4232, "step": 5369 }, { "epoch": 13.12958435207824, "grad_norm": 0.22310877988146, "learning_rate": 1.8938365505429546e-07, "loss": 0.4586, "step": 5370 }, { "epoch": 13.1320293398533, "grad_norm": 0.2781800836351792, "learning_rate": 1.8889518535408408e-07, "loss": 0.4173, "step": 5371 }, { "epoch": 13.134474327628363, "grad_norm": 0.1858373991184152, "learning_rate": 1.884073216784593e-07, "loss": 0.421, "step": 5372 }, { "epoch": 13.136919315403423, "grad_norm": 0.26055126789028776, "learning_rate": 1.879200641553508e-07, "loss": 0.4762, "step": 5373 }, { "epoch": 13.139364303178484, "grad_norm": 0.2586943160775621, "learning_rate": 1.874334129125277e-07, "loss": 0.8741, "step": 5374 }, { "epoch": 13.141809290953546, "grad_norm": 0.27012592786365863, "learning_rate": 1.8694736807760182e-07, "loss": 1.148, "step": 5375 }, { "epoch": 13.144254278728607, "grad_norm": 0.24833506044983888, "learning_rate": 1.8646192977802512e-07, "loss": 1.0364, "step": 5376 }, { "epoch": 13.146699266503667, "grad_norm": 0.2627093372286388, "learning_rate": 1.8597709814108994e-07, "loss": 0.441, "step": 5377 }, { "epoch": 13.14914425427873, "grad_norm": 0.33038045302867636, "learning_rate": 1.8549287329393152e-07, "loss": 0.3233, "step": 5378 }, { "epoch": 13.15158924205379, "grad_norm": 0.2294171361742112, "learning_rate": 1.85009255363523e-07, "loss": 0.5905, "step": 5379 }, { "epoch": 13.15403422982885, "grad_norm": 0.3288735135583903, "learning_rate": 1.8452624447668183e-07, "loss": 0.6526, "step": 5380 }, { "epoch": 13.156479217603913, "grad_norm": 0.33241961435341855, "learning_rate": 1.8404384076006337e-07, "loss": 0.5513, "step": 5381 }, { "epoch": 13.158924205378973, "grad_norm": 0.3128579067905457, "learning_rate": 1.8356204434016506e-07, "loss": 0.6678, "step": 5382 }, { "epoch": 13.161369193154034, "grad_norm": 0.30575572484832775, "learning_rate": 1.8308085534332583e-07, "loss": 0.9598, "step": 5383 }, { "epoch": 13.163814180929096, "grad_norm": 0.30994106215091816, "learning_rate": 1.8260027389572304e-07, "loss": 0.496, "step": 5384 }, { "epoch": 13.166259168704157, "grad_norm": 0.25375397853172155, "learning_rate": 1.8212030012337707e-07, "loss": 0.4537, "step": 5385 }, { "epoch": 13.168704156479217, "grad_norm": 0.23529172021408684, "learning_rate": 1.816409341521483e-07, "loss": 0.5188, "step": 5386 }, { "epoch": 13.17114914425428, "grad_norm": 0.37632655701366396, "learning_rate": 1.8116217610773674e-07, "loss": 0.5368, "step": 5387 }, { "epoch": 13.17359413202934, "grad_norm": 0.28815332515151965, "learning_rate": 1.806840261156842e-07, "loss": 0.4294, "step": 5388 }, { "epoch": 13.1760391198044, "grad_norm": 0.3420186153916135, "learning_rate": 1.8020648430137178e-07, "loss": 0.4324, "step": 5389 }, { "epoch": 13.178484107579463, "grad_norm": 0.23125020863378182, "learning_rate": 1.7972955079002264e-07, "loss": 0.6136, "step": 5390 }, { "epoch": 13.180929095354523, "grad_norm": 0.2108950475877583, "learning_rate": 1.7925322570669922e-07, "loss": 0.3598, "step": 5391 }, { "epoch": 13.183374083129584, "grad_norm": 0.23539782124721473, "learning_rate": 1.7877750917630526e-07, "loss": 0.4782, "step": 5392 }, { "epoch": 13.185819070904646, "grad_norm": 0.2570534941175301, "learning_rate": 1.7830240132358483e-07, "loss": 0.3959, "step": 5393 }, { "epoch": 13.188264058679707, "grad_norm": 0.19278439001622985, "learning_rate": 1.7782790227312107e-07, "loss": 0.4183, "step": 5394 }, { "epoch": 13.190709046454767, "grad_norm": 0.22205381240663455, "learning_rate": 1.7735401214933917e-07, "loss": 0.5167, "step": 5395 }, { "epoch": 13.19315403422983, "grad_norm": 0.3077070213800318, "learning_rate": 1.768807310765039e-07, "loss": 0.9318, "step": 5396 }, { "epoch": 13.19559902200489, "grad_norm": 0.24152133875360682, "learning_rate": 1.7640805917872012e-07, "loss": 0.5515, "step": 5397 }, { "epoch": 13.19804400977995, "grad_norm": 0.3141975585054514, "learning_rate": 1.7593599657993398e-07, "loss": 0.4818, "step": 5398 }, { "epoch": 13.200488997555013, "grad_norm": 0.18260588498724392, "learning_rate": 1.7546454340392982e-07, "loss": 0.4238, "step": 5399 }, { "epoch": 13.202933985330073, "grad_norm": 0.34983530438151883, "learning_rate": 1.7499369977433455e-07, "loss": 0.5123, "step": 5400 }, { "epoch": 13.205378973105134, "grad_norm": 0.3478786911167567, "learning_rate": 1.7452346581461332e-07, "loss": 0.6526, "step": 5401 }, { "epoch": 13.207823960880196, "grad_norm": 0.23796683240605365, "learning_rate": 1.7405384164807304e-07, "loss": 0.6726, "step": 5402 }, { "epoch": 13.210268948655257, "grad_norm": 0.3156136031891151, "learning_rate": 1.7358482739785994e-07, "loss": 0.3996, "step": 5403 }, { "epoch": 13.212713936430317, "grad_norm": 0.2829726352384672, "learning_rate": 1.731164231869595e-07, "loss": 0.4556, "step": 5404 }, { "epoch": 13.21515892420538, "grad_norm": 0.289498340957733, "learning_rate": 1.726486291381982e-07, "loss": 0.7063, "step": 5405 }, { "epoch": 13.21760391198044, "grad_norm": 0.2864270363966545, "learning_rate": 1.721814453742432e-07, "loss": 0.5701, "step": 5406 }, { "epoch": 13.2200488997555, "grad_norm": 0.2604235273711713, "learning_rate": 1.7171487201759985e-07, "loss": 0.4336, "step": 5407 }, { "epoch": 13.222493887530563, "grad_norm": 0.2901454280928388, "learning_rate": 1.7124890919061499e-07, "loss": 0.3852, "step": 5408 }, { "epoch": 13.224938875305623, "grad_norm": 0.2739360748475561, "learning_rate": 1.7078355701547395e-07, "loss": 0.5694, "step": 5409 }, { "epoch": 13.227383863080684, "grad_norm": 0.29789762494675154, "learning_rate": 1.7031881561420328e-07, "loss": 0.8256, "step": 5410 }, { "epoch": 13.229828850855746, "grad_norm": 0.2996639185880644, "learning_rate": 1.698546851086691e-07, "loss": 0.5712, "step": 5411 }, { "epoch": 13.232273838630807, "grad_norm": 0.2114121850825437, "learning_rate": 1.6939116562057602e-07, "loss": 0.5365, "step": 5412 }, { "epoch": 13.234718826405867, "grad_norm": 0.24672784008966578, "learning_rate": 1.6892825727147093e-07, "loss": 0.8515, "step": 5413 }, { "epoch": 13.23716381418093, "grad_norm": 0.22342188053126433, "learning_rate": 1.6846596018273765e-07, "loss": 0.727, "step": 5414 }, { "epoch": 13.23960880195599, "grad_norm": 0.19140015142261502, "learning_rate": 1.6800427447560163e-07, "loss": 0.5274, "step": 5415 }, { "epoch": 13.24205378973105, "grad_norm": 0.26030001943765374, "learning_rate": 1.6754320027112804e-07, "loss": 0.5158, "step": 5416 }, { "epoch": 13.244498777506113, "grad_norm": 0.31538007506023785, "learning_rate": 1.6708273769021982e-07, "loss": 0.4878, "step": 5417 }, { "epoch": 13.246943765281173, "grad_norm": 0.26249968645003663, "learning_rate": 1.666228868536218e-07, "loss": 0.9716, "step": 5418 }, { "epoch": 13.249388753056234, "grad_norm": 0.29411296807861387, "learning_rate": 1.6616364788191701e-07, "loss": 0.5538, "step": 5419 }, { "epoch": 13.251833740831296, "grad_norm": 0.3100028316246551, "learning_rate": 1.6570502089552797e-07, "loss": 0.3993, "step": 5420 }, { "epoch": 13.254278728606357, "grad_norm": 0.28717284408981897, "learning_rate": 1.652470060147185e-07, "loss": 0.3507, "step": 5421 }, { "epoch": 13.256723716381417, "grad_norm": 0.2728850107273424, "learning_rate": 1.647896033595886e-07, "loss": 0.3709, "step": 5422 }, { "epoch": 13.25916870415648, "grad_norm": 0.24961662446534846, "learning_rate": 1.643328130500818e-07, "loss": 0.8345, "step": 5423 }, { "epoch": 13.26161369193154, "grad_norm": 0.23503284287766496, "learning_rate": 1.6387663520597759e-07, "loss": 0.8496, "step": 5424 }, { "epoch": 13.2640586797066, "grad_norm": 0.20369153136981225, "learning_rate": 1.634210699468969e-07, "loss": 0.3983, "step": 5425 }, { "epoch": 13.266503667481663, "grad_norm": 0.35033394533814466, "learning_rate": 1.6296611739229916e-07, "loss": 0.6519, "step": 5426 }, { "epoch": 13.268948655256724, "grad_norm": 0.3200013353050572, "learning_rate": 1.6251177766148319e-07, "loss": 0.6848, "step": 5427 }, { "epoch": 13.271393643031784, "grad_norm": 0.16736278824731854, "learning_rate": 1.6205805087358778e-07, "loss": 0.4779, "step": 5428 }, { "epoch": 13.273838630806846, "grad_norm": 0.21544050507727647, "learning_rate": 1.6160493714758973e-07, "loss": 0.363, "step": 5429 }, { "epoch": 13.276283618581907, "grad_norm": 0.3471500597252846, "learning_rate": 1.6115243660230624e-07, "loss": 0.4872, "step": 5430 }, { "epoch": 13.278728606356967, "grad_norm": 0.2169835441942447, "learning_rate": 1.6070054935639373e-07, "loss": 0.4654, "step": 5431 }, { "epoch": 13.28117359413203, "grad_norm": 0.31038049465885664, "learning_rate": 1.6024927552834634e-07, "loss": 0.4282, "step": 5432 }, { "epoch": 13.28361858190709, "grad_norm": 0.2591053837566953, "learning_rate": 1.5979861523649914e-07, "loss": 0.3958, "step": 5433 }, { "epoch": 13.28606356968215, "grad_norm": 0.3058535559941627, "learning_rate": 1.5934856859902535e-07, "loss": 0.663, "step": 5434 }, { "epoch": 13.288508557457213, "grad_norm": 0.2505403970219653, "learning_rate": 1.5889913573393784e-07, "loss": 0.7135, "step": 5435 }, { "epoch": 13.290953545232274, "grad_norm": 0.2753032421602159, "learning_rate": 1.584503167590881e-07, "loss": 0.9208, "step": 5436 }, { "epoch": 13.293398533007334, "grad_norm": 0.2441708607650385, "learning_rate": 1.5800211179216622e-07, "loss": 0.475, "step": 5437 }, { "epoch": 13.295843520782396, "grad_norm": 0.26648614781931507, "learning_rate": 1.5755452095070262e-07, "loss": 0.5796, "step": 5438 }, { "epoch": 13.298288508557457, "grad_norm": 0.23103480182826988, "learning_rate": 1.5710754435206476e-07, "loss": 0.5013, "step": 5439 }, { "epoch": 13.30073349633252, "grad_norm": 0.42689724115997363, "learning_rate": 1.5666118211346115e-07, "loss": 0.4449, "step": 5440 }, { "epoch": 13.30317848410758, "grad_norm": 0.20371610254782882, "learning_rate": 1.5621543435193815e-07, "loss": 1.0173, "step": 5441 }, { "epoch": 13.30562347188264, "grad_norm": 0.3845821748326833, "learning_rate": 1.5577030118438002e-07, "loss": 0.5033, "step": 5442 }, { "epoch": 13.3080684596577, "grad_norm": 0.26362582317648026, "learning_rate": 1.5532578272751176e-07, "loss": 0.5584, "step": 5443 }, { "epoch": 13.310513447432763, "grad_norm": 0.4045725625627985, "learning_rate": 1.5488187909789592e-07, "loss": 0.4591, "step": 5444 }, { "epoch": 13.312958435207824, "grad_norm": 0.24794199515767687, "learning_rate": 1.5443859041193443e-07, "loss": 1.285, "step": 5445 }, { "epoch": 13.315403422982886, "grad_norm": 0.2700159370747307, "learning_rate": 1.5399591678586818e-07, "loss": 0.4318, "step": 5446 }, { "epoch": 13.317848410757946, "grad_norm": 0.288098459304985, "learning_rate": 1.5355385833577542e-07, "loss": 0.6339, "step": 5447 }, { "epoch": 13.320293398533007, "grad_norm": 0.2733879970621616, "learning_rate": 1.5311241517757453e-07, "loss": 0.3282, "step": 5448 }, { "epoch": 13.32273838630807, "grad_norm": 0.24433091524197376, "learning_rate": 1.5267158742702154e-07, "loss": 0.7697, "step": 5449 }, { "epoch": 13.32518337408313, "grad_norm": 0.2552396165325618, "learning_rate": 1.5223137519971175e-07, "loss": 0.6787, "step": 5450 }, { "epoch": 13.32762836185819, "grad_norm": 0.19813482855934522, "learning_rate": 1.5179177861107953e-07, "loss": 0.6919, "step": 5451 }, { "epoch": 13.330073349633253, "grad_norm": 0.32427617012288973, "learning_rate": 1.5135279777639622e-07, "loss": 0.8107, "step": 5452 }, { "epoch": 13.332518337408313, "grad_norm": 0.2844620406411681, "learning_rate": 1.509144328107731e-07, "loss": 0.4028, "step": 5453 }, { "epoch": 13.334963325183374, "grad_norm": 0.40510371931792327, "learning_rate": 1.5047668382915965e-07, "loss": 0.4015, "step": 5454 }, { "epoch": 13.337408312958436, "grad_norm": 0.23784702829753648, "learning_rate": 1.5003955094634316e-07, "loss": 0.5732, "step": 5455 }, { "epoch": 13.339853300733497, "grad_norm": 0.3776919732760176, "learning_rate": 1.4960303427695083e-07, "loss": 0.7404, "step": 5456 }, { "epoch": 13.342298288508557, "grad_norm": 0.2629738149289269, "learning_rate": 1.491671339354464e-07, "loss": 0.6322, "step": 5457 }, { "epoch": 13.34474327628362, "grad_norm": 0.4048695479502515, "learning_rate": 1.4873185003613344e-07, "loss": 0.5214, "step": 5458 }, { "epoch": 13.34718826405868, "grad_norm": 0.16627672781917818, "learning_rate": 1.4829718269315286e-07, "loss": 0.3935, "step": 5459 }, { "epoch": 13.34963325183374, "grad_norm": 0.18038169680147564, "learning_rate": 1.4786313202048457e-07, "loss": 0.3242, "step": 5460 }, { "epoch": 13.352078239608803, "grad_norm": 0.25426090928832334, "learning_rate": 1.474296981319473e-07, "loss": 0.6927, "step": 5461 }, { "epoch": 13.354523227383863, "grad_norm": 0.27156593608876545, "learning_rate": 1.4699688114119591e-07, "loss": 1.21, "step": 5462 }, { "epoch": 13.356968215158924, "grad_norm": 0.26794482778588635, "learning_rate": 1.4656468116172606e-07, "loss": 0.5235, "step": 5463 }, { "epoch": 13.359413202933986, "grad_norm": 0.262087658132917, "learning_rate": 1.4613309830687068e-07, "loss": 0.7544, "step": 5464 }, { "epoch": 13.361858190709047, "grad_norm": 0.22971049088929713, "learning_rate": 1.4570213268979922e-07, "loss": 0.6002, "step": 5465 }, { "epoch": 13.364303178484107, "grad_norm": 0.2545606803547083, "learning_rate": 1.4527178442352242e-07, "loss": 0.5226, "step": 5466 }, { "epoch": 13.36674816625917, "grad_norm": 0.2798879064892046, "learning_rate": 1.448420536208861e-07, "loss": 0.6584, "step": 5467 }, { "epoch": 13.36919315403423, "grad_norm": 0.31484292006273545, "learning_rate": 1.4441294039457683e-07, "loss": 0.4756, "step": 5468 }, { "epoch": 13.37163814180929, "grad_norm": 0.22179126495474688, "learning_rate": 1.4398444485711677e-07, "loss": 0.4373, "step": 5469 }, { "epoch": 13.374083129584353, "grad_norm": 0.48191616107992935, "learning_rate": 1.4355656712086746e-07, "loss": 0.3548, "step": 5470 }, { "epoch": 13.376528117359413, "grad_norm": 0.2245549080198216, "learning_rate": 1.4312930729802887e-07, "loss": 0.7357, "step": 5471 }, { "epoch": 13.378973105134474, "grad_norm": 0.26151041781822554, "learning_rate": 1.4270266550063777e-07, "loss": 1.2821, "step": 5472 }, { "epoch": 13.381418092909536, "grad_norm": 0.21174856190448904, "learning_rate": 1.422766418405694e-07, "loss": 0.5342, "step": 5473 }, { "epoch": 13.383863080684597, "grad_norm": 0.3366622616466138, "learning_rate": 1.4185123642953742e-07, "loss": 0.8511, "step": 5474 }, { "epoch": 13.386308068459657, "grad_norm": 0.32167258081763, "learning_rate": 1.4142644937909206e-07, "loss": 0.6384, "step": 5475 }, { "epoch": 13.38875305623472, "grad_norm": 0.34986141558194417, "learning_rate": 1.4100228080062284e-07, "loss": 0.7247, "step": 5476 }, { "epoch": 13.39119804400978, "grad_norm": 0.3124050914274767, "learning_rate": 1.4057873080535605e-07, "loss": 0.7543, "step": 5477 }, { "epoch": 13.39364303178484, "grad_norm": 0.342714201933395, "learning_rate": 1.4015579950435676e-07, "loss": 0.5818, "step": 5478 }, { "epoch": 13.396088019559903, "grad_norm": 0.34807431439368264, "learning_rate": 1.3973348700852678e-07, "loss": 0.4377, "step": 5479 }, { "epoch": 13.398533007334963, "grad_norm": 0.33408715591217425, "learning_rate": 1.3931179342860618e-07, "loss": 0.4069, "step": 5480 }, { "epoch": 13.400977995110024, "grad_norm": 0.2824110238634206, "learning_rate": 1.3889071887517287e-07, "loss": 0.8315, "step": 5481 }, { "epoch": 13.403422982885086, "grad_norm": 0.17699753469013663, "learning_rate": 1.3847026345864184e-07, "loss": 0.5688, "step": 5482 }, { "epoch": 13.405867970660147, "grad_norm": 0.3802514115381719, "learning_rate": 1.380504272892666e-07, "loss": 0.4387, "step": 5483 }, { "epoch": 13.408312958435207, "grad_norm": 0.29446160182884545, "learning_rate": 1.3763121047713767e-07, "loss": 0.4712, "step": 5484 }, { "epoch": 13.41075794621027, "grad_norm": 0.18475740564409524, "learning_rate": 1.3721261313218298e-07, "loss": 0.7188, "step": 5485 }, { "epoch": 13.41320293398533, "grad_norm": 0.2833861629393022, "learning_rate": 1.3679463536416887e-07, "loss": 0.796, "step": 5486 }, { "epoch": 13.41564792176039, "grad_norm": 0.267642139077156, "learning_rate": 1.363772772826974e-07, "loss": 0.4476, "step": 5487 }, { "epoch": 13.418092909535453, "grad_norm": 0.28880346818983627, "learning_rate": 1.3596053899721157e-07, "loss": 0.4699, "step": 5488 }, { "epoch": 13.420537897310513, "grad_norm": 0.22870223697992778, "learning_rate": 1.3554442061698785e-07, "loss": 0.9342, "step": 5489 }, { "epoch": 13.422982885085574, "grad_norm": 0.2816303351311041, "learning_rate": 1.351289222511426e-07, "loss": 0.4656, "step": 5490 }, { "epoch": 13.425427872860636, "grad_norm": 0.4071698967253718, "learning_rate": 1.3471404400862976e-07, "loss": 0.7067, "step": 5491 }, { "epoch": 13.427872860635697, "grad_norm": 0.1893772907896548, "learning_rate": 1.342997859982384e-07, "loss": 0.6452, "step": 5492 }, { "epoch": 13.430317848410757, "grad_norm": 0.26997642285480666, "learning_rate": 1.3388614832859774e-07, "loss": 0.6897, "step": 5493 }, { "epoch": 13.43276283618582, "grad_norm": 0.24348747411767688, "learning_rate": 1.3347313110817272e-07, "loss": 0.8249, "step": 5494 }, { "epoch": 13.43520782396088, "grad_norm": 0.32386295527794956, "learning_rate": 1.330607344452653e-07, "loss": 0.4036, "step": 5495 }, { "epoch": 13.43765281173594, "grad_norm": 0.256221278276785, "learning_rate": 1.3264895844801618e-07, "loss": 0.7382, "step": 5496 }, { "epoch": 13.440097799511003, "grad_norm": 0.31016210968045405, "learning_rate": 1.322378032244015e-07, "loss": 0.5195, "step": 5497 }, { "epoch": 13.442542787286063, "grad_norm": 0.25013645072378415, "learning_rate": 1.3182726888223667e-07, "loss": 0.7298, "step": 5498 }, { "epoch": 13.444987775061124, "grad_norm": 0.1846426342626662, "learning_rate": 1.3141735552917228e-07, "loss": 0.435, "step": 5499 }, { "epoch": 13.447432762836186, "grad_norm": 0.32572540090788976, "learning_rate": 1.3100806327269733e-07, "loss": 0.4677, "step": 5500 }, { "epoch": 13.449877750611247, "grad_norm": 0.22450410603233825, "learning_rate": 1.3059939222013818e-07, "loss": 0.6829, "step": 5501 }, { "epoch": 13.452322738386307, "grad_norm": 0.23357462477144836, "learning_rate": 1.301913424786569e-07, "loss": 0.4415, "step": 5502 }, { "epoch": 13.45476772616137, "grad_norm": 0.2727292281769691, "learning_rate": 1.297839141552537e-07, "loss": 0.4654, "step": 5503 }, { "epoch": 13.45721271393643, "grad_norm": 0.24544774538442626, "learning_rate": 1.2937710735676596e-07, "loss": 0.8823, "step": 5504 }, { "epoch": 13.45965770171149, "grad_norm": 0.34006390174217604, "learning_rate": 1.2897092218986718e-07, "loss": 0.4274, "step": 5505 }, { "epoch": 13.462102689486553, "grad_norm": 0.26371114119596045, "learning_rate": 1.2856535876106886e-07, "loss": 0.9728, "step": 5506 }, { "epoch": 13.464547677261614, "grad_norm": 0.2151636074054762, "learning_rate": 1.281604171767187e-07, "loss": 0.4613, "step": 5507 }, { "epoch": 13.466992665036674, "grad_norm": 0.38804716176156856, "learning_rate": 1.27756097543002e-07, "loss": 0.8472, "step": 5508 }, { "epoch": 13.469437652811736, "grad_norm": 0.2536928639402621, "learning_rate": 1.2735239996594013e-07, "loss": 0.37, "step": 5509 }, { "epoch": 13.471882640586797, "grad_norm": 0.20325778689076104, "learning_rate": 1.2694932455139226e-07, "loss": 0.4524, "step": 5510 }, { "epoch": 13.474327628361857, "grad_norm": 0.28878343008149404, "learning_rate": 1.2654687140505413e-07, "loss": 0.3618, "step": 5511 }, { "epoch": 13.47677261613692, "grad_norm": 0.30093682869050103, "learning_rate": 1.261450406324577e-07, "loss": 0.4545, "step": 5512 }, { "epoch": 13.47921760391198, "grad_norm": 0.19270717204212476, "learning_rate": 1.257438323389723e-07, "loss": 0.3434, "step": 5513 }, { "epoch": 13.48166259168704, "grad_norm": 0.2420352155070497, "learning_rate": 1.2534324662980434e-07, "loss": 0.4591, "step": 5514 }, { "epoch": 13.484107579462103, "grad_norm": 0.40176290576989615, "learning_rate": 1.2494328360999592e-07, "loss": 0.3896, "step": 5515 }, { "epoch": 13.486552567237164, "grad_norm": 0.2998968031832042, "learning_rate": 1.2454394338442726e-07, "loss": 0.5822, "step": 5516 }, { "epoch": 13.488997555012224, "grad_norm": 0.3503253198656912, "learning_rate": 1.2414522605781355e-07, "loss": 0.6744, "step": 5517 }, { "epoch": 13.491442542787286, "grad_norm": 0.23791028559403393, "learning_rate": 1.237471317347083e-07, "loss": 0.3759, "step": 5518 }, { "epoch": 13.493887530562347, "grad_norm": 0.27728368245119683, "learning_rate": 1.2334966051950086e-07, "loss": 0.4714, "step": 5519 }, { "epoch": 13.496332518337407, "grad_norm": 0.24732886272790217, "learning_rate": 1.22952812516417e-07, "loss": 0.3861, "step": 5520 }, { "epoch": 13.49877750611247, "grad_norm": 0.20110125569955892, "learning_rate": 1.2255658782952013e-07, "loss": 0.6863, "step": 5521 }, { "epoch": 13.50122249388753, "grad_norm": 0.20074626203303475, "learning_rate": 1.2216098656270824e-07, "loss": 0.4338, "step": 5522 }, { "epoch": 13.503667481662593, "grad_norm": 0.169916250263693, "learning_rate": 1.2176600881971784e-07, "loss": 0.4296, "step": 5523 }, { "epoch": 13.506112469437653, "grad_norm": 0.24214297617312927, "learning_rate": 1.2137165470412134e-07, "loss": 0.4364, "step": 5524 }, { "epoch": 13.508557457212714, "grad_norm": 0.3218353993077113, "learning_rate": 1.2097792431932653e-07, "loss": 0.591, "step": 5525 }, { "epoch": 13.511002444987774, "grad_norm": 0.2882503381878655, "learning_rate": 1.205848177685795e-07, "loss": 0.3309, "step": 5526 }, { "epoch": 13.513447432762836, "grad_norm": 0.32826182968853174, "learning_rate": 1.201923351549608e-07, "loss": 0.6806, "step": 5527 }, { "epoch": 13.515892420537897, "grad_norm": 0.3016361991693213, "learning_rate": 1.1980047658138926e-07, "loss": 0.3373, "step": 5528 }, { "epoch": 13.51833740831296, "grad_norm": 0.23492412532991028, "learning_rate": 1.1940924215061878e-07, "loss": 0.9251, "step": 5529 }, { "epoch": 13.52078239608802, "grad_norm": 0.35464190228328957, "learning_rate": 1.1901863196523921e-07, "loss": 0.6157, "step": 5530 }, { "epoch": 13.52322738386308, "grad_norm": 0.27884521536947593, "learning_rate": 1.1862864612767921e-07, "loss": 0.4315, "step": 5531 }, { "epoch": 13.525672371638143, "grad_norm": 0.3335432288201711, "learning_rate": 1.1823928474020057e-07, "loss": 0.3951, "step": 5532 }, { "epoch": 13.528117359413203, "grad_norm": 0.20952527514343117, "learning_rate": 1.178505479049033e-07, "loss": 0.4901, "step": 5533 }, { "epoch": 13.530562347188264, "grad_norm": 0.20416209376344643, "learning_rate": 1.1746243572372335e-07, "loss": 0.6897, "step": 5534 }, { "epoch": 13.533007334963326, "grad_norm": 0.37810003170379586, "learning_rate": 1.1707494829843208e-07, "loss": 0.7884, "step": 5535 }, { "epoch": 13.535452322738386, "grad_norm": 0.2757127167778145, "learning_rate": 1.1668808573063795e-07, "loss": 0.6171, "step": 5536 }, { "epoch": 13.537897310513447, "grad_norm": 0.281349043444956, "learning_rate": 1.1630184812178453e-07, "loss": 0.5106, "step": 5537 }, { "epoch": 13.54034229828851, "grad_norm": 0.27364883590576317, "learning_rate": 1.1591623557315302e-07, "loss": 0.3692, "step": 5538 }, { "epoch": 13.54278728606357, "grad_norm": 0.21645327035397532, "learning_rate": 1.1553124818585948e-07, "loss": 0.7495, "step": 5539 }, { "epoch": 13.54523227383863, "grad_norm": 0.22542112697996644, "learning_rate": 1.1514688606085594e-07, "loss": 0.43, "step": 5540 }, { "epoch": 13.547677261613693, "grad_norm": 0.2595556855489822, "learning_rate": 1.1476314929893207e-07, "loss": 0.3733, "step": 5541 }, { "epoch": 13.550122249388753, "grad_norm": 0.20784251358452624, "learning_rate": 1.1438003800071123e-07, "loss": 0.7624, "step": 5542 }, { "epoch": 13.552567237163814, "grad_norm": 0.3736623667846649, "learning_rate": 1.1399755226665476e-07, "loss": 0.4928, "step": 5543 }, { "epoch": 13.555012224938876, "grad_norm": 0.22931721144500014, "learning_rate": 1.1361569219705909e-07, "loss": 0.4382, "step": 5544 }, { "epoch": 13.557457212713937, "grad_norm": 0.27195582440396815, "learning_rate": 1.1323445789205633e-07, "loss": 0.5064, "step": 5545 }, { "epoch": 13.559902200488997, "grad_norm": 0.15860479836460462, "learning_rate": 1.1285384945161542e-07, "loss": 0.4119, "step": 5546 }, { "epoch": 13.56234718826406, "grad_norm": 0.3413168392662795, "learning_rate": 1.1247386697553986e-07, "loss": 0.3358, "step": 5547 }, { "epoch": 13.56479217603912, "grad_norm": 0.2540496734017776, "learning_rate": 1.120945105634702e-07, "loss": 0.8551, "step": 5548 }, { "epoch": 13.56723716381418, "grad_norm": 0.2834539217337119, "learning_rate": 1.1171578031488245e-07, "loss": 0.6504, "step": 5549 }, { "epoch": 13.569682151589243, "grad_norm": 0.16213642286649776, "learning_rate": 1.1133767632908799e-07, "loss": 0.3996, "step": 5550 }, { "epoch": 13.572127139364303, "grad_norm": 0.2401672763681036, "learning_rate": 1.1096019870523444e-07, "loss": 0.7987, "step": 5551 }, { "epoch": 13.574572127139364, "grad_norm": 0.22753802013516647, "learning_rate": 1.1058334754230515e-07, "loss": 0.3622, "step": 5552 }, { "epoch": 13.577017114914426, "grad_norm": 0.19888001714476328, "learning_rate": 1.1020712293911912e-07, "loss": 0.5267, "step": 5553 }, { "epoch": 13.579462102689487, "grad_norm": 0.4040812928179129, "learning_rate": 1.0983152499433159e-07, "loss": 0.6383, "step": 5554 }, { "epoch": 13.581907090464547, "grad_norm": 0.4122965750207208, "learning_rate": 1.0945655380643183e-07, "loss": 0.7265, "step": 5555 }, { "epoch": 13.58435207823961, "grad_norm": 0.24515409049586184, "learning_rate": 1.0908220947374704e-07, "loss": 0.6433, "step": 5556 }, { "epoch": 13.58679706601467, "grad_norm": 0.3209985402865371, "learning_rate": 1.0870849209443784e-07, "loss": 0.424, "step": 5557 }, { "epoch": 13.58924205378973, "grad_norm": 0.3491010832431343, "learning_rate": 1.0833540176650197e-07, "loss": 0.386, "step": 5558 }, { "epoch": 13.591687041564793, "grad_norm": 0.26611475726174433, "learning_rate": 1.0796293858777285e-07, "loss": 0.8779, "step": 5559 }, { "epoch": 13.594132029339853, "grad_norm": 0.26785842469949483, "learning_rate": 1.075911026559176e-07, "loss": 0.7852, "step": 5560 }, { "epoch": 13.596577017114914, "grad_norm": 0.3196723739359268, "learning_rate": 1.0721989406844074e-07, "loss": 0.7188, "step": 5561 }, { "epoch": 13.599022004889976, "grad_norm": 0.3377698865709029, "learning_rate": 1.0684931292268163e-07, "loss": 0.3626, "step": 5562 }, { "epoch": 13.601466992665037, "grad_norm": 0.37309932370865584, "learning_rate": 1.0647935931581532e-07, "loss": 1.1618, "step": 5563 }, { "epoch": 13.603911980440097, "grad_norm": 0.3178922148360617, "learning_rate": 1.0611003334485226e-07, "loss": 0.7084, "step": 5564 }, { "epoch": 13.60635696821516, "grad_norm": 0.3275162778672461, "learning_rate": 1.0574133510663748e-07, "loss": 0.7152, "step": 5565 }, { "epoch": 13.60880195599022, "grad_norm": 0.3201526475463653, "learning_rate": 1.0537326469785281e-07, "loss": 0.5989, "step": 5566 }, { "epoch": 13.61124694376528, "grad_norm": 0.21027932118077458, "learning_rate": 1.050058222150141e-07, "loss": 0.7291, "step": 5567 }, { "epoch": 13.613691931540343, "grad_norm": 0.29295512146782515, "learning_rate": 1.0463900775447345e-07, "loss": 0.6523, "step": 5568 }, { "epoch": 13.616136919315403, "grad_norm": 0.22941105746364399, "learning_rate": 1.0427282141241862e-07, "loss": 0.7076, "step": 5569 }, { "epoch": 13.618581907090464, "grad_norm": 0.2277477790862266, "learning_rate": 1.0390726328487088e-07, "loss": 0.9967, "step": 5570 }, { "epoch": 13.621026894865526, "grad_norm": 0.3537175954855979, "learning_rate": 1.0354233346768855e-07, "loss": 0.6843, "step": 5571 }, { "epoch": 13.623471882640587, "grad_norm": 0.214071385457428, "learning_rate": 1.0317803205656479e-07, "loss": 0.6996, "step": 5572 }, { "epoch": 13.625916870415647, "grad_norm": 0.24350129312577748, "learning_rate": 1.0281435914702736e-07, "loss": 0.7302, "step": 5573 }, { "epoch": 13.62836185819071, "grad_norm": 0.35748596914359504, "learning_rate": 1.0245131483444026e-07, "loss": 0.7052, "step": 5574 }, { "epoch": 13.63080684596577, "grad_norm": 0.2464543130820322, "learning_rate": 1.0208889921400122e-07, "loss": 0.3268, "step": 5575 }, { "epoch": 13.63325183374083, "grad_norm": 0.2530724922034853, "learning_rate": 1.017271123807445e-07, "loss": 0.3802, "step": 5576 }, { "epoch": 13.635696821515893, "grad_norm": 0.2547182461894175, "learning_rate": 1.0136595442953839e-07, "loss": 0.3817, "step": 5577 }, { "epoch": 13.638141809290953, "grad_norm": 0.2347048562609496, "learning_rate": 1.0100542545508713e-07, "loss": 0.6289, "step": 5578 }, { "epoch": 13.640586797066014, "grad_norm": 0.40699033551254127, "learning_rate": 1.006455255519298e-07, "loss": 0.4444, "step": 5579 }, { "epoch": 13.643031784841076, "grad_norm": 0.253408994334587, "learning_rate": 1.0028625481443982e-07, "loss": 0.372, "step": 5580 }, { "epoch": 13.645476772616137, "grad_norm": 0.2700962927274832, "learning_rate": 9.992761333682683e-08, "loss": 0.6272, "step": 5581 }, { "epoch": 13.647921760391197, "grad_norm": 0.2586830769537017, "learning_rate": 9.956960121313447e-08, "loss": 0.5185, "step": 5582 }, { "epoch": 13.65036674816626, "grad_norm": 0.3032139142960979, "learning_rate": 9.921221853724155e-08, "loss": 0.6208, "step": 5583 }, { "epoch": 13.65281173594132, "grad_norm": 0.2910797918523374, "learning_rate": 9.885546540286255e-08, "loss": 0.4145, "step": 5584 }, { "epoch": 13.65525672371638, "grad_norm": 0.3131412803201063, "learning_rate": 9.849934190354599e-08, "loss": 0.4628, "step": 5585 }, { "epoch": 13.657701711491443, "grad_norm": 0.18618772916662946, "learning_rate": 9.814384813267575e-08, "loss": 0.7337, "step": 5586 }, { "epoch": 13.660146699266503, "grad_norm": 0.3064789270112802, "learning_rate": 9.778898418347005e-08, "loss": 0.8502, "step": 5587 }, { "epoch": 13.662591687041564, "grad_norm": 0.2457187182585951, "learning_rate": 9.743475014898251e-08, "loss": 0.5088, "step": 5588 }, { "epoch": 13.665036674816626, "grad_norm": 0.2747033736230123, "learning_rate": 9.70811461221019e-08, "loss": 1.0551, "step": 5589 }, { "epoch": 13.667481662591687, "grad_norm": 0.18071413014314702, "learning_rate": 9.672817219555041e-08, "loss": 0.3956, "step": 5590 }, { "epoch": 13.669926650366747, "grad_norm": 0.2002353052001889, "learning_rate": 9.637582846188654e-08, "loss": 0.41, "step": 5591 }, { "epoch": 13.67237163814181, "grad_norm": 0.2655963786380672, "learning_rate": 9.602411501350273e-08, "loss": 0.5863, "step": 5592 }, { "epoch": 13.67481662591687, "grad_norm": 0.26722078536378874, "learning_rate": 9.567303194262634e-08, "loss": 0.5135, "step": 5593 }, { "epoch": 13.67726161369193, "grad_norm": 0.44084598116593093, "learning_rate": 9.5322579341319e-08, "loss": 0.5545, "step": 5594 }, { "epoch": 13.679706601466993, "grad_norm": 0.22793067029525646, "learning_rate": 9.497275730147776e-08, "loss": 0.4857, "step": 5595 }, { "epoch": 13.682151589242054, "grad_norm": 0.3366078425956773, "learning_rate": 9.462356591483424e-08, "loss": 0.692, "step": 5596 }, { "epoch": 13.684596577017114, "grad_norm": 0.3016154508879463, "learning_rate": 9.427500527295381e-08, "loss": 0.4885, "step": 5597 }, { "epoch": 13.687041564792176, "grad_norm": 0.2949481875266221, "learning_rate": 9.392707546723751e-08, "loss": 0.5066, "step": 5598 }, { "epoch": 13.689486552567237, "grad_norm": 0.2743711463609389, "learning_rate": 9.357977658892043e-08, "loss": 0.7031, "step": 5599 }, { "epoch": 13.6919315403423, "grad_norm": 0.3975870256365274, "learning_rate": 9.323310872907165e-08, "loss": 0.6523, "step": 5600 }, { "epoch": 13.69437652811736, "grad_norm": 0.3006610452872104, "learning_rate": 9.288707197859626e-08, "loss": 0.6622, "step": 5601 }, { "epoch": 13.69682151589242, "grad_norm": 0.28910606408622425, "learning_rate": 9.254166642823303e-08, "loss": 0.3731, "step": 5602 }, { "epoch": 13.69926650366748, "grad_norm": 0.24739031345940038, "learning_rate": 9.219689216855427e-08, "loss": 0.5051, "step": 5603 }, { "epoch": 13.701711491442543, "grad_norm": 0.24729099273927424, "learning_rate": 9.185274928996901e-08, "loss": 1.0197, "step": 5604 }, { "epoch": 13.704156479217604, "grad_norm": 0.2586187907145074, "learning_rate": 9.150923788271788e-08, "loss": 0.599, "step": 5605 }, { "epoch": 13.706601466992666, "grad_norm": 0.26733276592245103, "learning_rate": 9.11663580368788e-08, "loss": 0.6108, "step": 5606 }, { "epoch": 13.709046454767726, "grad_norm": 0.16962873030933348, "learning_rate": 9.082410984236179e-08, "loss": 0.4192, "step": 5607 }, { "epoch": 13.711491442542787, "grad_norm": 0.19109742641524302, "learning_rate": 9.048249338891257e-08, "loss": 0.5796, "step": 5608 }, { "epoch": 13.713936430317847, "grad_norm": 0.2575927257921119, "learning_rate": 9.014150876611089e-08, "loss": 0.6919, "step": 5609 }, { "epoch": 13.71638141809291, "grad_norm": 0.21605336601919534, "learning_rate": 8.980115606337047e-08, "loss": 0.5767, "step": 5610 }, { "epoch": 13.71882640586797, "grad_norm": 0.27651660457344274, "learning_rate": 8.94614353699394e-08, "loss": 0.651, "step": 5611 }, { "epoch": 13.721271393643033, "grad_norm": 0.37645635866485083, "learning_rate": 8.912234677490111e-08, "loss": 0.4514, "step": 5612 }, { "epoch": 13.723716381418093, "grad_norm": 0.2070683807925209, "learning_rate": 8.878389036717144e-08, "loss": 0.552, "step": 5613 }, { "epoch": 13.726161369193154, "grad_norm": 0.3094229333854138, "learning_rate": 8.844606623550189e-08, "loss": 0.4939, "step": 5614 }, { "epoch": 13.728606356968216, "grad_norm": 0.17967611499043526, "learning_rate": 8.810887446847688e-08, "loss": 0.4057, "step": 5615 }, { "epoch": 13.731051344743276, "grad_norm": 0.26738429494402016, "learning_rate": 8.777231515451707e-08, "loss": 0.6242, "step": 5616 }, { "epoch": 13.733496332518337, "grad_norm": 0.232749654494077, "learning_rate": 8.743638838187545e-08, "loss": 0.6963, "step": 5617 }, { "epoch": 13.7359413202934, "grad_norm": 0.23981390645506026, "learning_rate": 8.710109423863966e-08, "loss": 0.4963, "step": 5618 }, { "epoch": 13.73838630806846, "grad_norm": 0.1939563806979734, "learning_rate": 8.676643281273156e-08, "loss": 0.4204, "step": 5619 }, { "epoch": 13.74083129584352, "grad_norm": 0.27966211203387115, "learning_rate": 8.643240419190679e-08, "loss": 0.6107, "step": 5620 }, { "epoch": 13.743276283618583, "grad_norm": 0.2558240384902012, "learning_rate": 8.60990084637553e-08, "loss": 0.9455, "step": 5621 }, { "epoch": 13.745721271393643, "grad_norm": 0.2561290626797861, "learning_rate": 8.576624571570186e-08, "loss": 0.352, "step": 5622 }, { "epoch": 13.748166259168704, "grad_norm": 0.29951124682271113, "learning_rate": 8.543411603500362e-08, "loss": 0.4775, "step": 5623 }, { "epoch": 13.750611246943766, "grad_norm": 0.2008154712772762, "learning_rate": 8.510261950875287e-08, "loss": 0.706, "step": 5624 }, { "epoch": 13.753056234718827, "grad_norm": 0.18568914027581912, "learning_rate": 8.477175622387563e-08, "loss": 0.5576, "step": 5625 }, { "epoch": 13.755501222493887, "grad_norm": 0.26797044014583915, "learning_rate": 8.444152626713137e-08, "loss": 0.4845, "step": 5626 }, { "epoch": 13.75794621026895, "grad_norm": 0.1876005338849139, "learning_rate": 8.411192972511473e-08, "loss": 0.4242, "step": 5627 }, { "epoch": 13.76039119804401, "grad_norm": 0.2607193734334943, "learning_rate": 8.378296668425295e-08, "loss": 0.4002, "step": 5628 }, { "epoch": 13.76283618581907, "grad_norm": 0.3065621075125215, "learning_rate": 8.345463723080816e-08, "loss": 0.4738, "step": 5629 }, { "epoch": 13.765281173594133, "grad_norm": 0.2567670482454851, "learning_rate": 8.312694145087536e-08, "loss": 0.394, "step": 5630 }, { "epoch": 13.767726161369193, "grad_norm": 0.3002023303836253, "learning_rate": 8.279987943038387e-08, "loss": 1.0003, "step": 5631 }, { "epoch": 13.770171149144254, "grad_norm": 0.29251310064018576, "learning_rate": 8.247345125509731e-08, "loss": 0.4632, "step": 5632 }, { "epoch": 13.772616136919316, "grad_norm": 0.38975783466576047, "learning_rate": 8.214765701061195e-08, "loss": 0.493, "step": 5633 }, { "epoch": 13.775061124694377, "grad_norm": 0.3813235569340884, "learning_rate": 8.182249678235915e-08, "loss": 0.7381, "step": 5634 }, { "epoch": 13.777506112469437, "grad_norm": 0.19287406815628588, "learning_rate": 8.149797065560295e-08, "loss": 0.3958, "step": 5635 }, { "epoch": 13.7799511002445, "grad_norm": 0.3013083609596784, "learning_rate": 8.117407871544164e-08, "loss": 0.3714, "step": 5636 }, { "epoch": 13.78239608801956, "grad_norm": 0.1922372539946887, "learning_rate": 8.085082104680703e-08, "loss": 0.6339, "step": 5637 }, { "epoch": 13.78484107579462, "grad_norm": 0.2560376188748304, "learning_rate": 8.052819773446491e-08, "loss": 0.3944, "step": 5638 }, { "epoch": 13.787286063569683, "grad_norm": 0.2649681491755055, "learning_rate": 8.020620886301455e-08, "loss": 1.0834, "step": 5639 }, { "epoch": 13.789731051344743, "grad_norm": 0.23791159543852178, "learning_rate": 7.988485451688816e-08, "loss": 0.6581, "step": 5640 }, { "epoch": 13.792176039119804, "grad_norm": 0.2576480953647651, "learning_rate": 7.956413478035302e-08, "loss": 0.3729, "step": 5641 }, { "epoch": 13.794621026894866, "grad_norm": 0.19345502612854307, "learning_rate": 7.924404973750882e-08, "loss": 0.4756, "step": 5642 }, { "epoch": 13.797066014669927, "grad_norm": 0.33689254297536514, "learning_rate": 7.892459947228898e-08, "loss": 0.4492, "step": 5643 }, { "epoch": 13.799511002444987, "grad_norm": 0.2661417476710071, "learning_rate": 7.860578406846119e-08, "loss": 0.402, "step": 5644 }, { "epoch": 13.80195599022005, "grad_norm": 0.21273356311174985, "learning_rate": 7.828760360962557e-08, "loss": 0.4415, "step": 5645 }, { "epoch": 13.80440097799511, "grad_norm": 0.16459861997507153, "learning_rate": 7.797005817921671e-08, "loss": 0.5038, "step": 5646 }, { "epoch": 13.80684596577017, "grad_norm": 0.28082257240417136, "learning_rate": 7.765314786050193e-08, "loss": 0.5834, "step": 5647 }, { "epoch": 13.809290953545233, "grad_norm": 0.20134198057690336, "learning_rate": 7.73368727365828e-08, "loss": 0.3513, "step": 5648 }, { "epoch": 13.811735941320293, "grad_norm": 0.22177385520647788, "learning_rate": 7.702123289039381e-08, "loss": 0.4342, "step": 5649 }, { "epoch": 13.814180929095354, "grad_norm": 0.258314550031699, "learning_rate": 7.670622840470266e-08, "loss": 0.9537, "step": 5650 }, { "epoch": 13.816625916870416, "grad_norm": 0.275645109158862, "learning_rate": 7.639185936211102e-08, "loss": 0.7908, "step": 5651 }, { "epoch": 13.819070904645477, "grad_norm": 0.3100151195823987, "learning_rate": 7.607812584505381e-08, "loss": 0.3885, "step": 5652 }, { "epoch": 13.821515892420537, "grad_norm": 0.2501411160924779, "learning_rate": 7.576502793579854e-08, "loss": 0.3672, "step": 5653 }, { "epoch": 13.8239608801956, "grad_norm": 0.26189423340120893, "learning_rate": 7.545256571644732e-08, "loss": 0.4584, "step": 5654 }, { "epoch": 13.82640586797066, "grad_norm": 0.24923541158079357, "learning_rate": 7.514073926893434e-08, "loss": 0.7607, "step": 5655 }, { "epoch": 13.82885085574572, "grad_norm": 0.22822968553625902, "learning_rate": 7.482954867502778e-08, "loss": 0.3692, "step": 5656 }, { "epoch": 13.831295843520783, "grad_norm": 0.31626538773344065, "learning_rate": 7.451899401632934e-08, "loss": 0.4656, "step": 5657 }, { "epoch": 13.833740831295843, "grad_norm": 0.37669617118209225, "learning_rate": 7.420907537427274e-08, "loss": 0.5186, "step": 5658 }, { "epoch": 13.836185819070904, "grad_norm": 0.2725636513235219, "learning_rate": 7.38997928301266e-08, "loss": 0.71, "step": 5659 }, { "epoch": 13.838630806845966, "grad_norm": 0.19050397559028576, "learning_rate": 7.359114646499132e-08, "loss": 0.4655, "step": 5660 }, { "epoch": 13.841075794621027, "grad_norm": 0.29809880828616875, "learning_rate": 7.328313635980133e-08, "loss": 0.3895, "step": 5661 }, { "epoch": 13.843520782396087, "grad_norm": 0.26127137286759944, "learning_rate": 7.297576259532424e-08, "loss": 0.5182, "step": 5662 }, { "epoch": 13.84596577017115, "grad_norm": 0.2221476818756455, "learning_rate": 7.266902525215975e-08, "loss": 0.3993, "step": 5663 }, { "epoch": 13.84841075794621, "grad_norm": 0.24812101016135327, "learning_rate": 7.236292441074183e-08, "loss": 1.3562, "step": 5664 }, { "epoch": 13.85085574572127, "grad_norm": 0.23690908231347022, "learning_rate": 7.205746015133714e-08, "loss": 0.5625, "step": 5665 }, { "epoch": 13.853300733496333, "grad_norm": 0.2254165759273135, "learning_rate": 7.175263255404519e-08, "loss": 0.4106, "step": 5666 }, { "epoch": 13.855745721271393, "grad_norm": 0.27894230635927325, "learning_rate": 7.1448441698799e-08, "loss": 0.4381, "step": 5667 }, { "epoch": 13.858190709046454, "grad_norm": 0.2970428587120057, "learning_rate": 7.11448876653642e-08, "loss": 0.4672, "step": 5668 }, { "epoch": 13.860635696821516, "grad_norm": 0.2188016856271119, "learning_rate": 7.084197053333935e-08, "loss": 0.4248, "step": 5669 }, { "epoch": 13.863080684596577, "grad_norm": 0.22295590564714088, "learning_rate": 7.053969038215675e-08, "loss": 0.3437, "step": 5670 }, { "epoch": 13.865525672371637, "grad_norm": 0.2658431040469962, "learning_rate": 7.02380472910813e-08, "loss": 0.7699, "step": 5671 }, { "epoch": 13.8679706601467, "grad_norm": 0.24060295044355764, "learning_rate": 6.993704133921031e-08, "loss": 0.456, "step": 5672 }, { "epoch": 13.87041564792176, "grad_norm": 0.2504773059857715, "learning_rate": 6.963667260547424e-08, "loss": 0.3673, "step": 5673 }, { "epoch": 13.87286063569682, "grad_norm": 0.2320020000388462, "learning_rate": 6.933694116863732e-08, "loss": 0.768, "step": 5674 }, { "epoch": 13.875305623471883, "grad_norm": 0.35343926487640603, "learning_rate": 6.903784710729556e-08, "loss": 0.9156, "step": 5675 }, { "epoch": 13.877750611246944, "grad_norm": 0.14781233743998162, "learning_rate": 6.873939049987815e-08, "loss": 0.4237, "step": 5676 }, { "epoch": 13.880195599022004, "grad_norm": 0.2405914601348876, "learning_rate": 6.84415714246478e-08, "loss": 1.1149, "step": 5677 }, { "epoch": 13.882640586797066, "grad_norm": 0.19366228739159255, "learning_rate": 6.814438995969869e-08, "loss": 0.7345, "step": 5678 }, { "epoch": 13.885085574572127, "grad_norm": 0.3126684624303472, "learning_rate": 6.784784618295931e-08, "loss": 0.3926, "step": 5679 }, { "epoch": 13.887530562347187, "grad_norm": 0.19913135887303235, "learning_rate": 6.755194017218969e-08, "loss": 0.6636, "step": 5680 }, { "epoch": 13.88997555012225, "grad_norm": 0.2764871408509584, "learning_rate": 6.725667200498359e-08, "loss": 0.424, "step": 5681 }, { "epoch": 13.89242053789731, "grad_norm": 0.27110874004422386, "learning_rate": 6.696204175876686e-08, "loss": 0.8935, "step": 5682 }, { "epoch": 13.894865525672373, "grad_norm": 0.36660469013234076, "learning_rate": 6.666804951079825e-08, "loss": 0.442, "step": 5683 }, { "epoch": 13.897310513447433, "grad_norm": 0.23605112407126835, "learning_rate": 6.637469533816942e-08, "loss": 0.5302, "step": 5684 }, { "epoch": 13.899755501222494, "grad_norm": 0.19421516023246746, "learning_rate": 6.608197931780496e-08, "loss": 0.3436, "step": 5685 }, { "epoch": 13.902200488997554, "grad_norm": 0.2557163795818338, "learning_rate": 6.57899015264607e-08, "loss": 0.4836, "step": 5686 }, { "epoch": 13.904645476772616, "grad_norm": 0.24336404126836905, "learning_rate": 6.549846204072674e-08, "loss": 0.8938, "step": 5687 }, { "epoch": 13.907090464547677, "grad_norm": 0.21311180509306366, "learning_rate": 6.52076609370253e-08, "loss": 0.7128, "step": 5688 }, { "epoch": 13.90953545232274, "grad_norm": 0.3119674325525338, "learning_rate": 6.491749829161064e-08, "loss": 0.6151, "step": 5689 }, { "epoch": 13.9119804400978, "grad_norm": 0.22805670842899523, "learning_rate": 6.462797418057026e-08, "loss": 0.555, "step": 5690 }, { "epoch": 13.91442542787286, "grad_norm": 0.19712025178994977, "learning_rate": 6.433908867982396e-08, "loss": 0.7656, "step": 5691 }, { "epoch": 13.91687041564792, "grad_norm": 0.2783258649754196, "learning_rate": 6.40508418651245e-08, "loss": 0.4223, "step": 5692 }, { "epoch": 13.919315403422983, "grad_norm": 0.30330294826347237, "learning_rate": 6.37632338120564e-08, "loss": 0.4482, "step": 5693 }, { "epoch": 13.921760391198044, "grad_norm": 0.23712160942872165, "learning_rate": 6.347626459603712e-08, "loss": 0.4084, "step": 5694 }, { "epoch": 13.924205378973106, "grad_norm": 0.30705814404516824, "learning_rate": 6.318993429231701e-08, "loss": 0.3655, "step": 5695 }, { "epoch": 13.926650366748166, "grad_norm": 0.3847960084801866, "learning_rate": 6.29042429759777e-08, "loss": 0.5673, "step": 5696 }, { "epoch": 13.929095354523227, "grad_norm": 0.3173446863891032, "learning_rate": 6.26191907219345e-08, "loss": 0.4031, "step": 5697 }, { "epoch": 13.93154034229829, "grad_norm": 0.28772152516507377, "learning_rate": 6.233477760493434e-08, "loss": 0.325, "step": 5698 }, { "epoch": 13.93398533007335, "grad_norm": 0.4109835014227078, "learning_rate": 6.205100369955697e-08, "loss": 0.3943, "step": 5699 }, { "epoch": 13.93643031784841, "grad_norm": 0.1813254542501591, "learning_rate": 6.176786908021454e-08, "loss": 0.7648, "step": 5700 }, { "epoch": 13.938875305623473, "grad_norm": 0.19124120902598046, "learning_rate": 6.148537382115072e-08, "loss": 0.6457, "step": 5701 }, { "epoch": 13.941320293398533, "grad_norm": 0.289905686606393, "learning_rate": 6.120351799644348e-08, "loss": 0.7174, "step": 5702 }, { "epoch": 13.943765281173594, "grad_norm": 0.1961814892727553, "learning_rate": 6.092230168000091e-08, "loss": 0.4523, "step": 5703 }, { "epoch": 13.946210268948656, "grad_norm": 0.1814368265122953, "learning_rate": 6.064172494556486e-08, "loss": 0.3739, "step": 5704 }, { "epoch": 13.948655256723717, "grad_norm": 0.25653401971813755, "learning_rate": 6.03617878667087e-08, "loss": 1.2176, "step": 5705 }, { "epoch": 13.951100244498777, "grad_norm": 0.26187902538561575, "learning_rate": 6.008249051683846e-08, "loss": 0.7135, "step": 5706 }, { "epoch": 13.95354523227384, "grad_norm": 0.2238991789221987, "learning_rate": 5.980383296919246e-08, "loss": 0.4338, "step": 5707 }, { "epoch": 13.9559902200489, "grad_norm": 0.2924162212545847, "learning_rate": 5.952581529684032e-08, "loss": 0.4481, "step": 5708 }, { "epoch": 13.95843520782396, "grad_norm": 0.25905106056044247, "learning_rate": 5.9248437572685394e-08, "loss": 0.572, "step": 5709 }, { "epoch": 13.960880195599023, "grad_norm": 0.3037916468831237, "learning_rate": 5.897169986946255e-08, "loss": 0.6279, "step": 5710 }, { "epoch": 13.963325183374083, "grad_norm": 0.3758677418363692, "learning_rate": 5.869560225973847e-08, "loss": 0.6665, "step": 5711 }, { "epoch": 13.965770171149144, "grad_norm": 0.3517825453987885, "learning_rate": 5.842014481591191e-08, "loss": 0.3757, "step": 5712 }, { "epoch": 13.968215158924206, "grad_norm": 0.24231342837645276, "learning_rate": 5.8145327610214517e-08, "loss": 0.4617, "step": 5713 }, { "epoch": 13.970660146699267, "grad_norm": 0.42769162915816133, "learning_rate": 5.7871150714709775e-08, "loss": 0.6231, "step": 5714 }, { "epoch": 13.973105134474327, "grad_norm": 0.38546248623982654, "learning_rate": 5.759761420129323e-08, "loss": 0.3858, "step": 5715 }, { "epoch": 13.97555012224939, "grad_norm": 0.22900211024519998, "learning_rate": 5.732471814169194e-08, "loss": 0.4308, "step": 5716 }, { "epoch": 13.97799511002445, "grad_norm": 0.22073803768835076, "learning_rate": 5.705246260746561e-08, "loss": 0.8944, "step": 5717 }, { "epoch": 13.98044009779951, "grad_norm": 0.24392192024486242, "learning_rate": 5.6780847670006024e-08, "loss": 0.4954, "step": 5718 }, { "epoch": 13.982885085574573, "grad_norm": 0.2282555014573532, "learning_rate": 5.650987340053704e-08, "loss": 0.805, "step": 5719 }, { "epoch": 13.985330073349633, "grad_norm": 0.3478591864448769, "learning_rate": 5.623953987011405e-08, "loss": 0.5645, "step": 5720 }, { "epoch": 13.987775061124694, "grad_norm": 0.19278412553455818, "learning_rate": 5.5969847149624504e-08, "loss": 0.4398, "step": 5721 }, { "epoch": 13.990220048899756, "grad_norm": 0.18703972728150048, "learning_rate": 5.5700795309788504e-08, "loss": 0.5281, "step": 5722 }, { "epoch": 13.992665036674817, "grad_norm": 0.30102840640489964, "learning_rate": 5.543238442115684e-08, "loss": 1.2257, "step": 5723 }, { "epoch": 13.995110024449877, "grad_norm": 0.23766909768482902, "learning_rate": 5.516461455411348e-08, "loss": 0.6255, "step": 5724 }, { "epoch": 13.99755501222494, "grad_norm": 0.26714910233036704, "learning_rate": 5.489748577887449e-08, "loss": 0.8751, "step": 5725 }, { "epoch": 14.0, "grad_norm": 0.3298800115212367, "learning_rate": 5.463099816548578e-08, "loss": 0.3495, "step": 5726 }, { "epoch": 14.00244498777506, "grad_norm": 0.22192010782224986, "learning_rate": 5.436515178382784e-08, "loss": 0.6872, "step": 5727 }, { "epoch": 14.004889975550123, "grad_norm": 0.2879347368121287, "learning_rate": 5.409994670361046e-08, "loss": 0.551, "step": 5728 }, { "epoch": 14.007334963325183, "grad_norm": 0.2504483246657384, "learning_rate": 5.383538299437718e-08, "loss": 0.5138, "step": 5729 }, { "epoch": 14.009779951100244, "grad_norm": 0.35599323991963117, "learning_rate": 5.357146072550279e-08, "loss": 0.3974, "step": 5730 }, { "epoch": 14.012224938875306, "grad_norm": 0.23943331782735083, "learning_rate": 5.330817996619331e-08, "loss": 0.5203, "step": 5731 }, { "epoch": 14.014669926650367, "grad_norm": 0.22062892944837517, "learning_rate": 5.3045540785487126e-08, "loss": 0.8056, "step": 5732 }, { "epoch": 14.017114914425427, "grad_norm": 0.2828544134546111, "learning_rate": 5.2783543252254155e-08, "loss": 0.7433, "step": 5733 }, { "epoch": 14.01955990220049, "grad_norm": 0.21081910692921516, "learning_rate": 5.252218743519638e-08, "loss": 0.5019, "step": 5734 }, { "epoch": 14.02200488997555, "grad_norm": 0.2669155030292534, "learning_rate": 5.226147340284732e-08, "loss": 0.8479, "step": 5735 }, { "epoch": 14.02444987775061, "grad_norm": 0.2790013268955985, "learning_rate": 5.2001401223572e-08, "loss": 0.6977, "step": 5736 }, { "epoch": 14.026894865525673, "grad_norm": 0.32816481984915846, "learning_rate": 5.1741970965567265e-08, "loss": 0.5888, "step": 5737 }, { "epoch": 14.029339853300733, "grad_norm": 0.28714087407627525, "learning_rate": 5.148318269686176e-08, "loss": 0.5297, "step": 5738 }, { "epoch": 14.031784841075794, "grad_norm": 0.18498835416486617, "learning_rate": 5.122503648531535e-08, "loss": 0.3846, "step": 5739 }, { "epoch": 14.034229828850856, "grad_norm": 0.3183783486535158, "learning_rate": 5.096753239862057e-08, "loss": 0.955, "step": 5740 }, { "epoch": 14.036674816625917, "grad_norm": 0.34311010252291824, "learning_rate": 5.071067050430034e-08, "loss": 0.6715, "step": 5741 }, { "epoch": 14.039119804400977, "grad_norm": 0.24539892462208537, "learning_rate": 5.045445086970996e-08, "loss": 0.9961, "step": 5742 }, { "epoch": 14.04156479217604, "grad_norm": 0.2516239191165433, "learning_rate": 5.0198873562035955e-08, "loss": 0.4897, "step": 5743 }, { "epoch": 14.0440097799511, "grad_norm": 0.3866157286017683, "learning_rate": 4.9943938648296387e-08, "loss": 0.8461, "step": 5744 }, { "epoch": 14.04645476772616, "grad_norm": 0.2361703214114367, "learning_rate": 4.968964619534139e-08, "loss": 0.4465, "step": 5745 }, { "epoch": 14.048899755501223, "grad_norm": 0.238516751797963, "learning_rate": 4.943599626985207e-08, "loss": 0.4396, "step": 5746 }, { "epoch": 14.051344743276283, "grad_norm": 0.307040114652679, "learning_rate": 4.918298893834161e-08, "loss": 0.4036, "step": 5747 }, { "epoch": 14.053789731051344, "grad_norm": 0.176317695365481, "learning_rate": 4.893062426715361e-08, "loss": 0.573, "step": 5748 }, { "epoch": 14.056234718826406, "grad_norm": 0.25214111050787685, "learning_rate": 4.867890232246431e-08, "loss": 0.3647, "step": 5749 }, { "epoch": 14.058679706601467, "grad_norm": 0.22568228292484724, "learning_rate": 4.8427823170281166e-08, "loss": 0.6943, "step": 5750 }, { "epoch": 14.061124694376527, "grad_norm": 0.25393174608989827, "learning_rate": 4.8177386876442354e-08, "loss": 0.837, "step": 5751 }, { "epoch": 14.06356968215159, "grad_norm": 0.16866394044734753, "learning_rate": 4.792759350661841e-08, "loss": 0.5495, "step": 5752 }, { "epoch": 14.06601466992665, "grad_norm": 0.31340198118609325, "learning_rate": 4.7678443126310814e-08, "loss": 0.4447, "step": 5753 }, { "epoch": 14.06845965770171, "grad_norm": 0.218632707539735, "learning_rate": 4.742993580085231e-08, "loss": 0.696, "step": 5754 }, { "epoch": 14.070904645476773, "grad_norm": 0.2582786896053266, "learning_rate": 4.7182071595407164e-08, "loss": 0.6007, "step": 5755 }, { "epoch": 14.073349633251834, "grad_norm": 0.22280107086787024, "learning_rate": 4.693485057497144e-08, "loss": 0.7417, "step": 5756 }, { "epoch": 14.075794621026894, "grad_norm": 0.2979382944684062, "learning_rate": 4.668827280437188e-08, "loss": 0.7841, "step": 5757 }, { "epoch": 14.078239608801956, "grad_norm": 0.2839969661401148, "learning_rate": 4.644233834826678e-08, "loss": 0.5157, "step": 5758 }, { "epoch": 14.080684596577017, "grad_norm": 0.19479949942606686, "learning_rate": 4.619704727114593e-08, "loss": 0.4837, "step": 5759 }, { "epoch": 14.083129584352077, "grad_norm": 0.20972120580148954, "learning_rate": 4.5952399637330114e-08, "loss": 0.7799, "step": 5760 }, { "epoch": 14.08557457212714, "grad_norm": 0.26987748208919193, "learning_rate": 4.570839551097162e-08, "loss": 0.5006, "step": 5761 }, { "epoch": 14.0880195599022, "grad_norm": 0.27305830778720164, "learning_rate": 4.546503495605398e-08, "loss": 0.8407, "step": 5762 }, { "epoch": 14.09046454767726, "grad_norm": 0.35038717884448706, "learning_rate": 4.5222318036391985e-08, "loss": 0.4351, "step": 5763 }, { "epoch": 14.092909535452323, "grad_norm": 0.32793521890272304, "learning_rate": 4.498024481563112e-08, "loss": 0.3932, "step": 5764 }, { "epoch": 14.095354523227384, "grad_norm": 0.222753086369839, "learning_rate": 4.4738815357248664e-08, "loss": 0.544, "step": 5765 }, { "epoch": 14.097799511002446, "grad_norm": 0.4282622543133458, "learning_rate": 4.449802972455342e-08, "loss": 0.5178, "step": 5766 }, { "epoch": 14.100244498777506, "grad_norm": 0.3569361387002512, "learning_rate": 4.425788798068459e-08, "loss": 0.663, "step": 5767 }, { "epoch": 14.102689486552567, "grad_norm": 0.2548868236955468, "learning_rate": 4.401839018861265e-08, "loss": 0.7352, "step": 5768 }, { "epoch": 14.10513447432763, "grad_norm": 0.3939596972060168, "learning_rate": 4.3779536411139856e-08, "loss": 0.8855, "step": 5769 }, { "epoch": 14.10757946210269, "grad_norm": 0.24702342476808542, "learning_rate": 4.354132671089861e-08, "loss": 0.72, "step": 5770 }, { "epoch": 14.11002444987775, "grad_norm": 0.32146872624906675, "learning_rate": 4.33037611503534e-08, "loss": 0.4705, "step": 5771 }, { "epoch": 14.112469437652813, "grad_norm": 0.22284450075668427, "learning_rate": 4.306683979179882e-08, "loss": 0.378, "step": 5772 }, { "epoch": 14.114914425427873, "grad_norm": 0.25090653870013874, "learning_rate": 4.2830562697361856e-08, "loss": 0.3851, "step": 5773 }, { "epoch": 14.117359413202934, "grad_norm": 0.33274588254089627, "learning_rate": 4.2594929928999054e-08, "loss": 0.5454, "step": 5774 }, { "epoch": 14.119804400977996, "grad_norm": 0.3419712093315197, "learning_rate": 4.235994154849904e-08, "loss": 0.4075, "step": 5775 }, { "epoch": 14.122249388753056, "grad_norm": 0.28474404662421293, "learning_rate": 4.212559761748086e-08, "loss": 0.522, "step": 5776 }, { "epoch": 14.124694376528117, "grad_norm": 0.21447609684088564, "learning_rate": 4.1891898197395367e-08, "loss": 0.3952, "step": 5777 }, { "epoch": 14.12713936430318, "grad_norm": 0.24304734405654124, "learning_rate": 4.165884334952352e-08, "loss": 0.8457, "step": 5778 }, { "epoch": 14.12958435207824, "grad_norm": 0.26982376550642384, "learning_rate": 4.142643313497785e-08, "loss": 0.3908, "step": 5779 }, { "epoch": 14.1320293398533, "grad_norm": 0.17896362938785879, "learning_rate": 4.119466761470153e-08, "loss": 0.598, "step": 5780 }, { "epoch": 14.134474327628363, "grad_norm": 0.18992077251970388, "learning_rate": 4.096354684946846e-08, "loss": 0.5713, "step": 5781 }, { "epoch": 14.136919315403423, "grad_norm": 0.26972108640436354, "learning_rate": 4.0733070899884316e-08, "loss": 0.5261, "step": 5782 }, { "epoch": 14.139364303178484, "grad_norm": 0.3468339580445499, "learning_rate": 4.050323982638521e-08, "loss": 0.9662, "step": 5783 }, { "epoch": 14.141809290953546, "grad_norm": 0.1732534930877171, "learning_rate": 4.0274053689237645e-08, "loss": 0.5134, "step": 5784 }, { "epoch": 14.144254278728607, "grad_norm": 0.28928495557866196, "learning_rate": 4.0045512548539665e-08, "loss": 0.843, "step": 5785 }, { "epoch": 14.146699266503667, "grad_norm": 0.3186053722928193, "learning_rate": 3.9817616464219996e-08, "loss": 0.6506, "step": 5786 }, { "epoch": 14.14914425427873, "grad_norm": 0.25442512894081876, "learning_rate": 3.959036549603834e-08, "loss": 0.7672, "step": 5787 }, { "epoch": 14.15158924205379, "grad_norm": 0.2339412941506138, "learning_rate": 3.93637597035848e-08, "loss": 0.5266, "step": 5788 }, { "epoch": 14.15403422982885, "grad_norm": 0.2474792418936597, "learning_rate": 3.913779914628074e-08, "loss": 0.3655, "step": 5789 }, { "epoch": 14.156479217603913, "grad_norm": 0.2871799981087508, "learning_rate": 3.8912483883378486e-08, "loss": 0.8848, "step": 5790 }, { "epoch": 14.158924205378973, "grad_norm": 0.19100529412154701, "learning_rate": 3.868781397396049e-08, "loss": 0.4238, "step": 5791 }, { "epoch": 14.161369193154034, "grad_norm": 0.28444728316033757, "learning_rate": 3.846378947694046e-08, "loss": 0.7894, "step": 5792 }, { "epoch": 14.163814180929096, "grad_norm": 0.17480592619873836, "learning_rate": 3.824041045106308e-08, "loss": 0.6868, "step": 5793 }, { "epoch": 14.166259168704157, "grad_norm": 0.2916003025169568, "learning_rate": 3.8017676954902884e-08, "loss": 1.0485, "step": 5794 }, { "epoch": 14.168704156479217, "grad_norm": 0.24219757604746475, "learning_rate": 3.7795589046865924e-08, "loss": 0.607, "step": 5795 }, { "epoch": 14.17114914425428, "grad_norm": 0.282070639694405, "learning_rate": 3.757414678518867e-08, "loss": 0.5482, "step": 5796 }, { "epoch": 14.17359413202934, "grad_norm": 0.32637063388131127, "learning_rate": 3.735335022793857e-08, "loss": 0.4665, "step": 5797 }, { "epoch": 14.1760391198044, "grad_norm": 0.32643344181320255, "learning_rate": 3.713319943301319e-08, "loss": 0.7588, "step": 5798 }, { "epoch": 14.178484107579463, "grad_norm": 0.2772353649858115, "learning_rate": 3.691369445814136e-08, "loss": 0.643, "step": 5799 }, { "epoch": 14.180929095354523, "grad_norm": 0.3227630973400612, "learning_rate": 3.669483536088231e-08, "loss": 1.5477, "step": 5800 }, { "epoch": 14.183374083129584, "grad_norm": 0.272374224596172, "learning_rate": 3.6476622198625965e-08, "loss": 0.4688, "step": 5801 }, { "epoch": 14.185819070904646, "grad_norm": 0.2921798931564791, "learning_rate": 3.625905502859239e-08, "loss": 0.4324, "step": 5802 }, { "epoch": 14.188264058679707, "grad_norm": 0.24301959083423358, "learning_rate": 3.6042133907833176e-08, "loss": 0.4456, "step": 5803 }, { "epoch": 14.190709046454767, "grad_norm": 0.2509397771046098, "learning_rate": 3.582585889322976e-08, "loss": 0.8093, "step": 5804 }, { "epoch": 14.19315403422983, "grad_norm": 0.21047896067486083, "learning_rate": 3.5610230041494834e-08, "loss": 0.9347, "step": 5805 }, { "epoch": 14.19559902200489, "grad_norm": 0.2997670530108449, "learning_rate": 3.5395247409170385e-08, "loss": 1.1111, "step": 5806 }, { "epoch": 14.19804400977995, "grad_norm": 0.22629703693309292, "learning_rate": 3.518091105263022e-08, "loss": 0.4266, "step": 5807 }, { "epoch": 14.200488997555013, "grad_norm": 0.2430767257480146, "learning_rate": 3.496722102807853e-08, "loss": 0.9914, "step": 5808 }, { "epoch": 14.202933985330073, "grad_norm": 0.2543193292456342, "learning_rate": 3.475417739154913e-08, "loss": 0.4999, "step": 5809 }, { "epoch": 14.205378973105134, "grad_norm": 0.31183675571629743, "learning_rate": 3.4541780198907584e-08, "loss": 0.7114, "step": 5810 }, { "epoch": 14.207823960880196, "grad_norm": 0.31855643673328343, "learning_rate": 3.4330029505848794e-08, "loss": 0.6549, "step": 5811 }, { "epoch": 14.210268948655257, "grad_norm": 0.32168105736284275, "learning_rate": 3.41189253678989e-08, "loss": 0.4952, "step": 5812 }, { "epoch": 14.212713936430317, "grad_norm": 0.17470943079997792, "learning_rate": 3.390846784041418e-08, "loss": 0.4032, "step": 5813 }, { "epoch": 14.21515892420538, "grad_norm": 0.1983161791480925, "learning_rate": 3.3698656978581325e-08, "loss": 0.4699, "step": 5814 }, { "epoch": 14.21760391198044, "grad_norm": 0.2887570403266946, "learning_rate": 3.348949283741798e-08, "loss": 0.5431, "step": 5815 }, { "epoch": 14.2200488997555, "grad_norm": 0.23896624940573702, "learning_rate": 3.3280975471771114e-08, "loss": 0.4592, "step": 5816 }, { "epoch": 14.222493887530563, "grad_norm": 0.2542155307981523, "learning_rate": 3.3073104936318936e-08, "loss": 0.4584, "step": 5817 }, { "epoch": 14.224938875305623, "grad_norm": 0.2481052644371504, "learning_rate": 3.2865881285570054e-08, "loss": 0.401, "step": 5818 }, { "epoch": 14.227383863080684, "grad_norm": 0.20676258494954353, "learning_rate": 3.265930457386296e-08, "loss": 0.6551, "step": 5819 }, { "epoch": 14.229828850855746, "grad_norm": 0.2849613847337719, "learning_rate": 3.245337485536737e-08, "loss": 0.5869, "step": 5820 }, { "epoch": 14.232273838630807, "grad_norm": 0.3626632418726902, "learning_rate": 3.224809218408231e-08, "loss": 0.7387, "step": 5821 }, { "epoch": 14.234718826405867, "grad_norm": 0.19636986069558876, "learning_rate": 3.204345661383751e-08, "loss": 0.4553, "step": 5822 }, { "epoch": 14.23716381418093, "grad_norm": 0.2222285707980301, "learning_rate": 3.1839468198293644e-08, "loss": 0.3726, "step": 5823 }, { "epoch": 14.23960880195599, "grad_norm": 0.28211746781872943, "learning_rate": 3.163612699094043e-08, "loss": 0.6344, "step": 5824 }, { "epoch": 14.24205378973105, "grad_norm": 0.3583086516114762, "learning_rate": 3.14334330450991e-08, "loss": 0.6171, "step": 5825 }, { "epoch": 14.244498777506113, "grad_norm": 0.33559944457654917, "learning_rate": 3.1231386413920485e-08, "loss": 0.7111, "step": 5826 }, { "epoch": 14.246943765281173, "grad_norm": 0.23220668053602828, "learning_rate": 3.1029987150385796e-08, "loss": 0.4841, "step": 5827 }, { "epoch": 14.249388753056234, "grad_norm": 0.22459419682948784, "learning_rate": 3.082923530730669e-08, "loss": 0.7535, "step": 5828 }, { "epoch": 14.251833740831296, "grad_norm": 0.267867488759652, "learning_rate": 3.062913093732467e-08, "loss": 0.5828, "step": 5829 }, { "epoch": 14.254278728606357, "grad_norm": 0.19388296447291925, "learning_rate": 3.042967409291192e-08, "loss": 0.4412, "step": 5830 }, { "epoch": 14.256723716381417, "grad_norm": 0.28650977167044434, "learning_rate": 3.023086482637022e-08, "loss": 0.4488, "step": 5831 }, { "epoch": 14.25916870415648, "grad_norm": 0.3235625557495168, "learning_rate": 3.0032703189832315e-08, "loss": 0.5247, "step": 5832 }, { "epoch": 14.26161369193154, "grad_norm": 1.6108017619387143, "learning_rate": 2.983518923526052e-08, "loss": 0.805, "step": 5833 }, { "epoch": 14.2640586797066, "grad_norm": 0.31446525579947765, "learning_rate": 2.9638323014447577e-08, "loss": 0.7721, "step": 5834 }, { "epoch": 14.266503667481663, "grad_norm": 0.24807680695131962, "learning_rate": 2.944210457901636e-08, "loss": 0.5272, "step": 5835 }, { "epoch": 14.268948655256724, "grad_norm": 0.2745691982755478, "learning_rate": 2.9246533980419878e-08, "loss": 0.5274, "step": 5836 }, { "epoch": 14.271393643031784, "grad_norm": 0.3368195697390879, "learning_rate": 2.9051611269940718e-08, "loss": 0.6726, "step": 5837 }, { "epoch": 14.273838630806846, "grad_norm": 0.1933341907948257, "learning_rate": 2.8857336498692712e-08, "loss": 0.3729, "step": 5838 }, { "epoch": 14.276283618581907, "grad_norm": 0.2594667788203325, "learning_rate": 2.8663709717619003e-08, "loss": 0.4295, "step": 5839 }, { "epoch": 14.278728606356967, "grad_norm": 0.2639111743020132, "learning_rate": 2.847073097749259e-08, "loss": 0.383, "step": 5840 }, { "epoch": 14.28117359413203, "grad_norm": 0.26811570536981266, "learning_rate": 2.8278400328917434e-08, "loss": 0.4062, "step": 5841 }, { "epoch": 14.28361858190709, "grad_norm": 0.256179869294639, "learning_rate": 2.8086717822326526e-08, "loss": 0.6226, "step": 5842 }, { "epoch": 14.28606356968215, "grad_norm": 0.21704994850977963, "learning_rate": 2.7895683507984107e-08, "loss": 0.4964, "step": 5843 }, { "epoch": 14.288508557457213, "grad_norm": 0.16805759807252935, "learning_rate": 2.770529743598316e-08, "loss": 0.321, "step": 5844 }, { "epoch": 14.290953545232274, "grad_norm": 0.25032701111503786, "learning_rate": 2.751555965624736e-08, "loss": 0.5232, "step": 5845 }, { "epoch": 14.293398533007334, "grad_norm": 0.3030503931391864, "learning_rate": 2.7326470218530242e-08, "loss": 0.3471, "step": 5846 }, { "epoch": 14.295843520782396, "grad_norm": 0.2639972823139374, "learning_rate": 2.7138029172415757e-08, "loss": 0.3927, "step": 5847 }, { "epoch": 14.298288508557457, "grad_norm": 0.31790158040174044, "learning_rate": 2.6950236567317157e-08, "loss": 0.3144, "step": 5848 }, { "epoch": 14.30073349633252, "grad_norm": 0.28752766629175236, "learning_rate": 2.6763092452477823e-08, "loss": 0.5689, "step": 5849 }, { "epoch": 14.30317848410758, "grad_norm": 0.20613251497739601, "learning_rate": 2.6576596876971562e-08, "loss": 0.9234, "step": 5850 }, { "epoch": 14.30562347188264, "grad_norm": 0.2960427100855193, "learning_rate": 2.639074988970175e-08, "loss": 0.3269, "step": 5851 }, { "epoch": 14.3080684596577, "grad_norm": 0.25671179313754044, "learning_rate": 2.6205551539401353e-08, "loss": 0.4295, "step": 5852 }, { "epoch": 14.310513447432763, "grad_norm": 0.2496847246034851, "learning_rate": 2.602100187463402e-08, "loss": 0.6114, "step": 5853 }, { "epoch": 14.312958435207824, "grad_norm": 0.41377589381240654, "learning_rate": 2.5837100943792714e-08, "loss": 0.9258, "step": 5854 }, { "epoch": 14.315403422982886, "grad_norm": 0.24918178144569356, "learning_rate": 2.56538487951008e-08, "loss": 0.4429, "step": 5855 }, { "epoch": 14.317848410757946, "grad_norm": 0.3600099652689131, "learning_rate": 2.547124547661067e-08, "loss": 0.4993, "step": 5856 }, { "epoch": 14.320293398533007, "grad_norm": 0.317314336296521, "learning_rate": 2.5289291036205142e-08, "loss": 0.329, "step": 5857 }, { "epoch": 14.32273838630807, "grad_norm": 0.24780148755521586, "learning_rate": 2.5107985521597434e-08, "loss": 0.8495, "step": 5858 }, { "epoch": 14.32518337408313, "grad_norm": 0.19761235090892323, "learning_rate": 2.4927328980329236e-08, "loss": 0.6029, "step": 5859 }, { "epoch": 14.32762836185819, "grad_norm": 0.24276681499798555, "learning_rate": 2.474732145977321e-08, "loss": 0.6108, "step": 5860 }, { "epoch": 14.330073349633253, "grad_norm": 0.4499166168837541, "learning_rate": 2.45679630071316e-08, "loss": 0.7908, "step": 5861 }, { "epoch": 14.332518337408313, "grad_norm": 0.23121864585343233, "learning_rate": 2.4389253669435674e-08, "loss": 0.4192, "step": 5862 }, { "epoch": 14.334963325183374, "grad_norm": 0.21463803044609225, "learning_rate": 2.4211193493547946e-08, "loss": 0.8216, "step": 5863 }, { "epoch": 14.337408312958436, "grad_norm": 0.32413040044665375, "learning_rate": 2.40337825261594e-08, "loss": 0.4791, "step": 5864 }, { "epoch": 14.339853300733497, "grad_norm": 0.23079069001757216, "learning_rate": 2.3857020813791433e-08, "loss": 0.4953, "step": 5865 }, { "epoch": 14.342298288508557, "grad_norm": 0.2447484706211273, "learning_rate": 2.3680908402794745e-08, "loss": 0.8017, "step": 5866 }, { "epoch": 14.34474327628362, "grad_norm": 0.27288029653247675, "learning_rate": 2.3505445339350175e-08, "loss": 0.6673, "step": 5867 }, { "epoch": 14.34718826405868, "grad_norm": 0.24957128080161445, "learning_rate": 2.333063166946842e-08, "loss": 0.6841, "step": 5868 }, { "epoch": 14.34963325183374, "grad_norm": 0.2174832054818101, "learning_rate": 2.3156467438989194e-08, "loss": 0.6158, "step": 5869 }, { "epoch": 14.352078239608803, "grad_norm": 0.32288480461399, "learning_rate": 2.2982952693582638e-08, "loss": 0.4772, "step": 5870 }, { "epoch": 14.354523227383863, "grad_norm": 0.23771156325053922, "learning_rate": 2.281008747874819e-08, "loss": 0.4998, "step": 5871 }, { "epoch": 14.356968215158924, "grad_norm": 0.24868999418083254, "learning_rate": 2.2637871839814863e-08, "loss": 1.0037, "step": 5872 }, { "epoch": 14.359413202933986, "grad_norm": 0.2281428282696578, "learning_rate": 2.2466305821941815e-08, "loss": 0.4357, "step": 5873 }, { "epoch": 14.361858190709047, "grad_norm": 0.3040265399616088, "learning_rate": 2.22953894701175e-08, "loss": 0.3662, "step": 5874 }, { "epoch": 14.364303178484107, "grad_norm": 0.22099601567100147, "learning_rate": 2.2125122829160516e-08, "loss": 0.4357, "step": 5875 }, { "epoch": 14.36674816625917, "grad_norm": 0.3326749355981543, "learning_rate": 2.1955505943717926e-08, "loss": 0.5771, "step": 5876 }, { "epoch": 14.36919315403423, "grad_norm": 0.28093276359216685, "learning_rate": 2.178653885826748e-08, "loss": 0.5564, "step": 5877 }, { "epoch": 14.37163814180929, "grad_norm": 0.19460678368847734, "learning_rate": 2.1618221617116243e-08, "loss": 0.4142, "step": 5878 }, { "epoch": 14.374083129584353, "grad_norm": 0.1666924558980049, "learning_rate": 2.1450554264400846e-08, "loss": 0.5622, "step": 5879 }, { "epoch": 14.376528117359413, "grad_norm": 0.4720935220821703, "learning_rate": 2.1283536844087514e-08, "loss": 0.4856, "step": 5880 }, { "epoch": 14.378973105134474, "grad_norm": 0.18700294343653207, "learning_rate": 2.1117169399972327e-08, "loss": 0.6129, "step": 5881 }, { "epoch": 14.381418092909536, "grad_norm": 0.36678118975137125, "learning_rate": 2.0951451975679827e-08, "loss": 0.4568, "step": 5882 }, { "epoch": 14.383863080684597, "grad_norm": 0.28995097553404914, "learning_rate": 2.078638461466581e-08, "loss": 0.756, "step": 5883 }, { "epoch": 14.386308068459657, "grad_norm": 0.19667664652681321, "learning_rate": 2.062196736021399e-08, "loss": 0.421, "step": 5884 }, { "epoch": 14.38875305623472, "grad_norm": 0.2908773173423205, "learning_rate": 2.0458200255439042e-08, "loss": 0.6971, "step": 5885 }, { "epoch": 14.39119804400978, "grad_norm": 0.22692559188191663, "learning_rate": 2.029508334328384e-08, "loss": 0.388, "step": 5886 }, { "epoch": 14.39364303178484, "grad_norm": 0.21265110816814112, "learning_rate": 2.0132616666521666e-08, "loss": 0.4936, "step": 5887 }, { "epoch": 14.396088019559903, "grad_norm": 0.21045318807801613, "learning_rate": 1.9970800267755108e-08, "loss": 0.6917, "step": 5888 }, { "epoch": 14.398533007334963, "grad_norm": 0.21180267442415515, "learning_rate": 1.980963418941606e-08, "loss": 0.3705, "step": 5889 }, { "epoch": 14.400977995110024, "grad_norm": 0.3685732318156828, "learning_rate": 1.964911847376544e-08, "loss": 0.859, "step": 5890 }, { "epoch": 14.403422982885086, "grad_norm": 0.33655485843873795, "learning_rate": 1.9489253162895127e-08, "loss": 0.3365, "step": 5891 }, { "epoch": 14.405867970660147, "grad_norm": 0.26800385093450046, "learning_rate": 1.9330038298724652e-08, "loss": 0.9645, "step": 5892 }, { "epoch": 14.408312958435207, "grad_norm": 0.23234158208899475, "learning_rate": 1.9171473923003946e-08, "loss": 0.6413, "step": 5893 }, { "epoch": 14.41075794621027, "grad_norm": 0.28703713000881426, "learning_rate": 1.9013560077312253e-08, "loss": 0.3595, "step": 5894 }, { "epoch": 14.41320293398533, "grad_norm": 0.3136894419786609, "learning_rate": 1.885629680305867e-08, "loss": 0.4776, "step": 5895 }, { "epoch": 14.41564792176039, "grad_norm": 0.19211267149853545, "learning_rate": 1.869968414148049e-08, "loss": 0.4327, "step": 5896 }, { "epoch": 14.418092909535453, "grad_norm": 0.2236724758160594, "learning_rate": 1.8543722133645693e-08, "loss": 0.4481, "step": 5897 }, { "epoch": 14.420537897310513, "grad_norm": 0.23997687927456546, "learning_rate": 1.838841082045073e-08, "loss": 0.7161, "step": 5898 }, { "epoch": 14.422982885085574, "grad_norm": 0.2470977475350639, "learning_rate": 1.8233750242621917e-08, "loss": 0.5961, "step": 5899 }, { "epoch": 14.425427872860636, "grad_norm": 0.3582095420789469, "learning_rate": 1.8079740440714866e-08, "loss": 0.7068, "step": 5900 }, { "epoch": 14.427872860635697, "grad_norm": 0.2367564665371195, "learning_rate": 1.7926381455114495e-08, "loss": 0.8264, "step": 5901 }, { "epoch": 14.430317848410757, "grad_norm": 0.20600421315518097, "learning_rate": 1.7773673326035024e-08, "loss": 0.644, "step": 5902 }, { "epoch": 14.43276283618582, "grad_norm": 0.3341196720073081, "learning_rate": 1.762161609351998e-08, "loss": 0.3405, "step": 5903 }, { "epoch": 14.43520782396088, "grad_norm": 0.1995247221169888, "learning_rate": 1.7470209797441905e-08, "loss": 0.4356, "step": 5904 }, { "epoch": 14.43765281173594, "grad_norm": 0.24691762455068278, "learning_rate": 1.7319454477503495e-08, "loss": 0.8613, "step": 5905 }, { "epoch": 14.440097799511003, "grad_norm": 0.23092519815290935, "learning_rate": 1.7169350173236176e-08, "loss": 0.7153, "step": 5906 }, { "epoch": 14.442542787286063, "grad_norm": 0.264879717034351, "learning_rate": 1.7019896924000412e-08, "loss": 0.3733, "step": 5907 }, { "epoch": 14.444987775061124, "grad_norm": 0.23277367272197552, "learning_rate": 1.6871094768986795e-08, "loss": 0.5023, "step": 5908 }, { "epoch": 14.447432762836186, "grad_norm": 0.3917808881196965, "learning_rate": 1.6722943747214117e-08, "loss": 0.4251, "step": 5909 }, { "epoch": 14.449877750611247, "grad_norm": 0.28663015315239915, "learning_rate": 1.6575443897531295e-08, "loss": 0.4217, "step": 5910 }, { "epoch": 14.452322738386307, "grad_norm": 0.21802247534454908, "learning_rate": 1.6428595258616008e-08, "loss": 0.4989, "step": 5911 }, { "epoch": 14.45476772616137, "grad_norm": 0.30452590965971515, "learning_rate": 1.628239786897551e-08, "loss": 0.6905, "step": 5912 }, { "epoch": 14.45721271393643, "grad_norm": 0.23974505361225917, "learning_rate": 1.6136851766945793e-08, "loss": 0.5973, "step": 5913 }, { "epoch": 14.45965770171149, "grad_norm": 0.3599449784250477, "learning_rate": 1.5991956990692724e-08, "loss": 0.4503, "step": 5914 }, { "epoch": 14.462102689486553, "grad_norm": 0.40809627687685907, "learning_rate": 1.5847713578210632e-08, "loss": 0.4662, "step": 5915 }, { "epoch": 14.464547677261614, "grad_norm": 0.17465008783350863, "learning_rate": 1.5704121567323704e-08, "loss": 0.4139, "step": 5916 }, { "epoch": 14.466992665036674, "grad_norm": 0.42271568944935967, "learning_rate": 1.5561180995685156e-08, "loss": 0.3809, "step": 5917 }, { "epoch": 14.469437652811736, "grad_norm": 0.24442652435875195, "learning_rate": 1.5418891900777223e-08, "loss": 0.325, "step": 5918 }, { "epoch": 14.471882640586797, "grad_norm": 0.14954539407623077, "learning_rate": 1.527725431991145e-08, "loss": 0.479, "step": 5919 }, { "epoch": 14.474327628361857, "grad_norm": 0.27725361238971163, "learning_rate": 1.5136268290227852e-08, "loss": 0.5152, "step": 5920 }, { "epoch": 14.47677261613692, "grad_norm": 0.28664971567029324, "learning_rate": 1.499593384869713e-08, "loss": 0.3667, "step": 5921 }, { "epoch": 14.47921760391198, "grad_norm": 0.20328984180663284, "learning_rate": 1.4856251032117353e-08, "loss": 0.8143, "step": 5922 }, { "epoch": 14.48166259168704, "grad_norm": 0.3158681674724048, "learning_rate": 1.4717219877116996e-08, "loss": 0.3009, "step": 5923 }, { "epoch": 14.484107579462103, "grad_norm": 0.5917968853268613, "learning_rate": 1.4578840420153006e-08, "loss": 0.831, "step": 5924 }, { "epoch": 14.486552567237164, "grad_norm": 0.2422789309488887, "learning_rate": 1.4441112697511638e-08, "loss": 0.5739, "step": 5925 }, { "epoch": 14.488997555012224, "grad_norm": 0.1697282485955006, "learning_rate": 1.4304036745308724e-08, "loss": 0.4268, "step": 5926 }, { "epoch": 14.491442542787286, "grad_norm": 0.22723014484916346, "learning_rate": 1.4167612599488012e-08, "loss": 0.6064, "step": 5927 }, { "epoch": 14.493887530562347, "grad_norm": 0.3569433971771532, "learning_rate": 1.4031840295823663e-08, "loss": 0.566, "step": 5928 }, { "epoch": 14.496332518337407, "grad_norm": 0.20195702490881626, "learning_rate": 1.3896719869917753e-08, "loss": 0.4743, "step": 5929 }, { "epoch": 14.49877750611247, "grad_norm": 0.2197320089290472, "learning_rate": 1.3762251357202216e-08, "loss": 0.5578, "step": 5930 }, { "epoch": 14.50122249388753, "grad_norm": 0.20256403039679488, "learning_rate": 1.3628434792938016e-08, "loss": 0.3571, "step": 5931 }, { "epoch": 14.503667481662593, "grad_norm": 0.2059181372928567, "learning_rate": 1.3495270212214584e-08, "loss": 0.9066, "step": 5932 }, { "epoch": 14.506112469437653, "grad_norm": 0.2903535547845707, "learning_rate": 1.3362757649950652e-08, "loss": 0.8555, "step": 5933 }, { "epoch": 14.508557457212714, "grad_norm": 0.22822853438766613, "learning_rate": 1.323089714089426e-08, "loss": 0.4462, "step": 5934 }, { "epoch": 14.511002444987774, "grad_norm": 0.18205427653832412, "learning_rate": 1.3099688719622195e-08, "loss": 0.4886, "step": 5935 }, { "epoch": 14.513447432762836, "grad_norm": 0.28864277506262753, "learning_rate": 1.2969132420540265e-08, "loss": 0.4744, "step": 5936 }, { "epoch": 14.515892420537897, "grad_norm": 0.3287538193052468, "learning_rate": 1.2839228277883032e-08, "loss": 0.3179, "step": 5937 }, { "epoch": 14.51833740831296, "grad_norm": 0.2629643344453778, "learning_rate": 1.2709976325715189e-08, "loss": 1.4419, "step": 5938 }, { "epoch": 14.52078239608802, "grad_norm": 0.24686352942596163, "learning_rate": 1.2581376597928796e-08, "loss": 0.5284, "step": 5939 }, { "epoch": 14.52322738386308, "grad_norm": 0.23014420988141793, "learning_rate": 1.2453429128245765e-08, "loss": 0.6586, "step": 5940 }, { "epoch": 14.525672371638143, "grad_norm": 0.3216063361305703, "learning_rate": 1.2326133950217034e-08, "loss": 0.655, "step": 5941 }, { "epoch": 14.528117359413203, "grad_norm": 0.18130755911801785, "learning_rate": 1.2199491097222293e-08, "loss": 0.6401, "step": 5942 }, { "epoch": 14.530562347188264, "grad_norm": 0.3403314851621884, "learning_rate": 1.2073500602470257e-08, "loss": 0.458, "step": 5943 }, { "epoch": 14.533007334963326, "grad_norm": 0.23245002733190495, "learning_rate": 1.1948162498998107e-08, "loss": 0.4228, "step": 5944 }, { "epoch": 14.535452322738386, "grad_norm": 0.24536593026426867, "learning_rate": 1.1823476819672885e-08, "loss": 0.8562, "step": 5945 }, { "epoch": 14.537897310513447, "grad_norm": 0.3546474440078184, "learning_rate": 1.169944359718983e-08, "loss": 0.3927, "step": 5946 }, { "epoch": 14.54034229828851, "grad_norm": 0.3157771654121839, "learning_rate": 1.1576062864073201e-08, "loss": 0.4022, "step": 5947 }, { "epoch": 14.54278728606357, "grad_norm": 0.43788749389038056, "learning_rate": 1.1453334652676285e-08, "loss": 0.4188, "step": 5948 }, { "epoch": 14.54523227383863, "grad_norm": 0.24826591774214785, "learning_rate": 1.1331258995181116e-08, "loss": 0.741, "step": 5949 }, { "epoch": 14.547677261613693, "grad_norm": 0.2724595057026709, "learning_rate": 1.1209835923599033e-08, "loss": 0.6084, "step": 5950 }, { "epoch": 14.550122249388753, "grad_norm": 0.3328890855915936, "learning_rate": 1.1089065469769844e-08, "loss": 0.4137, "step": 5951 }, { "epoch": 14.552567237163814, "grad_norm": 0.2815242918215007, "learning_rate": 1.0968947665362107e-08, "loss": 0.7544, "step": 5952 }, { "epoch": 14.555012224938876, "grad_norm": 0.3134859748126106, "learning_rate": 1.084948254187368e-08, "loss": 0.8459, "step": 5953 }, { "epoch": 14.557457212713937, "grad_norm": 0.2871337806962544, "learning_rate": 1.0730670130630893e-08, "loss": 0.6503, "step": 5954 }, { "epoch": 14.559902200488997, "grad_norm": 0.4192824698489224, "learning_rate": 1.0612510462789383e-08, "loss": 0.3846, "step": 5955 }, { "epoch": 14.56234718826406, "grad_norm": 0.44153812040528045, "learning_rate": 1.0495003569332974e-08, "loss": 0.5928, "step": 5956 }, { "epoch": 14.56479217603912, "grad_norm": 0.33955729192970996, "learning_rate": 1.0378149481074796e-08, "loss": 0.6242, "step": 5957 }, { "epoch": 14.56723716381418, "grad_norm": 0.22637009065739055, "learning_rate": 1.0261948228656726e-08, "loss": 0.8407, "step": 5958 }, { "epoch": 14.569682151589243, "grad_norm": 0.314303991489773, "learning_rate": 1.0146399842549116e-08, "loss": 0.5836, "step": 5959 }, { "epoch": 14.572127139364303, "grad_norm": 0.2849968244545076, "learning_rate": 1.0031504353051891e-08, "loss": 0.5092, "step": 5960 }, { "epoch": 14.574572127139364, "grad_norm": 0.31104698716324186, "learning_rate": 9.917261790292898e-09, "loss": 1.1197, "step": 5961 }, { "epoch": 14.577017114914426, "grad_norm": 0.24736896285181478, "learning_rate": 9.803672184229285e-09, "loss": 0.7574, "step": 5962 }, { "epoch": 14.579462102689487, "grad_norm": 0.2938954362423323, "learning_rate": 9.690735564646946e-09, "loss": 0.5853, "step": 5963 }, { "epoch": 14.581907090464547, "grad_norm": 0.2550767577779828, "learning_rate": 9.57845196116025e-09, "loss": 0.4612, "step": 5964 }, { "epoch": 14.58435207823961, "grad_norm": 0.1789110933635988, "learning_rate": 9.46682140321259e-09, "loss": 0.7292, "step": 5965 }, { "epoch": 14.58679706601467, "grad_norm": 0.27553399752375046, "learning_rate": 9.355843920076101e-09, "loss": 0.6208, "step": 5966 }, { "epoch": 14.58924205378973, "grad_norm": 0.24773899640474598, "learning_rate": 9.2455195408514e-09, "loss": 0.8385, "step": 5967 }, { "epoch": 14.591687041564793, "grad_norm": 0.3733501862142067, "learning_rate": 9.135848294468396e-09, "loss": 0.8394, "step": 5968 }, { "epoch": 14.594132029339853, "grad_norm": 0.3578934411237306, "learning_rate": 9.0268302096852e-09, "loss": 0.7831, "step": 5969 }, { "epoch": 14.596577017114914, "grad_norm": 0.27935597065238066, "learning_rate": 8.918465315088943e-09, "loss": 0.4194, "step": 5970 }, { "epoch": 14.599022004889976, "grad_norm": 0.34961499210108704, "learning_rate": 8.810753639095227e-09, "loss": 0.8387, "step": 5971 }, { "epoch": 14.601466992665037, "grad_norm": 0.26205186074461606, "learning_rate": 8.70369520994896e-09, "loss": 0.3376, "step": 5972 }, { "epoch": 14.603911980440097, "grad_norm": 0.3454017371791375, "learning_rate": 8.597290055722685e-09, "loss": 0.4305, "step": 5973 }, { "epoch": 14.60635696821516, "grad_norm": 0.2654511539156003, "learning_rate": 8.491538204318805e-09, "loss": 0.605, "step": 5974 }, { "epoch": 14.60880195599022, "grad_norm": 0.22827230052113323, "learning_rate": 8.386439683467362e-09, "loss": 0.5083, "step": 5975 }, { "epoch": 14.61124694376528, "grad_norm": 0.21119746496374842, "learning_rate": 8.281994520727976e-09, "loss": 0.6664, "step": 5976 }, { "epoch": 14.613691931540343, "grad_norm": 0.15489263894510524, "learning_rate": 8.178202743488462e-09, "loss": 0.4218, "step": 5977 }, { "epoch": 14.616136919315403, "grad_norm": 0.19392786515917243, "learning_rate": 8.075064378965381e-09, "loss": 0.5175, "step": 5978 }, { "epoch": 14.618581907090464, "grad_norm": 0.18273882377754652, "learning_rate": 7.972579454204043e-09, "loss": 0.4118, "step": 5979 }, { "epoch": 14.621026894865526, "grad_norm": 0.22618739440532434, "learning_rate": 7.870747996078231e-09, "loss": 0.7905, "step": 5980 }, { "epoch": 14.623471882640587, "grad_norm": 0.23847204498152716, "learning_rate": 7.769570031290752e-09, "loss": 0.4172, "step": 5981 }, { "epoch": 14.625916870415647, "grad_norm": 0.3629398630770936, "learning_rate": 7.669045586372325e-09, "loss": 0.3507, "step": 5982 }, { "epoch": 14.62836185819071, "grad_norm": 0.21479144620517518, "learning_rate": 7.569174687683257e-09, "loss": 0.3935, "step": 5983 }, { "epoch": 14.63080684596577, "grad_norm": 0.23743563284813463, "learning_rate": 7.469957361411483e-09, "loss": 0.4452, "step": 5984 }, { "epoch": 14.63325183374083, "grad_norm": 0.20815162666555456, "learning_rate": 7.371393633574253e-09, "loss": 0.4598, "step": 5985 }, { "epoch": 14.635696821515893, "grad_norm": 0.2942837952267298, "learning_rate": 7.273483530017556e-09, "loss": 0.7701, "step": 5986 }, { "epoch": 14.638141809290953, "grad_norm": 0.24847949017980506, "learning_rate": 7.176227076415299e-09, "loss": 0.9498, "step": 5987 }, { "epoch": 14.640586797066014, "grad_norm": 0.2886962492048405, "learning_rate": 7.0796242982704155e-09, "loss": 1.057, "step": 5988 }, { "epoch": 14.643031784841076, "grad_norm": 0.2407338622916156, "learning_rate": 6.98367522091431e-09, "loss": 0.7743, "step": 5989 }, { "epoch": 14.645476772616137, "grad_norm": 0.35345018489614627, "learning_rate": 6.888379869507134e-09, "loss": 0.4915, "step": 5990 }, { "epoch": 14.647921760391197, "grad_norm": 0.2588222064637483, "learning_rate": 6.7937382690375085e-09, "loss": 0.6053, "step": 5991 }, { "epoch": 14.65036674816626, "grad_norm": 0.19407202420345154, "learning_rate": 6.699750444322806e-09, "loss": 0.7861, "step": 5992 }, { "epoch": 14.65281173594132, "grad_norm": 0.38056007830827043, "learning_rate": 6.6064164200083105e-09, "loss": 0.4325, "step": 5993 }, { "epoch": 14.65525672371638, "grad_norm": 0.324321971744957, "learning_rate": 6.513736220568889e-09, "loss": 0.9293, "step": 5994 }, { "epoch": 14.657701711491443, "grad_norm": 0.3170154986244328, "learning_rate": 6.421709870306769e-09, "loss": 0.4786, "step": 5995 }, { "epoch": 14.660146699266503, "grad_norm": 0.24505145820311475, "learning_rate": 6.330337393354036e-09, "loss": 0.5915, "step": 5996 }, { "epoch": 14.662591687041564, "grad_norm": 0.2775314860635019, "learning_rate": 6.239618813670412e-09, "loss": 0.5127, "step": 5997 }, { "epoch": 14.665036674816626, "grad_norm": 0.22930119755436948, "learning_rate": 6.149554155044091e-09, "loss": 0.8323, "step": 5998 }, { "epoch": 14.667481662591687, "grad_norm": 0.2309457425243145, "learning_rate": 6.060143441092292e-09, "loss": 0.4913, "step": 5999 }, { "epoch": 14.669926650366747, "grad_norm": 0.2421342613995642, "learning_rate": 5.971386695260706e-09, "loss": 0.9249, "step": 6000 }, { "epoch": 14.67237163814181, "grad_norm": 0.2290265428212619, "learning_rate": 5.8832839408232144e-09, "loss": 0.3409, "step": 6001 }, { "epoch": 14.67481662591687, "grad_norm": 0.19172885839247258, "learning_rate": 5.795835200882449e-09, "loss": 0.6352, "step": 6002 }, { "epoch": 14.67726161369193, "grad_norm": 0.2390363370720583, "learning_rate": 5.709040498369512e-09, "loss": 0.8822, "step": 6003 }, { "epoch": 14.679706601466993, "grad_norm": 0.21259312329093594, "learning_rate": 5.6228998560436955e-09, "loss": 0.6894, "step": 6004 }, { "epoch": 14.682151589242054, "grad_norm": 0.3373322950250815, "learning_rate": 5.537413296493321e-09, "loss": 0.6568, "step": 6005 }, { "epoch": 14.684596577017114, "grad_norm": 0.24686164502993627, "learning_rate": 5.4525808421349025e-09, "loss": 0.5377, "step": 6006 }, { "epoch": 14.687041564792176, "grad_norm": 0.3449344032626107, "learning_rate": 5.368402515213422e-09, "loss": 0.4257, "step": 6007 }, { "epoch": 14.689486552567237, "grad_norm": 0.2896496601183963, "learning_rate": 5.284878337802057e-09, "loss": 0.3719, "step": 6008 }, { "epoch": 14.6919315403423, "grad_norm": 0.22205103984039853, "learning_rate": 5.202008331803287e-09, "loss": 0.3607, "step": 6009 }, { "epoch": 14.69437652811736, "grad_norm": 0.21567865837402447, "learning_rate": 5.119792518947231e-09, "loss": 0.3643, "step": 6010 }, { "epoch": 14.69682151589242, "grad_norm": 0.2768312071846796, "learning_rate": 5.038230920792752e-09, "loss": 0.7869, "step": 6011 }, { "epoch": 14.69926650366748, "grad_norm": 0.24095109061649272, "learning_rate": 4.95732355872719e-09, "loss": 0.6822, "step": 6012 }, { "epoch": 14.701711491442543, "grad_norm": 0.31400601391504207, "learning_rate": 4.877070453966348e-09, "loss": 0.6928, "step": 6013 }, { "epoch": 14.704156479217604, "grad_norm": 0.2530379183141433, "learning_rate": 4.797471627554507e-09, "loss": 0.3742, "step": 6014 }, { "epoch": 14.706601466992666, "grad_norm": 0.34061381163783283, "learning_rate": 4.7185271003641346e-09, "loss": 0.7075, "step": 6015 }, { "epoch": 14.709046454767726, "grad_norm": 0.3105218567832549, "learning_rate": 4.640236893096173e-09, "loss": 0.4934, "step": 6016 }, { "epoch": 14.711491442542787, "grad_norm": 0.15443151716892953, "learning_rate": 4.562601026280589e-09, "loss": 0.3425, "step": 6017 }, { "epoch": 14.713936430317847, "grad_norm": 0.1744557264533319, "learning_rate": 4.485619520274709e-09, "loss": 0.4387, "step": 6018 }, { "epoch": 14.71638141809291, "grad_norm": 0.24883032921484904, "learning_rate": 4.409292395265441e-09, "loss": 0.6523, "step": 6019 }, { "epoch": 14.71882640586797, "grad_norm": 0.15763629075769425, "learning_rate": 4.333619671267053e-09, "loss": 0.4129, "step": 6020 }, { "epoch": 14.721271393643033, "grad_norm": 0.2573036895064007, "learning_rate": 4.258601368122839e-09, "loss": 0.4781, "step": 6021 }, { "epoch": 14.723716381418093, "grad_norm": 0.2235907691225696, "learning_rate": 4.184237505504008e-09, "loss": 0.5225, "step": 6022 }, { "epoch": 14.726161369193154, "grad_norm": 0.2521904329130875, "learning_rate": 4.110528102911071e-09, "loss": 0.4445, "step": 6023 }, { "epoch": 14.728606356968216, "grad_norm": 0.3550970617349348, "learning_rate": 4.037473179671625e-09, "loss": 0.5881, "step": 6024 }, { "epoch": 14.731051344743276, "grad_norm": 0.262785714811522, "learning_rate": 3.9650727549431224e-09, "loss": 0.4914, "step": 6025 }, { "epoch": 14.733496332518337, "grad_norm": 0.19715560985937872, "learning_rate": 3.893326847710099e-09, "loss": 0.5852, "step": 6026 }, { "epoch": 14.7359413202934, "grad_norm": 0.19421441157057798, "learning_rate": 3.82223547678584e-09, "loss": 0.3984, "step": 6027 }, { "epoch": 14.73838630806846, "grad_norm": 0.2331120945539375, "learning_rate": 3.751798660812378e-09, "loss": 0.4183, "step": 6028 }, { "epoch": 14.74083129584352, "grad_norm": 0.2167657792224364, "learning_rate": 3.682016418260215e-09, "loss": 0.6365, "step": 6029 }, { "epoch": 14.743276283618583, "grad_norm": 0.22012217801081127, "learning_rate": 3.612888767427214e-09, "loss": 0.5316, "step": 6030 }, { "epoch": 14.745721271393643, "grad_norm": 0.33133075114151056, "learning_rate": 3.544415726440542e-09, "loss": 0.5028, "step": 6031 }, { "epoch": 14.748166259168704, "grad_norm": 0.2052462827748625, "learning_rate": 3.476597313255281e-09, "loss": 0.5279, "step": 6032 }, { "epoch": 14.750611246943766, "grad_norm": 0.29823741499673084, "learning_rate": 3.4094335456549834e-09, "loss": 0.5983, "step": 6033 }, { "epoch": 14.753056234718827, "grad_norm": 0.23820070159138937, "learning_rate": 3.3429244412516713e-09, "loss": 0.3798, "step": 6034 }, { "epoch": 14.755501222493887, "grad_norm": 0.29170791712317556, "learning_rate": 3.2770700174852843e-09, "loss": 0.6791, "step": 6035 }, { "epoch": 14.75794621026895, "grad_norm": 0.2825423108756363, "learning_rate": 3.2118702916247856e-09, "loss": 0.7253, "step": 6036 }, { "epoch": 14.76039119804401, "grad_norm": 0.2890878537413693, "learning_rate": 3.1473252807667776e-09, "loss": 0.4895, "step": 6037 }, { "epoch": 14.76283618581907, "grad_norm": 0.29379187125581036, "learning_rate": 3.083435001836332e-09, "loss": 0.3434, "step": 6038 }, { "epoch": 14.765281173594133, "grad_norm": 0.2579358018145425, "learning_rate": 3.0201994715872706e-09, "loss": 0.3923, "step": 6039 }, { "epoch": 14.767726161369193, "grad_norm": 0.2391035060441323, "learning_rate": 2.957618706601051e-09, "loss": 0.4424, "step": 6040 }, { "epoch": 14.770171149144254, "grad_norm": 0.19576696855516323, "learning_rate": 2.8956927232878796e-09, "loss": 0.4877, "step": 6041 }, { "epoch": 14.772616136919316, "grad_norm": 0.240593856185608, "learning_rate": 2.834421537886156e-09, "loss": 0.6814, "step": 6042 }, { "epoch": 14.775061124694377, "grad_norm": 0.2772221945451591, "learning_rate": 2.773805166462751e-09, "loss": 0.8086, "step": 6043 }, { "epoch": 14.777506112469437, "grad_norm": 0.2100573816261956, "learning_rate": 2.713843624912449e-09, "loss": 0.507, "step": 6044 }, { "epoch": 14.7799511002445, "grad_norm": 0.34050871174936265, "learning_rate": 2.6545369289587843e-09, "loss": 0.487, "step": 6045 }, { "epoch": 14.78239608801956, "grad_norm": 0.27647510912626844, "learning_rate": 2.5958850941532053e-09, "loss": 1.0872, "step": 6046 }, { "epoch": 14.78484107579462, "grad_norm": 0.4111356064594576, "learning_rate": 2.537888135875355e-09, "loss": 0.8439, "step": 6047 }, { "epoch": 14.787286063569683, "grad_norm": 0.1906839892483114, "learning_rate": 2.4805460693336226e-09, "loss": 0.4373, "step": 6048 }, { "epoch": 14.789731051344743, "grad_norm": 0.24808603343107538, "learning_rate": 2.423858909564314e-09, "loss": 0.406, "step": 6049 }, { "epoch": 14.792176039119804, "grad_norm": 0.37771878750139837, "learning_rate": 2.3678266714322053e-09, "loss": 0.3198, "step": 6050 }, { "epoch": 14.794621026894866, "grad_norm": 0.2160390329535855, "learning_rate": 2.3124493696302653e-09, "loss": 0.4314, "step": 6051 }, { "epoch": 14.797066014669927, "grad_norm": 0.3051128816311609, "learning_rate": 2.2577270186793786e-09, "loss": 0.4283, "step": 6052 }, { "epoch": 14.799511002444987, "grad_norm": 0.3763204378554294, "learning_rate": 2.2036596329294556e-09, "loss": 0.631, "step": 6053 }, { "epoch": 14.80195599022005, "grad_norm": 0.35005814685659103, "learning_rate": 2.150247226557767e-09, "loss": 0.4814, "step": 6054 }, { "epoch": 14.80440097799511, "grad_norm": 0.2929301930980776, "learning_rate": 2.0974898135703305e-09, "loss": 0.7775, "step": 6055 }, { "epoch": 14.80684596577017, "grad_norm": 0.2049185745490653, "learning_rate": 2.0453874078016355e-09, "loss": 0.4771, "step": 6056 }, { "epoch": 14.809290953545233, "grad_norm": 0.1649797453771252, "learning_rate": 1.993940022914087e-09, "loss": 0.6443, "step": 6057 }, { "epoch": 14.811735941320293, "grad_norm": 0.16624824826202417, "learning_rate": 1.943147672398005e-09, "loss": 0.4902, "step": 6058 }, { "epoch": 14.814180929095354, "grad_norm": 0.28322847142945035, "learning_rate": 1.8930103695727344e-09, "loss": 0.5741, "step": 6059 }, { "epoch": 14.816625916870416, "grad_norm": 0.3099886287964264, "learning_rate": 1.8435281275849814e-09, "loss": 0.4538, "step": 6060 }, { "epoch": 14.819070904645477, "grad_norm": 0.29517200474282757, "learning_rate": 1.794700959410478e-09, "loss": 0.4519, "step": 6061 }, { "epoch": 14.821515892420537, "grad_norm": 0.24404313300535843, "learning_rate": 1.746528877852871e-09, "loss": 0.3422, "step": 6062 }, { "epoch": 14.8239608801956, "grad_norm": 0.22328362712086952, "learning_rate": 1.699011895543723e-09, "loss": 0.7422, "step": 6063 }, { "epoch": 14.82640586797066, "grad_norm": 0.19718418400246815, "learning_rate": 1.6521500249430667e-09, "loss": 0.5134, "step": 6064 }, { "epoch": 14.82885085574572, "grad_norm": 0.1940561726468138, "learning_rate": 1.6059432783394058e-09, "loss": 0.5373, "step": 6065 }, { "epoch": 14.831295843520783, "grad_norm": 0.26077652921015565, "learning_rate": 1.5603916678488817e-09, "loss": 0.8625, "step": 6066 }, { "epoch": 14.833740831295843, "grad_norm": 0.31417874742777985, "learning_rate": 1.5154952054166617e-09, "loss": 0.5852, "step": 6067 }, { "epoch": 14.836185819070904, "grad_norm": 0.3209894253250868, "learning_rate": 1.4712539028149952e-09, "loss": 0.5541, "step": 6068 }, { "epoch": 14.838630806845966, "grad_norm": 0.21047144209334923, "learning_rate": 1.4276677716457131e-09, "loss": 0.3689, "step": 6069 }, { "epoch": 14.841075794621027, "grad_norm": 0.21693428824831038, "learning_rate": 1.3847368233374515e-09, "loss": 0.9993, "step": 6070 }, { "epoch": 14.843520782396087, "grad_norm": 0.4590276494781457, "learning_rate": 1.3424610691478711e-09, "loss": 0.4425, "step": 6071 }, { "epoch": 14.84596577017115, "grad_norm": 0.2578953441546614, "learning_rate": 1.3008405201625495e-09, "loss": 0.4734, "step": 6072 }, { "epoch": 14.84841075794621, "grad_norm": 0.33917636190556705, "learning_rate": 1.2598751872958114e-09, "loss": 0.6688, "step": 6073 }, { "epoch": 14.85085574572127, "grad_norm": 0.27530160112989455, "learning_rate": 1.2195650812890648e-09, "loss": 0.5289, "step": 6074 }, { "epoch": 14.853300733496333, "grad_norm": 0.2154966682765608, "learning_rate": 1.1799102127130203e-09, "loss": 0.417, "step": 6075 }, { "epoch": 14.855745721271393, "grad_norm": 0.30745640063522156, "learning_rate": 1.1409105919660268e-09, "loss": 0.3396, "step": 6076 }, { "epoch": 14.858190709046454, "grad_norm": 0.3223174564683795, "learning_rate": 1.1025662292743488e-09, "loss": 0.4648, "step": 6077 }, { "epoch": 14.860635696821516, "grad_norm": 0.22441185036931252, "learning_rate": 1.0648771346929986e-09, "loss": 0.4567, "step": 6078 }, { "epoch": 14.863080684596577, "grad_norm": 0.35753099593101356, "learning_rate": 1.0278433181051817e-09, "loss": 0.5927, "step": 6079 }, { "epoch": 14.865525672371637, "grad_norm": 0.3359838507410903, "learning_rate": 9.914647892214634e-10, "loss": 0.5264, "step": 6080 }, { "epoch": 14.8679706601467, "grad_norm": 0.20156700931920143, "learning_rate": 9.557415575814355e-10, "loss": 0.6415, "step": 6081 }, { "epoch": 14.87041564792176, "grad_norm": 0.14308754578036692, "learning_rate": 9.206736325526044e-10, "loss": 0.2489, "step": 6082 }, { "epoch": 14.87286063569682, "grad_norm": 0.278511609295102, "learning_rate": 8.862610233303925e-10, "loss": 0.9581, "step": 6083 }, { "epoch": 14.875305623471883, "grad_norm": 0.24077294849432818, "learning_rate": 8.525037389386925e-10, "loss": 0.35, "step": 6084 }, { "epoch": 14.877750611246944, "grad_norm": 0.2193195410966309, "learning_rate": 8.194017882295902e-10, "loss": 0.4227, "step": 6085 }, { "epoch": 14.880195599022004, "grad_norm": 0.1941903427998383, "learning_rate": 7.869551798828091e-10, "loss": 0.6199, "step": 6086 }, { "epoch": 14.882640586797066, "grad_norm": 0.3156006195839875, "learning_rate": 7.551639224068207e-10, "loss": 0.4067, "step": 6087 }, { "epoch": 14.885085574572127, "grad_norm": 0.28527074687708986, "learning_rate": 7.240280241380127e-10, "loss": 0.7832, "step": 6088 }, { "epoch": 14.887530562347187, "grad_norm": 0.3521589981447914, "learning_rate": 6.935474932406872e-10, "loss": 0.4085, "step": 6089 }, { "epoch": 14.88997555012225, "grad_norm": 0.37967704995307855, "learning_rate": 6.63722337707895e-10, "loss": 0.3486, "step": 6090 }, { "epoch": 14.89242053789731, "grad_norm": 0.24573600595054293, "learning_rate": 6.345525653606021e-10, "loss": 0.8806, "step": 6091 }, { "epoch": 14.894865525672373, "grad_norm": 0.27177497279807844, "learning_rate": 6.060381838471352e-10, "loss": 0.6773, "step": 6092 }, { "epoch": 14.897310513447433, "grad_norm": 0.3140673910233011, "learning_rate": 5.781792006451237e-10, "loss": 0.807, "step": 6093 }, { "epoch": 14.899755501222494, "grad_norm": 0.20959808951064923, "learning_rate": 5.509756230598351e-10, "loss": 0.4013, "step": 6094 }, { "epoch": 14.902200488997554, "grad_norm": 0.25672289390834235, "learning_rate": 5.244274582244524e-10, "loss": 0.4963, "step": 6095 }, { "epoch": 14.904645476772616, "grad_norm": 0.17681462007478535, "learning_rate": 4.985347131006291e-10, "loss": 0.447, "step": 6096 }, { "epoch": 14.907090464547677, "grad_norm": 0.2779137446642931, "learning_rate": 4.73297394477934e-10, "loss": 0.6167, "step": 6097 }, { "epoch": 14.90953545232274, "grad_norm": 0.35764417329776804, "learning_rate": 4.487155089741291e-10, "loss": 0.4435, "step": 6098 }, { "epoch": 14.9119804400978, "grad_norm": 0.39071684985120936, "learning_rate": 4.247890630354468e-10, "loss": 0.6944, "step": 6099 }, { "epoch": 14.91442542787286, "grad_norm": 0.24672364232590013, "learning_rate": 4.0151806293575733e-10, "loss": 0.5014, "step": 6100 }, { "epoch": 14.91687041564792, "grad_norm": 0.3305020093427622, "learning_rate": 3.789025147774017e-10, "loss": 0.4721, "step": 6101 }, { "epoch": 14.919315403422983, "grad_norm": 0.21254176497695207, "learning_rate": 3.569424244903585e-10, "loss": 0.5032, "step": 6102 }, { "epoch": 14.921760391198044, "grad_norm": 0.3303295121425075, "learning_rate": 3.3563779783307715e-10, "loss": 0.5623, "step": 6103 }, { "epoch": 14.924205378973106, "grad_norm": 0.25570965246055927, "learning_rate": 3.1498864039275487e-10, "loss": 0.4422, "step": 6104 }, { "epoch": 14.926650366748166, "grad_norm": 0.21943883857846808, "learning_rate": 2.949949575833944e-10, "loss": 0.7702, "step": 6105 }, { "epoch": 14.929095354523227, "grad_norm": 0.35506875983878006, "learning_rate": 2.756567546477462e-10, "loss": 0.6513, "step": 6106 }, { "epoch": 14.93154034229829, "grad_norm": 0.43100986641863454, "learning_rate": 2.569740366573092e-10, "loss": 0.6179, "step": 6107 }, { "epoch": 14.93398533007335, "grad_norm": 0.33852455104193246, "learning_rate": 2.3894680851094254e-10, "loss": 0.4555, "step": 6108 }, { "epoch": 14.93643031784841, "grad_norm": 0.2997287660461397, "learning_rate": 2.2157507493542062e-10, "loss": 0.4912, "step": 6109 }, { "epoch": 14.938875305623473, "grad_norm": 0.36365554388015014, "learning_rate": 2.0485884048626614e-10, "loss": 0.4247, "step": 6110 }, { "epoch": 14.941320293398533, "grad_norm": 0.21761490860903127, "learning_rate": 1.8879810954691714e-10, "loss": 0.726, "step": 6111 }, { "epoch": 14.943765281173594, "grad_norm": 0.26260422366110636, "learning_rate": 1.733928863287271e-10, "loss": 0.6425, "step": 6112 }, { "epoch": 14.946210268948656, "grad_norm": 0.33083469470162913, "learning_rate": 1.5864317487152004e-10, "loss": 0.464, "step": 6113 }, { "epoch": 14.948655256723717, "grad_norm": 0.39038244878432726, "learning_rate": 1.445489790427579e-10, "loss": 0.4409, "step": 6114 }, { "epoch": 14.951100244498777, "grad_norm": 0.29973846111304453, "learning_rate": 1.311103025383731e-10, "loss": 0.3724, "step": 6115 }, { "epoch": 14.95354523227384, "grad_norm": 0.13253995018446735, "learning_rate": 1.1832714888221353e-10, "loss": 0.547, "step": 6116 }, { "epoch": 14.9559902200489, "grad_norm": 0.34153815086086503, "learning_rate": 1.0619952142659761e-10, "loss": 0.4168, "step": 6117 }, { "epoch": 14.95843520782396, "grad_norm": 0.20630446108208855, "learning_rate": 9.472742335120411e-11, "loss": 0.5538, "step": 6118 }, { "epoch": 14.960880195599023, "grad_norm": 0.28672511549056245, "learning_rate": 8.391085766473739e-11, "loss": 0.3125, "step": 6119 }, { "epoch": 14.963325183374083, "grad_norm": 0.6869203370630521, "learning_rate": 7.374982720326218e-11, "loss": 0.6003, "step": 6120 }, { "epoch": 14.965770171149144, "grad_norm": 0.362939994202542, "learning_rate": 6.424433463131374e-11, "loss": 0.4988, "step": 6121 }, { "epoch": 14.968215158924206, "grad_norm": 0.4673443001120997, "learning_rate": 5.5394382441342764e-11, "loss": 0.4976, "step": 6122 }, { "epoch": 14.970660146699267, "grad_norm": 0.24007421082167638, "learning_rate": 4.7199972954270476e-11, "loss": 0.8173, "step": 6123 }, { "epoch": 14.973105134474327, "grad_norm": 0.29326816384038545, "learning_rate": 3.966110831865599e-11, "loss": 0.429, "step": 6124 }, { "epoch": 14.97555012224939, "grad_norm": 0.2177326378161841, "learning_rate": 3.2777790511251404e-11, "loss": 0.7403, "step": 6125 }, { "epoch": 14.97799511002445, "grad_norm": 0.27485976870018675, "learning_rate": 2.6550021337279354e-11, "loss": 0.5434, "step": 6126 }, { "epoch": 14.98044009779951, "grad_norm": 0.21829330167434846, "learning_rate": 2.0977802429877902e-11, "loss": 0.5768, "step": 6127 }, { "epoch": 14.982885085574573, "grad_norm": 0.17184937449556248, "learning_rate": 1.6061135249823e-11, "loss": 0.5274, "step": 6128 }, { "epoch": 14.985330073349633, "grad_norm": 0.2751475564433603, "learning_rate": 1.180002108663869e-11, "loss": 0.6488, "step": 6129 }, { "epoch": 14.987775061124694, "grad_norm": 0.2576255837945489, "learning_rate": 8.194461057764447e-12, "loss": 0.846, "step": 6130 }, { "epoch": 14.990220048899756, "grad_norm": 0.273057326128053, "learning_rate": 5.244456108277618e-12, "loss": 0.5486, "step": 6131 }, { "epoch": 14.992665036674817, "grad_norm": 0.45194041858522627, "learning_rate": 2.9500070122812085e-12, "loss": 0.3828, "step": 6132 }, { "epoch": 14.995110024449877, "grad_norm": 0.3662560694585631, "learning_rate": 1.3111143709609863e-12, "loss": 0.6831, "step": 6133 }, { "epoch": 14.99755501222494, "grad_norm": 0.19005317783635392, "learning_rate": 3.277786142508177e-13, "loss": 0.5808, "step": 6134 }, { "epoch": 15.0, "grad_norm": 0.2616668832139272, "learning_rate": 0.0, "loss": 0.4434, "step": 6135 } ], "logging_steps": 1, "max_steps": 6135, "num_input_tokens_seen": 0, "num_train_epochs": 15, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 9138056050737152.0, "train_batch_size": 1, "trial_name": null, "trial_params": null }