{ "best_global_step": 5900, "best_metric": 1.0105234384536743, "best_model_checkpoint": "./outputs/checkpoint-5900", "epoch": 2.176326530612245, "eval_steps": 100, "global_step": 6000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0007256235827664399, "grad_norm": 0.88671875, "learning_rate": 2.0000000000000003e-06, "loss": 1.5116, "step": 2 }, { "epoch": 0.0014512471655328798, "grad_norm": 1.8046875, "learning_rate": 6e-06, "loss": 1.7699, "step": 4 }, { "epoch": 0.0021768707482993197, "grad_norm": 0.7890625, "learning_rate": 1e-05, "loss": 1.4788, "step": 6 }, { "epoch": 0.0029024943310657597, "grad_norm": 1.3671875, "learning_rate": 1.4000000000000001e-05, "loss": 1.6688, "step": 8 }, { "epoch": 0.0036281179138321997, "grad_norm": 1.7421875, "learning_rate": 1.8e-05, "loss": 1.798, "step": 10 }, { "epoch": 0.004353741496598639, "grad_norm": 1.171875, "learning_rate": 2.2000000000000003e-05, "loss": 1.6686, "step": 12 }, { "epoch": 0.005079365079365079, "grad_norm": 1.1484375, "learning_rate": 2.6000000000000002e-05, "loss": 1.8021, "step": 14 }, { "epoch": 0.005804988662131519, "grad_norm": 0.7890625, "learning_rate": 3e-05, "loss": 1.6768, "step": 16 }, { "epoch": 0.006530612244897959, "grad_norm": 0.83203125, "learning_rate": 3.4000000000000007e-05, "loss": 1.4667, "step": 18 }, { "epoch": 0.0072562358276643995, "grad_norm": 0.98828125, "learning_rate": 3.8e-05, "loss": 1.6074, "step": 20 }, { "epoch": 0.00798185941043084, "grad_norm": 0.90234375, "learning_rate": 4.2e-05, "loss": 1.6353, "step": 22 }, { "epoch": 0.008707482993197279, "grad_norm": 0.92578125, "learning_rate": 4.600000000000001e-05, "loss": 1.5452, "step": 24 }, { "epoch": 0.00943310657596372, "grad_norm": 0.87109375, "learning_rate": 5e-05, "loss": 1.6205, "step": 26 }, { "epoch": 0.010158730158730159, "grad_norm": 0.61328125, "learning_rate": 5.4000000000000005e-05, "loss": 1.4568, "step": 28 }, { "epoch": 0.010884353741496598, "grad_norm": 0.69921875, "learning_rate": 5.8e-05, "loss": 1.4901, "step": 30 }, { "epoch": 0.011609977324263039, "grad_norm": 0.98046875, "learning_rate": 6.2e-05, "loss": 1.422, "step": 32 }, { "epoch": 0.012335600907029478, "grad_norm": 0.984375, "learning_rate": 6.6e-05, "loss": 1.406, "step": 34 }, { "epoch": 0.013061224489795919, "grad_norm": 0.69921875, "learning_rate": 7e-05, "loss": 1.2822, "step": 36 }, { "epoch": 0.013786848072562358, "grad_norm": 0.86328125, "learning_rate": 7.4e-05, "loss": 1.3677, "step": 38 }, { "epoch": 0.014512471655328799, "grad_norm": 0.78125, "learning_rate": 7.800000000000001e-05, "loss": 1.3597, "step": 40 }, { "epoch": 0.015238095238095238, "grad_norm": 0.828125, "learning_rate": 8.2e-05, "loss": 1.263, "step": 42 }, { "epoch": 0.01596371882086168, "grad_norm": 0.75390625, "learning_rate": 8.6e-05, "loss": 1.314, "step": 44 }, { "epoch": 0.016689342403628116, "grad_norm": 0.46484375, "learning_rate": 9e-05, "loss": 1.2784, "step": 46 }, { "epoch": 0.017414965986394557, "grad_norm": 0.65625, "learning_rate": 9.4e-05, "loss": 1.1945, "step": 48 }, { "epoch": 0.018140589569160998, "grad_norm": 0.5703125, "learning_rate": 9.8e-05, "loss": 1.2637, "step": 50 }, { "epoch": 0.01886621315192744, "grad_norm": 0.625, "learning_rate": 0.00010200000000000001, "loss": 1.1792, "step": 52 }, { "epoch": 0.019591836734693877, "grad_norm": 0.349609375, "learning_rate": 0.00010600000000000002, "loss": 1.1901, "step": 54 }, { "epoch": 0.020317460317460317, "grad_norm": 0.392578125, "learning_rate": 0.00011000000000000002, "loss": 1.1808, "step": 56 }, { "epoch": 0.02104308390022676, "grad_norm": 0.34765625, "learning_rate": 0.00011399999999999999, "loss": 1.134, "step": 58 }, { "epoch": 0.021768707482993196, "grad_norm": 0.326171875, "learning_rate": 0.000118, "loss": 1.1071, "step": 60 }, { "epoch": 0.022494331065759637, "grad_norm": 0.255859375, "learning_rate": 0.000122, "loss": 1.1066, "step": 62 }, { "epoch": 0.023219954648526078, "grad_norm": 0.30859375, "learning_rate": 0.000126, "loss": 1.1922, "step": 64 }, { "epoch": 0.02394557823129252, "grad_norm": 0.2578125, "learning_rate": 0.00013000000000000002, "loss": 1.1028, "step": 66 }, { "epoch": 0.024671201814058956, "grad_norm": 0.369140625, "learning_rate": 0.000134, "loss": 1.0819, "step": 68 }, { "epoch": 0.025396825396825397, "grad_norm": 0.275390625, "learning_rate": 0.000138, "loss": 1.1744, "step": 70 }, { "epoch": 0.026122448979591838, "grad_norm": 0.26171875, "learning_rate": 0.000142, "loss": 1.0784, "step": 72 }, { "epoch": 0.026848072562358275, "grad_norm": 0.21875, "learning_rate": 0.000146, "loss": 1.0396, "step": 74 }, { "epoch": 0.027573696145124716, "grad_norm": 0.2333984375, "learning_rate": 0.00015000000000000001, "loss": 1.2014, "step": 76 }, { "epoch": 0.028299319727891157, "grad_norm": 0.23828125, "learning_rate": 0.000154, "loss": 1.1024, "step": 78 }, { "epoch": 0.029024943310657598, "grad_norm": 0.28125, "learning_rate": 0.00015800000000000002, "loss": 0.9752, "step": 80 }, { "epoch": 0.029750566893424035, "grad_norm": 0.2373046875, "learning_rate": 0.000162, "loss": 1.0902, "step": 82 }, { "epoch": 0.030476190476190476, "grad_norm": 0.19921875, "learning_rate": 0.000166, "loss": 1.0058, "step": 84 }, { "epoch": 0.031201814058956917, "grad_norm": 0.2197265625, "learning_rate": 0.00017, "loss": 1.0682, "step": 86 }, { "epoch": 0.03192743764172336, "grad_norm": 0.2119140625, "learning_rate": 0.000174, "loss": 1.0874, "step": 88 }, { "epoch": 0.0326530612244898, "grad_norm": 0.458984375, "learning_rate": 0.00017800000000000002, "loss": 1.0739, "step": 90 }, { "epoch": 0.03337868480725623, "grad_norm": 0.212890625, "learning_rate": 0.000182, "loss": 1.0259, "step": 92 }, { "epoch": 0.034104308390022674, "grad_norm": 0.2470703125, "learning_rate": 0.00018600000000000002, "loss": 1.0279, "step": 94 }, { "epoch": 0.034829931972789115, "grad_norm": 0.2001953125, "learning_rate": 0.00019, "loss": 1.0263, "step": 96 }, { "epoch": 0.035555555555555556, "grad_norm": 0.251953125, "learning_rate": 0.000194, "loss": 1.0206, "step": 98 }, { "epoch": 0.036281179138321996, "grad_norm": 0.2353515625, "learning_rate": 0.00019800000000000002, "loss": 1.1275, "step": 100 }, { "epoch": 0.036281179138321996, "eval_loss": 1.0672887563705444, "eval_runtime": 99.6091, "eval_samples_per_second": 18.071, "eval_steps_per_second": 0.572, "step": 100 }, { "epoch": 0.03700680272108844, "grad_norm": 0.267578125, "learning_rate": 0.0001999755231917758, "loss": 1.1141, "step": 102 }, { "epoch": 0.03773242630385488, "grad_norm": 0.2080078125, "learning_rate": 0.00019992656957532737, "loss": 1.0543, "step": 104 }, { "epoch": 0.03845804988662131, "grad_norm": 0.20703125, "learning_rate": 0.00019987761595887898, "loss": 1.0389, "step": 106 }, { "epoch": 0.03918367346938775, "grad_norm": 0.1982421875, "learning_rate": 0.00019982866234243056, "loss": 1.0699, "step": 108 }, { "epoch": 0.039909297052154194, "grad_norm": 0.20703125, "learning_rate": 0.00019977970872598214, "loss": 1.0756, "step": 110 }, { "epoch": 0.040634920634920635, "grad_norm": 0.1845703125, "learning_rate": 0.00019973075510953373, "loss": 1.0309, "step": 112 }, { "epoch": 0.041360544217687076, "grad_norm": 0.1884765625, "learning_rate": 0.0001996818014930853, "loss": 0.9821, "step": 114 }, { "epoch": 0.04208616780045352, "grad_norm": 0.2353515625, "learning_rate": 0.0001996328478766369, "loss": 1.0897, "step": 116 }, { "epoch": 0.04281179138321996, "grad_norm": 0.2080078125, "learning_rate": 0.0001995838942601885, "loss": 1.0576, "step": 118 }, { "epoch": 0.04353741496598639, "grad_norm": 0.2001953125, "learning_rate": 0.00019953494064374008, "loss": 1.0482, "step": 120 }, { "epoch": 0.04426303854875283, "grad_norm": 0.259765625, "learning_rate": 0.00019948598702729164, "loss": 1.0018, "step": 122 }, { "epoch": 0.04498866213151927, "grad_norm": 0.1865234375, "learning_rate": 0.00019943703341084322, "loss": 1.0637, "step": 124 }, { "epoch": 0.045714285714285714, "grad_norm": 0.1806640625, "learning_rate": 0.00019938807979439483, "loss": 1.0471, "step": 126 }, { "epoch": 0.046439909297052155, "grad_norm": 0.1875, "learning_rate": 0.0001993391261779464, "loss": 0.9958, "step": 128 }, { "epoch": 0.047165532879818596, "grad_norm": 0.201171875, "learning_rate": 0.000199290172561498, "loss": 1.0396, "step": 130 }, { "epoch": 0.04789115646258504, "grad_norm": 0.2333984375, "learning_rate": 0.00019924121894504958, "loss": 1.0407, "step": 132 }, { "epoch": 0.04861678004535147, "grad_norm": 0.201171875, "learning_rate": 0.00019919226532860116, "loss": 1.043, "step": 134 }, { "epoch": 0.04934240362811791, "grad_norm": 0.193359375, "learning_rate": 0.00019914331171215274, "loss": 1.0859, "step": 136 }, { "epoch": 0.05006802721088435, "grad_norm": 0.181640625, "learning_rate": 0.00019909435809570435, "loss": 1.1251, "step": 138 }, { "epoch": 0.050793650793650794, "grad_norm": 0.2431640625, "learning_rate": 0.0001990454044792559, "loss": 1.0666, "step": 140 }, { "epoch": 0.051519274376417235, "grad_norm": 0.203125, "learning_rate": 0.0001989964508628075, "loss": 0.9962, "step": 142 }, { "epoch": 0.052244897959183675, "grad_norm": 0.1923828125, "learning_rate": 0.00019894749724635907, "loss": 1.1262, "step": 144 }, { "epoch": 0.052970521541950116, "grad_norm": 0.1875, "learning_rate": 0.00019889854362991068, "loss": 1.0174, "step": 146 }, { "epoch": 0.05369614512471655, "grad_norm": 0.181640625, "learning_rate": 0.00019884959001346226, "loss": 1.0509, "step": 148 }, { "epoch": 0.05442176870748299, "grad_norm": 0.21875, "learning_rate": 0.00019880063639701382, "loss": 1.05, "step": 150 }, { "epoch": 0.05514739229024943, "grad_norm": 0.2109375, "learning_rate": 0.00019875168278056543, "loss": 1.134, "step": 152 }, { "epoch": 0.05587301587301587, "grad_norm": 0.1650390625, "learning_rate": 0.000198702729164117, "loss": 0.9511, "step": 154 }, { "epoch": 0.056598639455782314, "grad_norm": 0.1748046875, "learning_rate": 0.0001986537755476686, "loss": 0.9989, "step": 156 }, { "epoch": 0.057324263038548755, "grad_norm": 0.162109375, "learning_rate": 0.0001986048219312202, "loss": 0.9466, "step": 158 }, { "epoch": 0.058049886621315196, "grad_norm": 0.162109375, "learning_rate": 0.00019855586831477175, "loss": 0.9605, "step": 160 }, { "epoch": 0.05877551020408163, "grad_norm": 0.2001953125, "learning_rate": 0.00019850691469832334, "loss": 1.0511, "step": 162 }, { "epoch": 0.05950113378684807, "grad_norm": 0.205078125, "learning_rate": 0.00019845796108187492, "loss": 1.0575, "step": 164 }, { "epoch": 0.06022675736961451, "grad_norm": 0.189453125, "learning_rate": 0.00019840900746542653, "loss": 1.0211, "step": 166 }, { "epoch": 0.06095238095238095, "grad_norm": 0.1923828125, "learning_rate": 0.0001983600538489781, "loss": 1.0435, "step": 168 }, { "epoch": 0.06167800453514739, "grad_norm": 0.2021484375, "learning_rate": 0.00019831110023252967, "loss": 0.9651, "step": 170 }, { "epoch": 0.062403628117913834, "grad_norm": 0.17578125, "learning_rate": 0.00019826214661608128, "loss": 0.9683, "step": 172 }, { "epoch": 0.06312925170068028, "grad_norm": 0.1796875, "learning_rate": 0.00019821319299963286, "loss": 1.0064, "step": 174 }, { "epoch": 0.06385487528344672, "grad_norm": 0.1875, "learning_rate": 0.00019816423938318444, "loss": 1.0669, "step": 176 }, { "epoch": 0.06458049886621316, "grad_norm": 0.1796875, "learning_rate": 0.00019811528576673605, "loss": 1.0931, "step": 178 }, { "epoch": 0.0653061224489796, "grad_norm": 0.19921875, "learning_rate": 0.0001980663321502876, "loss": 1.04, "step": 180 }, { "epoch": 0.06603174603174604, "grad_norm": 0.1884765625, "learning_rate": 0.0001980173785338392, "loss": 0.9917, "step": 182 }, { "epoch": 0.06675736961451247, "grad_norm": 0.166015625, "learning_rate": 0.00019796842491739077, "loss": 0.9985, "step": 184 }, { "epoch": 0.0674829931972789, "grad_norm": 0.1728515625, "learning_rate": 0.00019791947130094238, "loss": 0.9511, "step": 186 }, { "epoch": 0.06820861678004535, "grad_norm": 0.169921875, "learning_rate": 0.00019787051768449396, "loss": 0.9866, "step": 188 }, { "epoch": 0.06893424036281179, "grad_norm": 0.1982421875, "learning_rate": 0.00019782156406804552, "loss": 1.0879, "step": 190 }, { "epoch": 0.06965986394557823, "grad_norm": 0.181640625, "learning_rate": 0.00019777261045159713, "loss": 1.006, "step": 192 }, { "epoch": 0.07038548752834467, "grad_norm": 0.1787109375, "learning_rate": 0.0001977236568351487, "loss": 1.0737, "step": 194 }, { "epoch": 0.07111111111111111, "grad_norm": 0.203125, "learning_rate": 0.0001976747032187003, "loss": 0.9719, "step": 196 }, { "epoch": 0.07183673469387755, "grad_norm": 0.21484375, "learning_rate": 0.0001976257496022519, "loss": 1.0539, "step": 198 }, { "epoch": 0.07256235827664399, "grad_norm": 0.19921875, "learning_rate": 0.00019757679598580345, "loss": 1.0863, "step": 200 }, { "epoch": 0.07256235827664399, "eval_loss": 1.0431190729141235, "eval_runtime": 99.2841, "eval_samples_per_second": 18.13, "eval_steps_per_second": 0.574, "step": 200 }, { "epoch": 0.07328798185941043, "grad_norm": 0.2021484375, "learning_rate": 0.00019752784236935504, "loss": 1.1022, "step": 202 }, { "epoch": 0.07401360544217687, "grad_norm": 0.177734375, "learning_rate": 0.00019747888875290662, "loss": 0.9781, "step": 204 }, { "epoch": 0.07473922902494332, "grad_norm": 0.21875, "learning_rate": 0.00019742993513645823, "loss": 1.0735, "step": 206 }, { "epoch": 0.07546485260770976, "grad_norm": 0.1865234375, "learning_rate": 0.0001973809815200098, "loss": 0.957, "step": 208 }, { "epoch": 0.0761904761904762, "grad_norm": 0.158203125, "learning_rate": 0.00019733202790356137, "loss": 1.0229, "step": 210 }, { "epoch": 0.07691609977324262, "grad_norm": 0.1962890625, "learning_rate": 0.00019728307428711298, "loss": 0.9876, "step": 212 }, { "epoch": 0.07764172335600907, "grad_norm": 0.1884765625, "learning_rate": 0.00019723412067066456, "loss": 1.0047, "step": 214 }, { "epoch": 0.0783673469387755, "grad_norm": 0.1875, "learning_rate": 0.00019718516705421614, "loss": 0.9995, "step": 216 }, { "epoch": 0.07909297052154195, "grad_norm": 0.1923828125, "learning_rate": 0.00019713621343776772, "loss": 1.0447, "step": 218 }, { "epoch": 0.07981859410430839, "grad_norm": 0.193359375, "learning_rate": 0.0001970872598213193, "loss": 1.0969, "step": 220 }, { "epoch": 0.08054421768707483, "grad_norm": 0.203125, "learning_rate": 0.0001970383062048709, "loss": 0.9948, "step": 222 }, { "epoch": 0.08126984126984127, "grad_norm": 0.1591796875, "learning_rate": 0.00019698935258842247, "loss": 0.9642, "step": 224 }, { "epoch": 0.08199546485260771, "grad_norm": 0.19140625, "learning_rate": 0.00019694039897197408, "loss": 1.0414, "step": 226 }, { "epoch": 0.08272108843537415, "grad_norm": 0.1943359375, "learning_rate": 0.00019689144535552563, "loss": 0.9293, "step": 228 }, { "epoch": 0.08344671201814059, "grad_norm": 0.2021484375, "learning_rate": 0.00019684249173907722, "loss": 0.9743, "step": 230 }, { "epoch": 0.08417233560090703, "grad_norm": 0.189453125, "learning_rate": 0.00019679353812262883, "loss": 1.0206, "step": 232 }, { "epoch": 0.08489795918367347, "grad_norm": 0.1787109375, "learning_rate": 0.0001967445845061804, "loss": 0.9956, "step": 234 }, { "epoch": 0.08562358276643992, "grad_norm": 0.185546875, "learning_rate": 0.000196695630889732, "loss": 1.0204, "step": 236 }, { "epoch": 0.08634920634920636, "grad_norm": 0.177734375, "learning_rate": 0.00019664667727328357, "loss": 0.9644, "step": 238 }, { "epoch": 0.08707482993197278, "grad_norm": 0.1748046875, "learning_rate": 0.00019659772365683515, "loss": 0.9574, "step": 240 }, { "epoch": 0.08780045351473922, "grad_norm": 0.181640625, "learning_rate": 0.00019654877004038674, "loss": 1.0699, "step": 242 }, { "epoch": 0.08852607709750566, "grad_norm": 0.193359375, "learning_rate": 0.00019649981642393832, "loss": 1.0396, "step": 244 }, { "epoch": 0.0892517006802721, "grad_norm": 0.1611328125, "learning_rate": 0.00019645086280748993, "loss": 0.9685, "step": 246 }, { "epoch": 0.08997732426303855, "grad_norm": 0.162109375, "learning_rate": 0.00019640190919104148, "loss": 0.9963, "step": 248 }, { "epoch": 0.09070294784580499, "grad_norm": 0.181640625, "learning_rate": 0.00019635295557459307, "loss": 1.0085, "step": 250 }, { "epoch": 0.09142857142857143, "grad_norm": 0.216796875, "learning_rate": 0.00019630400195814468, "loss": 1.0444, "step": 252 }, { "epoch": 0.09215419501133787, "grad_norm": 0.1591796875, "learning_rate": 0.00019625504834169626, "loss": 0.9959, "step": 254 }, { "epoch": 0.09287981859410431, "grad_norm": 0.189453125, "learning_rate": 0.00019620609472524784, "loss": 1.0705, "step": 256 }, { "epoch": 0.09360544217687075, "grad_norm": 0.1796875, "learning_rate": 0.00019615714110879942, "loss": 1.0762, "step": 258 }, { "epoch": 0.09433106575963719, "grad_norm": 0.1826171875, "learning_rate": 0.000196108187492351, "loss": 1.0227, "step": 260 }, { "epoch": 0.09505668934240363, "grad_norm": 0.2080078125, "learning_rate": 0.00019605923387590259, "loss": 1.0503, "step": 262 }, { "epoch": 0.09578231292517007, "grad_norm": 0.1796875, "learning_rate": 0.00019601028025945417, "loss": 1.0013, "step": 264 }, { "epoch": 0.09650793650793651, "grad_norm": 0.1689453125, "learning_rate": 0.00019596132664300578, "loss": 1.0844, "step": 266 }, { "epoch": 0.09723356009070294, "grad_norm": 0.1767578125, "learning_rate": 0.00019591237302655733, "loss": 0.9824, "step": 268 }, { "epoch": 0.09795918367346938, "grad_norm": 0.1533203125, "learning_rate": 0.00019586341941010892, "loss": 0.9688, "step": 270 }, { "epoch": 0.09868480725623582, "grad_norm": 0.169921875, "learning_rate": 0.00019581446579366052, "loss": 0.923, "step": 272 }, { "epoch": 0.09941043083900226, "grad_norm": 0.2197265625, "learning_rate": 0.0001957655121772121, "loss": 0.9817, "step": 274 }, { "epoch": 0.1001360544217687, "grad_norm": 0.1728515625, "learning_rate": 0.0001957165585607637, "loss": 0.9691, "step": 276 }, { "epoch": 0.10086167800453515, "grad_norm": 0.1806640625, "learning_rate": 0.00019566760494431527, "loss": 1.019, "step": 278 }, { "epoch": 0.10158730158730159, "grad_norm": 0.1767578125, "learning_rate": 0.00019561865132786685, "loss": 1.008, "step": 280 }, { "epoch": 0.10231292517006803, "grad_norm": 0.162109375, "learning_rate": 0.00019556969771141844, "loss": 0.9477, "step": 282 }, { "epoch": 0.10303854875283447, "grad_norm": 0.1748046875, "learning_rate": 0.00019552074409497002, "loss": 0.962, "step": 284 }, { "epoch": 0.10376417233560091, "grad_norm": 0.169921875, "learning_rate": 0.0001954717904785216, "loss": 0.9685, "step": 286 }, { "epoch": 0.10448979591836735, "grad_norm": 0.1650390625, "learning_rate": 0.00019542283686207318, "loss": 1.015, "step": 288 }, { "epoch": 0.10521541950113379, "grad_norm": 0.166015625, "learning_rate": 0.00019537388324562477, "loss": 0.9902, "step": 290 }, { "epoch": 0.10594104308390023, "grad_norm": 0.16796875, "learning_rate": 0.00019532492962917637, "loss": 1.0309, "step": 292 }, { "epoch": 0.10666666666666667, "grad_norm": 0.1953125, "learning_rate": 0.00019527597601272796, "loss": 1.0179, "step": 294 }, { "epoch": 0.1073922902494331, "grad_norm": 0.16796875, "learning_rate": 0.0001952270223962795, "loss": 1.0208, "step": 296 }, { "epoch": 0.10811791383219954, "grad_norm": 0.16796875, "learning_rate": 0.00019517806877983112, "loss": 1.0252, "step": 298 }, { "epoch": 0.10884353741496598, "grad_norm": 0.171875, "learning_rate": 0.0001951291151633827, "loss": 1.0185, "step": 300 }, { "epoch": 0.10884353741496598, "eval_loss": 1.036349892616272, "eval_runtime": 91.1673, "eval_samples_per_second": 19.744, "eval_steps_per_second": 0.625, "step": 300 }, { "epoch": 0.10956916099773242, "grad_norm": 0.177734375, "learning_rate": 0.00019508016154693429, "loss": 0.9886, "step": 302 }, { "epoch": 0.11029478458049886, "grad_norm": 0.1689453125, "learning_rate": 0.00019503120793048587, "loss": 1.0495, "step": 304 }, { "epoch": 0.1110204081632653, "grad_norm": 0.1533203125, "learning_rate": 0.00019498225431403745, "loss": 1.0057, "step": 306 }, { "epoch": 0.11174603174603175, "grad_norm": 0.1728515625, "learning_rate": 0.00019493330069758903, "loss": 1.042, "step": 308 }, { "epoch": 0.11247165532879819, "grad_norm": 0.1767578125, "learning_rate": 0.00019488434708114062, "loss": 1.0619, "step": 310 }, { "epoch": 0.11319727891156463, "grad_norm": 0.150390625, "learning_rate": 0.00019483539346469222, "loss": 1.0055, "step": 312 }, { "epoch": 0.11392290249433107, "grad_norm": 0.1826171875, "learning_rate": 0.0001947864398482438, "loss": 0.9609, "step": 314 }, { "epoch": 0.11464852607709751, "grad_norm": 0.1689453125, "learning_rate": 0.00019473748623179536, "loss": 1.0478, "step": 316 }, { "epoch": 0.11537414965986395, "grad_norm": 0.1826171875, "learning_rate": 0.00019468853261534697, "loss": 1.0008, "step": 318 }, { "epoch": 0.11609977324263039, "grad_norm": 0.15625, "learning_rate": 0.00019463957899889855, "loss": 0.9464, "step": 320 }, { "epoch": 0.11682539682539683, "grad_norm": 0.16015625, "learning_rate": 0.00019459062538245014, "loss": 1.0368, "step": 322 }, { "epoch": 0.11755102040816326, "grad_norm": 0.171875, "learning_rate": 0.00019454167176600172, "loss": 1.0134, "step": 324 }, { "epoch": 0.1182766439909297, "grad_norm": 0.15625, "learning_rate": 0.0001944927181495533, "loss": 0.9533, "step": 326 }, { "epoch": 0.11900226757369614, "grad_norm": 0.1689453125, "learning_rate": 0.00019444376453310488, "loss": 0.9526, "step": 328 }, { "epoch": 0.11972789115646258, "grad_norm": 0.18359375, "learning_rate": 0.00019439481091665647, "loss": 0.9875, "step": 330 }, { "epoch": 0.12045351473922902, "grad_norm": 0.1787109375, "learning_rate": 0.00019434585730020807, "loss": 1.0052, "step": 332 }, { "epoch": 0.12117913832199546, "grad_norm": 0.1943359375, "learning_rate": 0.00019429690368375966, "loss": 1.0672, "step": 334 }, { "epoch": 0.1219047619047619, "grad_norm": 0.171875, "learning_rate": 0.0001942479500673112, "loss": 0.9938, "step": 336 }, { "epoch": 0.12263038548752835, "grad_norm": 0.1669921875, "learning_rate": 0.00019419899645086282, "loss": 0.9938, "step": 338 }, { "epoch": 0.12335600907029479, "grad_norm": 0.1748046875, "learning_rate": 0.0001941500428344144, "loss": 0.9733, "step": 340 }, { "epoch": 0.12408163265306123, "grad_norm": 0.166015625, "learning_rate": 0.00019410108921796599, "loss": 0.9917, "step": 342 }, { "epoch": 0.12480725623582767, "grad_norm": 0.2041015625, "learning_rate": 0.00019405213560151757, "loss": 0.9594, "step": 344 }, { "epoch": 0.1255328798185941, "grad_norm": 0.177734375, "learning_rate": 0.00019400318198506915, "loss": 1.0255, "step": 346 }, { "epoch": 0.12625850340136055, "grad_norm": 0.2294921875, "learning_rate": 0.00019395422836862073, "loss": 1.0643, "step": 348 }, { "epoch": 0.12698412698412698, "grad_norm": 0.1669921875, "learning_rate": 0.00019390527475217232, "loss": 0.9907, "step": 350 }, { "epoch": 0.12770975056689343, "grad_norm": 0.1640625, "learning_rate": 0.00019385632113572392, "loss": 1.0455, "step": 352 }, { "epoch": 0.12843537414965986, "grad_norm": 0.1611328125, "learning_rate": 0.0001938073675192755, "loss": 0.9612, "step": 354 }, { "epoch": 0.1291609977324263, "grad_norm": 0.146484375, "learning_rate": 0.00019375841390282706, "loss": 0.9235, "step": 356 }, { "epoch": 0.12988662131519274, "grad_norm": 0.193359375, "learning_rate": 0.00019370946028637867, "loss": 1.009, "step": 358 }, { "epoch": 0.1306122448979592, "grad_norm": 0.1728515625, "learning_rate": 0.00019366050666993025, "loss": 1.0014, "step": 360 }, { "epoch": 0.13133786848072562, "grad_norm": 0.173828125, "learning_rate": 0.00019361155305348184, "loss": 0.9666, "step": 362 }, { "epoch": 0.13206349206349208, "grad_norm": 0.1728515625, "learning_rate": 0.00019356259943703342, "loss": 1.0091, "step": 364 }, { "epoch": 0.1327891156462585, "grad_norm": 0.166015625, "learning_rate": 0.000193513645820585, "loss": 1.0238, "step": 366 }, { "epoch": 0.13351473922902493, "grad_norm": 0.1611328125, "learning_rate": 0.00019346469220413658, "loss": 1.0511, "step": 368 }, { "epoch": 0.13424036281179139, "grad_norm": 0.1513671875, "learning_rate": 0.00019341573858768817, "loss": 0.986, "step": 370 }, { "epoch": 0.1349659863945578, "grad_norm": 0.16796875, "learning_rate": 0.00019336678497123977, "loss": 1.0094, "step": 372 }, { "epoch": 0.13569160997732427, "grad_norm": 0.15234375, "learning_rate": 0.00019331783135479133, "loss": 1.0596, "step": 374 }, { "epoch": 0.1364172335600907, "grad_norm": 0.171875, "learning_rate": 0.0001932688777383429, "loss": 0.9845, "step": 376 }, { "epoch": 0.13714285714285715, "grad_norm": 0.158203125, "learning_rate": 0.00019321992412189452, "loss": 0.9855, "step": 378 }, { "epoch": 0.13786848072562358, "grad_norm": 0.1806640625, "learning_rate": 0.0001931709705054461, "loss": 1.039, "step": 380 }, { "epoch": 0.13859410430839003, "grad_norm": 0.1728515625, "learning_rate": 0.00019312201688899769, "loss": 1.0108, "step": 382 }, { "epoch": 0.13931972789115646, "grad_norm": 0.18359375, "learning_rate": 0.00019307306327254927, "loss": 0.977, "step": 384 }, { "epoch": 0.1400453514739229, "grad_norm": 0.17578125, "learning_rate": 0.00019302410965610085, "loss": 0.9754, "step": 386 }, { "epoch": 0.14077097505668934, "grad_norm": 0.1728515625, "learning_rate": 0.00019297515603965243, "loss": 1.1274, "step": 388 }, { "epoch": 0.1414965986394558, "grad_norm": 0.1689453125, "learning_rate": 0.00019292620242320401, "loss": 1.0126, "step": 390 }, { "epoch": 0.14222222222222222, "grad_norm": 0.1572265625, "learning_rate": 0.00019287724880675562, "loss": 1.0009, "step": 392 }, { "epoch": 0.14294784580498865, "grad_norm": 0.1806640625, "learning_rate": 0.00019282829519030718, "loss": 0.9729, "step": 394 }, { "epoch": 0.1436734693877551, "grad_norm": 0.171875, "learning_rate": 0.00019277934157385876, "loss": 1.0157, "step": 396 }, { "epoch": 0.14439909297052153, "grad_norm": 0.1689453125, "learning_rate": 0.00019273038795741037, "loss": 0.9222, "step": 398 }, { "epoch": 0.14512471655328799, "grad_norm": 0.1669921875, "learning_rate": 0.00019268143434096195, "loss": 0.9869, "step": 400 }, { "epoch": 0.14512471655328799, "eval_loss": 1.033414363861084, "eval_runtime": 94.5288, "eval_samples_per_second": 19.042, "eval_steps_per_second": 0.603, "step": 400 }, { "epoch": 0.1458503401360544, "grad_norm": 0.173828125, "learning_rate": 0.00019263248072451354, "loss": 1.0265, "step": 402 }, { "epoch": 0.14657596371882087, "grad_norm": 0.1748046875, "learning_rate": 0.00019258352710806512, "loss": 1.0718, "step": 404 }, { "epoch": 0.1473015873015873, "grad_norm": 0.1728515625, "learning_rate": 0.0001925345734916167, "loss": 1.0709, "step": 406 }, { "epoch": 0.14802721088435375, "grad_norm": 0.1689453125, "learning_rate": 0.00019248561987516828, "loss": 1.0135, "step": 408 }, { "epoch": 0.14875283446712018, "grad_norm": 0.169921875, "learning_rate": 0.00019243666625871986, "loss": 0.9483, "step": 410 }, { "epoch": 0.14947845804988663, "grad_norm": 0.16015625, "learning_rate": 0.00019238771264227147, "loss": 1.011, "step": 412 }, { "epoch": 0.15020408163265306, "grad_norm": 0.158203125, "learning_rate": 0.00019233875902582303, "loss": 1.0434, "step": 414 }, { "epoch": 0.1509297052154195, "grad_norm": 0.185546875, "learning_rate": 0.0001922898054093746, "loss": 1.0594, "step": 416 }, { "epoch": 0.15165532879818594, "grad_norm": 0.18359375, "learning_rate": 0.00019224085179292622, "loss": 0.998, "step": 418 }, { "epoch": 0.1523809523809524, "grad_norm": 0.173828125, "learning_rate": 0.0001921918981764778, "loss": 1.0772, "step": 420 }, { "epoch": 0.15310657596371882, "grad_norm": 0.1435546875, "learning_rate": 0.00019214294456002939, "loss": 0.9089, "step": 422 }, { "epoch": 0.15383219954648525, "grad_norm": 0.203125, "learning_rate": 0.00019209399094358097, "loss": 1.0361, "step": 424 }, { "epoch": 0.1545578231292517, "grad_norm": 0.1865234375, "learning_rate": 0.00019204503732713255, "loss": 1.0325, "step": 426 }, { "epoch": 0.15528344671201813, "grad_norm": 0.1552734375, "learning_rate": 0.00019199608371068413, "loss": 1.0237, "step": 428 }, { "epoch": 0.15600907029478459, "grad_norm": 0.166015625, "learning_rate": 0.00019194713009423571, "loss": 0.9513, "step": 430 }, { "epoch": 0.156734693877551, "grad_norm": 0.15625, "learning_rate": 0.0001918981764777873, "loss": 0.9614, "step": 432 }, { "epoch": 0.15746031746031747, "grad_norm": 0.1552734375, "learning_rate": 0.00019184922286133888, "loss": 0.9507, "step": 434 }, { "epoch": 0.1581859410430839, "grad_norm": 0.1533203125, "learning_rate": 0.00019180026924489046, "loss": 0.9842, "step": 436 }, { "epoch": 0.15891156462585035, "grad_norm": 0.146484375, "learning_rate": 0.00019175131562844207, "loss": 1.0938, "step": 438 }, { "epoch": 0.15963718820861678, "grad_norm": 0.171875, "learning_rate": 0.00019170236201199365, "loss": 0.9905, "step": 440 }, { "epoch": 0.16036281179138323, "grad_norm": 0.1748046875, "learning_rate": 0.0001916534083955452, "loss": 1.049, "step": 442 }, { "epoch": 0.16108843537414966, "grad_norm": 0.15234375, "learning_rate": 0.00019160445477909682, "loss": 0.9727, "step": 444 }, { "epoch": 0.1618140589569161, "grad_norm": 0.1513671875, "learning_rate": 0.0001915555011626484, "loss": 1.0255, "step": 446 }, { "epoch": 0.16253968253968254, "grad_norm": 0.2197265625, "learning_rate": 0.00019150654754619998, "loss": 0.9454, "step": 448 }, { "epoch": 0.16326530612244897, "grad_norm": 0.1806640625, "learning_rate": 0.00019145759392975156, "loss": 1.0037, "step": 450 }, { "epoch": 0.16399092970521542, "grad_norm": 0.1533203125, "learning_rate": 0.00019140864031330315, "loss": 0.9791, "step": 452 }, { "epoch": 0.16471655328798185, "grad_norm": 0.1455078125, "learning_rate": 0.00019135968669685473, "loss": 1.0033, "step": 454 }, { "epoch": 0.1654421768707483, "grad_norm": 0.1689453125, "learning_rate": 0.0001913107330804063, "loss": 1.0124, "step": 456 }, { "epoch": 0.16616780045351473, "grad_norm": 0.173828125, "learning_rate": 0.00019126177946395792, "loss": 1.0341, "step": 458 }, { "epoch": 0.16689342403628118, "grad_norm": 0.15234375, "learning_rate": 0.0001912128258475095, "loss": 0.999, "step": 460 }, { "epoch": 0.1676190476190476, "grad_norm": 0.166015625, "learning_rate": 0.00019116387223106106, "loss": 1.0343, "step": 462 }, { "epoch": 0.16834467120181407, "grad_norm": 0.1767578125, "learning_rate": 0.00019111491861461267, "loss": 1.1115, "step": 464 }, { "epoch": 0.1690702947845805, "grad_norm": 0.1640625, "learning_rate": 0.00019106596499816425, "loss": 1.0324, "step": 466 }, { "epoch": 0.16979591836734695, "grad_norm": 0.171875, "learning_rate": 0.00019101701138171583, "loss": 1.0122, "step": 468 }, { "epoch": 0.17052154195011338, "grad_norm": 0.18359375, "learning_rate": 0.00019096805776526741, "loss": 1.075, "step": 470 }, { "epoch": 0.17124716553287983, "grad_norm": 0.1748046875, "learning_rate": 0.000190919104148819, "loss": 1.117, "step": 472 }, { "epoch": 0.17197278911564626, "grad_norm": 0.1826171875, "learning_rate": 0.00019087015053237058, "loss": 0.9855, "step": 474 }, { "epoch": 0.1726984126984127, "grad_norm": 0.1533203125, "learning_rate": 0.00019082119691592216, "loss": 0.9953, "step": 476 }, { "epoch": 0.17342403628117914, "grad_norm": 0.1953125, "learning_rate": 0.00019077224329947377, "loss": 1.0023, "step": 478 }, { "epoch": 0.17414965986394557, "grad_norm": 0.16796875, "learning_rate": 0.00019072328968302535, "loss": 1.0291, "step": 480 }, { "epoch": 0.17487528344671202, "grad_norm": 0.146484375, "learning_rate": 0.0001906743360665769, "loss": 0.9518, "step": 482 }, { "epoch": 0.17560090702947845, "grad_norm": 0.162109375, "learning_rate": 0.00019062538245012852, "loss": 1.0413, "step": 484 }, { "epoch": 0.1763265306122449, "grad_norm": 0.1484375, "learning_rate": 0.0001905764288336801, "loss": 1.0636, "step": 486 }, { "epoch": 0.17705215419501133, "grad_norm": 0.1630859375, "learning_rate": 0.00019052747521723168, "loss": 1.041, "step": 488 }, { "epoch": 0.17777777777777778, "grad_norm": 0.1689453125, "learning_rate": 0.00019047852160078326, "loss": 1.0237, "step": 490 }, { "epoch": 0.1785034013605442, "grad_norm": 0.1630859375, "learning_rate": 0.00019042956798433485, "loss": 0.9516, "step": 492 }, { "epoch": 0.17922902494331067, "grad_norm": 0.1708984375, "learning_rate": 0.00019038061436788643, "loss": 1.0947, "step": 494 }, { "epoch": 0.1799546485260771, "grad_norm": 0.1708984375, "learning_rate": 0.000190331660751438, "loss": 1.1305, "step": 496 }, { "epoch": 0.18068027210884355, "grad_norm": 0.140625, "learning_rate": 0.00019028270713498962, "loss": 1.0146, "step": 498 }, { "epoch": 0.18140589569160998, "grad_norm": 0.1591796875, "learning_rate": 0.0001902337535185412, "loss": 0.9402, "step": 500 }, { "epoch": 0.18140589569160998, "eval_loss": 1.030946969985962, "eval_runtime": 94.9838, "eval_samples_per_second": 18.951, "eval_steps_per_second": 0.6, "step": 500 }, { "epoch": 0.18213151927437643, "grad_norm": 0.177734375, "learning_rate": 0.00019018479990209276, "loss": 0.9476, "step": 502 }, { "epoch": 0.18285714285714286, "grad_norm": 0.1669921875, "learning_rate": 0.00019013584628564437, "loss": 1.1158, "step": 504 }, { "epoch": 0.18358276643990928, "grad_norm": 0.1640625, "learning_rate": 0.00019008689266919595, "loss": 1.0539, "step": 506 }, { "epoch": 0.18430839002267574, "grad_norm": 0.1572265625, "learning_rate": 0.00019003793905274753, "loss": 1.0797, "step": 508 }, { "epoch": 0.18503401360544217, "grad_norm": 0.1533203125, "learning_rate": 0.00018998898543629911, "loss": 0.9487, "step": 510 }, { "epoch": 0.18575963718820862, "grad_norm": 0.14453125, "learning_rate": 0.0001899400318198507, "loss": 1.0215, "step": 512 }, { "epoch": 0.18648526077097505, "grad_norm": 0.1708984375, "learning_rate": 0.00018989107820340228, "loss": 1.0063, "step": 514 }, { "epoch": 0.1872108843537415, "grad_norm": 0.1748046875, "learning_rate": 0.00018984212458695386, "loss": 1.108, "step": 516 }, { "epoch": 0.18793650793650793, "grad_norm": 0.1552734375, "learning_rate": 0.00018979317097050547, "loss": 1.0557, "step": 518 }, { "epoch": 0.18866213151927438, "grad_norm": 0.169921875, "learning_rate": 0.00018974421735405703, "loss": 0.977, "step": 520 }, { "epoch": 0.1893877551020408, "grad_norm": 0.1640625, "learning_rate": 0.0001896952637376086, "loss": 1.0308, "step": 522 }, { "epoch": 0.19011337868480727, "grad_norm": 0.1767578125, "learning_rate": 0.00018964631012116022, "loss": 1.0537, "step": 524 }, { "epoch": 0.1908390022675737, "grad_norm": 0.189453125, "learning_rate": 0.0001895973565047118, "loss": 1.0358, "step": 526 }, { "epoch": 0.19156462585034015, "grad_norm": 0.1435546875, "learning_rate": 0.00018954840288826338, "loss": 1.0018, "step": 528 }, { "epoch": 0.19229024943310657, "grad_norm": 0.1787109375, "learning_rate": 0.00018949944927181496, "loss": 1.0255, "step": 530 }, { "epoch": 0.19301587301587303, "grad_norm": 0.140625, "learning_rate": 0.00018945049565536655, "loss": 0.9378, "step": 532 }, { "epoch": 0.19374149659863946, "grad_norm": 0.1728515625, "learning_rate": 0.00018940154203891813, "loss": 1.0373, "step": 534 }, { "epoch": 0.19446712018140588, "grad_norm": 0.1640625, "learning_rate": 0.0001893525884224697, "loss": 1.0026, "step": 536 }, { "epoch": 0.19519274376417234, "grad_norm": 0.1826171875, "learning_rate": 0.00018930363480602132, "loss": 1.0897, "step": 538 }, { "epoch": 0.19591836734693877, "grad_norm": 0.15234375, "learning_rate": 0.00018925468118957288, "loss": 0.9296, "step": 540 }, { "epoch": 0.19664399092970522, "grad_norm": 0.1806640625, "learning_rate": 0.00018920572757312446, "loss": 1.0485, "step": 542 }, { "epoch": 0.19736961451247165, "grad_norm": 0.1875, "learning_rate": 0.00018915677395667607, "loss": 0.9125, "step": 544 }, { "epoch": 0.1980952380952381, "grad_norm": 0.150390625, "learning_rate": 0.00018910782034022765, "loss": 1.0227, "step": 546 }, { "epoch": 0.19882086167800453, "grad_norm": 0.158203125, "learning_rate": 0.00018905886672377923, "loss": 0.9335, "step": 548 }, { "epoch": 0.19954648526077098, "grad_norm": 0.1787109375, "learning_rate": 0.00018900991310733081, "loss": 1.0549, "step": 550 }, { "epoch": 0.2002721088435374, "grad_norm": 0.1962890625, "learning_rate": 0.0001889609594908824, "loss": 0.9836, "step": 552 }, { "epoch": 0.20099773242630387, "grad_norm": 0.15234375, "learning_rate": 0.00018891200587443398, "loss": 1.0058, "step": 554 }, { "epoch": 0.2017233560090703, "grad_norm": 0.154296875, "learning_rate": 0.00018886305225798556, "loss": 1.0077, "step": 556 }, { "epoch": 0.20244897959183675, "grad_norm": 0.150390625, "learning_rate": 0.00018881409864153717, "loss": 0.9855, "step": 558 }, { "epoch": 0.20317460317460317, "grad_norm": 0.1611328125, "learning_rate": 0.00018876514502508873, "loss": 0.9711, "step": 560 }, { "epoch": 0.2039002267573696, "grad_norm": 0.142578125, "learning_rate": 0.0001887161914086403, "loss": 0.9793, "step": 562 }, { "epoch": 0.20462585034013606, "grad_norm": 0.171875, "learning_rate": 0.00018866723779219192, "loss": 1.0408, "step": 564 }, { "epoch": 0.20535147392290248, "grad_norm": 0.1787109375, "learning_rate": 0.0001886182841757435, "loss": 1.0226, "step": 566 }, { "epoch": 0.20607709750566894, "grad_norm": 0.169921875, "learning_rate": 0.00018856933055929508, "loss": 0.9565, "step": 568 }, { "epoch": 0.20680272108843537, "grad_norm": 0.1611328125, "learning_rate": 0.00018852037694284666, "loss": 1.0553, "step": 570 }, { "epoch": 0.20752834467120182, "grad_norm": 0.162109375, "learning_rate": 0.00018847142332639825, "loss": 1.0416, "step": 572 }, { "epoch": 0.20825396825396825, "grad_norm": 0.181640625, "learning_rate": 0.00018842246970994983, "loss": 1.0078, "step": 574 }, { "epoch": 0.2089795918367347, "grad_norm": 0.154296875, "learning_rate": 0.0001883735160935014, "loss": 0.9935, "step": 576 }, { "epoch": 0.20970521541950113, "grad_norm": 0.1708984375, "learning_rate": 0.000188324562477053, "loss": 1.0732, "step": 578 }, { "epoch": 0.21043083900226758, "grad_norm": 0.1728515625, "learning_rate": 0.00018827560886060458, "loss": 1.013, "step": 580 }, { "epoch": 0.211156462585034, "grad_norm": 0.1494140625, "learning_rate": 0.00018822665524415616, "loss": 1.0011, "step": 582 }, { "epoch": 0.21188208616780047, "grad_norm": 0.1435546875, "learning_rate": 0.00018817770162770777, "loss": 1.0494, "step": 584 }, { "epoch": 0.2126077097505669, "grad_norm": 0.1513671875, "learning_rate": 0.00018812874801125935, "loss": 0.9439, "step": 586 }, { "epoch": 0.21333333333333335, "grad_norm": 0.169921875, "learning_rate": 0.0001880797943948109, "loss": 1.0595, "step": 588 }, { "epoch": 0.21405895691609977, "grad_norm": 0.171875, "learning_rate": 0.00018803084077836251, "loss": 0.9564, "step": 590 }, { "epoch": 0.2147845804988662, "grad_norm": 0.1708984375, "learning_rate": 0.0001879818871619141, "loss": 1.1101, "step": 592 }, { "epoch": 0.21551020408163266, "grad_norm": 0.1513671875, "learning_rate": 0.00018793293354546568, "loss": 0.9792, "step": 594 }, { "epoch": 0.21623582766439908, "grad_norm": 0.1396484375, "learning_rate": 0.00018788397992901726, "loss": 0.9356, "step": 596 }, { "epoch": 0.21696145124716554, "grad_norm": 0.1318359375, "learning_rate": 0.00018783502631256884, "loss": 0.9553, "step": 598 }, { "epoch": 0.21768707482993196, "grad_norm": 0.1650390625, "learning_rate": 0.00018778607269612043, "loss": 0.9454, "step": 600 }, { "epoch": 0.21768707482993196, "eval_loss": 1.0292859077453613, "eval_runtime": 95.3967, "eval_samples_per_second": 18.869, "eval_steps_per_second": 0.598, "step": 600 }, { "epoch": 0.21841269841269842, "grad_norm": 0.1572265625, "learning_rate": 0.000187737119079672, "loss": 1.0517, "step": 602 }, { "epoch": 0.21913832199546485, "grad_norm": 0.1484375, "learning_rate": 0.00018768816546322362, "loss": 1.0417, "step": 604 }, { "epoch": 0.2198639455782313, "grad_norm": 0.1484375, "learning_rate": 0.0001876392118467752, "loss": 0.9824, "step": 606 }, { "epoch": 0.22058956916099773, "grad_norm": 0.158203125, "learning_rate": 0.00018759025823032675, "loss": 1.0043, "step": 608 }, { "epoch": 0.22131519274376418, "grad_norm": 0.1611328125, "learning_rate": 0.00018754130461387836, "loss": 0.94, "step": 610 }, { "epoch": 0.2220408163265306, "grad_norm": 0.185546875, "learning_rate": 0.00018749235099742995, "loss": 0.9731, "step": 612 }, { "epoch": 0.22276643990929706, "grad_norm": 0.162109375, "learning_rate": 0.00018744339738098153, "loss": 1.0113, "step": 614 }, { "epoch": 0.2234920634920635, "grad_norm": 0.1708984375, "learning_rate": 0.0001873944437645331, "loss": 1.0831, "step": 616 }, { "epoch": 0.22421768707482992, "grad_norm": 0.1630859375, "learning_rate": 0.0001873454901480847, "loss": 1.0206, "step": 618 }, { "epoch": 0.22494331065759637, "grad_norm": 0.15625, "learning_rate": 0.00018729653653163628, "loss": 1.0225, "step": 620 }, { "epoch": 0.2256689342403628, "grad_norm": 0.16015625, "learning_rate": 0.00018724758291518786, "loss": 0.9866, "step": 622 }, { "epoch": 0.22639455782312926, "grad_norm": 0.154296875, "learning_rate": 0.00018719862929873947, "loss": 1.0209, "step": 624 }, { "epoch": 0.22712018140589568, "grad_norm": 0.173828125, "learning_rate": 0.00018714967568229105, "loss": 0.9636, "step": 626 }, { "epoch": 0.22784580498866214, "grad_norm": 0.1552734375, "learning_rate": 0.0001871007220658426, "loss": 0.9683, "step": 628 }, { "epoch": 0.22857142857142856, "grad_norm": 0.2099609375, "learning_rate": 0.00018705176844939421, "loss": 0.9994, "step": 630 }, { "epoch": 0.22929705215419502, "grad_norm": 0.1611328125, "learning_rate": 0.0001870028148329458, "loss": 1.0061, "step": 632 }, { "epoch": 0.23002267573696145, "grad_norm": 0.1591796875, "learning_rate": 0.00018695386121649738, "loss": 0.9647, "step": 634 }, { "epoch": 0.2307482993197279, "grad_norm": 0.181640625, "learning_rate": 0.00018690490760004896, "loss": 1.0194, "step": 636 }, { "epoch": 0.23147392290249433, "grad_norm": 0.171875, "learning_rate": 0.00018685595398360054, "loss": 1.0266, "step": 638 }, { "epoch": 0.23219954648526078, "grad_norm": 0.16015625, "learning_rate": 0.00018680700036715213, "loss": 1.011, "step": 640 }, { "epoch": 0.2329251700680272, "grad_norm": 0.1494140625, "learning_rate": 0.0001867580467507037, "loss": 0.9558, "step": 642 }, { "epoch": 0.23365079365079366, "grad_norm": 0.15625, "learning_rate": 0.00018670909313425532, "loss": 1.0053, "step": 644 }, { "epoch": 0.2343764172335601, "grad_norm": 0.1494140625, "learning_rate": 0.0001866601395178069, "loss": 1.054, "step": 646 }, { "epoch": 0.23510204081632652, "grad_norm": 0.158203125, "learning_rate": 0.00018661118590135845, "loss": 0.9628, "step": 648 }, { "epoch": 0.23582766439909297, "grad_norm": 0.1435546875, "learning_rate": 0.00018656223228491006, "loss": 1.0158, "step": 650 }, { "epoch": 0.2365532879818594, "grad_norm": 0.15234375, "learning_rate": 0.00018651327866846165, "loss": 0.9922, "step": 652 }, { "epoch": 0.23727891156462586, "grad_norm": 0.154296875, "learning_rate": 0.00018646432505201323, "loss": 0.9373, "step": 654 }, { "epoch": 0.23800453514739228, "grad_norm": 0.1455078125, "learning_rate": 0.0001864153714355648, "loss": 0.993, "step": 656 }, { "epoch": 0.23873015873015874, "grad_norm": 0.1533203125, "learning_rate": 0.0001863664178191164, "loss": 1.0862, "step": 658 }, { "epoch": 0.23945578231292516, "grad_norm": 0.1611328125, "learning_rate": 0.00018631746420266798, "loss": 0.9666, "step": 660 }, { "epoch": 0.24018140589569162, "grad_norm": 0.1484375, "learning_rate": 0.00018626851058621956, "loss": 0.9805, "step": 662 }, { "epoch": 0.24090702947845805, "grad_norm": 0.15625, "learning_rate": 0.00018621955696977117, "loss": 0.9948, "step": 664 }, { "epoch": 0.2416326530612245, "grad_norm": 0.1435546875, "learning_rate": 0.00018617060335332272, "loss": 1.0119, "step": 666 }, { "epoch": 0.24235827664399093, "grad_norm": 0.154296875, "learning_rate": 0.0001861216497368743, "loss": 1.041, "step": 668 }, { "epoch": 0.24308390022675738, "grad_norm": 0.1728515625, "learning_rate": 0.00018607269612042591, "loss": 0.9779, "step": 670 }, { "epoch": 0.2438095238095238, "grad_norm": 0.16796875, "learning_rate": 0.0001860237425039775, "loss": 1.0314, "step": 672 }, { "epoch": 0.24453514739229024, "grad_norm": 0.1650390625, "learning_rate": 0.00018597478888752908, "loss": 1.0456, "step": 674 }, { "epoch": 0.2452607709750567, "grad_norm": 0.1396484375, "learning_rate": 0.00018592583527108066, "loss": 1.0908, "step": 676 }, { "epoch": 0.24598639455782312, "grad_norm": 0.2216796875, "learning_rate": 0.00018587688165463224, "loss": 0.9864, "step": 678 }, { "epoch": 0.24671201814058957, "grad_norm": 0.25390625, "learning_rate": 0.00018582792803818383, "loss": 1.0069, "step": 680 }, { "epoch": 0.247437641723356, "grad_norm": 0.1787109375, "learning_rate": 0.0001857789744217354, "loss": 1.0085, "step": 682 }, { "epoch": 0.24816326530612245, "grad_norm": 0.158203125, "learning_rate": 0.00018573002080528702, "loss": 0.9846, "step": 684 }, { "epoch": 0.24888888888888888, "grad_norm": 0.1845703125, "learning_rate": 0.00018568106718883857, "loss": 1.0301, "step": 686 }, { "epoch": 0.24961451247165534, "grad_norm": 0.1865234375, "learning_rate": 0.00018563211357239015, "loss": 1.031, "step": 688 }, { "epoch": 0.2503401360544218, "grad_norm": 0.1669921875, "learning_rate": 0.00018558315995594176, "loss": 1.043, "step": 690 }, { "epoch": 0.2510657596371882, "grad_norm": 0.15625, "learning_rate": 0.00018553420633949335, "loss": 0.9125, "step": 692 }, { "epoch": 0.25179138321995465, "grad_norm": 0.158203125, "learning_rate": 0.00018548525272304493, "loss": 1.0135, "step": 694 }, { "epoch": 0.2525170068027211, "grad_norm": 0.15625, "learning_rate": 0.0001854362991065965, "loss": 0.9936, "step": 696 }, { "epoch": 0.25324263038548755, "grad_norm": 0.142578125, "learning_rate": 0.0001853873454901481, "loss": 0.9794, "step": 698 }, { "epoch": 0.25396825396825395, "grad_norm": 0.1640625, "learning_rate": 0.00018533839187369968, "loss": 1.0237, "step": 700 }, { "epoch": 0.25396825396825395, "eval_loss": 1.0280758142471313, "eval_runtime": 92.7311, "eval_samples_per_second": 19.411, "eval_steps_per_second": 0.615, "step": 700 }, { "epoch": 0.2546938775510204, "grad_norm": 0.150390625, "learning_rate": 0.00018528943825725126, "loss": 1.069, "step": 702 }, { "epoch": 0.25541950113378686, "grad_norm": 0.1591796875, "learning_rate": 0.00018524048464080287, "loss": 0.902, "step": 704 }, { "epoch": 0.25614512471655326, "grad_norm": 0.171875, "learning_rate": 0.00018519153102435442, "loss": 1.0793, "step": 706 }, { "epoch": 0.2568707482993197, "grad_norm": 0.1572265625, "learning_rate": 0.000185142577407906, "loss": 1.0871, "step": 708 }, { "epoch": 0.2575963718820862, "grad_norm": 0.1494140625, "learning_rate": 0.0001850936237914576, "loss": 0.9444, "step": 710 }, { "epoch": 0.2583219954648526, "grad_norm": 0.169921875, "learning_rate": 0.0001850446701750092, "loss": 1.045, "step": 712 }, { "epoch": 0.259047619047619, "grad_norm": 0.1513671875, "learning_rate": 0.00018499571655856078, "loss": 1.0274, "step": 714 }, { "epoch": 0.2597732426303855, "grad_norm": 0.1611328125, "learning_rate": 0.00018494676294211236, "loss": 1.0513, "step": 716 }, { "epoch": 0.26049886621315194, "grad_norm": 0.16015625, "learning_rate": 0.00018489780932566394, "loss": 0.9794, "step": 718 }, { "epoch": 0.2612244897959184, "grad_norm": 0.1533203125, "learning_rate": 0.00018484885570921552, "loss": 0.9927, "step": 720 }, { "epoch": 0.2619501133786848, "grad_norm": 0.162109375, "learning_rate": 0.0001847999020927671, "loss": 0.9424, "step": 722 }, { "epoch": 0.26267573696145124, "grad_norm": 0.1689453125, "learning_rate": 0.0001847509484763187, "loss": 0.9492, "step": 724 }, { "epoch": 0.2634013605442177, "grad_norm": 0.1708984375, "learning_rate": 0.00018470199485987027, "loss": 1.0054, "step": 726 }, { "epoch": 0.26412698412698415, "grad_norm": 0.287109375, "learning_rate": 0.00018465304124342185, "loss": 1.0401, "step": 728 }, { "epoch": 0.26485260770975055, "grad_norm": 0.140625, "learning_rate": 0.00018460408762697346, "loss": 0.9785, "step": 730 }, { "epoch": 0.265578231292517, "grad_norm": 0.169921875, "learning_rate": 0.00018455513401052505, "loss": 1.0867, "step": 732 }, { "epoch": 0.26630385487528346, "grad_norm": 0.2001953125, "learning_rate": 0.0001845061803940766, "loss": 0.9375, "step": 734 }, { "epoch": 0.26702947845804986, "grad_norm": 0.1455078125, "learning_rate": 0.0001844572267776282, "loss": 0.9576, "step": 736 }, { "epoch": 0.2677551020408163, "grad_norm": 0.1630859375, "learning_rate": 0.0001844082731611798, "loss": 1.0314, "step": 738 }, { "epoch": 0.26848072562358277, "grad_norm": 0.142578125, "learning_rate": 0.00018435931954473137, "loss": 0.9826, "step": 740 }, { "epoch": 0.2692063492063492, "grad_norm": 0.201171875, "learning_rate": 0.00018431036592828296, "loss": 1.0339, "step": 742 }, { "epoch": 0.2699319727891156, "grad_norm": 0.13671875, "learning_rate": 0.00018426141231183454, "loss": 1.0327, "step": 744 }, { "epoch": 0.2706575963718821, "grad_norm": 0.150390625, "learning_rate": 0.00018421245869538612, "loss": 1.029, "step": 746 }, { "epoch": 0.27138321995464854, "grad_norm": 0.15625, "learning_rate": 0.0001841635050789377, "loss": 1.0364, "step": 748 }, { "epoch": 0.272108843537415, "grad_norm": 0.166015625, "learning_rate": 0.0001841145514624893, "loss": 0.925, "step": 750 }, { "epoch": 0.2728344671201814, "grad_norm": 0.16796875, "learning_rate": 0.0001840655978460409, "loss": 0.9998, "step": 752 }, { "epoch": 0.27356009070294784, "grad_norm": 0.1826171875, "learning_rate": 0.00018401664422959245, "loss": 1.0414, "step": 754 }, { "epoch": 0.2742857142857143, "grad_norm": 0.1494140625, "learning_rate": 0.00018396769061314406, "loss": 1.0585, "step": 756 }, { "epoch": 0.2750113378684807, "grad_norm": 0.1533203125, "learning_rate": 0.00018391873699669564, "loss": 0.945, "step": 758 }, { "epoch": 0.27573696145124715, "grad_norm": 0.1826171875, "learning_rate": 0.00018386978338024722, "loss": 1.0171, "step": 760 }, { "epoch": 0.2764625850340136, "grad_norm": 0.154296875, "learning_rate": 0.0001838208297637988, "loss": 1.0176, "step": 762 }, { "epoch": 0.27718820861678006, "grad_norm": 0.1484375, "learning_rate": 0.0001837718761473504, "loss": 0.9225, "step": 764 }, { "epoch": 0.27791383219954646, "grad_norm": 0.146484375, "learning_rate": 0.00018372292253090197, "loss": 1.0185, "step": 766 }, { "epoch": 0.2786394557823129, "grad_norm": 0.150390625, "learning_rate": 0.00018367396891445355, "loss": 1.0516, "step": 768 }, { "epoch": 0.27936507936507937, "grad_norm": 0.1474609375, "learning_rate": 0.00018362501529800516, "loss": 0.972, "step": 770 }, { "epoch": 0.2800907029478458, "grad_norm": 0.1552734375, "learning_rate": 0.00018357606168155675, "loss": 1.0903, "step": 772 }, { "epoch": 0.2808163265306122, "grad_norm": 0.1484375, "learning_rate": 0.0001835271080651083, "loss": 1.0122, "step": 774 }, { "epoch": 0.2815419501133787, "grad_norm": 0.1728515625, "learning_rate": 0.0001834781544486599, "loss": 1.0724, "step": 776 }, { "epoch": 0.28226757369614514, "grad_norm": 0.1640625, "learning_rate": 0.0001834292008322115, "loss": 1.02, "step": 778 }, { "epoch": 0.2829931972789116, "grad_norm": 0.1357421875, "learning_rate": 0.00018338024721576307, "loss": 0.9654, "step": 780 }, { "epoch": 0.283718820861678, "grad_norm": 0.173828125, "learning_rate": 0.00018333129359931466, "loss": 1.0622, "step": 782 }, { "epoch": 0.28444444444444444, "grad_norm": 0.16015625, "learning_rate": 0.00018328233998286624, "loss": 0.9866, "step": 784 }, { "epoch": 0.2851700680272109, "grad_norm": 0.1611328125, "learning_rate": 0.00018323338636641782, "loss": 1.048, "step": 786 }, { "epoch": 0.2858956916099773, "grad_norm": 0.1533203125, "learning_rate": 0.0001831844327499694, "loss": 0.9802, "step": 788 }, { "epoch": 0.28662131519274375, "grad_norm": 0.17578125, "learning_rate": 0.000183135479133521, "loss": 0.9756, "step": 790 }, { "epoch": 0.2873469387755102, "grad_norm": 0.1474609375, "learning_rate": 0.0001830865255170726, "loss": 1.0143, "step": 792 }, { "epoch": 0.28807256235827666, "grad_norm": 0.1787109375, "learning_rate": 0.00018303757190062415, "loss": 1.0164, "step": 794 }, { "epoch": 0.28879818594104306, "grad_norm": 0.169921875, "learning_rate": 0.00018298861828417576, "loss": 0.9813, "step": 796 }, { "epoch": 0.2895238095238095, "grad_norm": 0.166015625, "learning_rate": 0.00018293966466772734, "loss": 1.0568, "step": 798 }, { "epoch": 0.29024943310657597, "grad_norm": 0.1474609375, "learning_rate": 0.00018289071105127892, "loss": 0.9557, "step": 800 }, { "epoch": 0.29024943310657597, "eval_loss": 1.0269455909729004, "eval_runtime": 94.2392, "eval_samples_per_second": 19.1, "eval_steps_per_second": 0.605, "step": 800 }, { "epoch": 0.2909750566893424, "grad_norm": 0.162109375, "learning_rate": 0.0001828417574348305, "loss": 1.0491, "step": 802 }, { "epoch": 0.2917006802721088, "grad_norm": 0.1533203125, "learning_rate": 0.0001827928038183821, "loss": 0.9705, "step": 804 }, { "epoch": 0.2924263038548753, "grad_norm": 0.15234375, "learning_rate": 0.00018274385020193367, "loss": 1.0566, "step": 806 }, { "epoch": 0.29315192743764173, "grad_norm": 0.1611328125, "learning_rate": 0.00018269489658548525, "loss": 0.9782, "step": 808 }, { "epoch": 0.2938775510204082, "grad_norm": 0.154296875, "learning_rate": 0.00018264594296903686, "loss": 0.9967, "step": 810 }, { "epoch": 0.2946031746031746, "grad_norm": 0.1630859375, "learning_rate": 0.00018259698935258842, "loss": 1.0371, "step": 812 }, { "epoch": 0.29532879818594104, "grad_norm": 0.158203125, "learning_rate": 0.00018254803573614, "loss": 0.9803, "step": 814 }, { "epoch": 0.2960544217687075, "grad_norm": 0.158203125, "learning_rate": 0.0001824990821196916, "loss": 0.9969, "step": 816 }, { "epoch": 0.2967800453514739, "grad_norm": 0.154296875, "learning_rate": 0.0001824501285032432, "loss": 1.0483, "step": 818 }, { "epoch": 0.29750566893424035, "grad_norm": 0.1669921875, "learning_rate": 0.00018240117488679477, "loss": 1.0164, "step": 820 }, { "epoch": 0.2982312925170068, "grad_norm": 0.1630859375, "learning_rate": 0.00018235222127034636, "loss": 0.9407, "step": 822 }, { "epoch": 0.29895691609977326, "grad_norm": 0.158203125, "learning_rate": 0.00018230326765389794, "loss": 1.0397, "step": 824 }, { "epoch": 0.29968253968253966, "grad_norm": 0.138671875, "learning_rate": 0.00018225431403744952, "loss": 0.9518, "step": 826 }, { "epoch": 0.3004081632653061, "grad_norm": 0.154296875, "learning_rate": 0.0001822053604210011, "loss": 1.0551, "step": 828 }, { "epoch": 0.30113378684807257, "grad_norm": 0.16796875, "learning_rate": 0.0001821564068045527, "loss": 0.9826, "step": 830 }, { "epoch": 0.301859410430839, "grad_norm": 0.14453125, "learning_rate": 0.00018210745318810427, "loss": 0.945, "step": 832 }, { "epoch": 0.3025850340136054, "grad_norm": 0.1875, "learning_rate": 0.00018205849957165585, "loss": 0.9464, "step": 834 }, { "epoch": 0.3033106575963719, "grad_norm": 0.1611328125, "learning_rate": 0.00018200954595520746, "loss": 0.982, "step": 836 }, { "epoch": 0.30403628117913833, "grad_norm": 0.173828125, "learning_rate": 0.00018196059233875904, "loss": 1.0529, "step": 838 }, { "epoch": 0.3047619047619048, "grad_norm": 0.162109375, "learning_rate": 0.00018191163872231062, "loss": 0.9982, "step": 840 }, { "epoch": 0.3054875283446712, "grad_norm": 0.1533203125, "learning_rate": 0.0001818626851058622, "loss": 0.946, "step": 842 }, { "epoch": 0.30621315192743764, "grad_norm": 0.1396484375, "learning_rate": 0.0001818137314894138, "loss": 1.0053, "step": 844 }, { "epoch": 0.3069387755102041, "grad_norm": 0.1669921875, "learning_rate": 0.00018176477787296537, "loss": 0.9307, "step": 846 }, { "epoch": 0.3076643990929705, "grad_norm": 0.162109375, "learning_rate": 0.00018171582425651695, "loss": 0.9697, "step": 848 }, { "epoch": 0.30839002267573695, "grad_norm": 0.146484375, "learning_rate": 0.00018166687064006856, "loss": 0.9617, "step": 850 }, { "epoch": 0.3091156462585034, "grad_norm": 0.15234375, "learning_rate": 0.00018161791702362012, "loss": 0.9534, "step": 852 }, { "epoch": 0.30984126984126986, "grad_norm": 0.15234375, "learning_rate": 0.0001815689634071717, "loss": 1.0296, "step": 854 }, { "epoch": 0.31056689342403626, "grad_norm": 0.1494140625, "learning_rate": 0.0001815200097907233, "loss": 1.0481, "step": 856 }, { "epoch": 0.3112925170068027, "grad_norm": 0.154296875, "learning_rate": 0.0001814710561742749, "loss": 1.0118, "step": 858 }, { "epoch": 0.31201814058956917, "grad_norm": 0.1533203125, "learning_rate": 0.00018142210255782647, "loss": 1.0305, "step": 860 }, { "epoch": 0.3127437641723356, "grad_norm": 0.1591796875, "learning_rate": 0.00018137314894137806, "loss": 1.0597, "step": 862 }, { "epoch": 0.313469387755102, "grad_norm": 0.1591796875, "learning_rate": 0.00018132419532492964, "loss": 1.0554, "step": 864 }, { "epoch": 0.3141950113378685, "grad_norm": 0.134765625, "learning_rate": 0.00018127524170848122, "loss": 0.9902, "step": 866 }, { "epoch": 0.31492063492063493, "grad_norm": 0.1435546875, "learning_rate": 0.0001812262880920328, "loss": 0.9705, "step": 868 }, { "epoch": 0.31564625850340133, "grad_norm": 0.173828125, "learning_rate": 0.00018117733447558439, "loss": 1.0712, "step": 870 }, { "epoch": 0.3163718820861678, "grad_norm": 0.14453125, "learning_rate": 0.00018112838085913597, "loss": 1.0368, "step": 872 }, { "epoch": 0.31709750566893424, "grad_norm": 0.14453125, "learning_rate": 0.00018107942724268755, "loss": 0.9667, "step": 874 }, { "epoch": 0.3178231292517007, "grad_norm": 0.17578125, "learning_rate": 0.00018103047362623916, "loss": 0.9888, "step": 876 }, { "epoch": 0.3185487528344671, "grad_norm": 0.1689453125, "learning_rate": 0.00018098152000979074, "loss": 0.9979, "step": 878 }, { "epoch": 0.31927437641723355, "grad_norm": 0.154296875, "learning_rate": 0.0001809325663933423, "loss": 0.9947, "step": 880 }, { "epoch": 0.32, "grad_norm": 0.1552734375, "learning_rate": 0.0001808836127768939, "loss": 1.0054, "step": 882 }, { "epoch": 0.32072562358276646, "grad_norm": 0.1591796875, "learning_rate": 0.0001808346591604455, "loss": 1.0014, "step": 884 }, { "epoch": 0.32145124716553286, "grad_norm": 0.2021484375, "learning_rate": 0.00018078570554399707, "loss": 0.9482, "step": 886 }, { "epoch": 0.3221768707482993, "grad_norm": 0.146484375, "learning_rate": 0.00018073675192754865, "loss": 1.0337, "step": 888 }, { "epoch": 0.32290249433106577, "grad_norm": 0.150390625, "learning_rate": 0.00018068779831110024, "loss": 0.9075, "step": 890 }, { "epoch": 0.3236281179138322, "grad_norm": 0.1650390625, "learning_rate": 0.00018063884469465182, "loss": 0.9731, "step": 892 }, { "epoch": 0.3243537414965986, "grad_norm": 0.1513671875, "learning_rate": 0.0001805898910782034, "loss": 1.0288, "step": 894 }, { "epoch": 0.3250793650793651, "grad_norm": 0.1611328125, "learning_rate": 0.000180540937461755, "loss": 0.9984, "step": 896 }, { "epoch": 0.32580498866213153, "grad_norm": 0.1455078125, "learning_rate": 0.0001804919838453066, "loss": 0.9527, "step": 898 }, { "epoch": 0.32653061224489793, "grad_norm": 0.150390625, "learning_rate": 0.00018044303022885815, "loss": 0.9399, "step": 900 }, { "epoch": 0.32653061224489793, "eval_loss": 1.0258883237838745, "eval_runtime": 95.9536, "eval_samples_per_second": 18.759, "eval_steps_per_second": 0.594, "step": 900 }, { "epoch": 0.3272562358276644, "grad_norm": 0.1630859375, "learning_rate": 0.00018039407661240976, "loss": 0.9891, "step": 902 }, { "epoch": 0.32798185941043084, "grad_norm": 0.16015625, "learning_rate": 0.00018034512299596134, "loss": 0.9403, "step": 904 }, { "epoch": 0.3287074829931973, "grad_norm": 0.1787109375, "learning_rate": 0.00018029616937951292, "loss": 1.0682, "step": 906 }, { "epoch": 0.3294331065759637, "grad_norm": 0.1435546875, "learning_rate": 0.0001802472157630645, "loss": 1.0294, "step": 908 }, { "epoch": 0.33015873015873015, "grad_norm": 0.1435546875, "learning_rate": 0.00018019826214661609, "loss": 0.9923, "step": 910 }, { "epoch": 0.3308843537414966, "grad_norm": 0.1533203125, "learning_rate": 0.00018014930853016767, "loss": 0.9585, "step": 912 }, { "epoch": 0.33160997732426306, "grad_norm": 0.2041015625, "learning_rate": 0.00018010035491371925, "loss": 1.0097, "step": 914 }, { "epoch": 0.33233560090702946, "grad_norm": 0.1533203125, "learning_rate": 0.00018005140129727086, "loss": 0.9935, "step": 916 }, { "epoch": 0.3330612244897959, "grad_norm": 0.14453125, "learning_rate": 0.00018000244768082244, "loss": 0.9699, "step": 918 }, { "epoch": 0.33378684807256237, "grad_norm": 0.146484375, "learning_rate": 0.000179953494064374, "loss": 1.0012, "step": 920 }, { "epoch": 0.3345124716553288, "grad_norm": 0.1728515625, "learning_rate": 0.0001799045404479256, "loss": 0.9984, "step": 922 }, { "epoch": 0.3352380952380952, "grad_norm": 0.154296875, "learning_rate": 0.0001798555868314772, "loss": 0.978, "step": 924 }, { "epoch": 0.3359637188208617, "grad_norm": 0.146484375, "learning_rate": 0.00017980663321502877, "loss": 1.0188, "step": 926 }, { "epoch": 0.33668934240362813, "grad_norm": 0.169921875, "learning_rate": 0.00017975767959858035, "loss": 0.9777, "step": 928 }, { "epoch": 0.33741496598639453, "grad_norm": 0.1416015625, "learning_rate": 0.00017970872598213194, "loss": 1.042, "step": 930 }, { "epoch": 0.338140589569161, "grad_norm": 0.1455078125, "learning_rate": 0.00017965977236568352, "loss": 1.0074, "step": 932 }, { "epoch": 0.33886621315192744, "grad_norm": 0.1494140625, "learning_rate": 0.0001796108187492351, "loss": 1.0228, "step": 934 }, { "epoch": 0.3395918367346939, "grad_norm": 0.169921875, "learning_rate": 0.0001795618651327867, "loss": 1.0906, "step": 936 }, { "epoch": 0.3403174603174603, "grad_norm": 0.1494140625, "learning_rate": 0.0001795129115163383, "loss": 1.0117, "step": 938 }, { "epoch": 0.34104308390022675, "grad_norm": 0.185546875, "learning_rate": 0.00017946395789988985, "loss": 1.0274, "step": 940 }, { "epoch": 0.3417687074829932, "grad_norm": 0.15234375, "learning_rate": 0.00017941500428344146, "loss": 1.0271, "step": 942 }, { "epoch": 0.34249433106575966, "grad_norm": 0.181640625, "learning_rate": 0.00017936605066699304, "loss": 1.0064, "step": 944 }, { "epoch": 0.34321995464852606, "grad_norm": 0.13671875, "learning_rate": 0.00017931709705054462, "loss": 0.9696, "step": 946 }, { "epoch": 0.3439455782312925, "grad_norm": 0.1552734375, "learning_rate": 0.0001792681434340962, "loss": 1.0467, "step": 948 }, { "epoch": 0.34467120181405897, "grad_norm": 0.1748046875, "learning_rate": 0.00017921918981764779, "loss": 1.0357, "step": 950 }, { "epoch": 0.3453968253968254, "grad_norm": 0.1357421875, "learning_rate": 0.00017917023620119937, "loss": 0.9136, "step": 952 }, { "epoch": 0.3461224489795918, "grad_norm": 0.1630859375, "learning_rate": 0.00017912128258475095, "loss": 1.036, "step": 954 }, { "epoch": 0.3468480725623583, "grad_norm": 0.1484375, "learning_rate": 0.00017907232896830256, "loss": 0.9946, "step": 956 }, { "epoch": 0.34757369614512473, "grad_norm": 0.146484375, "learning_rate": 0.00017902337535185411, "loss": 0.9912, "step": 958 }, { "epoch": 0.34829931972789113, "grad_norm": 0.138671875, "learning_rate": 0.0001789744217354057, "loss": 0.9606, "step": 960 }, { "epoch": 0.3490249433106576, "grad_norm": 0.1396484375, "learning_rate": 0.0001789254681189573, "loss": 0.9763, "step": 962 }, { "epoch": 0.34975056689342404, "grad_norm": 0.1337890625, "learning_rate": 0.0001788765145025089, "loss": 0.9389, "step": 964 }, { "epoch": 0.3504761904761905, "grad_norm": 0.14453125, "learning_rate": 0.00017882756088606047, "loss": 0.9526, "step": 966 }, { "epoch": 0.3512018140589569, "grad_norm": 0.1728515625, "learning_rate": 0.00017877860726961205, "loss": 0.9523, "step": 968 }, { "epoch": 0.35192743764172335, "grad_norm": 0.1533203125, "learning_rate": 0.00017872965365316364, "loss": 1.0858, "step": 970 }, { "epoch": 0.3526530612244898, "grad_norm": 0.1552734375, "learning_rate": 0.00017868070003671522, "loss": 0.9991, "step": 972 }, { "epoch": 0.35337868480725626, "grad_norm": 0.1494140625, "learning_rate": 0.0001786317464202668, "loss": 0.9704, "step": 974 }, { "epoch": 0.35410430839002266, "grad_norm": 0.138671875, "learning_rate": 0.0001785827928038184, "loss": 0.9472, "step": 976 }, { "epoch": 0.3548299319727891, "grad_norm": 0.140625, "learning_rate": 0.00017853383918736996, "loss": 1.0397, "step": 978 }, { "epoch": 0.35555555555555557, "grad_norm": 0.154296875, "learning_rate": 0.00017848488557092155, "loss": 1.0412, "step": 980 }, { "epoch": 0.35628117913832197, "grad_norm": 0.1591796875, "learning_rate": 0.00017843593195447316, "loss": 1.07, "step": 982 }, { "epoch": 0.3570068027210884, "grad_norm": 0.15625, "learning_rate": 0.00017838697833802474, "loss": 0.947, "step": 984 }, { "epoch": 0.3577324263038549, "grad_norm": 0.1494140625, "learning_rate": 0.00017833802472157632, "loss": 0.9768, "step": 986 }, { "epoch": 0.35845804988662133, "grad_norm": 0.154296875, "learning_rate": 0.0001782890711051279, "loss": 1.0229, "step": 988 }, { "epoch": 0.35918367346938773, "grad_norm": 0.1611328125, "learning_rate": 0.00017824011748867949, "loss": 1.0429, "step": 990 }, { "epoch": 0.3599092970521542, "grad_norm": 0.154296875, "learning_rate": 0.00017819116387223107, "loss": 0.9842, "step": 992 }, { "epoch": 0.36063492063492064, "grad_norm": 0.1416015625, "learning_rate": 0.00017814221025578265, "loss": 0.981, "step": 994 }, { "epoch": 0.3613605442176871, "grad_norm": 0.1494140625, "learning_rate": 0.00017809325663933426, "loss": 1.0028, "step": 996 }, { "epoch": 0.3620861678004535, "grad_norm": 0.1572265625, "learning_rate": 0.00017804430302288581, "loss": 1.061, "step": 998 }, { "epoch": 0.36281179138321995, "grad_norm": 0.1767578125, "learning_rate": 0.0001779953494064374, "loss": 1.0082, "step": 1000 }, { "epoch": 0.36281179138321995, "eval_loss": 1.0250780582427979, "eval_runtime": 92.3729, "eval_samples_per_second": 19.486, "eval_steps_per_second": 0.617, "step": 1000 }, { "epoch": 0.3635374149659864, "grad_norm": 0.158203125, "learning_rate": 0.000177946395789989, "loss": 0.9744, "step": 1002 }, { "epoch": 0.36426303854875286, "grad_norm": 0.1669921875, "learning_rate": 0.0001778974421735406, "loss": 1.0106, "step": 1004 }, { "epoch": 0.36498866213151926, "grad_norm": 0.1455078125, "learning_rate": 0.00017784848855709217, "loss": 0.8815, "step": 1006 }, { "epoch": 0.3657142857142857, "grad_norm": 0.1435546875, "learning_rate": 0.00017779953494064375, "loss": 0.9473, "step": 1008 }, { "epoch": 0.36643990929705217, "grad_norm": 0.15625, "learning_rate": 0.00017775058132419534, "loss": 0.9883, "step": 1010 }, { "epoch": 0.36716553287981857, "grad_norm": 0.12890625, "learning_rate": 0.00017770162770774692, "loss": 1.0715, "step": 1012 }, { "epoch": 0.367891156462585, "grad_norm": 0.1630859375, "learning_rate": 0.0001776526740912985, "loss": 0.9768, "step": 1014 }, { "epoch": 0.3686167800453515, "grad_norm": 0.1484375, "learning_rate": 0.00017760372047485008, "loss": 0.9155, "step": 1016 }, { "epoch": 0.36934240362811793, "grad_norm": 0.154296875, "learning_rate": 0.00017755476685840166, "loss": 1.051, "step": 1018 }, { "epoch": 0.37006802721088433, "grad_norm": 0.14453125, "learning_rate": 0.00017750581324195325, "loss": 1.0361, "step": 1020 }, { "epoch": 0.3707936507936508, "grad_norm": 0.138671875, "learning_rate": 0.00017745685962550486, "loss": 1.024, "step": 1022 }, { "epoch": 0.37151927437641724, "grad_norm": 0.1484375, "learning_rate": 0.00017740790600905644, "loss": 1.0117, "step": 1024 }, { "epoch": 0.3722448979591837, "grad_norm": 0.1650390625, "learning_rate": 0.000177358952392608, "loss": 0.992, "step": 1026 }, { "epoch": 0.3729705215419501, "grad_norm": 0.1728515625, "learning_rate": 0.0001773099987761596, "loss": 1.0872, "step": 1028 }, { "epoch": 0.37369614512471655, "grad_norm": 0.1435546875, "learning_rate": 0.00017726104515971118, "loss": 0.9335, "step": 1030 }, { "epoch": 0.374421768707483, "grad_norm": 0.1494140625, "learning_rate": 0.00017721209154326277, "loss": 1.017, "step": 1032 }, { "epoch": 0.37514739229024946, "grad_norm": 0.1455078125, "learning_rate": 0.00017716313792681435, "loss": 0.9909, "step": 1034 }, { "epoch": 0.37587301587301586, "grad_norm": 0.1484375, "learning_rate": 0.00017711418431036593, "loss": 1.0097, "step": 1036 }, { "epoch": 0.3765986394557823, "grad_norm": 0.1455078125, "learning_rate": 0.00017706523069391751, "loss": 0.9652, "step": 1038 }, { "epoch": 0.37732426303854877, "grad_norm": 0.1708984375, "learning_rate": 0.0001770162770774691, "loss": 0.9893, "step": 1040 }, { "epoch": 0.37804988662131517, "grad_norm": 0.13671875, "learning_rate": 0.0001769673234610207, "loss": 1.0294, "step": 1042 }, { "epoch": 0.3787755102040816, "grad_norm": 0.1611328125, "learning_rate": 0.0001769183698445723, "loss": 1.0322, "step": 1044 }, { "epoch": 0.3795011337868481, "grad_norm": 0.146484375, "learning_rate": 0.00017686941622812384, "loss": 0.908, "step": 1046 }, { "epoch": 0.38022675736961453, "grad_norm": 0.169921875, "learning_rate": 0.00017682046261167545, "loss": 1.0802, "step": 1048 }, { "epoch": 0.38095238095238093, "grad_norm": 0.1748046875, "learning_rate": 0.00017677150899522703, "loss": 1.1756, "step": 1050 }, { "epoch": 0.3816780045351474, "grad_norm": 0.15234375, "learning_rate": 0.00017672255537877862, "loss": 0.9534, "step": 1052 }, { "epoch": 0.38240362811791384, "grad_norm": 0.177734375, "learning_rate": 0.0001766736017623302, "loss": 0.9512, "step": 1054 }, { "epoch": 0.3831292517006803, "grad_norm": 0.1513671875, "learning_rate": 0.00017662464814588178, "loss": 1.0466, "step": 1056 }, { "epoch": 0.3838548752834467, "grad_norm": 0.173828125, "learning_rate": 0.00017657569452943336, "loss": 1.0233, "step": 1058 }, { "epoch": 0.38458049886621315, "grad_norm": 0.1591796875, "learning_rate": 0.00017652674091298495, "loss": 1.0718, "step": 1060 }, { "epoch": 0.3853061224489796, "grad_norm": 0.1689453125, "learning_rate": 0.00017647778729653656, "loss": 1.0447, "step": 1062 }, { "epoch": 0.38603174603174606, "grad_norm": 0.16796875, "learning_rate": 0.00017642883368008814, "loss": 1.0104, "step": 1064 }, { "epoch": 0.38675736961451246, "grad_norm": 0.1357421875, "learning_rate": 0.0001763798800636397, "loss": 0.9774, "step": 1066 }, { "epoch": 0.3874829931972789, "grad_norm": 0.158203125, "learning_rate": 0.0001763309264471913, "loss": 0.9544, "step": 1068 }, { "epoch": 0.38820861678004537, "grad_norm": 0.1884765625, "learning_rate": 0.00017628197283074288, "loss": 1.0006, "step": 1070 }, { "epoch": 0.38893424036281177, "grad_norm": 0.138671875, "learning_rate": 0.00017623301921429447, "loss": 0.9301, "step": 1072 }, { "epoch": 0.3896598639455782, "grad_norm": 0.1533203125, "learning_rate": 0.00017618406559784605, "loss": 1.0312, "step": 1074 }, { "epoch": 0.3903854875283447, "grad_norm": 0.138671875, "learning_rate": 0.00017613511198139763, "loss": 0.9978, "step": 1076 }, { "epoch": 0.39111111111111113, "grad_norm": 0.158203125, "learning_rate": 0.00017608615836494921, "loss": 0.9708, "step": 1078 }, { "epoch": 0.39183673469387753, "grad_norm": 0.1376953125, "learning_rate": 0.0001760372047485008, "loss": 0.9779, "step": 1080 }, { "epoch": 0.392562358276644, "grad_norm": 0.1494140625, "learning_rate": 0.0001759882511320524, "loss": 0.9583, "step": 1082 }, { "epoch": 0.39328798185941044, "grad_norm": 0.1591796875, "learning_rate": 0.000175939297515604, "loss": 1.0361, "step": 1084 }, { "epoch": 0.3940136054421769, "grad_norm": 0.1650390625, "learning_rate": 0.00017589034389915554, "loss": 1.061, "step": 1086 }, { "epoch": 0.3947392290249433, "grad_norm": 0.177734375, "learning_rate": 0.00017584139028270715, "loss": 1.0351, "step": 1088 }, { "epoch": 0.39546485260770975, "grad_norm": 0.1533203125, "learning_rate": 0.00017579243666625873, "loss": 0.9836, "step": 1090 }, { "epoch": 0.3961904761904762, "grad_norm": 0.1376953125, "learning_rate": 0.00017574348304981032, "loss": 0.9366, "step": 1092 }, { "epoch": 0.3969160997732426, "grad_norm": 0.1552734375, "learning_rate": 0.0001756945294333619, "loss": 0.9386, "step": 1094 }, { "epoch": 0.39764172335600906, "grad_norm": 0.134765625, "learning_rate": 0.00017564557581691348, "loss": 0.9872, "step": 1096 }, { "epoch": 0.3983673469387755, "grad_norm": 0.1572265625, "learning_rate": 0.00017559662220046506, "loss": 0.962, "step": 1098 }, { "epoch": 0.39909297052154197, "grad_norm": 0.1611328125, "learning_rate": 0.00017554766858401665, "loss": 1.0224, "step": 1100 }, { "epoch": 0.39909297052154197, "eval_loss": 1.024340033531189, "eval_runtime": 96.4085, "eval_samples_per_second": 18.671, "eval_steps_per_second": 0.591, "step": 1100 }, { "epoch": 0.39981859410430837, "grad_norm": 0.146484375, "learning_rate": 0.00017549871496756826, "loss": 0.9655, "step": 1102 }, { "epoch": 0.4005442176870748, "grad_norm": 0.158203125, "learning_rate": 0.0001754497613511198, "loss": 0.9732, "step": 1104 }, { "epoch": 0.4012698412698413, "grad_norm": 0.1630859375, "learning_rate": 0.0001754008077346714, "loss": 1.0525, "step": 1106 }, { "epoch": 0.40199546485260773, "grad_norm": 0.1591796875, "learning_rate": 0.000175351854118223, "loss": 0.9244, "step": 1108 }, { "epoch": 0.40272108843537413, "grad_norm": 0.162109375, "learning_rate": 0.00017530290050177458, "loss": 0.9976, "step": 1110 }, { "epoch": 0.4034467120181406, "grad_norm": 0.1533203125, "learning_rate": 0.00017525394688532617, "loss": 0.9802, "step": 1112 }, { "epoch": 0.40417233560090704, "grad_norm": 0.158203125, "learning_rate": 0.00017520499326887775, "loss": 0.9932, "step": 1114 }, { "epoch": 0.4048979591836735, "grad_norm": 0.15234375, "learning_rate": 0.00017515603965242933, "loss": 1.0189, "step": 1116 }, { "epoch": 0.4056235827664399, "grad_norm": 0.142578125, "learning_rate": 0.00017510708603598091, "loss": 0.908, "step": 1118 }, { "epoch": 0.40634920634920635, "grad_norm": 0.1591796875, "learning_rate": 0.0001750581324195325, "loss": 0.9909, "step": 1120 }, { "epoch": 0.4070748299319728, "grad_norm": 0.166015625, "learning_rate": 0.0001750091788030841, "loss": 1.0264, "step": 1122 }, { "epoch": 0.4078004535147392, "grad_norm": 0.1865234375, "learning_rate": 0.00017496022518663566, "loss": 1.0606, "step": 1124 }, { "epoch": 0.40852607709750566, "grad_norm": 0.1552734375, "learning_rate": 0.00017491127157018724, "loss": 0.9924, "step": 1126 }, { "epoch": 0.4092517006802721, "grad_norm": 0.14453125, "learning_rate": 0.00017486231795373885, "loss": 0.9948, "step": 1128 }, { "epoch": 0.40997732426303857, "grad_norm": 0.134765625, "learning_rate": 0.00017481336433729043, "loss": 0.9702, "step": 1130 }, { "epoch": 0.41070294784580497, "grad_norm": 0.15234375, "learning_rate": 0.00017476441072084202, "loss": 1.0459, "step": 1132 }, { "epoch": 0.4114285714285714, "grad_norm": 0.150390625, "learning_rate": 0.0001747154571043936, "loss": 0.9733, "step": 1134 }, { "epoch": 0.4121541950113379, "grad_norm": 0.14453125, "learning_rate": 0.00017466650348794518, "loss": 0.9991, "step": 1136 }, { "epoch": 0.41287981859410433, "grad_norm": 0.1630859375, "learning_rate": 0.00017461754987149676, "loss": 0.9936, "step": 1138 }, { "epoch": 0.41360544217687073, "grad_norm": 0.166015625, "learning_rate": 0.00017456859625504835, "loss": 0.9779, "step": 1140 }, { "epoch": 0.4143310657596372, "grad_norm": 0.1669921875, "learning_rate": 0.00017451964263859996, "loss": 0.996, "step": 1142 }, { "epoch": 0.41505668934240364, "grad_norm": 0.140625, "learning_rate": 0.0001744706890221515, "loss": 1.0079, "step": 1144 }, { "epoch": 0.4157823129251701, "grad_norm": 0.1767578125, "learning_rate": 0.0001744217354057031, "loss": 0.9805, "step": 1146 }, { "epoch": 0.4165079365079365, "grad_norm": 0.146484375, "learning_rate": 0.0001743727817892547, "loss": 0.9657, "step": 1148 }, { "epoch": 0.41723356009070295, "grad_norm": 0.1416015625, "learning_rate": 0.00017432382817280628, "loss": 1.0011, "step": 1150 }, { "epoch": 0.4179591836734694, "grad_norm": 0.14453125, "learning_rate": 0.00017427487455635787, "loss": 1.0128, "step": 1152 }, { "epoch": 0.4186848072562358, "grad_norm": 0.154296875, "learning_rate": 0.00017422592093990945, "loss": 0.97, "step": 1154 }, { "epoch": 0.41941043083900226, "grad_norm": 0.1611328125, "learning_rate": 0.00017417696732346103, "loss": 1.0179, "step": 1156 }, { "epoch": 0.4201360544217687, "grad_norm": 0.1513671875, "learning_rate": 0.0001741280137070126, "loss": 1.0459, "step": 1158 }, { "epoch": 0.42086167800453517, "grad_norm": 0.1845703125, "learning_rate": 0.0001740790600905642, "loss": 1.0039, "step": 1160 }, { "epoch": 0.42158730158730157, "grad_norm": 0.1484375, "learning_rate": 0.00017403010647411578, "loss": 1.0498, "step": 1162 }, { "epoch": 0.422312925170068, "grad_norm": 0.146484375, "learning_rate": 0.00017398115285766736, "loss": 1.0111, "step": 1164 }, { "epoch": 0.4230385487528345, "grad_norm": 0.1513671875, "learning_rate": 0.00017393219924121894, "loss": 0.9945, "step": 1166 }, { "epoch": 0.42376417233560093, "grad_norm": 0.1474609375, "learning_rate": 0.00017388324562477055, "loss": 0.9439, "step": 1168 }, { "epoch": 0.42448979591836733, "grad_norm": 0.146484375, "learning_rate": 0.00017383429200832213, "loss": 0.9019, "step": 1170 }, { "epoch": 0.4252154195011338, "grad_norm": 0.1787109375, "learning_rate": 0.0001737853383918737, "loss": 0.9996, "step": 1172 }, { "epoch": 0.42594104308390024, "grad_norm": 0.171875, "learning_rate": 0.0001737363847754253, "loss": 1.0023, "step": 1174 }, { "epoch": 0.4266666666666667, "grad_norm": 0.181640625, "learning_rate": 0.00017368743115897688, "loss": 0.9721, "step": 1176 }, { "epoch": 0.4273922902494331, "grad_norm": 0.154296875, "learning_rate": 0.00017363847754252846, "loss": 1.0203, "step": 1178 }, { "epoch": 0.42811791383219955, "grad_norm": 0.16015625, "learning_rate": 0.00017358952392608005, "loss": 0.9972, "step": 1180 }, { "epoch": 0.428843537414966, "grad_norm": 0.1728515625, "learning_rate": 0.00017354057030963163, "loss": 0.9853, "step": 1182 }, { "epoch": 0.4295691609977324, "grad_norm": 0.1474609375, "learning_rate": 0.0001734916166931832, "loss": 0.9784, "step": 1184 }, { "epoch": 0.43029478458049886, "grad_norm": 0.166015625, "learning_rate": 0.0001734426630767348, "loss": 1.0201, "step": 1186 }, { "epoch": 0.4310204081632653, "grad_norm": 0.1552734375, "learning_rate": 0.0001733937094602864, "loss": 1.0091, "step": 1188 }, { "epoch": 0.43174603174603177, "grad_norm": 0.158203125, "learning_rate": 0.00017334475584383798, "loss": 1.0313, "step": 1190 }, { "epoch": 0.43247165532879817, "grad_norm": 0.1474609375, "learning_rate": 0.00017329580222738954, "loss": 1.0487, "step": 1192 }, { "epoch": 0.4331972789115646, "grad_norm": 0.1455078125, "learning_rate": 0.00017324684861094115, "loss": 1.0059, "step": 1194 }, { "epoch": 0.4339229024943311, "grad_norm": 0.1484375, "learning_rate": 0.00017319789499449273, "loss": 0.9841, "step": 1196 }, { "epoch": 0.43464852607709753, "grad_norm": 0.1484375, "learning_rate": 0.0001731489413780443, "loss": 0.9891, "step": 1198 }, { "epoch": 0.43537414965986393, "grad_norm": 0.154296875, "learning_rate": 0.0001730999877615959, "loss": 1.0482, "step": 1200 }, { "epoch": 0.43537414965986393, "eval_loss": 1.023897647857666, "eval_runtime": 91.99, "eval_samples_per_second": 19.567, "eval_steps_per_second": 0.62, "step": 1200 }, { "epoch": 0.4360997732426304, "grad_norm": 0.1630859375, "learning_rate": 0.00017305103414514748, "loss": 0.9969, "step": 1202 }, { "epoch": 0.43682539682539684, "grad_norm": 0.1435546875, "learning_rate": 0.00017300208052869906, "loss": 0.9979, "step": 1204 }, { "epoch": 0.43755102040816324, "grad_norm": 0.158203125, "learning_rate": 0.00017295312691225064, "loss": 0.9974, "step": 1206 }, { "epoch": 0.4382766439909297, "grad_norm": 0.16015625, "learning_rate": 0.00017290417329580225, "loss": 1.0932, "step": 1208 }, { "epoch": 0.43900226757369615, "grad_norm": 0.1474609375, "learning_rate": 0.00017285521967935383, "loss": 0.9845, "step": 1210 }, { "epoch": 0.4397278911564626, "grad_norm": 0.1513671875, "learning_rate": 0.0001728062660629054, "loss": 1.0405, "step": 1212 }, { "epoch": 0.440453514739229, "grad_norm": 0.1396484375, "learning_rate": 0.000172757312446457, "loss": 0.9646, "step": 1214 }, { "epoch": 0.44117913832199546, "grad_norm": 0.13671875, "learning_rate": 0.00017270835883000858, "loss": 0.9535, "step": 1216 }, { "epoch": 0.4419047619047619, "grad_norm": 0.166015625, "learning_rate": 0.00017265940521356016, "loss": 1.0106, "step": 1218 }, { "epoch": 0.44263038548752837, "grad_norm": 0.1787109375, "learning_rate": 0.00017261045159711175, "loss": 0.9388, "step": 1220 }, { "epoch": 0.44335600907029477, "grad_norm": 0.142578125, "learning_rate": 0.00017256149798066333, "loss": 0.8825, "step": 1222 }, { "epoch": 0.4440816326530612, "grad_norm": 0.150390625, "learning_rate": 0.0001725125443642149, "loss": 0.98, "step": 1224 }, { "epoch": 0.4448072562358277, "grad_norm": 0.1484375, "learning_rate": 0.0001724635907477665, "loss": 0.996, "step": 1226 }, { "epoch": 0.44553287981859413, "grad_norm": 0.146484375, "learning_rate": 0.0001724146371313181, "loss": 0.9907, "step": 1228 }, { "epoch": 0.44625850340136053, "grad_norm": 0.154296875, "learning_rate": 0.00017236568351486968, "loss": 1.0013, "step": 1230 }, { "epoch": 0.446984126984127, "grad_norm": 0.1748046875, "learning_rate": 0.00017231672989842124, "loss": 0.9736, "step": 1232 }, { "epoch": 0.44770975056689344, "grad_norm": 0.1484375, "learning_rate": 0.00017226777628197285, "loss": 1.0385, "step": 1234 }, { "epoch": 0.44843537414965984, "grad_norm": 0.1748046875, "learning_rate": 0.00017221882266552443, "loss": 0.9977, "step": 1236 }, { "epoch": 0.4491609977324263, "grad_norm": 0.166015625, "learning_rate": 0.000172169869049076, "loss": 1.0035, "step": 1238 }, { "epoch": 0.44988662131519275, "grad_norm": 0.162109375, "learning_rate": 0.0001721209154326276, "loss": 1.0349, "step": 1240 }, { "epoch": 0.4506122448979592, "grad_norm": 0.1650390625, "learning_rate": 0.00017207196181617918, "loss": 1.0325, "step": 1242 }, { "epoch": 0.4513378684807256, "grad_norm": 0.1796875, "learning_rate": 0.00017202300819973076, "loss": 1.0148, "step": 1244 }, { "epoch": 0.45206349206349206, "grad_norm": 0.1533203125, "learning_rate": 0.00017197405458328234, "loss": 0.997, "step": 1246 }, { "epoch": 0.4527891156462585, "grad_norm": 0.1962890625, "learning_rate": 0.00017192510096683395, "loss": 0.913, "step": 1248 }, { "epoch": 0.45351473922902497, "grad_norm": 0.1708984375, "learning_rate": 0.0001718761473503855, "loss": 0.9991, "step": 1250 }, { "epoch": 0.45424036281179136, "grad_norm": 0.146484375, "learning_rate": 0.0001718271937339371, "loss": 1.0035, "step": 1252 }, { "epoch": 0.4549659863945578, "grad_norm": 0.130859375, "learning_rate": 0.0001717782401174887, "loss": 0.9321, "step": 1254 }, { "epoch": 0.4556916099773243, "grad_norm": 0.142578125, "learning_rate": 0.00017172928650104028, "loss": 0.9722, "step": 1256 }, { "epoch": 0.45641723356009073, "grad_norm": 0.1533203125, "learning_rate": 0.00017168033288459186, "loss": 1.0485, "step": 1258 }, { "epoch": 0.45714285714285713, "grad_norm": 0.14453125, "learning_rate": 0.00017163137926814345, "loss": 0.953, "step": 1260 }, { "epoch": 0.4578684807256236, "grad_norm": 0.1884765625, "learning_rate": 0.00017158242565169503, "loss": 1.0478, "step": 1262 }, { "epoch": 0.45859410430839004, "grad_norm": 0.1337890625, "learning_rate": 0.0001715334720352466, "loss": 0.8863, "step": 1264 }, { "epoch": 0.45931972789115644, "grad_norm": 0.1513671875, "learning_rate": 0.0001714845184187982, "loss": 0.9771, "step": 1266 }, { "epoch": 0.4600453514739229, "grad_norm": 0.1611328125, "learning_rate": 0.0001714355648023498, "loss": 0.9525, "step": 1268 }, { "epoch": 0.46077097505668935, "grad_norm": 0.1533203125, "learning_rate": 0.00017138661118590136, "loss": 1.0151, "step": 1270 }, { "epoch": 0.4614965986394558, "grad_norm": 0.16015625, "learning_rate": 0.00017133765756945294, "loss": 1.0558, "step": 1272 }, { "epoch": 0.4622222222222222, "grad_norm": 0.138671875, "learning_rate": 0.00017128870395300455, "loss": 1.0136, "step": 1274 }, { "epoch": 0.46294784580498866, "grad_norm": 0.1474609375, "learning_rate": 0.00017123975033655613, "loss": 0.9975, "step": 1276 }, { "epoch": 0.4636734693877551, "grad_norm": 0.1708984375, "learning_rate": 0.0001711907967201077, "loss": 1.028, "step": 1278 }, { "epoch": 0.46439909297052157, "grad_norm": 0.1611328125, "learning_rate": 0.0001711418431036593, "loss": 0.9967, "step": 1280 }, { "epoch": 0.46512471655328796, "grad_norm": 0.1650390625, "learning_rate": 0.00017109288948721088, "loss": 0.939, "step": 1282 }, { "epoch": 0.4658503401360544, "grad_norm": 0.15234375, "learning_rate": 0.00017104393587076246, "loss": 1.011, "step": 1284 }, { "epoch": 0.4665759637188209, "grad_norm": 0.146484375, "learning_rate": 0.00017099498225431404, "loss": 1.0439, "step": 1286 }, { "epoch": 0.46730158730158733, "grad_norm": 0.1748046875, "learning_rate": 0.00017094602863786565, "loss": 1.0081, "step": 1288 }, { "epoch": 0.46802721088435373, "grad_norm": 0.1708984375, "learning_rate": 0.0001708970750214172, "loss": 0.9553, "step": 1290 }, { "epoch": 0.4687528344671202, "grad_norm": 0.1640625, "learning_rate": 0.0001708481214049688, "loss": 0.9824, "step": 1292 }, { "epoch": 0.46947845804988664, "grad_norm": 0.1767578125, "learning_rate": 0.0001707991677885204, "loss": 1.045, "step": 1294 }, { "epoch": 0.47020408163265304, "grad_norm": 0.1611328125, "learning_rate": 0.00017075021417207198, "loss": 1.0157, "step": 1296 }, { "epoch": 0.4709297052154195, "grad_norm": 0.185546875, "learning_rate": 0.00017070126055562356, "loss": 0.9727, "step": 1298 }, { "epoch": 0.47165532879818595, "grad_norm": 0.1650390625, "learning_rate": 0.00017065230693917515, "loss": 0.9486, "step": 1300 }, { "epoch": 0.47165532879818595, "eval_loss": 1.0229827165603638, "eval_runtime": 91.9183, "eval_samples_per_second": 19.583, "eval_steps_per_second": 0.62, "step": 1300 }, { "epoch": 0.4723809523809524, "grad_norm": 0.1455078125, "learning_rate": 0.00017060335332272673, "loss": 1.0029, "step": 1302 }, { "epoch": 0.4731065759637188, "grad_norm": 0.1650390625, "learning_rate": 0.0001705543997062783, "loss": 1.0371, "step": 1304 }, { "epoch": 0.47383219954648526, "grad_norm": 0.1533203125, "learning_rate": 0.0001705054460898299, "loss": 0.9764, "step": 1306 }, { "epoch": 0.4745578231292517, "grad_norm": 0.169921875, "learning_rate": 0.00017045649247338147, "loss": 0.9555, "step": 1308 }, { "epoch": 0.47528344671201816, "grad_norm": 0.171875, "learning_rate": 0.00017040753885693306, "loss": 1.0557, "step": 1310 }, { "epoch": 0.47600907029478456, "grad_norm": 0.18359375, "learning_rate": 0.00017035858524048464, "loss": 0.9758, "step": 1312 }, { "epoch": 0.476734693877551, "grad_norm": 0.1611328125, "learning_rate": 0.00017030963162403625, "loss": 1.0147, "step": 1314 }, { "epoch": 0.4774603174603175, "grad_norm": 0.1630859375, "learning_rate": 0.00017026067800758783, "loss": 1.0386, "step": 1316 }, { "epoch": 0.4781859410430839, "grad_norm": 0.1435546875, "learning_rate": 0.00017021172439113939, "loss": 0.9744, "step": 1318 }, { "epoch": 0.47891156462585033, "grad_norm": 0.1611328125, "learning_rate": 0.000170162770774691, "loss": 0.9863, "step": 1320 }, { "epoch": 0.4796371882086168, "grad_norm": 0.171875, "learning_rate": 0.00017011381715824258, "loss": 1.0313, "step": 1322 }, { "epoch": 0.48036281179138324, "grad_norm": 0.1484375, "learning_rate": 0.00017006486354179416, "loss": 0.9102, "step": 1324 }, { "epoch": 0.48108843537414964, "grad_norm": 0.1650390625, "learning_rate": 0.00017001590992534574, "loss": 1.0352, "step": 1326 }, { "epoch": 0.4818140589569161, "grad_norm": 0.140625, "learning_rate": 0.00016996695630889732, "loss": 1.011, "step": 1328 }, { "epoch": 0.48253968253968255, "grad_norm": 0.1474609375, "learning_rate": 0.0001699180026924489, "loss": 0.9854, "step": 1330 }, { "epoch": 0.483265306122449, "grad_norm": 0.15234375, "learning_rate": 0.0001698690490760005, "loss": 0.9816, "step": 1332 }, { "epoch": 0.4839909297052154, "grad_norm": 0.140625, "learning_rate": 0.0001698200954595521, "loss": 0.9738, "step": 1334 }, { "epoch": 0.48471655328798186, "grad_norm": 0.19140625, "learning_rate": 0.00016977114184310368, "loss": 1.0348, "step": 1336 }, { "epoch": 0.4854421768707483, "grad_norm": 0.1669921875, "learning_rate": 0.00016972218822665524, "loss": 1.0062, "step": 1338 }, { "epoch": 0.48616780045351476, "grad_norm": 0.1357421875, "learning_rate": 0.00016967323461020684, "loss": 0.9665, "step": 1340 }, { "epoch": 0.48689342403628116, "grad_norm": 0.140625, "learning_rate": 0.00016962428099375843, "loss": 0.978, "step": 1342 }, { "epoch": 0.4876190476190476, "grad_norm": 0.1484375, "learning_rate": 0.00016957532737731, "loss": 1.0053, "step": 1344 }, { "epoch": 0.4883446712018141, "grad_norm": 0.162109375, "learning_rate": 0.0001695263737608616, "loss": 1.0164, "step": 1346 }, { "epoch": 0.4890702947845805, "grad_norm": 0.154296875, "learning_rate": 0.00016947742014441317, "loss": 1.0096, "step": 1348 }, { "epoch": 0.4897959183673469, "grad_norm": 0.146484375, "learning_rate": 0.00016942846652796476, "loss": 0.9065, "step": 1350 }, { "epoch": 0.4905215419501134, "grad_norm": 0.1494140625, "learning_rate": 0.00016937951291151634, "loss": 0.9667, "step": 1352 }, { "epoch": 0.49124716553287984, "grad_norm": 0.15234375, "learning_rate": 0.00016933055929506795, "loss": 1.0211, "step": 1354 }, { "epoch": 0.49197278911564624, "grad_norm": 0.1640625, "learning_rate": 0.00016928160567861953, "loss": 1.1053, "step": 1356 }, { "epoch": 0.4926984126984127, "grad_norm": 0.1455078125, "learning_rate": 0.00016923265206217109, "loss": 1.0523, "step": 1358 }, { "epoch": 0.49342403628117915, "grad_norm": 0.1611328125, "learning_rate": 0.0001691836984457227, "loss": 0.9851, "step": 1360 }, { "epoch": 0.4941496598639456, "grad_norm": 0.1611328125, "learning_rate": 0.00016913474482927428, "loss": 1.0227, "step": 1362 }, { "epoch": 0.494875283446712, "grad_norm": 0.1474609375, "learning_rate": 0.00016908579121282586, "loss": 0.9792, "step": 1364 }, { "epoch": 0.49560090702947845, "grad_norm": 0.15234375, "learning_rate": 0.00016903683759637744, "loss": 0.9671, "step": 1366 }, { "epoch": 0.4963265306122449, "grad_norm": 0.154296875, "learning_rate": 0.00016898788397992902, "loss": 0.9982, "step": 1368 }, { "epoch": 0.49705215419501136, "grad_norm": 0.1484375, "learning_rate": 0.0001689389303634806, "loss": 0.9745, "step": 1370 }, { "epoch": 0.49777777777777776, "grad_norm": 0.166015625, "learning_rate": 0.0001688899767470322, "loss": 1.0174, "step": 1372 }, { "epoch": 0.4985034013605442, "grad_norm": 0.150390625, "learning_rate": 0.0001688410231305838, "loss": 1.0175, "step": 1374 }, { "epoch": 0.4992290249433107, "grad_norm": 0.1708984375, "learning_rate": 0.00016879206951413538, "loss": 0.9615, "step": 1376 }, { "epoch": 0.4999546485260771, "grad_norm": 0.1572265625, "learning_rate": 0.00016874311589768694, "loss": 0.9905, "step": 1378 }, { "epoch": 0.5006802721088436, "grad_norm": 0.2158203125, "learning_rate": 0.00016869416228123854, "loss": 0.9977, "step": 1380 }, { "epoch": 0.50140589569161, "grad_norm": 0.1552734375, "learning_rate": 0.00016864520866479013, "loss": 0.9374, "step": 1382 }, { "epoch": 0.5021315192743764, "grad_norm": 0.1533203125, "learning_rate": 0.0001685962550483417, "loss": 1.0049, "step": 1384 }, { "epoch": 0.5028571428571429, "grad_norm": 0.158203125, "learning_rate": 0.0001685473014318933, "loss": 0.9839, "step": 1386 }, { "epoch": 0.5035827664399093, "grad_norm": 0.162109375, "learning_rate": 0.00016849834781544487, "loss": 0.9599, "step": 1388 }, { "epoch": 0.5043083900226757, "grad_norm": 0.1767578125, "learning_rate": 0.00016844939419899646, "loss": 0.9604, "step": 1390 }, { "epoch": 0.5050340136054422, "grad_norm": 0.142578125, "learning_rate": 0.00016840044058254804, "loss": 0.9192, "step": 1392 }, { "epoch": 0.5057596371882086, "grad_norm": 0.142578125, "learning_rate": 0.00016835148696609965, "loss": 0.9461, "step": 1394 }, { "epoch": 0.5064852607709751, "grad_norm": 0.17578125, "learning_rate": 0.0001683025333496512, "loss": 0.9682, "step": 1396 }, { "epoch": 0.5072108843537415, "grad_norm": 0.1455078125, "learning_rate": 0.00016825357973320279, "loss": 1.0531, "step": 1398 }, { "epoch": 0.5079365079365079, "grad_norm": 0.1845703125, "learning_rate": 0.0001682046261167544, "loss": 0.998, "step": 1400 }, { "epoch": 0.5079365079365079, "eval_loss": 1.0224273204803467, "eval_runtime": 90.8291, "eval_samples_per_second": 19.817, "eval_steps_per_second": 0.628, "step": 1400 }, { "epoch": 0.5086621315192744, "grad_norm": 0.16796875, "learning_rate": 0.00016815567250030598, "loss": 1.0047, "step": 1402 }, { "epoch": 0.5093877551020408, "grad_norm": 0.1591796875, "learning_rate": 0.00016810671888385756, "loss": 0.9845, "step": 1404 }, { "epoch": 0.5101133786848072, "grad_norm": 0.1669921875, "learning_rate": 0.00016805776526740914, "loss": 0.9928, "step": 1406 }, { "epoch": 0.5108390022675737, "grad_norm": 0.1552734375, "learning_rate": 0.00016800881165096072, "loss": 0.971, "step": 1408 }, { "epoch": 0.5115646258503401, "grad_norm": 0.1474609375, "learning_rate": 0.0001679598580345123, "loss": 0.9549, "step": 1410 }, { "epoch": 0.5122902494331065, "grad_norm": 0.1455078125, "learning_rate": 0.0001679109044180639, "loss": 1.056, "step": 1412 }, { "epoch": 0.513015873015873, "grad_norm": 0.1572265625, "learning_rate": 0.0001678619508016155, "loss": 1.037, "step": 1414 }, { "epoch": 0.5137414965986394, "grad_norm": 0.150390625, "learning_rate": 0.00016781299718516705, "loss": 1.0098, "step": 1416 }, { "epoch": 0.514467120181406, "grad_norm": 0.1416015625, "learning_rate": 0.00016776404356871864, "loss": 0.9179, "step": 1418 }, { "epoch": 0.5151927437641723, "grad_norm": 0.154296875, "learning_rate": 0.00016771508995227024, "loss": 1.0747, "step": 1420 }, { "epoch": 0.5159183673469387, "grad_norm": 0.142578125, "learning_rate": 0.00016766613633582183, "loss": 0.9267, "step": 1422 }, { "epoch": 0.5166439909297053, "grad_norm": 0.146484375, "learning_rate": 0.0001676171827193734, "loss": 0.9201, "step": 1424 }, { "epoch": 0.5173696145124717, "grad_norm": 0.15234375, "learning_rate": 0.000167568229102925, "loss": 1.0046, "step": 1426 }, { "epoch": 0.518095238095238, "grad_norm": 0.158203125, "learning_rate": 0.00016751927548647657, "loss": 1.0547, "step": 1428 }, { "epoch": 0.5188208616780046, "grad_norm": 0.158203125, "learning_rate": 0.00016747032187002816, "loss": 0.9742, "step": 1430 }, { "epoch": 0.519546485260771, "grad_norm": 0.162109375, "learning_rate": 0.00016742136825357974, "loss": 1.0289, "step": 1432 }, { "epoch": 0.5202721088435374, "grad_norm": 0.1435546875, "learning_rate": 0.00016737241463713135, "loss": 0.9975, "step": 1434 }, { "epoch": 0.5209977324263039, "grad_norm": 0.1708984375, "learning_rate": 0.0001673234610206829, "loss": 1.0441, "step": 1436 }, { "epoch": 0.5217233560090703, "grad_norm": 0.1669921875, "learning_rate": 0.00016727450740423449, "loss": 1.0447, "step": 1438 }, { "epoch": 0.5224489795918368, "grad_norm": 0.1845703125, "learning_rate": 0.0001672255537877861, "loss": 1.083, "step": 1440 }, { "epoch": 0.5231746031746032, "grad_norm": 0.1416015625, "learning_rate": 0.00016717660017133768, "loss": 0.9575, "step": 1442 }, { "epoch": 0.5239002267573696, "grad_norm": 0.16796875, "learning_rate": 0.00016712764655488926, "loss": 1.0298, "step": 1444 }, { "epoch": 0.5246258503401361, "grad_norm": 0.166015625, "learning_rate": 0.00016707869293844084, "loss": 1.0205, "step": 1446 }, { "epoch": 0.5253514739229025, "grad_norm": 0.171875, "learning_rate": 0.00016702973932199242, "loss": 0.9916, "step": 1448 }, { "epoch": 0.5260770975056689, "grad_norm": 0.1572265625, "learning_rate": 0.000166980785705544, "loss": 1.1116, "step": 1450 }, { "epoch": 0.5268027210884354, "grad_norm": 0.44140625, "learning_rate": 0.0001669318320890956, "loss": 1.072, "step": 1452 }, { "epoch": 0.5275283446712018, "grad_norm": 0.1943359375, "learning_rate": 0.00016688287847264717, "loss": 0.9555, "step": 1454 }, { "epoch": 0.5282539682539683, "grad_norm": 0.2177734375, "learning_rate": 0.00016683392485619875, "loss": 0.9393, "step": 1456 }, { "epoch": 0.5289795918367347, "grad_norm": 0.162109375, "learning_rate": 0.00016678497123975033, "loss": 0.9854, "step": 1458 }, { "epoch": 0.5297052154195011, "grad_norm": 0.1708984375, "learning_rate": 0.00016673601762330194, "loss": 1.0133, "step": 1460 }, { "epoch": 0.5304308390022676, "grad_norm": 0.177734375, "learning_rate": 0.00016668706400685353, "loss": 1.0171, "step": 1462 }, { "epoch": 0.531156462585034, "grad_norm": 0.1767578125, "learning_rate": 0.00016663811039040508, "loss": 0.9813, "step": 1464 }, { "epoch": 0.5318820861678004, "grad_norm": 0.166015625, "learning_rate": 0.0001665891567739567, "loss": 1.0612, "step": 1466 }, { "epoch": 0.5326077097505669, "grad_norm": 0.1669921875, "learning_rate": 0.00016654020315750827, "loss": 0.9901, "step": 1468 }, { "epoch": 0.5333333333333333, "grad_norm": 0.1513671875, "learning_rate": 0.00016649124954105986, "loss": 1.0232, "step": 1470 }, { "epoch": 0.5340589569160997, "grad_norm": 0.17578125, "learning_rate": 0.00016644229592461144, "loss": 1.0505, "step": 1472 }, { "epoch": 0.5347845804988662, "grad_norm": 0.1494140625, "learning_rate": 0.00016639334230816302, "loss": 0.9528, "step": 1474 }, { "epoch": 0.5355102040816326, "grad_norm": 0.1552734375, "learning_rate": 0.0001663443886917146, "loss": 1.0136, "step": 1476 }, { "epoch": 0.5362358276643991, "grad_norm": 0.1337890625, "learning_rate": 0.00016629543507526618, "loss": 0.9827, "step": 1478 }, { "epoch": 0.5369614512471655, "grad_norm": 0.16796875, "learning_rate": 0.0001662464814588178, "loss": 0.9911, "step": 1480 }, { "epoch": 0.5376870748299319, "grad_norm": 0.16015625, "learning_rate": 0.00016619752784236938, "loss": 0.9756, "step": 1482 }, { "epoch": 0.5384126984126985, "grad_norm": 0.162109375, "learning_rate": 0.00016614857422592093, "loss": 0.9821, "step": 1484 }, { "epoch": 0.5391383219954649, "grad_norm": 0.1591796875, "learning_rate": 0.00016609962060947254, "loss": 0.9637, "step": 1486 }, { "epoch": 0.5398639455782313, "grad_norm": 0.1591796875, "learning_rate": 0.00016605066699302412, "loss": 1.0195, "step": 1488 }, { "epoch": 0.5405895691609978, "grad_norm": 0.1572265625, "learning_rate": 0.0001660017133765757, "loss": 1.0409, "step": 1490 }, { "epoch": 0.5413151927437642, "grad_norm": 0.1748046875, "learning_rate": 0.0001659527597601273, "loss": 1.0361, "step": 1492 }, { "epoch": 0.5420408163265306, "grad_norm": 0.15625, "learning_rate": 0.00016590380614367887, "loss": 1.0197, "step": 1494 }, { "epoch": 0.5427664399092971, "grad_norm": 0.177734375, "learning_rate": 0.00016585485252723045, "loss": 0.9975, "step": 1496 }, { "epoch": 0.5434920634920635, "grad_norm": 0.154296875, "learning_rate": 0.00016580589891078203, "loss": 0.9776, "step": 1498 }, { "epoch": 0.54421768707483, "grad_norm": 0.1474609375, "learning_rate": 0.00016575694529433364, "loss": 0.9366, "step": 1500 }, { "epoch": 0.54421768707483, "eval_loss": 1.0217424631118774, "eval_runtime": 92.5726, "eval_samples_per_second": 19.444, "eval_steps_per_second": 0.616, "step": 1500 }, { "epoch": 0.5449433106575964, "grad_norm": 0.1689453125, "learning_rate": 0.00016570799167788523, "loss": 1.0161, "step": 1502 }, { "epoch": 0.5456689342403628, "grad_norm": 0.1640625, "learning_rate": 0.00016565903806143678, "loss": 1.07, "step": 1504 }, { "epoch": 0.5463945578231293, "grad_norm": 0.138671875, "learning_rate": 0.0001656100844449884, "loss": 0.8992, "step": 1506 }, { "epoch": 0.5471201814058957, "grad_norm": 0.1396484375, "learning_rate": 0.00016556113082853997, "loss": 1.123, "step": 1508 }, { "epoch": 0.5478458049886621, "grad_norm": 0.1533203125, "learning_rate": 0.00016551217721209156, "loss": 1.1036, "step": 1510 }, { "epoch": 0.5485714285714286, "grad_norm": 0.1630859375, "learning_rate": 0.00016546322359564314, "loss": 0.8912, "step": 1512 }, { "epoch": 0.549297052154195, "grad_norm": 0.1552734375, "learning_rate": 0.00016541426997919472, "loss": 0.9678, "step": 1514 }, { "epoch": 0.5500226757369614, "grad_norm": 0.15234375, "learning_rate": 0.0001653653163627463, "loss": 0.9695, "step": 1516 }, { "epoch": 0.5507482993197279, "grad_norm": 0.1474609375, "learning_rate": 0.00016531636274629788, "loss": 0.9806, "step": 1518 }, { "epoch": 0.5514739229024943, "grad_norm": 0.1767578125, "learning_rate": 0.0001652674091298495, "loss": 1.0622, "step": 1520 }, { "epoch": 0.5521995464852608, "grad_norm": 0.1435546875, "learning_rate": 0.00016521845551340108, "loss": 0.9587, "step": 1522 }, { "epoch": 0.5529251700680272, "grad_norm": 0.1728515625, "learning_rate": 0.00016516950189695263, "loss": 0.9624, "step": 1524 }, { "epoch": 0.5536507936507936, "grad_norm": 0.1435546875, "learning_rate": 0.00016512054828050424, "loss": 0.9404, "step": 1526 }, { "epoch": 0.5543764172335601, "grad_norm": 0.1650390625, "learning_rate": 0.00016507159466405582, "loss": 1.0409, "step": 1528 }, { "epoch": 0.5551020408163265, "grad_norm": 0.15625, "learning_rate": 0.0001650226410476074, "loss": 1.0547, "step": 1530 }, { "epoch": 0.5558276643990929, "grad_norm": 0.150390625, "learning_rate": 0.00016497368743115896, "loss": 0.9696, "step": 1532 }, { "epoch": 0.5565532879818594, "grad_norm": 0.1435546875, "learning_rate": 0.00016492473381471057, "loss": 1.032, "step": 1534 }, { "epoch": 0.5572789115646258, "grad_norm": 0.1484375, "learning_rate": 0.00016487578019826215, "loss": 1.0726, "step": 1536 }, { "epoch": 0.5580045351473923, "grad_norm": 0.1552734375, "learning_rate": 0.00016482682658181373, "loss": 1.0213, "step": 1538 }, { "epoch": 0.5587301587301587, "grad_norm": 0.1337890625, "learning_rate": 0.00016477787296536534, "loss": 0.9804, "step": 1540 }, { "epoch": 0.5594557823129251, "grad_norm": 0.162109375, "learning_rate": 0.0001647289193489169, "loss": 1.0306, "step": 1542 }, { "epoch": 0.5601814058956917, "grad_norm": 0.1650390625, "learning_rate": 0.00016467996573246848, "loss": 1.0537, "step": 1544 }, { "epoch": 0.560907029478458, "grad_norm": 0.146484375, "learning_rate": 0.0001646310121160201, "loss": 0.928, "step": 1546 }, { "epoch": 0.5616326530612245, "grad_norm": 0.1572265625, "learning_rate": 0.00016458205849957167, "loss": 1.0409, "step": 1548 }, { "epoch": 0.562358276643991, "grad_norm": 0.154296875, "learning_rate": 0.00016453310488312326, "loss": 0.966, "step": 1550 }, { "epoch": 0.5630839002267574, "grad_norm": 0.1611328125, "learning_rate": 0.0001644841512666748, "loss": 1.0074, "step": 1552 }, { "epoch": 0.5638095238095238, "grad_norm": 0.166015625, "learning_rate": 0.00016443519765022642, "loss": 0.9354, "step": 1554 }, { "epoch": 0.5645351473922903, "grad_norm": 0.154296875, "learning_rate": 0.000164386244033778, "loss": 1.0507, "step": 1556 }, { "epoch": 0.5652607709750567, "grad_norm": 0.166015625, "learning_rate": 0.00016433729041732958, "loss": 1.0014, "step": 1558 }, { "epoch": 0.5659863945578232, "grad_norm": 0.1357421875, "learning_rate": 0.0001642883368008812, "loss": 0.9745, "step": 1560 }, { "epoch": 0.5667120181405896, "grad_norm": 0.1533203125, "learning_rate": 0.00016423938318443275, "loss": 0.9293, "step": 1562 }, { "epoch": 0.567437641723356, "grad_norm": 0.1611328125, "learning_rate": 0.00016419042956798433, "loss": 1.1054, "step": 1564 }, { "epoch": 0.5681632653061225, "grad_norm": 0.150390625, "learning_rate": 0.00016414147595153594, "loss": 0.9644, "step": 1566 }, { "epoch": 0.5688888888888889, "grad_norm": 0.171875, "learning_rate": 0.00016409252233508752, "loss": 1.0018, "step": 1568 }, { "epoch": 0.5696145124716553, "grad_norm": 0.19140625, "learning_rate": 0.0001640435687186391, "loss": 1.0597, "step": 1570 }, { "epoch": 0.5703401360544218, "grad_norm": 0.1650390625, "learning_rate": 0.00016399461510219066, "loss": 1.0581, "step": 1572 }, { "epoch": 0.5710657596371882, "grad_norm": 0.1416015625, "learning_rate": 0.00016394566148574227, "loss": 1.0389, "step": 1574 }, { "epoch": 0.5717913832199546, "grad_norm": 0.1640625, "learning_rate": 0.00016389670786929385, "loss": 0.9944, "step": 1576 }, { "epoch": 0.5725170068027211, "grad_norm": 0.1435546875, "learning_rate": 0.00016384775425284543, "loss": 1.0125, "step": 1578 }, { "epoch": 0.5732426303854875, "grad_norm": 0.169921875, "learning_rate": 0.00016379880063639704, "loss": 1.0272, "step": 1580 }, { "epoch": 0.573968253968254, "grad_norm": 0.16796875, "learning_rate": 0.0001637498470199486, "loss": 1.0639, "step": 1582 }, { "epoch": 0.5746938775510204, "grad_norm": 0.1650390625, "learning_rate": 0.00016370089340350018, "loss": 0.9577, "step": 1584 }, { "epoch": 0.5754195011337868, "grad_norm": 0.154296875, "learning_rate": 0.0001636519397870518, "loss": 1.0295, "step": 1586 }, { "epoch": 0.5761451247165533, "grad_norm": 0.1640625, "learning_rate": 0.00016360298617060337, "loss": 1.0028, "step": 1588 }, { "epoch": 0.5768707482993197, "grad_norm": 0.1669921875, "learning_rate": 0.00016355403255415496, "loss": 1.0054, "step": 1590 }, { "epoch": 0.5775963718820861, "grad_norm": 0.162109375, "learning_rate": 0.00016350507893770654, "loss": 0.9994, "step": 1592 }, { "epoch": 0.5783219954648526, "grad_norm": 0.1435546875, "learning_rate": 0.00016345612532125812, "loss": 0.932, "step": 1594 }, { "epoch": 0.579047619047619, "grad_norm": 0.1474609375, "learning_rate": 0.0001634071717048097, "loss": 0.9427, "step": 1596 }, { "epoch": 0.5797732426303854, "grad_norm": 0.1455078125, "learning_rate": 0.00016335821808836128, "loss": 0.9373, "step": 1598 }, { "epoch": 0.5804988662131519, "grad_norm": 0.16015625, "learning_rate": 0.00016330926447191287, "loss": 1.0353, "step": 1600 }, { "epoch": 0.5804988662131519, "eval_loss": 1.0210474729537964, "eval_runtime": 93.6809, "eval_samples_per_second": 19.214, "eval_steps_per_second": 0.608, "step": 1600 }, { "epoch": 0.5812244897959183, "grad_norm": 0.1650390625, "learning_rate": 0.00016326031085546445, "loss": 0.9768, "step": 1602 }, { "epoch": 0.5819501133786849, "grad_norm": 0.1748046875, "learning_rate": 0.00016321135723901603, "loss": 0.9428, "step": 1604 }, { "epoch": 0.5826757369614513, "grad_norm": 0.1640625, "learning_rate": 0.00016316240362256764, "loss": 1.0214, "step": 1606 }, { "epoch": 0.5834013605442177, "grad_norm": 0.1513671875, "learning_rate": 0.00016311345000611922, "loss": 0.92, "step": 1608 }, { "epoch": 0.5841269841269842, "grad_norm": 0.18359375, "learning_rate": 0.00016306449638967078, "loss": 1.0035, "step": 1610 }, { "epoch": 0.5848526077097506, "grad_norm": 0.1611328125, "learning_rate": 0.0001630155427732224, "loss": 1.0347, "step": 1612 }, { "epoch": 0.585578231292517, "grad_norm": 0.185546875, "learning_rate": 0.00016296658915677397, "loss": 1.0052, "step": 1614 }, { "epoch": 0.5863038548752835, "grad_norm": 0.1435546875, "learning_rate": 0.00016291763554032555, "loss": 0.9912, "step": 1616 }, { "epoch": 0.5870294784580499, "grad_norm": 0.1669921875, "learning_rate": 0.00016286868192387713, "loss": 0.9186, "step": 1618 }, { "epoch": 0.5877551020408164, "grad_norm": 0.15234375, "learning_rate": 0.00016281972830742872, "loss": 1.0071, "step": 1620 }, { "epoch": 0.5884807256235828, "grad_norm": 0.1591796875, "learning_rate": 0.0001627707746909803, "loss": 0.9302, "step": 1622 }, { "epoch": 0.5892063492063492, "grad_norm": 0.1640625, "learning_rate": 0.00016272182107453188, "loss": 1.0591, "step": 1624 }, { "epoch": 0.5899319727891157, "grad_norm": 0.173828125, "learning_rate": 0.0001626728674580835, "loss": 0.9493, "step": 1626 }, { "epoch": 0.5906575963718821, "grad_norm": 0.1904296875, "learning_rate": 0.00016262391384163507, "loss": 1.0083, "step": 1628 }, { "epoch": 0.5913832199546485, "grad_norm": 0.16796875, "learning_rate": 0.00016257496022518663, "loss": 1.0483, "step": 1630 }, { "epoch": 0.592108843537415, "grad_norm": 0.1533203125, "learning_rate": 0.00016252600660873824, "loss": 1.0442, "step": 1632 }, { "epoch": 0.5928344671201814, "grad_norm": 0.1513671875, "learning_rate": 0.00016247705299228982, "loss": 1.0041, "step": 1634 }, { "epoch": 0.5935600907029478, "grad_norm": 0.1396484375, "learning_rate": 0.0001624280993758414, "loss": 0.8814, "step": 1636 }, { "epoch": 0.5942857142857143, "grad_norm": 0.154296875, "learning_rate": 0.00016237914575939298, "loss": 0.9601, "step": 1638 }, { "epoch": 0.5950113378684807, "grad_norm": 0.15234375, "learning_rate": 0.00016233019214294457, "loss": 1.0641, "step": 1640 }, { "epoch": 0.5957369614512472, "grad_norm": 0.1611328125, "learning_rate": 0.00016228123852649615, "loss": 0.928, "step": 1642 }, { "epoch": 0.5964625850340136, "grad_norm": 0.1484375, "learning_rate": 0.00016223228491004773, "loss": 0.9936, "step": 1644 }, { "epoch": 0.59718820861678, "grad_norm": 0.177734375, "learning_rate": 0.00016218333129359934, "loss": 1.0311, "step": 1646 }, { "epoch": 0.5979138321995465, "grad_norm": 0.134765625, "learning_rate": 0.00016213437767715092, "loss": 1.0138, "step": 1648 }, { "epoch": 0.5986394557823129, "grad_norm": 0.1337890625, "learning_rate": 0.00016208542406070248, "loss": 0.9511, "step": 1650 }, { "epoch": 0.5993650793650793, "grad_norm": 0.142578125, "learning_rate": 0.0001620364704442541, "loss": 0.9711, "step": 1652 }, { "epoch": 0.6000907029478458, "grad_norm": 0.1904296875, "learning_rate": 0.00016198751682780567, "loss": 0.9674, "step": 1654 }, { "epoch": 0.6008163265306122, "grad_norm": 0.1455078125, "learning_rate": 0.00016193856321135725, "loss": 1.0381, "step": 1656 }, { "epoch": 0.6015419501133786, "grad_norm": 0.1591796875, "learning_rate": 0.00016188960959490883, "loss": 0.9523, "step": 1658 }, { "epoch": 0.6022675736961451, "grad_norm": 0.166015625, "learning_rate": 0.00016184065597846042, "loss": 0.9676, "step": 1660 }, { "epoch": 0.6029931972789115, "grad_norm": 0.17578125, "learning_rate": 0.000161791702362012, "loss": 0.9896, "step": 1662 }, { "epoch": 0.603718820861678, "grad_norm": 0.16015625, "learning_rate": 0.00016174274874556358, "loss": 0.9777, "step": 1664 }, { "epoch": 0.6044444444444445, "grad_norm": 0.15625, "learning_rate": 0.0001616937951291152, "loss": 0.9839, "step": 1666 }, { "epoch": 0.6051700680272109, "grad_norm": 0.1396484375, "learning_rate": 0.00016164484151266677, "loss": 0.9354, "step": 1668 }, { "epoch": 0.6058956916099774, "grad_norm": 0.173828125, "learning_rate": 0.00016159588789621833, "loss": 0.9798, "step": 1670 }, { "epoch": 0.6066213151927438, "grad_norm": 0.171875, "learning_rate": 0.00016154693427976994, "loss": 0.99, "step": 1672 }, { "epoch": 0.6073469387755102, "grad_norm": 0.169921875, "learning_rate": 0.00016149798066332152, "loss": 1.0575, "step": 1674 }, { "epoch": 0.6080725623582767, "grad_norm": 0.1904296875, "learning_rate": 0.0001614490270468731, "loss": 1.0201, "step": 1676 }, { "epoch": 0.6087981859410431, "grad_norm": 0.162109375, "learning_rate": 0.00016140007343042466, "loss": 0.9888, "step": 1678 }, { "epoch": 0.6095238095238096, "grad_norm": 0.1650390625, "learning_rate": 0.00016135111981397627, "loss": 0.9927, "step": 1680 }, { "epoch": 0.610249433106576, "grad_norm": 0.1455078125, "learning_rate": 0.00016130216619752785, "loss": 1.0186, "step": 1682 }, { "epoch": 0.6109750566893424, "grad_norm": 0.162109375, "learning_rate": 0.00016125321258107943, "loss": 1.0128, "step": 1684 }, { "epoch": 0.6117006802721089, "grad_norm": 0.166015625, "learning_rate": 0.00016120425896463104, "loss": 1.0212, "step": 1686 }, { "epoch": 0.6124263038548753, "grad_norm": 0.15625, "learning_rate": 0.0001611553053481826, "loss": 0.9624, "step": 1688 }, { "epoch": 0.6131519274376417, "grad_norm": 0.1416015625, "learning_rate": 0.00016110635173173418, "loss": 1.0341, "step": 1690 }, { "epoch": 0.6138775510204082, "grad_norm": 0.1474609375, "learning_rate": 0.0001610573981152858, "loss": 0.9915, "step": 1692 }, { "epoch": 0.6146031746031746, "grad_norm": 0.1591796875, "learning_rate": 0.00016100844449883737, "loss": 1.0157, "step": 1694 }, { "epoch": 0.615328798185941, "grad_norm": 0.1474609375, "learning_rate": 0.00016095949088238895, "loss": 1.0553, "step": 1696 }, { "epoch": 0.6160544217687075, "grad_norm": 0.185546875, "learning_rate": 0.0001609105372659405, "loss": 0.9294, "step": 1698 }, { "epoch": 0.6167800453514739, "grad_norm": 0.1748046875, "learning_rate": 0.00016086158364949212, "loss": 0.9272, "step": 1700 }, { "epoch": 0.6167800453514739, "eval_loss": 1.0209547281265259, "eval_runtime": 91.8934, "eval_samples_per_second": 19.588, "eval_steps_per_second": 0.62, "step": 1700 }, { "epoch": 0.6175056689342404, "grad_norm": 0.158203125, "learning_rate": 0.0001608126300330437, "loss": 0.976, "step": 1702 }, { "epoch": 0.6182312925170068, "grad_norm": 0.1435546875, "learning_rate": 0.00016076367641659528, "loss": 0.9733, "step": 1704 }, { "epoch": 0.6189569160997732, "grad_norm": 0.150390625, "learning_rate": 0.0001607147228001469, "loss": 1.0233, "step": 1706 }, { "epoch": 0.6196825396825397, "grad_norm": 0.158203125, "learning_rate": 0.00016066576918369845, "loss": 0.9805, "step": 1708 }, { "epoch": 0.6204081632653061, "grad_norm": 0.146484375, "learning_rate": 0.00016061681556725003, "loss": 0.9321, "step": 1710 }, { "epoch": 0.6211337868480725, "grad_norm": 0.1875, "learning_rate": 0.00016056786195080164, "loss": 1.0678, "step": 1712 }, { "epoch": 0.621859410430839, "grad_norm": 0.1533203125, "learning_rate": 0.00016051890833435322, "loss": 1.0146, "step": 1714 }, { "epoch": 0.6225850340136054, "grad_norm": 0.1474609375, "learning_rate": 0.0001604699547179048, "loss": 0.8791, "step": 1716 }, { "epoch": 0.6233106575963718, "grad_norm": 0.150390625, "learning_rate": 0.00016042100110145636, "loss": 0.9568, "step": 1718 }, { "epoch": 0.6240362811791383, "grad_norm": 0.1484375, "learning_rate": 0.00016037204748500797, "loss": 0.9577, "step": 1720 }, { "epoch": 0.6247619047619047, "grad_norm": 0.140625, "learning_rate": 0.00016032309386855955, "loss": 0.9506, "step": 1722 }, { "epoch": 0.6254875283446713, "grad_norm": 0.1455078125, "learning_rate": 0.00016027414025211113, "loss": 0.9523, "step": 1724 }, { "epoch": 0.6262131519274377, "grad_norm": 0.1484375, "learning_rate": 0.00016022518663566274, "loss": 0.997, "step": 1726 }, { "epoch": 0.626938775510204, "grad_norm": 0.1494140625, "learning_rate": 0.0001601762330192143, "loss": 0.9419, "step": 1728 }, { "epoch": 0.6276643990929706, "grad_norm": 0.1669921875, "learning_rate": 0.00016012727940276588, "loss": 0.9835, "step": 1730 }, { "epoch": 0.628390022675737, "grad_norm": 0.16015625, "learning_rate": 0.0001600783257863175, "loss": 0.9472, "step": 1732 }, { "epoch": 0.6291156462585034, "grad_norm": 0.177734375, "learning_rate": 0.00016002937216986907, "loss": 1.0306, "step": 1734 }, { "epoch": 0.6298412698412699, "grad_norm": 0.166015625, "learning_rate": 0.00015998041855342065, "loss": 1.0021, "step": 1736 }, { "epoch": 0.6305668934240363, "grad_norm": 0.1376953125, "learning_rate": 0.0001599314649369722, "loss": 0.9455, "step": 1738 }, { "epoch": 0.6312925170068027, "grad_norm": 0.173828125, "learning_rate": 0.00015988251132052382, "loss": 1.0614, "step": 1740 }, { "epoch": 0.6320181405895692, "grad_norm": 0.15234375, "learning_rate": 0.0001598335577040754, "loss": 0.9934, "step": 1742 }, { "epoch": 0.6327437641723356, "grad_norm": 0.1435546875, "learning_rate": 0.00015978460408762698, "loss": 1.0372, "step": 1744 }, { "epoch": 0.6334693877551021, "grad_norm": 0.14453125, "learning_rate": 0.00015973565047117856, "loss": 1.0382, "step": 1746 }, { "epoch": 0.6341950113378685, "grad_norm": 0.1650390625, "learning_rate": 0.00015968669685473015, "loss": 1.001, "step": 1748 }, { "epoch": 0.6349206349206349, "grad_norm": 0.1455078125, "learning_rate": 0.00015963774323828173, "loss": 0.9544, "step": 1750 }, { "epoch": 0.6356462585034014, "grad_norm": 0.140625, "learning_rate": 0.00015958878962183334, "loss": 0.9517, "step": 1752 }, { "epoch": 0.6363718820861678, "grad_norm": 0.169921875, "learning_rate": 0.00015953983600538492, "loss": 0.9706, "step": 1754 }, { "epoch": 0.6370975056689342, "grad_norm": 0.1376953125, "learning_rate": 0.00015949088238893647, "loss": 1.0153, "step": 1756 }, { "epoch": 0.6378231292517007, "grad_norm": 0.146484375, "learning_rate": 0.00015944192877248806, "loss": 0.9559, "step": 1758 }, { "epoch": 0.6385487528344671, "grad_norm": 0.15234375, "learning_rate": 0.00015939297515603967, "loss": 0.9877, "step": 1760 }, { "epoch": 0.6392743764172336, "grad_norm": 0.16015625, "learning_rate": 0.00015934402153959125, "loss": 0.9975, "step": 1762 }, { "epoch": 0.64, "grad_norm": 0.1640625, "learning_rate": 0.00015929506792314283, "loss": 0.9666, "step": 1764 }, { "epoch": 0.6407256235827664, "grad_norm": 0.15234375, "learning_rate": 0.0001592461143066944, "loss": 1.028, "step": 1766 }, { "epoch": 0.6414512471655329, "grad_norm": 0.1533203125, "learning_rate": 0.000159197160690246, "loss": 1.0233, "step": 1768 }, { "epoch": 0.6421768707482993, "grad_norm": 0.1513671875, "learning_rate": 0.00015914820707379758, "loss": 1.017, "step": 1770 }, { "epoch": 0.6429024943310657, "grad_norm": 0.14453125, "learning_rate": 0.0001590992534573492, "loss": 1.0094, "step": 1772 }, { "epoch": 0.6436281179138322, "grad_norm": 0.1689453125, "learning_rate": 0.00015905029984090077, "loss": 0.9779, "step": 1774 }, { "epoch": 0.6443537414965986, "grad_norm": 0.1806640625, "learning_rate": 0.00015900134622445232, "loss": 0.9917, "step": 1776 }, { "epoch": 0.645079365079365, "grad_norm": 0.1748046875, "learning_rate": 0.0001589523926080039, "loss": 0.9526, "step": 1778 }, { "epoch": 0.6458049886621315, "grad_norm": 0.158203125, "learning_rate": 0.00015890343899155552, "loss": 1.0131, "step": 1780 }, { "epoch": 0.6465306122448979, "grad_norm": 0.1572265625, "learning_rate": 0.0001588544853751071, "loss": 1.053, "step": 1782 }, { "epoch": 0.6472562358276645, "grad_norm": 0.1484375, "learning_rate": 0.00015880553175865868, "loss": 0.9761, "step": 1784 }, { "epoch": 0.6479818594104308, "grad_norm": 0.1640625, "learning_rate": 0.00015875657814221026, "loss": 0.9547, "step": 1786 }, { "epoch": 0.6487074829931972, "grad_norm": 0.1533203125, "learning_rate": 0.00015870762452576184, "loss": 1.0081, "step": 1788 }, { "epoch": 0.6494331065759638, "grad_norm": 0.1826171875, "learning_rate": 0.00015865867090931343, "loss": 1.007, "step": 1790 }, { "epoch": 0.6501587301587302, "grad_norm": 0.1513671875, "learning_rate": 0.00015860971729286504, "loss": 1.0005, "step": 1792 }, { "epoch": 0.6508843537414966, "grad_norm": 0.1298828125, "learning_rate": 0.00015856076367641662, "loss": 0.9823, "step": 1794 }, { "epoch": 0.6516099773242631, "grad_norm": 0.1396484375, "learning_rate": 0.00015851181005996817, "loss": 0.9324, "step": 1796 }, { "epoch": 0.6523356009070295, "grad_norm": 0.1640625, "learning_rate": 0.00015846285644351976, "loss": 1.0323, "step": 1798 }, { "epoch": 0.6530612244897959, "grad_norm": 0.1552734375, "learning_rate": 0.00015841390282707137, "loss": 0.942, "step": 1800 }, { "epoch": 0.6530612244897959, "eval_loss": 1.0201084613800049, "eval_runtime": 94.9174, "eval_samples_per_second": 18.964, "eval_steps_per_second": 0.601, "step": 1800 }, { "epoch": 0.6537868480725624, "grad_norm": 0.162109375, "learning_rate": 0.00015836494921062295, "loss": 0.9913, "step": 1802 }, { "epoch": 0.6545124716553288, "grad_norm": 0.1630859375, "learning_rate": 0.00015831599559417453, "loss": 1.098, "step": 1804 }, { "epoch": 0.6552380952380953, "grad_norm": 0.173828125, "learning_rate": 0.0001582670419777261, "loss": 1.0758, "step": 1806 }, { "epoch": 0.6559637188208617, "grad_norm": 0.1513671875, "learning_rate": 0.0001582180883612777, "loss": 0.9549, "step": 1808 }, { "epoch": 0.6566893424036281, "grad_norm": 0.171875, "learning_rate": 0.00015816913474482928, "loss": 1.0207, "step": 1810 }, { "epoch": 0.6574149659863946, "grad_norm": 0.158203125, "learning_rate": 0.0001581201811283809, "loss": 0.9206, "step": 1812 }, { "epoch": 0.658140589569161, "grad_norm": 0.1572265625, "learning_rate": 0.00015807122751193247, "loss": 0.9818, "step": 1814 }, { "epoch": 0.6588662131519274, "grad_norm": 0.1552734375, "learning_rate": 0.00015802227389548402, "loss": 1.0962, "step": 1816 }, { "epoch": 0.6595918367346939, "grad_norm": 0.1591796875, "learning_rate": 0.0001579733202790356, "loss": 0.9518, "step": 1818 }, { "epoch": 0.6603174603174603, "grad_norm": 0.154296875, "learning_rate": 0.00015792436666258722, "loss": 0.9924, "step": 1820 }, { "epoch": 0.6610430839002267, "grad_norm": 0.1728515625, "learning_rate": 0.0001578754130461388, "loss": 0.9963, "step": 1822 }, { "epoch": 0.6617687074829932, "grad_norm": 0.140625, "learning_rate": 0.00015782645942969035, "loss": 0.983, "step": 1824 }, { "epoch": 0.6624943310657596, "grad_norm": 0.162109375, "learning_rate": 0.00015777750581324196, "loss": 1.0437, "step": 1826 }, { "epoch": 0.6632199546485261, "grad_norm": 0.1689453125, "learning_rate": 0.00015772855219679354, "loss": 1.0547, "step": 1828 }, { "epoch": 0.6639455782312925, "grad_norm": 0.1572265625, "learning_rate": 0.00015767959858034513, "loss": 0.9898, "step": 1830 }, { "epoch": 0.6646712018140589, "grad_norm": 0.1435546875, "learning_rate": 0.00015763064496389674, "loss": 1.0343, "step": 1832 }, { "epoch": 0.6653968253968254, "grad_norm": 0.1767578125, "learning_rate": 0.0001575816913474483, "loss": 1.0069, "step": 1834 }, { "epoch": 0.6661224489795918, "grad_norm": 0.1611328125, "learning_rate": 0.00015753273773099987, "loss": 1.032, "step": 1836 }, { "epoch": 0.6668480725623582, "grad_norm": 0.15625, "learning_rate": 0.00015748378411455146, "loss": 0.9829, "step": 1838 }, { "epoch": 0.6675736961451247, "grad_norm": 0.14453125, "learning_rate": 0.00015743483049810307, "loss": 1.0174, "step": 1840 }, { "epoch": 0.6682993197278911, "grad_norm": 0.142578125, "learning_rate": 0.00015738587688165465, "loss": 1.0235, "step": 1842 }, { "epoch": 0.6690249433106576, "grad_norm": 0.154296875, "learning_rate": 0.0001573369232652062, "loss": 0.9988, "step": 1844 }, { "epoch": 0.669750566893424, "grad_norm": 0.1572265625, "learning_rate": 0.0001572879696487578, "loss": 0.9146, "step": 1846 }, { "epoch": 0.6704761904761904, "grad_norm": 0.158203125, "learning_rate": 0.0001572390160323094, "loss": 1.0098, "step": 1848 }, { "epoch": 0.671201814058957, "grad_norm": 0.16796875, "learning_rate": 0.00015719006241586098, "loss": 0.9822, "step": 1850 }, { "epoch": 0.6719274376417234, "grad_norm": 0.1484375, "learning_rate": 0.00015714110879941259, "loss": 1.0375, "step": 1852 }, { "epoch": 0.6726530612244898, "grad_norm": 0.171875, "learning_rate": 0.00015709215518296414, "loss": 1.0528, "step": 1854 }, { "epoch": 0.6733786848072563, "grad_norm": 0.1484375, "learning_rate": 0.00015704320156651572, "loss": 0.9579, "step": 1856 }, { "epoch": 0.6741043083900227, "grad_norm": 0.16796875, "learning_rate": 0.0001569942479500673, "loss": 1.0379, "step": 1858 }, { "epoch": 0.6748299319727891, "grad_norm": 0.1611328125, "learning_rate": 0.00015694529433361892, "loss": 1.0233, "step": 1860 }, { "epoch": 0.6755555555555556, "grad_norm": 0.1591796875, "learning_rate": 0.0001568963407171705, "loss": 0.9767, "step": 1862 }, { "epoch": 0.676281179138322, "grad_norm": 0.173828125, "learning_rate": 0.00015684738710072205, "loss": 0.9701, "step": 1864 }, { "epoch": 0.6770068027210885, "grad_norm": 0.162109375, "learning_rate": 0.00015679843348427366, "loss": 0.9249, "step": 1866 }, { "epoch": 0.6777324263038549, "grad_norm": 0.150390625, "learning_rate": 0.00015674947986782524, "loss": 0.9929, "step": 1868 }, { "epoch": 0.6784580498866213, "grad_norm": 0.1328125, "learning_rate": 0.00015670052625137683, "loss": 0.937, "step": 1870 }, { "epoch": 0.6791836734693878, "grad_norm": 0.1455078125, "learning_rate": 0.00015665157263492844, "loss": 0.9915, "step": 1872 }, { "epoch": 0.6799092970521542, "grad_norm": 0.1533203125, "learning_rate": 0.00015660261901848, "loss": 1.0443, "step": 1874 }, { "epoch": 0.6806349206349206, "grad_norm": 0.14453125, "learning_rate": 0.00015655366540203157, "loss": 1.051, "step": 1876 }, { "epoch": 0.6813605442176871, "grad_norm": 0.1337890625, "learning_rate": 0.00015650471178558316, "loss": 0.9451, "step": 1878 }, { "epoch": 0.6820861678004535, "grad_norm": 0.1591796875, "learning_rate": 0.00015645575816913477, "loss": 0.9998, "step": 1880 }, { "epoch": 0.6828117913832199, "grad_norm": 0.150390625, "learning_rate": 0.00015640680455268635, "loss": 0.9396, "step": 1882 }, { "epoch": 0.6835374149659864, "grad_norm": 0.158203125, "learning_rate": 0.0001563578509362379, "loss": 1.0181, "step": 1884 }, { "epoch": 0.6842630385487528, "grad_norm": 0.158203125, "learning_rate": 0.0001563088973197895, "loss": 1.0061, "step": 1886 }, { "epoch": 0.6849886621315193, "grad_norm": 0.1484375, "learning_rate": 0.0001562599437033411, "loss": 0.9953, "step": 1888 }, { "epoch": 0.6857142857142857, "grad_norm": 0.15625, "learning_rate": 0.00015621099008689268, "loss": 1.0486, "step": 1890 }, { "epoch": 0.6864399092970521, "grad_norm": 0.1806640625, "learning_rate": 0.00015616203647044426, "loss": 0.9057, "step": 1892 }, { "epoch": 0.6871655328798186, "grad_norm": 0.158203125, "learning_rate": 0.00015611308285399584, "loss": 0.9876, "step": 1894 }, { "epoch": 0.687891156462585, "grad_norm": 0.130859375, "learning_rate": 0.00015606412923754742, "loss": 0.9982, "step": 1896 }, { "epoch": 0.6886167800453514, "grad_norm": 0.173828125, "learning_rate": 0.000156015175621099, "loss": 0.9909, "step": 1898 }, { "epoch": 0.6893424036281179, "grad_norm": 0.1533203125, "learning_rate": 0.00015596622200465062, "loss": 0.9472, "step": 1900 }, { "epoch": 0.6893424036281179, "eval_loss": 1.019601821899414, "eval_runtime": 96.3922, "eval_samples_per_second": 18.674, "eval_steps_per_second": 0.591, "step": 1900 }, { "epoch": 0.6900680272108843, "grad_norm": 0.1572265625, "learning_rate": 0.00015591726838820217, "loss": 0.9808, "step": 1902 }, { "epoch": 0.6907936507936508, "grad_norm": 0.1474609375, "learning_rate": 0.00015586831477175375, "loss": 1.0067, "step": 1904 }, { "epoch": 0.6915192743764172, "grad_norm": 0.1640625, "learning_rate": 0.00015581936115530536, "loss": 0.941, "step": 1906 }, { "epoch": 0.6922448979591836, "grad_norm": 0.1630859375, "learning_rate": 0.00015577040753885694, "loss": 0.944, "step": 1908 }, { "epoch": 0.6929705215419502, "grad_norm": 0.15625, "learning_rate": 0.00015572145392240853, "loss": 0.9849, "step": 1910 }, { "epoch": 0.6936961451247166, "grad_norm": 0.1455078125, "learning_rate": 0.0001556725003059601, "loss": 0.9713, "step": 1912 }, { "epoch": 0.694421768707483, "grad_norm": 0.16015625, "learning_rate": 0.0001556235466895117, "loss": 1.0218, "step": 1914 }, { "epoch": 0.6951473922902495, "grad_norm": 0.1640625, "learning_rate": 0.00015557459307306327, "loss": 0.9998, "step": 1916 }, { "epoch": 0.6958730158730159, "grad_norm": 0.1435546875, "learning_rate": 0.00015552563945661486, "loss": 1.0357, "step": 1918 }, { "epoch": 0.6965986394557823, "grad_norm": 0.1533203125, "learning_rate": 0.00015547668584016647, "loss": 0.9338, "step": 1920 }, { "epoch": 0.6973242630385488, "grad_norm": 0.14453125, "learning_rate": 0.00015542773222371802, "loss": 1.0691, "step": 1922 }, { "epoch": 0.6980498866213152, "grad_norm": 0.146484375, "learning_rate": 0.0001553787786072696, "loss": 0.994, "step": 1924 }, { "epoch": 0.6987755102040817, "grad_norm": 0.150390625, "learning_rate": 0.0001553298249908212, "loss": 0.9554, "step": 1926 }, { "epoch": 0.6995011337868481, "grad_norm": 0.19921875, "learning_rate": 0.0001552808713743728, "loss": 1.0239, "step": 1928 }, { "epoch": 0.7002267573696145, "grad_norm": 0.1630859375, "learning_rate": 0.00015523191775792438, "loss": 1.0417, "step": 1930 }, { "epoch": 0.700952380952381, "grad_norm": 0.1474609375, "learning_rate": 0.00015518296414147596, "loss": 0.9733, "step": 1932 }, { "epoch": 0.7016780045351474, "grad_norm": 0.17578125, "learning_rate": 0.00015513401052502754, "loss": 1.0096, "step": 1934 }, { "epoch": 0.7024036281179138, "grad_norm": 0.1767578125, "learning_rate": 0.00015508505690857912, "loss": 0.9957, "step": 1936 }, { "epoch": 0.7031292517006803, "grad_norm": 0.1787109375, "learning_rate": 0.0001550361032921307, "loss": 0.9878, "step": 1938 }, { "epoch": 0.7038548752834467, "grad_norm": 0.1591796875, "learning_rate": 0.00015498714967568232, "loss": 0.9317, "step": 1940 }, { "epoch": 0.7045804988662131, "grad_norm": 0.1630859375, "learning_rate": 0.00015493819605923387, "loss": 1.0407, "step": 1942 }, { "epoch": 0.7053061224489796, "grad_norm": 0.15625, "learning_rate": 0.00015488924244278545, "loss": 1.0233, "step": 1944 }, { "epoch": 0.706031746031746, "grad_norm": 0.1630859375, "learning_rate": 0.00015484028882633706, "loss": 0.9254, "step": 1946 }, { "epoch": 0.7067573696145125, "grad_norm": 0.220703125, "learning_rate": 0.00015479133520988864, "loss": 0.9546, "step": 1948 }, { "epoch": 0.7074829931972789, "grad_norm": 0.208984375, "learning_rate": 0.00015474238159344023, "loss": 0.9609, "step": 1950 }, { "epoch": 0.7082086167800453, "grad_norm": 0.1533203125, "learning_rate": 0.0001546934279769918, "loss": 1.0335, "step": 1952 }, { "epoch": 0.7089342403628118, "grad_norm": 0.158203125, "learning_rate": 0.0001546444743605434, "loss": 0.952, "step": 1954 }, { "epoch": 0.7096598639455782, "grad_norm": 0.16015625, "learning_rate": 0.00015459552074409497, "loss": 0.9557, "step": 1956 }, { "epoch": 0.7103854875283446, "grad_norm": 0.142578125, "learning_rate": 0.00015454656712764656, "loss": 0.967, "step": 1958 }, { "epoch": 0.7111111111111111, "grad_norm": 0.1630859375, "learning_rate": 0.00015449761351119816, "loss": 1.0578, "step": 1960 }, { "epoch": 0.7118367346938775, "grad_norm": 0.134765625, "learning_rate": 0.00015444865989474972, "loss": 1.0424, "step": 1962 }, { "epoch": 0.7125623582766439, "grad_norm": 0.162109375, "learning_rate": 0.0001543997062783013, "loss": 1.0075, "step": 1964 }, { "epoch": 0.7132879818594104, "grad_norm": 0.162109375, "learning_rate": 0.0001543507526618529, "loss": 0.9869, "step": 1966 }, { "epoch": 0.7140136054421768, "grad_norm": 0.140625, "learning_rate": 0.0001543017990454045, "loss": 0.9439, "step": 1968 }, { "epoch": 0.7147392290249434, "grad_norm": 0.1669921875, "learning_rate": 0.00015425284542895605, "loss": 1.004, "step": 1970 }, { "epoch": 0.7154648526077098, "grad_norm": 0.169921875, "learning_rate": 0.00015420389181250766, "loss": 0.9863, "step": 1972 }, { "epoch": 0.7161904761904762, "grad_norm": 0.13671875, "learning_rate": 0.00015415493819605924, "loss": 0.9668, "step": 1974 }, { "epoch": 0.7169160997732427, "grad_norm": 0.2373046875, "learning_rate": 0.00015410598457961082, "loss": 0.9949, "step": 1976 }, { "epoch": 0.7176417233560091, "grad_norm": 0.1640625, "learning_rate": 0.0001540570309631624, "loss": 0.959, "step": 1978 }, { "epoch": 0.7183673469387755, "grad_norm": 0.16796875, "learning_rate": 0.000154008077346714, "loss": 1.0601, "step": 1980 }, { "epoch": 0.719092970521542, "grad_norm": 0.1552734375, "learning_rate": 0.00015395912373026557, "loss": 1.0484, "step": 1982 }, { "epoch": 0.7198185941043084, "grad_norm": 0.1435546875, "learning_rate": 0.00015391017011381715, "loss": 0.9602, "step": 1984 }, { "epoch": 0.7205442176870749, "grad_norm": 0.158203125, "learning_rate": 0.00015386121649736876, "loss": 0.9339, "step": 1986 }, { "epoch": 0.7212698412698413, "grad_norm": 0.1630859375, "learning_rate": 0.00015381226288092034, "loss": 1.0522, "step": 1988 }, { "epoch": 0.7219954648526077, "grad_norm": 0.1611328125, "learning_rate": 0.0001537633092644719, "loss": 0.9165, "step": 1990 }, { "epoch": 0.7227210884353742, "grad_norm": 0.146484375, "learning_rate": 0.0001537143556480235, "loss": 0.9953, "step": 1992 }, { "epoch": 0.7234467120181406, "grad_norm": 0.1591796875, "learning_rate": 0.0001536654020315751, "loss": 0.9712, "step": 1994 }, { "epoch": 0.724172335600907, "grad_norm": 0.1630859375, "learning_rate": 0.00015361644841512667, "loss": 1.0144, "step": 1996 }, { "epoch": 0.7248979591836735, "grad_norm": 0.1767578125, "learning_rate": 0.00015356749479867826, "loss": 1.0188, "step": 1998 }, { "epoch": 0.7256235827664399, "grad_norm": 0.1376953125, "learning_rate": 0.00015351854118222984, "loss": 1.0385, "step": 2000 }, { "epoch": 0.7256235827664399, "eval_loss": 1.0189679861068726, "eval_runtime": 100.5257, "eval_samples_per_second": 17.906, "eval_steps_per_second": 0.567, "step": 2000 }, { "epoch": 0.7263492063492063, "grad_norm": 0.1689453125, "learning_rate": 0.00015346958756578142, "loss": 0.9979, "step": 2002 }, { "epoch": 0.7270748299319728, "grad_norm": 0.146484375, "learning_rate": 0.000153420633949333, "loss": 0.9475, "step": 2004 }, { "epoch": 0.7278004535147392, "grad_norm": 0.1591796875, "learning_rate": 0.0001533716803328846, "loss": 1.0247, "step": 2006 }, { "epoch": 0.7285260770975057, "grad_norm": 0.2890625, "learning_rate": 0.0001533227267164362, "loss": 1.0989, "step": 2008 }, { "epoch": 0.7292517006802721, "grad_norm": 0.1552734375, "learning_rate": 0.00015327377309998775, "loss": 0.9775, "step": 2010 }, { "epoch": 0.7299773242630385, "grad_norm": 0.154296875, "learning_rate": 0.00015322481948353936, "loss": 1.0218, "step": 2012 }, { "epoch": 0.730702947845805, "grad_norm": 0.130859375, "learning_rate": 0.00015317586586709094, "loss": 0.943, "step": 2014 }, { "epoch": 0.7314285714285714, "grad_norm": 0.15234375, "learning_rate": 0.00015312691225064252, "loss": 0.9455, "step": 2016 }, { "epoch": 0.7321541950113378, "grad_norm": 0.177734375, "learning_rate": 0.0001530779586341941, "loss": 0.9995, "step": 2018 }, { "epoch": 0.7328798185941043, "grad_norm": 0.1513671875, "learning_rate": 0.0001530290050177457, "loss": 1.0052, "step": 2020 }, { "epoch": 0.7336054421768707, "grad_norm": 0.166015625, "learning_rate": 0.00015298005140129727, "loss": 1.0352, "step": 2022 }, { "epoch": 0.7343310657596371, "grad_norm": 0.1484375, "learning_rate": 0.00015293109778484885, "loss": 1.0545, "step": 2024 }, { "epoch": 0.7350566893424036, "grad_norm": 0.150390625, "learning_rate": 0.00015288214416840046, "loss": 1.0652, "step": 2026 }, { "epoch": 0.73578231292517, "grad_norm": 0.1669921875, "learning_rate": 0.00015283319055195204, "loss": 1.0658, "step": 2028 }, { "epoch": 0.7365079365079366, "grad_norm": 0.162109375, "learning_rate": 0.0001527842369355036, "loss": 1.0105, "step": 2030 }, { "epoch": 0.737233560090703, "grad_norm": 0.1552734375, "learning_rate": 0.0001527352833190552, "loss": 1.0369, "step": 2032 }, { "epoch": 0.7379591836734694, "grad_norm": 0.16015625, "learning_rate": 0.0001526863297026068, "loss": 1.0044, "step": 2034 }, { "epoch": 0.7386848072562359, "grad_norm": 0.154296875, "learning_rate": 0.00015263737608615837, "loss": 0.9738, "step": 2036 }, { "epoch": 0.7394104308390023, "grad_norm": 0.154296875, "learning_rate": 0.00015258842246970996, "loss": 1.0011, "step": 2038 }, { "epoch": 0.7401360544217687, "grad_norm": 0.1328125, "learning_rate": 0.00015253946885326154, "loss": 0.8985, "step": 2040 }, { "epoch": 0.7408616780045352, "grad_norm": 0.1630859375, "learning_rate": 0.00015249051523681312, "loss": 1.0394, "step": 2042 }, { "epoch": 0.7415873015873016, "grad_norm": 0.15625, "learning_rate": 0.0001524415616203647, "loss": 1.02, "step": 2044 }, { "epoch": 0.742312925170068, "grad_norm": 0.138671875, "learning_rate": 0.0001523926080039163, "loss": 0.9851, "step": 2046 }, { "epoch": 0.7430385487528345, "grad_norm": 0.1630859375, "learning_rate": 0.00015234365438746787, "loss": 1.0116, "step": 2048 }, { "epoch": 0.7437641723356009, "grad_norm": 0.16015625, "learning_rate": 0.00015229470077101945, "loss": 1.041, "step": 2050 }, { "epoch": 0.7444897959183674, "grad_norm": 0.1650390625, "learning_rate": 0.00015224574715457106, "loss": 0.9993, "step": 2052 }, { "epoch": 0.7452154195011338, "grad_norm": 0.138671875, "learning_rate": 0.00015219679353812264, "loss": 0.9596, "step": 2054 }, { "epoch": 0.7459410430839002, "grad_norm": 0.150390625, "learning_rate": 0.00015214783992167422, "loss": 1.0042, "step": 2056 }, { "epoch": 0.7466666666666667, "grad_norm": 0.158203125, "learning_rate": 0.0001520988863052258, "loss": 0.9341, "step": 2058 }, { "epoch": 0.7473922902494331, "grad_norm": 0.1728515625, "learning_rate": 0.0001520499326887774, "loss": 0.9923, "step": 2060 }, { "epoch": 0.7481179138321995, "grad_norm": 0.1435546875, "learning_rate": 0.00015200097907232897, "loss": 1.0059, "step": 2062 }, { "epoch": 0.748843537414966, "grad_norm": 0.1455078125, "learning_rate": 0.00015195202545588055, "loss": 0.9428, "step": 2064 }, { "epoch": 0.7495691609977324, "grad_norm": 0.1591796875, "learning_rate": 0.00015190307183943216, "loss": 0.9996, "step": 2066 }, { "epoch": 0.7502947845804989, "grad_norm": 0.1474609375, "learning_rate": 0.00015185411822298372, "loss": 0.9503, "step": 2068 }, { "epoch": 0.7510204081632653, "grad_norm": 0.1611328125, "learning_rate": 0.0001518051646065353, "loss": 1.0125, "step": 2070 }, { "epoch": 0.7517460317460317, "grad_norm": 0.169921875, "learning_rate": 0.0001517562109900869, "loss": 0.9977, "step": 2072 }, { "epoch": 0.7524716553287982, "grad_norm": 0.146484375, "learning_rate": 0.0001517072573736385, "loss": 1.0281, "step": 2074 }, { "epoch": 0.7531972789115646, "grad_norm": 0.1435546875, "learning_rate": 0.00015165830375719007, "loss": 0.9786, "step": 2076 }, { "epoch": 0.753922902494331, "grad_norm": 0.1611328125, "learning_rate": 0.00015160935014074166, "loss": 0.9559, "step": 2078 }, { "epoch": 0.7546485260770975, "grad_norm": 0.138671875, "learning_rate": 0.00015156039652429324, "loss": 0.9641, "step": 2080 }, { "epoch": 0.7553741496598639, "grad_norm": 0.1474609375, "learning_rate": 0.00015151144290784482, "loss": 0.9441, "step": 2082 }, { "epoch": 0.7560997732426303, "grad_norm": 0.1484375, "learning_rate": 0.0001514624892913964, "loss": 0.999, "step": 2084 }, { "epoch": 0.7568253968253968, "grad_norm": 0.2021484375, "learning_rate": 0.000151413535674948, "loss": 1.0128, "step": 2086 }, { "epoch": 0.7575510204081632, "grad_norm": 0.1640625, "learning_rate": 0.00015136458205849957, "loss": 1.0748, "step": 2088 }, { "epoch": 0.7582766439909298, "grad_norm": 0.1328125, "learning_rate": 0.00015131562844205115, "loss": 0.8835, "step": 2090 }, { "epoch": 0.7590022675736962, "grad_norm": 0.150390625, "learning_rate": 0.00015126667482560276, "loss": 0.9798, "step": 2092 }, { "epoch": 0.7597278911564626, "grad_norm": 0.173828125, "learning_rate": 0.00015121772120915434, "loss": 0.961, "step": 2094 }, { "epoch": 0.7604535147392291, "grad_norm": 0.1416015625, "learning_rate": 0.00015116876759270592, "loss": 0.9972, "step": 2096 }, { "epoch": 0.7611791383219955, "grad_norm": 0.1494140625, "learning_rate": 0.0001511198139762575, "loss": 0.9814, "step": 2098 }, { "epoch": 0.7619047619047619, "grad_norm": 0.1611328125, "learning_rate": 0.0001510708603598091, "loss": 0.9646, "step": 2100 }, { "epoch": 0.7619047619047619, "eval_loss": 1.0186907052993774, "eval_runtime": 95.5861, "eval_samples_per_second": 18.831, "eval_steps_per_second": 0.596, "step": 2100 }, { "epoch": 0.7626303854875284, "grad_norm": 0.1689453125, "learning_rate": 0.00015102190674336067, "loss": 1.0291, "step": 2102 }, { "epoch": 0.7633560090702948, "grad_norm": 0.1650390625, "learning_rate": 0.00015097295312691225, "loss": 0.944, "step": 2104 }, { "epoch": 0.7640816326530612, "grad_norm": 0.1533203125, "learning_rate": 0.00015092399951046386, "loss": 0.9971, "step": 2106 }, { "epoch": 0.7648072562358277, "grad_norm": 0.1494140625, "learning_rate": 0.00015087504589401542, "loss": 0.9816, "step": 2108 }, { "epoch": 0.7655328798185941, "grad_norm": 0.158203125, "learning_rate": 0.000150826092277567, "loss": 1.0105, "step": 2110 }, { "epoch": 0.7662585034013606, "grad_norm": 0.1455078125, "learning_rate": 0.0001507771386611186, "loss": 1.0059, "step": 2112 }, { "epoch": 0.766984126984127, "grad_norm": 0.1728515625, "learning_rate": 0.0001507281850446702, "loss": 0.9797, "step": 2114 }, { "epoch": 0.7677097505668934, "grad_norm": 0.1474609375, "learning_rate": 0.00015067923142822175, "loss": 0.9623, "step": 2116 }, { "epoch": 0.7684353741496599, "grad_norm": 0.17578125, "learning_rate": 0.00015063027781177335, "loss": 1.0352, "step": 2118 }, { "epoch": 0.7691609977324263, "grad_norm": 0.1689453125, "learning_rate": 0.00015058132419532494, "loss": 1.0687, "step": 2120 }, { "epoch": 0.7698866213151927, "grad_norm": 0.13671875, "learning_rate": 0.00015053237057887652, "loss": 0.9323, "step": 2122 }, { "epoch": 0.7706122448979592, "grad_norm": 0.1640625, "learning_rate": 0.0001504834169624281, "loss": 0.9436, "step": 2124 }, { "epoch": 0.7713378684807256, "grad_norm": 0.18359375, "learning_rate": 0.00015043446334597968, "loss": 1.0228, "step": 2126 }, { "epoch": 0.7720634920634921, "grad_norm": 0.150390625, "learning_rate": 0.00015038550972953127, "loss": 0.9885, "step": 2128 }, { "epoch": 0.7727891156462585, "grad_norm": 0.173828125, "learning_rate": 0.00015033655611308285, "loss": 0.9654, "step": 2130 }, { "epoch": 0.7735147392290249, "grad_norm": 0.3046875, "learning_rate": 0.00015028760249663446, "loss": 1.1141, "step": 2132 }, { "epoch": 0.7742403628117914, "grad_norm": 0.140625, "learning_rate": 0.00015023864888018604, "loss": 0.9624, "step": 2134 }, { "epoch": 0.7749659863945578, "grad_norm": 0.1513671875, "learning_rate": 0.0001501896952637376, "loss": 0.9589, "step": 2136 }, { "epoch": 0.7756916099773242, "grad_norm": 0.1826171875, "learning_rate": 0.0001501407416472892, "loss": 1.0223, "step": 2138 }, { "epoch": 0.7764172335600907, "grad_norm": 0.1640625, "learning_rate": 0.0001500917880308408, "loss": 0.9818, "step": 2140 }, { "epoch": 0.7771428571428571, "grad_norm": 0.14453125, "learning_rate": 0.00015004283441439237, "loss": 0.9146, "step": 2142 }, { "epoch": 0.7778684807256235, "grad_norm": 0.1357421875, "learning_rate": 0.00014999388079794395, "loss": 0.9904, "step": 2144 }, { "epoch": 0.77859410430839, "grad_norm": 0.1708984375, "learning_rate": 0.00014994492718149553, "loss": 0.9883, "step": 2146 }, { "epoch": 0.7793197278911564, "grad_norm": 0.1748046875, "learning_rate": 0.00014989597356504712, "loss": 1.0015, "step": 2148 }, { "epoch": 0.780045351473923, "grad_norm": 0.158203125, "learning_rate": 0.0001498470199485987, "loss": 1.0113, "step": 2150 }, { "epoch": 0.7807709750566894, "grad_norm": 0.1416015625, "learning_rate": 0.0001497980663321503, "loss": 0.9982, "step": 2152 }, { "epoch": 0.7814965986394558, "grad_norm": 0.1708984375, "learning_rate": 0.0001497491127157019, "loss": 1.0084, "step": 2154 }, { "epoch": 0.7822222222222223, "grad_norm": 0.146484375, "learning_rate": 0.00014970015909925345, "loss": 0.9613, "step": 2156 }, { "epoch": 0.7829478458049887, "grad_norm": 0.14453125, "learning_rate": 0.00014965120548280505, "loss": 1.0032, "step": 2158 }, { "epoch": 0.7836734693877551, "grad_norm": 0.1455078125, "learning_rate": 0.00014960225186635664, "loss": 0.979, "step": 2160 }, { "epoch": 0.7843990929705216, "grad_norm": 0.1591796875, "learning_rate": 0.00014955329824990822, "loss": 0.9816, "step": 2162 }, { "epoch": 0.785124716553288, "grad_norm": 0.1689453125, "learning_rate": 0.0001495043446334598, "loss": 0.9398, "step": 2164 }, { "epoch": 0.7858503401360544, "grad_norm": 0.150390625, "learning_rate": 0.00014945539101701138, "loss": 1.0336, "step": 2166 }, { "epoch": 0.7865759637188209, "grad_norm": 0.150390625, "learning_rate": 0.00014940643740056297, "loss": 0.9288, "step": 2168 }, { "epoch": 0.7873015873015873, "grad_norm": 0.1435546875, "learning_rate": 0.00014935748378411455, "loss": 1.0003, "step": 2170 }, { "epoch": 0.7880272108843538, "grad_norm": 0.138671875, "learning_rate": 0.00014930853016766616, "loss": 0.9418, "step": 2172 }, { "epoch": 0.7887528344671202, "grad_norm": 0.158203125, "learning_rate": 0.00014925957655121774, "loss": 1.0262, "step": 2174 }, { "epoch": 0.7894784580498866, "grad_norm": 0.146484375, "learning_rate": 0.0001492106229347693, "loss": 1.0496, "step": 2176 }, { "epoch": 0.7902040816326531, "grad_norm": 0.154296875, "learning_rate": 0.0001491616693183209, "loss": 0.937, "step": 2178 }, { "epoch": 0.7909297052154195, "grad_norm": 0.1474609375, "learning_rate": 0.0001491127157018725, "loss": 0.9422, "step": 2180 }, { "epoch": 0.7916553287981859, "grad_norm": 0.1728515625, "learning_rate": 0.00014906376208542407, "loss": 0.9921, "step": 2182 }, { "epoch": 0.7923809523809524, "grad_norm": 0.173828125, "learning_rate": 0.00014901480846897565, "loss": 0.9662, "step": 2184 }, { "epoch": 0.7931065759637188, "grad_norm": 0.1826171875, "learning_rate": 0.00014896585485252723, "loss": 1.0241, "step": 2186 }, { "epoch": 0.7938321995464852, "grad_norm": 0.16015625, "learning_rate": 0.00014891690123607882, "loss": 1.0144, "step": 2188 }, { "epoch": 0.7945578231292517, "grad_norm": 0.1650390625, "learning_rate": 0.0001488679476196304, "loss": 0.9861, "step": 2190 }, { "epoch": 0.7952834467120181, "grad_norm": 0.150390625, "learning_rate": 0.000148818994003182, "loss": 0.9249, "step": 2192 }, { "epoch": 0.7960090702947846, "grad_norm": 0.1435546875, "learning_rate": 0.00014877004038673356, "loss": 0.9389, "step": 2194 }, { "epoch": 0.796734693877551, "grad_norm": 0.1552734375, "learning_rate": 0.00014872108677028515, "loss": 0.9589, "step": 2196 }, { "epoch": 0.7974603174603174, "grad_norm": 0.1572265625, "learning_rate": 0.00014867213315383675, "loss": 1.0548, "step": 2198 }, { "epoch": 0.7981859410430839, "grad_norm": 0.189453125, "learning_rate": 0.00014862317953738834, "loss": 1.0526, "step": 2200 }, { "epoch": 0.7981859410430839, "eval_loss": 1.0182089805603027, "eval_runtime": 95.9846, "eval_samples_per_second": 18.753, "eval_steps_per_second": 0.594, "step": 2200 }, { "epoch": 0.7989115646258503, "grad_norm": 0.1572265625, "learning_rate": 0.00014857422592093992, "loss": 1.0364, "step": 2202 }, { "epoch": 0.7996371882086167, "grad_norm": 0.166015625, "learning_rate": 0.0001485252723044915, "loss": 0.9324, "step": 2204 }, { "epoch": 0.8003628117913832, "grad_norm": 0.1630859375, "learning_rate": 0.00014847631868804308, "loss": 0.936, "step": 2206 }, { "epoch": 0.8010884353741496, "grad_norm": 0.1728515625, "learning_rate": 0.00014842736507159467, "loss": 1.0559, "step": 2208 }, { "epoch": 0.8018140589569162, "grad_norm": 0.1484375, "learning_rate": 0.00014837841145514625, "loss": 1.0164, "step": 2210 }, { "epoch": 0.8025396825396826, "grad_norm": 0.1640625, "learning_rate": 0.00014832945783869786, "loss": 1.0198, "step": 2212 }, { "epoch": 0.803265306122449, "grad_norm": 0.1796875, "learning_rate": 0.0001482805042222494, "loss": 1.0224, "step": 2214 }, { "epoch": 0.8039909297052155, "grad_norm": 0.1396484375, "learning_rate": 0.000148231550605801, "loss": 0.9483, "step": 2216 }, { "epoch": 0.8047165532879819, "grad_norm": 0.1728515625, "learning_rate": 0.0001481825969893526, "loss": 1.0488, "step": 2218 }, { "epoch": 0.8054421768707483, "grad_norm": 0.1591796875, "learning_rate": 0.0001481336433729042, "loss": 0.9822, "step": 2220 }, { "epoch": 0.8061678004535148, "grad_norm": 0.1572265625, "learning_rate": 0.00014808468975645577, "loss": 0.969, "step": 2222 }, { "epoch": 0.8068934240362812, "grad_norm": 0.1669921875, "learning_rate": 0.00014803573614000735, "loss": 0.9803, "step": 2224 }, { "epoch": 0.8076190476190476, "grad_norm": 0.1708984375, "learning_rate": 0.00014798678252355893, "loss": 0.9859, "step": 2226 }, { "epoch": 0.8083446712018141, "grad_norm": 0.142578125, "learning_rate": 0.00014793782890711052, "loss": 1.0081, "step": 2228 }, { "epoch": 0.8090702947845805, "grad_norm": 0.1474609375, "learning_rate": 0.0001478888752906621, "loss": 0.9469, "step": 2230 }, { "epoch": 0.809795918367347, "grad_norm": 0.1474609375, "learning_rate": 0.0001478399216742137, "loss": 1.001, "step": 2232 }, { "epoch": 0.8105215419501134, "grad_norm": 0.1640625, "learning_rate": 0.00014779096805776526, "loss": 1.0429, "step": 2234 }, { "epoch": 0.8112471655328798, "grad_norm": 0.1640625, "learning_rate": 0.00014774201444131684, "loss": 0.9483, "step": 2236 }, { "epoch": 0.8119727891156463, "grad_norm": 0.142578125, "learning_rate": 0.00014769306082486845, "loss": 1.0031, "step": 2238 }, { "epoch": 0.8126984126984127, "grad_norm": 0.1513671875, "learning_rate": 0.00014764410720842004, "loss": 0.9813, "step": 2240 }, { "epoch": 0.8134240362811791, "grad_norm": 0.15625, "learning_rate": 0.00014759515359197162, "loss": 0.986, "step": 2242 }, { "epoch": 0.8141496598639456, "grad_norm": 0.146484375, "learning_rate": 0.0001475461999755232, "loss": 1.0328, "step": 2244 }, { "epoch": 0.814875283446712, "grad_norm": 0.1552734375, "learning_rate": 0.00014749724635907478, "loss": 1.032, "step": 2246 }, { "epoch": 0.8156009070294784, "grad_norm": 0.1376953125, "learning_rate": 0.00014744829274262637, "loss": 0.9657, "step": 2248 }, { "epoch": 0.8163265306122449, "grad_norm": 0.162109375, "learning_rate": 0.00014739933912617795, "loss": 0.9797, "step": 2250 }, { "epoch": 0.8170521541950113, "grad_norm": 0.1826171875, "learning_rate": 0.00014735038550972956, "loss": 0.9919, "step": 2252 }, { "epoch": 0.8177777777777778, "grad_norm": 0.1767578125, "learning_rate": 0.0001473014318932811, "loss": 1.0706, "step": 2254 }, { "epoch": 0.8185034013605442, "grad_norm": 0.1611328125, "learning_rate": 0.0001472524782768327, "loss": 0.9914, "step": 2256 }, { "epoch": 0.8192290249433106, "grad_norm": 0.1572265625, "learning_rate": 0.0001472035246603843, "loss": 1.0764, "step": 2258 }, { "epoch": 0.8199546485260771, "grad_norm": 0.1513671875, "learning_rate": 0.0001471545710439359, "loss": 0.973, "step": 2260 }, { "epoch": 0.8206802721088435, "grad_norm": 0.15234375, "learning_rate": 0.00014710561742748744, "loss": 0.9797, "step": 2262 }, { "epoch": 0.8214058956916099, "grad_norm": 0.150390625, "learning_rate": 0.00014705666381103905, "loss": 0.9685, "step": 2264 }, { "epoch": 0.8221315192743764, "grad_norm": 0.1591796875, "learning_rate": 0.00014700771019459063, "loss": 1.0145, "step": 2266 }, { "epoch": 0.8228571428571428, "grad_norm": 0.1787109375, "learning_rate": 0.00014695875657814222, "loss": 0.9714, "step": 2268 }, { "epoch": 0.8235827664399092, "grad_norm": 0.1572265625, "learning_rate": 0.0001469098029616938, "loss": 0.9954, "step": 2270 }, { "epoch": 0.8243083900226758, "grad_norm": 0.1728515625, "learning_rate": 0.00014686084934524538, "loss": 1.1002, "step": 2272 }, { "epoch": 0.8250340136054422, "grad_norm": 0.1474609375, "learning_rate": 0.00014681189572879696, "loss": 0.9418, "step": 2274 }, { "epoch": 0.8257596371882087, "grad_norm": 0.1357421875, "learning_rate": 0.00014676294211234854, "loss": 1.044, "step": 2276 }, { "epoch": 0.8264852607709751, "grad_norm": 0.1669921875, "learning_rate": 0.00014671398849590015, "loss": 1.0346, "step": 2278 }, { "epoch": 0.8272108843537415, "grad_norm": 0.1904296875, "learning_rate": 0.00014666503487945174, "loss": 1.027, "step": 2280 }, { "epoch": 0.827936507936508, "grad_norm": 0.154296875, "learning_rate": 0.0001466160812630033, "loss": 0.9886, "step": 2282 }, { "epoch": 0.8286621315192744, "grad_norm": 0.1884765625, "learning_rate": 0.0001465671276465549, "loss": 1.0345, "step": 2284 }, { "epoch": 0.8293877551020408, "grad_norm": 0.177734375, "learning_rate": 0.00014651817403010648, "loss": 1.0252, "step": 2286 }, { "epoch": 0.8301133786848073, "grad_norm": 0.1865234375, "learning_rate": 0.00014646922041365807, "loss": 1.0475, "step": 2288 }, { "epoch": 0.8308390022675737, "grad_norm": 0.1806640625, "learning_rate": 0.00014642026679720965, "loss": 1.0286, "step": 2290 }, { "epoch": 0.8315646258503402, "grad_norm": 0.1591796875, "learning_rate": 0.00014637131318076123, "loss": 0.9599, "step": 2292 }, { "epoch": 0.8322902494331066, "grad_norm": 0.166015625, "learning_rate": 0.0001463223595643128, "loss": 1.0277, "step": 2294 }, { "epoch": 0.833015873015873, "grad_norm": 0.15234375, "learning_rate": 0.0001462734059478644, "loss": 0.9699, "step": 2296 }, { "epoch": 0.8337414965986395, "grad_norm": 0.193359375, "learning_rate": 0.000146224452331416, "loss": 0.9157, "step": 2298 }, { "epoch": 0.8344671201814059, "grad_norm": 0.169921875, "learning_rate": 0.00014617549871496759, "loss": 1.0078, "step": 2300 }, { "epoch": 0.8344671201814059, "eval_loss": 1.018009901046753, "eval_runtime": 97.4465, "eval_samples_per_second": 18.472, "eval_steps_per_second": 0.585, "step": 2300 }, { "epoch": 0.8351927437641723, "grad_norm": 0.1376953125, "learning_rate": 0.00014612654509851914, "loss": 0.9975, "step": 2302 }, { "epoch": 0.8359183673469388, "grad_norm": 0.1552734375, "learning_rate": 0.00014607759148207075, "loss": 1.0193, "step": 2304 }, { "epoch": 0.8366439909297052, "grad_norm": 0.166015625, "learning_rate": 0.00014602863786562233, "loss": 0.9847, "step": 2306 }, { "epoch": 0.8373696145124716, "grad_norm": 0.1630859375, "learning_rate": 0.00014597968424917392, "loss": 1.0365, "step": 2308 }, { "epoch": 0.8380952380952381, "grad_norm": 0.1484375, "learning_rate": 0.0001459307306327255, "loss": 0.9204, "step": 2310 }, { "epoch": 0.8388208616780045, "grad_norm": 0.15234375, "learning_rate": 0.00014588177701627708, "loss": 0.973, "step": 2312 }, { "epoch": 0.839546485260771, "grad_norm": 0.1640625, "learning_rate": 0.00014583282339982866, "loss": 1.0135, "step": 2314 }, { "epoch": 0.8402721088435374, "grad_norm": 0.150390625, "learning_rate": 0.00014578386978338024, "loss": 0.9772, "step": 2316 }, { "epoch": 0.8409977324263038, "grad_norm": 0.1484375, "learning_rate": 0.00014573491616693185, "loss": 0.9281, "step": 2318 }, { "epoch": 0.8417233560090703, "grad_norm": 0.1572265625, "learning_rate": 0.00014568596255048344, "loss": 1.0221, "step": 2320 }, { "epoch": 0.8424489795918367, "grad_norm": 0.1826171875, "learning_rate": 0.000145637008934035, "loss": 0.972, "step": 2322 }, { "epoch": 0.8431746031746031, "grad_norm": 0.1728515625, "learning_rate": 0.0001455880553175866, "loss": 1.0392, "step": 2324 }, { "epoch": 0.8439002267573696, "grad_norm": 0.1845703125, "learning_rate": 0.00014553910170113818, "loss": 1.0201, "step": 2326 }, { "epoch": 0.844625850340136, "grad_norm": 0.158203125, "learning_rate": 0.00014549014808468977, "loss": 0.9544, "step": 2328 }, { "epoch": 0.8453514739229024, "grad_norm": 0.1533203125, "learning_rate": 0.00014544119446824135, "loss": 0.9848, "step": 2330 }, { "epoch": 0.846077097505669, "grad_norm": 0.166015625, "learning_rate": 0.00014539224085179293, "loss": 0.9847, "step": 2332 }, { "epoch": 0.8468027210884354, "grad_norm": 0.16796875, "learning_rate": 0.0001453432872353445, "loss": 1.0878, "step": 2334 }, { "epoch": 0.8475283446712019, "grad_norm": 0.162109375, "learning_rate": 0.0001452943336188961, "loss": 0.9742, "step": 2336 }, { "epoch": 0.8482539682539683, "grad_norm": 0.1484375, "learning_rate": 0.0001452453800024477, "loss": 1.06, "step": 2338 }, { "epoch": 0.8489795918367347, "grad_norm": 0.1435546875, "learning_rate": 0.00014519642638599926, "loss": 0.9542, "step": 2340 }, { "epoch": 0.8497052154195012, "grad_norm": 0.1611328125, "learning_rate": 0.00014514747276955084, "loss": 0.9095, "step": 2342 }, { "epoch": 0.8504308390022676, "grad_norm": 0.1689453125, "learning_rate": 0.00014509851915310245, "loss": 1.0327, "step": 2344 }, { "epoch": 0.851156462585034, "grad_norm": 0.14453125, "learning_rate": 0.00014504956553665403, "loss": 0.932, "step": 2346 }, { "epoch": 0.8518820861678005, "grad_norm": 0.1494140625, "learning_rate": 0.00014500061192020562, "loss": 0.9619, "step": 2348 }, { "epoch": 0.8526077097505669, "grad_norm": 0.1533203125, "learning_rate": 0.0001449516583037572, "loss": 0.9473, "step": 2350 }, { "epoch": 0.8533333333333334, "grad_norm": 0.1611328125, "learning_rate": 0.00014490270468730878, "loss": 1.0192, "step": 2352 }, { "epoch": 0.8540589569160998, "grad_norm": 0.1484375, "learning_rate": 0.00014485375107086036, "loss": 0.9686, "step": 2354 }, { "epoch": 0.8547845804988662, "grad_norm": 0.1611328125, "learning_rate": 0.00014480479745441194, "loss": 1.0355, "step": 2356 }, { "epoch": 0.8555102040816327, "grad_norm": 0.169921875, "learning_rate": 0.00014475584383796355, "loss": 0.9756, "step": 2358 }, { "epoch": 0.8562358276643991, "grad_norm": 0.1669921875, "learning_rate": 0.0001447068902215151, "loss": 1.0697, "step": 2360 }, { "epoch": 0.8569614512471655, "grad_norm": 0.1767578125, "learning_rate": 0.0001446579366050667, "loss": 0.9959, "step": 2362 }, { "epoch": 0.857687074829932, "grad_norm": 0.1689453125, "learning_rate": 0.0001446089829886183, "loss": 0.984, "step": 2364 }, { "epoch": 0.8584126984126984, "grad_norm": 0.16015625, "learning_rate": 0.00014456002937216988, "loss": 0.9598, "step": 2366 }, { "epoch": 0.8591383219954648, "grad_norm": 0.166015625, "learning_rate": 0.00014451107575572147, "loss": 1.0026, "step": 2368 }, { "epoch": 0.8598639455782313, "grad_norm": 0.138671875, "learning_rate": 0.00014446212213927305, "loss": 0.9883, "step": 2370 }, { "epoch": 0.8605895691609977, "grad_norm": 0.1611328125, "learning_rate": 0.00014441316852282463, "loss": 1.0565, "step": 2372 }, { "epoch": 0.8613151927437642, "grad_norm": 0.1513671875, "learning_rate": 0.0001443642149063762, "loss": 0.9496, "step": 2374 }, { "epoch": 0.8620408163265306, "grad_norm": 0.1650390625, "learning_rate": 0.0001443152612899278, "loss": 0.961, "step": 2376 }, { "epoch": 0.862766439909297, "grad_norm": 0.1572265625, "learning_rate": 0.0001442663076734794, "loss": 0.9518, "step": 2378 }, { "epoch": 0.8634920634920635, "grad_norm": 0.1455078125, "learning_rate": 0.00014421735405703096, "loss": 0.9264, "step": 2380 }, { "epoch": 0.8642176870748299, "grad_norm": 0.1533203125, "learning_rate": 0.00014416840044058254, "loss": 1.0389, "step": 2382 }, { "epoch": 0.8649433106575963, "grad_norm": 0.130859375, "learning_rate": 0.00014411944682413415, "loss": 1.0283, "step": 2384 }, { "epoch": 0.8656689342403628, "grad_norm": 0.193359375, "learning_rate": 0.00014407049320768573, "loss": 1.0456, "step": 2386 }, { "epoch": 0.8663945578231292, "grad_norm": 0.1484375, "learning_rate": 0.00014402153959123732, "loss": 1.007, "step": 2388 }, { "epoch": 0.8671201814058956, "grad_norm": 0.169921875, "learning_rate": 0.0001439725859747889, "loss": 1.0643, "step": 2390 }, { "epoch": 0.8678458049886622, "grad_norm": 0.14453125, "learning_rate": 0.00014392363235834048, "loss": 1.0052, "step": 2392 }, { "epoch": 0.8685714285714285, "grad_norm": 0.1533203125, "learning_rate": 0.00014387467874189206, "loss": 0.9164, "step": 2394 }, { "epoch": 0.8692970521541951, "grad_norm": 0.181640625, "learning_rate": 0.00014382572512544364, "loss": 0.9513, "step": 2396 }, { "epoch": 0.8700226757369615, "grad_norm": 0.1669921875, "learning_rate": 0.00014377677150899525, "loss": 0.9772, "step": 2398 }, { "epoch": 0.8707482993197279, "grad_norm": 0.1474609375, "learning_rate": 0.0001437278178925468, "loss": 1.0452, "step": 2400 }, { "epoch": 0.8707482993197279, "eval_loss": 1.0175837278366089, "eval_runtime": 93.47, "eval_samples_per_second": 19.258, "eval_steps_per_second": 0.61, "step": 2400 }, { "epoch": 0.8714739229024944, "grad_norm": 0.1650390625, "learning_rate": 0.0001436788642760984, "loss": 0.9611, "step": 2402 }, { "epoch": 0.8721995464852608, "grad_norm": 0.1455078125, "learning_rate": 0.00014362991065965, "loss": 0.9569, "step": 2404 }, { "epoch": 0.8729251700680272, "grad_norm": 0.1513671875, "learning_rate": 0.00014358095704320158, "loss": 1.0252, "step": 2406 }, { "epoch": 0.8736507936507937, "grad_norm": 0.1630859375, "learning_rate": 0.00014353200342675314, "loss": 1.0249, "step": 2408 }, { "epoch": 0.8743764172335601, "grad_norm": 0.1728515625, "learning_rate": 0.00014348304981030475, "loss": 1.034, "step": 2410 }, { "epoch": 0.8751020408163265, "grad_norm": 0.162109375, "learning_rate": 0.00014343409619385633, "loss": 0.9909, "step": 2412 }, { "epoch": 0.875827664399093, "grad_norm": 0.150390625, "learning_rate": 0.0001433851425774079, "loss": 0.9933, "step": 2414 }, { "epoch": 0.8765532879818594, "grad_norm": 0.21875, "learning_rate": 0.0001433361889609595, "loss": 0.8944, "step": 2416 }, { "epoch": 0.8772789115646259, "grad_norm": 0.154296875, "learning_rate": 0.00014328723534451108, "loss": 1.0102, "step": 2418 }, { "epoch": 0.8780045351473923, "grad_norm": 0.1474609375, "learning_rate": 0.00014323828172806266, "loss": 0.959, "step": 2420 }, { "epoch": 0.8787301587301587, "grad_norm": 0.1328125, "learning_rate": 0.00014318932811161424, "loss": 0.9532, "step": 2422 }, { "epoch": 0.8794557823129252, "grad_norm": 0.14453125, "learning_rate": 0.00014314037449516585, "loss": 0.9811, "step": 2424 }, { "epoch": 0.8801814058956916, "grad_norm": 0.154296875, "learning_rate": 0.00014309142087871743, "loss": 0.9697, "step": 2426 }, { "epoch": 0.880907029478458, "grad_norm": 0.18359375, "learning_rate": 0.000143042467262269, "loss": 1.0135, "step": 2428 }, { "epoch": 0.8816326530612245, "grad_norm": 0.1533203125, "learning_rate": 0.0001429935136458206, "loss": 1.0377, "step": 2430 }, { "epoch": 0.8823582766439909, "grad_norm": 0.1474609375, "learning_rate": 0.00014294456002937218, "loss": 0.9046, "step": 2432 }, { "epoch": 0.8830839002267574, "grad_norm": 0.1630859375, "learning_rate": 0.00014289560641292376, "loss": 0.9787, "step": 2434 }, { "epoch": 0.8838095238095238, "grad_norm": 0.15625, "learning_rate": 0.00014284665279647534, "loss": 1.0511, "step": 2436 }, { "epoch": 0.8845351473922902, "grad_norm": 0.15625, "learning_rate": 0.00014279769918002693, "loss": 0.9777, "step": 2438 }, { "epoch": 0.8852607709750567, "grad_norm": 0.1669921875, "learning_rate": 0.0001427487455635785, "loss": 0.9929, "step": 2440 }, { "epoch": 0.8859863945578231, "grad_norm": 0.1484375, "learning_rate": 0.0001426997919471301, "loss": 1.077, "step": 2442 }, { "epoch": 0.8867120181405895, "grad_norm": 0.1455078125, "learning_rate": 0.0001426508383306817, "loss": 1.0122, "step": 2444 }, { "epoch": 0.887437641723356, "grad_norm": 0.13671875, "learning_rate": 0.00014260188471423328, "loss": 0.9913, "step": 2446 }, { "epoch": 0.8881632653061224, "grad_norm": 0.15234375, "learning_rate": 0.00014255293109778484, "loss": 0.9736, "step": 2448 }, { "epoch": 0.8888888888888888, "grad_norm": 0.162109375, "learning_rate": 0.00014250397748133645, "loss": 0.9983, "step": 2450 }, { "epoch": 0.8896145124716553, "grad_norm": 0.1826171875, "learning_rate": 0.00014245502386488803, "loss": 1.033, "step": 2452 }, { "epoch": 0.8903401360544217, "grad_norm": 0.154296875, "learning_rate": 0.0001424060702484396, "loss": 1.0804, "step": 2454 }, { "epoch": 0.8910657596371883, "grad_norm": 0.17578125, "learning_rate": 0.0001423571166319912, "loss": 0.9584, "step": 2456 }, { "epoch": 0.8917913832199547, "grad_norm": 0.1728515625, "learning_rate": 0.00014230816301554278, "loss": 1.0151, "step": 2458 }, { "epoch": 0.8925170068027211, "grad_norm": 0.169921875, "learning_rate": 0.00014225920939909436, "loss": 1.0216, "step": 2460 }, { "epoch": 0.8932426303854876, "grad_norm": 0.1494140625, "learning_rate": 0.00014221025578264594, "loss": 1.0113, "step": 2462 }, { "epoch": 0.893968253968254, "grad_norm": 0.134765625, "learning_rate": 0.00014216130216619755, "loss": 1.0047, "step": 2464 }, { "epoch": 0.8946938775510204, "grad_norm": 0.162109375, "learning_rate": 0.00014211234854974913, "loss": 0.9717, "step": 2466 }, { "epoch": 0.8954195011337869, "grad_norm": 0.1484375, "learning_rate": 0.0001420633949333007, "loss": 1.0086, "step": 2468 }, { "epoch": 0.8961451247165533, "grad_norm": 0.1455078125, "learning_rate": 0.0001420144413168523, "loss": 0.9738, "step": 2470 }, { "epoch": 0.8968707482993197, "grad_norm": 0.1552734375, "learning_rate": 0.00014196548770040388, "loss": 1.0254, "step": 2472 }, { "epoch": 0.8975963718820862, "grad_norm": 0.154296875, "learning_rate": 0.00014191653408395546, "loss": 0.9363, "step": 2474 }, { "epoch": 0.8983219954648526, "grad_norm": 0.1591796875, "learning_rate": 0.00014186758046750704, "loss": 1.0181, "step": 2476 }, { "epoch": 0.8990476190476191, "grad_norm": 0.1650390625, "learning_rate": 0.00014181862685105863, "loss": 0.9986, "step": 2478 }, { "epoch": 0.8997732426303855, "grad_norm": 0.1767578125, "learning_rate": 0.0001417696732346102, "loss": 1.0503, "step": 2480 }, { "epoch": 0.9004988662131519, "grad_norm": 0.1591796875, "learning_rate": 0.0001417207196181618, "loss": 1.0875, "step": 2482 }, { "epoch": 0.9012244897959184, "grad_norm": 0.1611328125, "learning_rate": 0.0001416717660017134, "loss": 1.0045, "step": 2484 }, { "epoch": 0.9019501133786848, "grad_norm": 0.150390625, "learning_rate": 0.00014162281238526496, "loss": 1.0325, "step": 2486 }, { "epoch": 0.9026757369614512, "grad_norm": 0.1650390625, "learning_rate": 0.00014157385876881654, "loss": 0.9717, "step": 2488 }, { "epoch": 0.9034013605442177, "grad_norm": 0.1396484375, "learning_rate": 0.00014152490515236815, "loss": 1.0202, "step": 2490 }, { "epoch": 0.9041269841269841, "grad_norm": 0.1708984375, "learning_rate": 0.00014147595153591973, "loss": 0.9743, "step": 2492 }, { "epoch": 0.9048526077097505, "grad_norm": 0.162109375, "learning_rate": 0.0001414269979194713, "loss": 0.9956, "step": 2494 }, { "epoch": 0.905578231292517, "grad_norm": 0.16796875, "learning_rate": 0.0001413780443030229, "loss": 0.9782, "step": 2496 }, { "epoch": 0.9063038548752834, "grad_norm": 0.16796875, "learning_rate": 0.00014132909068657448, "loss": 1.0001, "step": 2498 }, { "epoch": 0.9070294784580499, "grad_norm": 0.1396484375, "learning_rate": 0.00014128013707012606, "loss": 0.9439, "step": 2500 }, { "epoch": 0.9070294784580499, "eval_loss": 1.0171304941177368, "eval_runtime": 93.0998, "eval_samples_per_second": 19.334, "eval_steps_per_second": 0.612, "step": 2500 }, { "epoch": 0.9077551020408163, "grad_norm": 0.181640625, "learning_rate": 0.00014123118345367764, "loss": 1.0005, "step": 2502 }, { "epoch": 0.9084807256235827, "grad_norm": 0.1591796875, "learning_rate": 0.00014118222983722925, "loss": 0.9762, "step": 2504 }, { "epoch": 0.9092063492063492, "grad_norm": 0.1455078125, "learning_rate": 0.0001411332762207808, "loss": 1.0189, "step": 2506 }, { "epoch": 0.9099319727891156, "grad_norm": 0.1591796875, "learning_rate": 0.0001410843226043324, "loss": 1.0226, "step": 2508 }, { "epoch": 0.910657596371882, "grad_norm": 0.1630859375, "learning_rate": 0.000141035368987884, "loss": 1.008, "step": 2510 }, { "epoch": 0.9113832199546485, "grad_norm": 0.1572265625, "learning_rate": 0.00014098641537143558, "loss": 0.9309, "step": 2512 }, { "epoch": 0.912108843537415, "grad_norm": 0.15625, "learning_rate": 0.00014093746175498716, "loss": 0.9321, "step": 2514 }, { "epoch": 0.9128344671201815, "grad_norm": 0.1689453125, "learning_rate": 0.00014088850813853874, "loss": 1.0389, "step": 2516 }, { "epoch": 0.9135600907029479, "grad_norm": 0.166015625, "learning_rate": 0.00014083955452209033, "loss": 0.9912, "step": 2518 }, { "epoch": 0.9142857142857143, "grad_norm": 0.1767578125, "learning_rate": 0.0001407906009056419, "loss": 1.0086, "step": 2520 }, { "epoch": 0.9150113378684808, "grad_norm": 0.1630859375, "learning_rate": 0.0001407416472891935, "loss": 1.0779, "step": 2522 }, { "epoch": 0.9157369614512472, "grad_norm": 0.1669921875, "learning_rate": 0.0001406926936727451, "loss": 1.0576, "step": 2524 }, { "epoch": 0.9164625850340136, "grad_norm": 0.1533203125, "learning_rate": 0.00014064374005629666, "loss": 1.0101, "step": 2526 }, { "epoch": 0.9171882086167801, "grad_norm": 0.162109375, "learning_rate": 0.00014059478643984824, "loss": 1.0382, "step": 2528 }, { "epoch": 0.9179138321995465, "grad_norm": 0.1748046875, "learning_rate": 0.00014054583282339985, "loss": 1.0193, "step": 2530 }, { "epoch": 0.9186394557823129, "grad_norm": 0.1484375, "learning_rate": 0.00014049687920695143, "loss": 1.0049, "step": 2532 }, { "epoch": 0.9193650793650794, "grad_norm": 0.162109375, "learning_rate": 0.000140447925590503, "loss": 0.9785, "step": 2534 }, { "epoch": 0.9200907029478458, "grad_norm": 0.1552734375, "learning_rate": 0.0001403989719740546, "loss": 0.9122, "step": 2536 }, { "epoch": 0.9208163265306123, "grad_norm": 0.146484375, "learning_rate": 0.00014035001835760618, "loss": 0.9331, "step": 2538 }, { "epoch": 0.9215419501133787, "grad_norm": 0.14453125, "learning_rate": 0.00014030106474115776, "loss": 1.0167, "step": 2540 }, { "epoch": 0.9222675736961451, "grad_norm": 0.154296875, "learning_rate": 0.00014025211112470934, "loss": 0.9788, "step": 2542 }, { "epoch": 0.9229931972789116, "grad_norm": 0.169921875, "learning_rate": 0.00014020315750826095, "loss": 1.0077, "step": 2544 }, { "epoch": 0.923718820861678, "grad_norm": 0.154296875, "learning_rate": 0.0001401542038918125, "loss": 1.0014, "step": 2546 }, { "epoch": 0.9244444444444444, "grad_norm": 0.15234375, "learning_rate": 0.0001401052502753641, "loss": 0.9479, "step": 2548 }, { "epoch": 0.9251700680272109, "grad_norm": 0.1708984375, "learning_rate": 0.0001400562966589157, "loss": 1.0187, "step": 2550 }, { "epoch": 0.9258956916099773, "grad_norm": 0.1494140625, "learning_rate": 0.00014000734304246728, "loss": 0.9687, "step": 2552 }, { "epoch": 0.9266213151927437, "grad_norm": 0.162109375, "learning_rate": 0.00013995838942601883, "loss": 0.9767, "step": 2554 }, { "epoch": 0.9273469387755102, "grad_norm": 0.146484375, "learning_rate": 0.00013990943580957044, "loss": 1.0263, "step": 2556 }, { "epoch": 0.9280725623582766, "grad_norm": 0.1591796875, "learning_rate": 0.00013986048219312203, "loss": 0.979, "step": 2558 }, { "epoch": 0.9287981859410431, "grad_norm": 0.1455078125, "learning_rate": 0.0001398115285766736, "loss": 1.0121, "step": 2560 }, { "epoch": 0.9295238095238095, "grad_norm": 0.169921875, "learning_rate": 0.0001397625749602252, "loss": 0.9887, "step": 2562 }, { "epoch": 0.9302494331065759, "grad_norm": 0.1484375, "learning_rate": 0.00013971362134377677, "loss": 0.949, "step": 2564 }, { "epoch": 0.9309750566893424, "grad_norm": 0.171875, "learning_rate": 0.00013966466772732835, "loss": 0.9792, "step": 2566 }, { "epoch": 0.9317006802721088, "grad_norm": 0.1708984375, "learning_rate": 0.00013961571411087994, "loss": 0.9836, "step": 2568 }, { "epoch": 0.9324263038548752, "grad_norm": 0.1513671875, "learning_rate": 0.00013956676049443155, "loss": 0.9845, "step": 2570 }, { "epoch": 0.9331519274376417, "grad_norm": 0.1533203125, "learning_rate": 0.00013951780687798313, "loss": 0.9387, "step": 2572 }, { "epoch": 0.9338775510204081, "grad_norm": 0.14453125, "learning_rate": 0.00013946885326153468, "loss": 1.0167, "step": 2574 }, { "epoch": 0.9346031746031747, "grad_norm": 0.1435546875, "learning_rate": 0.0001394198996450863, "loss": 1.0307, "step": 2576 }, { "epoch": 0.9353287981859411, "grad_norm": 0.173828125, "learning_rate": 0.00013937094602863788, "loss": 1.0411, "step": 2578 }, { "epoch": 0.9360544217687075, "grad_norm": 0.15234375, "learning_rate": 0.00013932199241218946, "loss": 0.9697, "step": 2580 }, { "epoch": 0.936780045351474, "grad_norm": 0.142578125, "learning_rate": 0.00013927303879574104, "loss": 0.9579, "step": 2582 }, { "epoch": 0.9375056689342404, "grad_norm": 0.1435546875, "learning_rate": 0.00013922408517929262, "loss": 0.9208, "step": 2584 }, { "epoch": 0.9382312925170068, "grad_norm": 0.158203125, "learning_rate": 0.0001391751315628442, "loss": 1.0053, "step": 2586 }, { "epoch": 0.9389569160997733, "grad_norm": 0.1455078125, "learning_rate": 0.0001391261779463958, "loss": 0.9873, "step": 2588 }, { "epoch": 0.9396825396825397, "grad_norm": 0.140625, "learning_rate": 0.0001390772243299474, "loss": 0.9857, "step": 2590 }, { "epoch": 0.9404081632653061, "grad_norm": 0.1484375, "learning_rate": 0.00013902827071349898, "loss": 1.0317, "step": 2592 }, { "epoch": 0.9411337868480726, "grad_norm": 0.1455078125, "learning_rate": 0.00013897931709705053, "loss": 0.939, "step": 2594 }, { "epoch": 0.941859410430839, "grad_norm": 0.1708984375, "learning_rate": 0.00013893036348060214, "loss": 1.006, "step": 2596 }, { "epoch": 0.9425850340136055, "grad_norm": 0.158203125, "learning_rate": 0.00013888140986415373, "loss": 1.083, "step": 2598 }, { "epoch": 0.9433106575963719, "grad_norm": 0.146484375, "learning_rate": 0.0001388324562477053, "loss": 1.0546, "step": 2600 }, { "epoch": 0.9433106575963719, "eval_loss": 1.016438603401184, "eval_runtime": 97.3894, "eval_samples_per_second": 18.482, "eval_steps_per_second": 0.585, "step": 2600 }, { "epoch": 0.9440362811791383, "grad_norm": 0.1591796875, "learning_rate": 0.0001387835026312569, "loss": 0.9987, "step": 2602 }, { "epoch": 0.9447619047619048, "grad_norm": 0.158203125, "learning_rate": 0.00013873454901480847, "loss": 0.9945, "step": 2604 }, { "epoch": 0.9454875283446712, "grad_norm": 0.15625, "learning_rate": 0.00013868559539836005, "loss": 1.0031, "step": 2606 }, { "epoch": 0.9462131519274376, "grad_norm": 0.1552734375, "learning_rate": 0.00013863664178191164, "loss": 1.0133, "step": 2608 }, { "epoch": 0.9469387755102041, "grad_norm": 0.142578125, "learning_rate": 0.00013858768816546325, "loss": 1.0281, "step": 2610 }, { "epoch": 0.9476643990929705, "grad_norm": 0.1689453125, "learning_rate": 0.00013853873454901483, "loss": 1.0062, "step": 2612 }, { "epoch": 0.9483900226757369, "grad_norm": 0.150390625, "learning_rate": 0.00013848978093256638, "loss": 0.9394, "step": 2614 }, { "epoch": 0.9491156462585034, "grad_norm": 0.1591796875, "learning_rate": 0.000138440827316118, "loss": 0.978, "step": 2616 }, { "epoch": 0.9498412698412698, "grad_norm": 0.1513671875, "learning_rate": 0.00013839187369966958, "loss": 0.9216, "step": 2618 }, { "epoch": 0.9505668934240363, "grad_norm": 0.140625, "learning_rate": 0.00013834292008322116, "loss": 0.9806, "step": 2620 }, { "epoch": 0.9512925170068027, "grad_norm": 0.16796875, "learning_rate": 0.00013829396646677274, "loss": 1.0237, "step": 2622 }, { "epoch": 0.9520181405895691, "grad_norm": 0.1455078125, "learning_rate": 0.00013824501285032432, "loss": 0.9156, "step": 2624 }, { "epoch": 0.9527437641723356, "grad_norm": 0.140625, "learning_rate": 0.0001381960592338759, "loss": 0.9625, "step": 2626 }, { "epoch": 0.953469387755102, "grad_norm": 0.169921875, "learning_rate": 0.0001381471056174275, "loss": 1.0259, "step": 2628 }, { "epoch": 0.9541950113378684, "grad_norm": 0.140625, "learning_rate": 0.0001380981520009791, "loss": 0.921, "step": 2630 }, { "epoch": 0.954920634920635, "grad_norm": 0.1533203125, "learning_rate": 0.00013804919838453065, "loss": 0.9542, "step": 2632 }, { "epoch": 0.9556462585034013, "grad_norm": 0.1806640625, "learning_rate": 0.00013800024476808223, "loss": 1.0174, "step": 2634 }, { "epoch": 0.9563718820861677, "grad_norm": 0.169921875, "learning_rate": 0.00013795129115163384, "loss": 1.0429, "step": 2636 }, { "epoch": 0.9570975056689343, "grad_norm": 0.15625, "learning_rate": 0.00013790233753518543, "loss": 0.9952, "step": 2638 }, { "epoch": 0.9578231292517007, "grad_norm": 0.1337890625, "learning_rate": 0.000137853383918737, "loss": 0.957, "step": 2640 }, { "epoch": 0.9585487528344672, "grad_norm": 0.1884765625, "learning_rate": 0.0001378044303022886, "loss": 1.0364, "step": 2642 }, { "epoch": 0.9592743764172336, "grad_norm": 0.150390625, "learning_rate": 0.00013775547668584017, "loss": 0.9722, "step": 2644 }, { "epoch": 0.96, "grad_norm": 0.130859375, "learning_rate": 0.00013770652306939175, "loss": 0.8664, "step": 2646 }, { "epoch": 0.9607256235827665, "grad_norm": 0.154296875, "learning_rate": 0.00013765756945294334, "loss": 1.0287, "step": 2648 }, { "epoch": 0.9614512471655329, "grad_norm": 0.1630859375, "learning_rate": 0.00013760861583649495, "loss": 0.9959, "step": 2650 }, { "epoch": 0.9621768707482993, "grad_norm": 0.15625, "learning_rate": 0.0001375596622200465, "loss": 0.9511, "step": 2652 }, { "epoch": 0.9629024943310658, "grad_norm": 0.150390625, "learning_rate": 0.00013751070860359808, "loss": 0.9145, "step": 2654 }, { "epoch": 0.9636281179138322, "grad_norm": 0.15234375, "learning_rate": 0.0001374617549871497, "loss": 1.0201, "step": 2656 }, { "epoch": 0.9643537414965987, "grad_norm": 0.1494140625, "learning_rate": 0.00013741280137070128, "loss": 0.9204, "step": 2658 }, { "epoch": 0.9650793650793651, "grad_norm": 0.1787109375, "learning_rate": 0.00013736384775425286, "loss": 0.9881, "step": 2660 }, { "epoch": 0.9658049886621315, "grad_norm": 0.14453125, "learning_rate": 0.00013731489413780444, "loss": 1.0235, "step": 2662 }, { "epoch": 0.966530612244898, "grad_norm": 0.142578125, "learning_rate": 0.00013726594052135602, "loss": 0.9197, "step": 2664 }, { "epoch": 0.9672562358276644, "grad_norm": 0.1650390625, "learning_rate": 0.0001372169869049076, "loss": 0.957, "step": 2666 }, { "epoch": 0.9679818594104308, "grad_norm": 0.1728515625, "learning_rate": 0.0001371680332884592, "loss": 1.0025, "step": 2668 }, { "epoch": 0.9687074829931973, "grad_norm": 0.1357421875, "learning_rate": 0.0001371190796720108, "loss": 0.9623, "step": 2670 }, { "epoch": 0.9694331065759637, "grad_norm": 0.1513671875, "learning_rate": 0.00013707012605556235, "loss": 0.9854, "step": 2672 }, { "epoch": 0.9701587301587301, "grad_norm": 0.1494140625, "learning_rate": 0.00013702117243911393, "loss": 0.9908, "step": 2674 }, { "epoch": 0.9708843537414966, "grad_norm": 0.1552734375, "learning_rate": 0.00013697221882266554, "loss": 0.9643, "step": 2676 }, { "epoch": 0.971609977324263, "grad_norm": 0.140625, "learning_rate": 0.00013692326520621713, "loss": 1.0681, "step": 2678 }, { "epoch": 0.9723356009070295, "grad_norm": 0.16015625, "learning_rate": 0.0001368743115897687, "loss": 0.9858, "step": 2680 }, { "epoch": 0.9730612244897959, "grad_norm": 0.1474609375, "learning_rate": 0.0001368253579733203, "loss": 0.9974, "step": 2682 }, { "epoch": 0.9737868480725623, "grad_norm": 0.150390625, "learning_rate": 0.00013677640435687187, "loss": 0.9672, "step": 2684 }, { "epoch": 0.9745124716553288, "grad_norm": 0.166015625, "learning_rate": 0.00013672745074042345, "loss": 0.9644, "step": 2686 }, { "epoch": 0.9752380952380952, "grad_norm": 0.142578125, "learning_rate": 0.00013667849712397504, "loss": 0.9687, "step": 2688 }, { "epoch": 0.9759637188208616, "grad_norm": 0.1435546875, "learning_rate": 0.00013662954350752665, "loss": 0.9365, "step": 2690 }, { "epoch": 0.9766893424036281, "grad_norm": 0.1494140625, "learning_rate": 0.0001365805898910782, "loss": 1.0098, "step": 2692 }, { "epoch": 0.9774149659863945, "grad_norm": 0.1767578125, "learning_rate": 0.00013653163627462978, "loss": 1.0377, "step": 2694 }, { "epoch": 0.978140589569161, "grad_norm": 0.1513671875, "learning_rate": 0.0001364826826581814, "loss": 0.9091, "step": 2696 }, { "epoch": 0.9788662131519275, "grad_norm": 0.1435546875, "learning_rate": 0.00013643372904173298, "loss": 1.0475, "step": 2698 }, { "epoch": 0.9795918367346939, "grad_norm": 0.1474609375, "learning_rate": 0.00013638477542528453, "loss": 0.9671, "step": 2700 }, { "epoch": 0.9795918367346939, "eval_loss": 1.0159815549850464, "eval_runtime": 97.2766, "eval_samples_per_second": 18.504, "eval_steps_per_second": 0.586, "step": 2700 }, { "epoch": 0.9803174603174604, "grad_norm": 0.1494140625, "learning_rate": 0.00013633582180883614, "loss": 0.9728, "step": 2702 }, { "epoch": 0.9810430839002268, "grad_norm": 0.1640625, "learning_rate": 0.00013628686819238772, "loss": 1.0255, "step": 2704 }, { "epoch": 0.9817687074829932, "grad_norm": 0.1552734375, "learning_rate": 0.0001362379145759393, "loss": 1.0246, "step": 2706 }, { "epoch": 0.9824943310657597, "grad_norm": 0.1640625, "learning_rate": 0.00013618896095949089, "loss": 0.9583, "step": 2708 }, { "epoch": 0.9832199546485261, "grad_norm": 0.158203125, "learning_rate": 0.00013614000734304247, "loss": 0.9655, "step": 2710 }, { "epoch": 0.9839455782312925, "grad_norm": 0.15234375, "learning_rate": 0.00013609105372659405, "loss": 0.9796, "step": 2712 }, { "epoch": 0.984671201814059, "grad_norm": 0.14453125, "learning_rate": 0.00013604210011014563, "loss": 0.9351, "step": 2714 }, { "epoch": 0.9853968253968254, "grad_norm": 0.201171875, "learning_rate": 0.00013599314649369724, "loss": 1.0364, "step": 2716 }, { "epoch": 0.9861224489795918, "grad_norm": 0.1357421875, "learning_rate": 0.00013594419287724882, "loss": 1.031, "step": 2718 }, { "epoch": 0.9868480725623583, "grad_norm": 0.13671875, "learning_rate": 0.00013589523926080038, "loss": 0.9863, "step": 2720 }, { "epoch": 0.9875736961451247, "grad_norm": 0.1640625, "learning_rate": 0.000135846285644352, "loss": 0.9365, "step": 2722 }, { "epoch": 0.9882993197278912, "grad_norm": 0.177734375, "learning_rate": 0.00013579733202790357, "loss": 1.0187, "step": 2724 }, { "epoch": 0.9890249433106576, "grad_norm": 0.138671875, "learning_rate": 0.00013574837841145515, "loss": 1.0717, "step": 2726 }, { "epoch": 0.989750566893424, "grad_norm": 0.162109375, "learning_rate": 0.00013569942479500674, "loss": 1.0162, "step": 2728 }, { "epoch": 0.9904761904761905, "grad_norm": 0.181640625, "learning_rate": 0.00013565047117855832, "loss": 1.0003, "step": 2730 }, { "epoch": 0.9912018140589569, "grad_norm": 0.14453125, "learning_rate": 0.0001356015175621099, "loss": 0.9634, "step": 2732 }, { "epoch": 0.9919274376417233, "grad_norm": 0.13671875, "learning_rate": 0.00013555256394566148, "loss": 0.9042, "step": 2734 }, { "epoch": 0.9926530612244898, "grad_norm": 0.1494140625, "learning_rate": 0.0001355036103292131, "loss": 1.0227, "step": 2736 }, { "epoch": 0.9933786848072562, "grad_norm": 0.154296875, "learning_rate": 0.00013545465671276467, "loss": 0.9887, "step": 2738 }, { "epoch": 0.9941043083900227, "grad_norm": 0.1748046875, "learning_rate": 0.00013540570309631623, "loss": 0.9897, "step": 2740 }, { "epoch": 0.9948299319727891, "grad_norm": 0.1650390625, "learning_rate": 0.00013535674947986784, "loss": 0.9965, "step": 2742 }, { "epoch": 0.9955555555555555, "grad_norm": 0.1630859375, "learning_rate": 0.00013530779586341942, "loss": 0.9902, "step": 2744 }, { "epoch": 0.996281179138322, "grad_norm": 0.1416015625, "learning_rate": 0.000135258842246971, "loss": 1.0522, "step": 2746 }, { "epoch": 0.9970068027210884, "grad_norm": 0.158203125, "learning_rate": 0.00013520988863052259, "loss": 0.9595, "step": 2748 }, { "epoch": 0.9977324263038548, "grad_norm": 0.1494140625, "learning_rate": 0.00013516093501407417, "loss": 0.9819, "step": 2750 }, { "epoch": 0.9984580498866213, "grad_norm": 0.185546875, "learning_rate": 0.00013511198139762575, "loss": 1.1001, "step": 2752 }, { "epoch": 0.9991836734693877, "grad_norm": 0.150390625, "learning_rate": 0.00013506302778117733, "loss": 0.9502, "step": 2754 }, { "epoch": 0.9999092970521541, "grad_norm": 0.15625, "learning_rate": 0.00013501407416472894, "loss": 0.9582, "step": 2756 }, { "epoch": 1.0003628117913832, "grad_norm": 0.1494140625, "learning_rate": 0.00013496512054828052, "loss": 0.9314, "step": 2758 }, { "epoch": 1.0010884353741496, "grad_norm": 0.162109375, "learning_rate": 0.00013491616693183208, "loss": 0.9523, "step": 2760 }, { "epoch": 1.001814058956916, "grad_norm": 0.1533203125, "learning_rate": 0.0001348672133153837, "loss": 1.0069, "step": 2762 }, { "epoch": 1.0025396825396826, "grad_norm": 0.1611328125, "learning_rate": 0.00013481825969893527, "loss": 0.9828, "step": 2764 }, { "epoch": 1.003265306122449, "grad_norm": 0.169921875, "learning_rate": 0.00013476930608248685, "loss": 1.0188, "step": 2766 }, { "epoch": 1.0039909297052154, "grad_norm": 0.1611328125, "learning_rate": 0.00013472035246603844, "loss": 0.9474, "step": 2768 }, { "epoch": 1.0047165532879818, "grad_norm": 0.1455078125, "learning_rate": 0.00013467139884959002, "loss": 0.9772, "step": 2770 }, { "epoch": 1.0054421768707482, "grad_norm": 0.154296875, "learning_rate": 0.0001346224452331416, "loss": 1.0774, "step": 2772 }, { "epoch": 1.0061678004535148, "grad_norm": 0.1435546875, "learning_rate": 0.00013457349161669318, "loss": 0.9454, "step": 2774 }, { "epoch": 1.0068934240362812, "grad_norm": 0.158203125, "learning_rate": 0.0001345245380002448, "loss": 0.9687, "step": 2776 }, { "epoch": 1.0076190476190476, "grad_norm": 0.1484375, "learning_rate": 0.00013447558438379635, "loss": 1.0119, "step": 2778 }, { "epoch": 1.008344671201814, "grad_norm": 0.1689453125, "learning_rate": 0.00013442663076734793, "loss": 1.0194, "step": 2780 }, { "epoch": 1.0090702947845804, "grad_norm": 0.1396484375, "learning_rate": 0.00013437767715089954, "loss": 0.9619, "step": 2782 }, { "epoch": 1.0097959183673468, "grad_norm": 0.1669921875, "learning_rate": 0.00013432872353445112, "loss": 0.9702, "step": 2784 }, { "epoch": 1.0105215419501135, "grad_norm": 0.1376953125, "learning_rate": 0.0001342797699180027, "loss": 0.9513, "step": 2786 }, { "epoch": 1.0112471655328799, "grad_norm": 0.1494140625, "learning_rate": 0.00013423081630155429, "loss": 1.0061, "step": 2788 }, { "epoch": 1.0119727891156463, "grad_norm": 0.1357421875, "learning_rate": 0.00013418186268510587, "loss": 0.9964, "step": 2790 }, { "epoch": 1.0126984126984127, "grad_norm": 0.1640625, "learning_rate": 0.00013413290906865745, "loss": 0.9677, "step": 2792 }, { "epoch": 1.013424036281179, "grad_norm": 0.1494140625, "learning_rate": 0.00013408395545220903, "loss": 0.9397, "step": 2794 }, { "epoch": 1.0141496598639457, "grad_norm": 0.1474609375, "learning_rate": 0.00013403500183576064, "loss": 0.9438, "step": 2796 }, { "epoch": 1.014875283446712, "grad_norm": 0.1591796875, "learning_rate": 0.0001339860482193122, "loss": 0.9445, "step": 2798 }, { "epoch": 1.0156009070294785, "grad_norm": 0.1591796875, "learning_rate": 0.00013393709460286378, "loss": 0.9987, "step": 2800 }, { "epoch": 1.0156009070294785, "eval_loss": 1.0157527923583984, "eval_runtime": 93.9867, "eval_samples_per_second": 19.152, "eval_steps_per_second": 0.606, "step": 2800 }, { "epoch": 1.0163265306122449, "grad_norm": 0.1748046875, "learning_rate": 0.0001338881409864154, "loss": 1.0336, "step": 2802 }, { "epoch": 1.0170521541950113, "grad_norm": 0.162109375, "learning_rate": 0.00013383918736996697, "loss": 1.0242, "step": 2804 }, { "epoch": 1.0177777777777777, "grad_norm": 0.154296875, "learning_rate": 0.00013379023375351855, "loss": 0.9604, "step": 2806 }, { "epoch": 1.0185034013605443, "grad_norm": 0.146484375, "learning_rate": 0.00013374128013707014, "loss": 1.0008, "step": 2808 }, { "epoch": 1.0192290249433107, "grad_norm": 0.162109375, "learning_rate": 0.00013369232652062172, "loss": 0.9853, "step": 2810 }, { "epoch": 1.019954648526077, "grad_norm": 0.1748046875, "learning_rate": 0.0001336433729041733, "loss": 1.1074, "step": 2812 }, { "epoch": 1.0206802721088435, "grad_norm": 0.1435546875, "learning_rate": 0.00013359441928772488, "loss": 1.0543, "step": 2814 }, { "epoch": 1.0214058956916099, "grad_norm": 0.1494140625, "learning_rate": 0.0001335454656712765, "loss": 0.9545, "step": 2816 }, { "epoch": 1.0221315192743765, "grad_norm": 0.162109375, "learning_rate": 0.00013349651205482805, "loss": 0.9777, "step": 2818 }, { "epoch": 1.022857142857143, "grad_norm": 0.150390625, "learning_rate": 0.00013344755843837963, "loss": 1.002, "step": 2820 }, { "epoch": 1.0235827664399093, "grad_norm": 0.158203125, "learning_rate": 0.00013339860482193124, "loss": 0.9998, "step": 2822 }, { "epoch": 1.0243083900226757, "grad_norm": 0.1669921875, "learning_rate": 0.00013334965120548282, "loss": 0.9978, "step": 2824 }, { "epoch": 1.025034013605442, "grad_norm": 0.1494140625, "learning_rate": 0.0001333006975890344, "loss": 1.0295, "step": 2826 }, { "epoch": 1.0257596371882087, "grad_norm": 0.1533203125, "learning_rate": 0.00013325174397258599, "loss": 1.0145, "step": 2828 }, { "epoch": 1.0264852607709751, "grad_norm": 0.1708984375, "learning_rate": 0.00013320279035613757, "loss": 0.9537, "step": 2830 }, { "epoch": 1.0272108843537415, "grad_norm": 0.1455078125, "learning_rate": 0.00013315383673968915, "loss": 1.091, "step": 2832 }, { "epoch": 1.027936507936508, "grad_norm": 0.150390625, "learning_rate": 0.00013310488312324073, "loss": 1.0578, "step": 2834 }, { "epoch": 1.0286621315192743, "grad_norm": 0.15234375, "learning_rate": 0.00013305592950679234, "loss": 0.9262, "step": 2836 }, { "epoch": 1.0293877551020407, "grad_norm": 0.1640625, "learning_rate": 0.0001330069758903439, "loss": 0.9779, "step": 2838 }, { "epoch": 1.0301133786848073, "grad_norm": 0.1806640625, "learning_rate": 0.00013295802227389548, "loss": 0.9869, "step": 2840 }, { "epoch": 1.0308390022675737, "grad_norm": 0.15625, "learning_rate": 0.0001329090686574471, "loss": 0.9595, "step": 2842 }, { "epoch": 1.0315646258503401, "grad_norm": 0.197265625, "learning_rate": 0.00013286011504099867, "loss": 1.0107, "step": 2844 }, { "epoch": 1.0322902494331065, "grad_norm": 0.1572265625, "learning_rate": 0.00013281116142455023, "loss": 1.0039, "step": 2846 }, { "epoch": 1.033015873015873, "grad_norm": 0.1728515625, "learning_rate": 0.00013276220780810184, "loss": 0.9353, "step": 2848 }, { "epoch": 1.0337414965986396, "grad_norm": 0.162109375, "learning_rate": 0.00013271325419165342, "loss": 0.9496, "step": 2850 }, { "epoch": 1.034467120181406, "grad_norm": 0.1416015625, "learning_rate": 0.000132664300575205, "loss": 1.0133, "step": 2852 }, { "epoch": 1.0351927437641724, "grad_norm": 0.1455078125, "learning_rate": 0.00013261534695875658, "loss": 0.9736, "step": 2854 }, { "epoch": 1.0359183673469388, "grad_norm": 0.1650390625, "learning_rate": 0.00013256639334230816, "loss": 1.0096, "step": 2856 }, { "epoch": 1.0366439909297052, "grad_norm": 0.1767578125, "learning_rate": 0.00013251743972585975, "loss": 1.0358, "step": 2858 }, { "epoch": 1.0373696145124716, "grad_norm": 0.1650390625, "learning_rate": 0.00013246848610941133, "loss": 1.0608, "step": 2860 }, { "epoch": 1.0380952380952382, "grad_norm": 0.16796875, "learning_rate": 0.00013241953249296294, "loss": 0.9904, "step": 2862 }, { "epoch": 1.0388208616780046, "grad_norm": 0.154296875, "learning_rate": 0.00013237057887651452, "loss": 0.9738, "step": 2864 }, { "epoch": 1.039546485260771, "grad_norm": 0.185546875, "learning_rate": 0.00013232162526006608, "loss": 1.0697, "step": 2866 }, { "epoch": 1.0402721088435374, "grad_norm": 0.146484375, "learning_rate": 0.00013227267164361769, "loss": 0.9685, "step": 2868 }, { "epoch": 1.0409977324263038, "grad_norm": 0.1669921875, "learning_rate": 0.00013222371802716927, "loss": 1.0041, "step": 2870 }, { "epoch": 1.0417233560090704, "grad_norm": 0.1513671875, "learning_rate": 0.00013217476441072085, "loss": 0.9809, "step": 2872 }, { "epoch": 1.0424489795918368, "grad_norm": 0.158203125, "learning_rate": 0.00013212581079427243, "loss": 0.9908, "step": 2874 }, { "epoch": 1.0431746031746032, "grad_norm": 0.16015625, "learning_rate": 0.00013207685717782401, "loss": 1.032, "step": 2876 }, { "epoch": 1.0439002267573696, "grad_norm": 0.146484375, "learning_rate": 0.0001320279035613756, "loss": 0.9646, "step": 2878 }, { "epoch": 1.044625850340136, "grad_norm": 0.166015625, "learning_rate": 0.00013197894994492718, "loss": 0.9965, "step": 2880 }, { "epoch": 1.0453514739229024, "grad_norm": 0.1640625, "learning_rate": 0.0001319299963284788, "loss": 0.9963, "step": 2882 }, { "epoch": 1.046077097505669, "grad_norm": 0.1650390625, "learning_rate": 0.00013188104271203037, "loss": 1.0057, "step": 2884 }, { "epoch": 1.0468027210884354, "grad_norm": 0.1572265625, "learning_rate": 0.00013183208909558193, "loss": 0.9545, "step": 2886 }, { "epoch": 1.0475283446712018, "grad_norm": 0.1640625, "learning_rate": 0.00013178313547913354, "loss": 0.9863, "step": 2888 }, { "epoch": 1.0482539682539682, "grad_norm": 0.15234375, "learning_rate": 0.00013173418186268512, "loss": 1.0419, "step": 2890 }, { "epoch": 1.0489795918367346, "grad_norm": 0.1796875, "learning_rate": 0.0001316852282462367, "loss": 1.013, "step": 2892 }, { "epoch": 1.0497052154195012, "grad_norm": 0.1533203125, "learning_rate": 0.00013163627462978828, "loss": 1.096, "step": 2894 }, { "epoch": 1.0504308390022676, "grad_norm": 0.146484375, "learning_rate": 0.00013158732101333986, "loss": 1.0023, "step": 2896 }, { "epoch": 1.051156462585034, "grad_norm": 0.1533203125, "learning_rate": 0.00013153836739689145, "loss": 0.9743, "step": 2898 }, { "epoch": 1.0518820861678004, "grad_norm": 0.1669921875, "learning_rate": 0.00013148941378044303, "loss": 1.0481, "step": 2900 }, { "epoch": 1.0518820861678004, "eval_loss": 1.0157631635665894, "eval_runtime": 98.609, "eval_samples_per_second": 18.254, "eval_steps_per_second": 0.578, "step": 2900 }, { "epoch": 1.0526077097505668, "grad_norm": 0.15625, "learning_rate": 0.00013144046016399464, "loss": 1.0343, "step": 2902 }, { "epoch": 1.0533333333333332, "grad_norm": 0.1513671875, "learning_rate": 0.00013139150654754622, "loss": 0.9971, "step": 2904 }, { "epoch": 1.0540589569160999, "grad_norm": 0.1689453125, "learning_rate": 0.00013134255293109778, "loss": 0.9762, "step": 2906 }, { "epoch": 1.0547845804988663, "grad_norm": 0.154296875, "learning_rate": 0.00013129359931464939, "loss": 1.1036, "step": 2908 }, { "epoch": 1.0555102040816327, "grad_norm": 0.169921875, "learning_rate": 0.00013124464569820097, "loss": 0.9814, "step": 2910 }, { "epoch": 1.056235827664399, "grad_norm": 0.1533203125, "learning_rate": 0.00013119569208175255, "loss": 0.9185, "step": 2912 }, { "epoch": 1.0569614512471655, "grad_norm": 0.1669921875, "learning_rate": 0.00013114673846530413, "loss": 0.9613, "step": 2914 }, { "epoch": 1.057687074829932, "grad_norm": 0.14453125, "learning_rate": 0.00013109778484885571, "loss": 0.9276, "step": 2916 }, { "epoch": 1.0584126984126985, "grad_norm": 0.1416015625, "learning_rate": 0.0001310488312324073, "loss": 0.9268, "step": 2918 }, { "epoch": 1.0591383219954649, "grad_norm": 0.1533203125, "learning_rate": 0.00013099987761595888, "loss": 1.0498, "step": 2920 }, { "epoch": 1.0598639455782313, "grad_norm": 0.158203125, "learning_rate": 0.0001309509239995105, "loss": 1.0223, "step": 2922 }, { "epoch": 1.0605895691609977, "grad_norm": 0.150390625, "learning_rate": 0.00013090197038306204, "loss": 0.9899, "step": 2924 }, { "epoch": 1.061315192743764, "grad_norm": 0.1533203125, "learning_rate": 0.00013085301676661363, "loss": 0.9639, "step": 2926 }, { "epoch": 1.0620408163265307, "grad_norm": 0.1513671875, "learning_rate": 0.00013080406315016524, "loss": 0.9369, "step": 2928 }, { "epoch": 1.062766439909297, "grad_norm": 0.1455078125, "learning_rate": 0.00013075510953371682, "loss": 0.93, "step": 2930 }, { "epoch": 1.0634920634920635, "grad_norm": 0.1611328125, "learning_rate": 0.0001307061559172684, "loss": 0.993, "step": 2932 }, { "epoch": 1.0642176870748299, "grad_norm": 0.181640625, "learning_rate": 0.00013065720230081998, "loss": 1.0847, "step": 2934 }, { "epoch": 1.0649433106575963, "grad_norm": 0.1552734375, "learning_rate": 0.00013060824868437156, "loss": 1.003, "step": 2936 }, { "epoch": 1.065668934240363, "grad_norm": 0.166015625, "learning_rate": 0.00013055929506792315, "loss": 1.0098, "step": 2938 }, { "epoch": 1.0663945578231293, "grad_norm": 0.1513671875, "learning_rate": 0.00013051034145147473, "loss": 0.9617, "step": 2940 }, { "epoch": 1.0671201814058957, "grad_norm": 0.15234375, "learning_rate": 0.00013046138783502634, "loss": 0.9282, "step": 2942 }, { "epoch": 1.067845804988662, "grad_norm": 0.1416015625, "learning_rate": 0.0001304124342185779, "loss": 0.9462, "step": 2944 }, { "epoch": 1.0685714285714285, "grad_norm": 0.1650390625, "learning_rate": 0.00013036348060212948, "loss": 1.0117, "step": 2946 }, { "epoch": 1.069297052154195, "grad_norm": 0.15625, "learning_rate": 0.00013031452698568109, "loss": 0.9634, "step": 2948 }, { "epoch": 1.0700226757369615, "grad_norm": 0.171875, "learning_rate": 0.00013026557336923267, "loss": 1.0243, "step": 2950 }, { "epoch": 1.070748299319728, "grad_norm": 0.142578125, "learning_rate": 0.00013021661975278425, "loss": 0.9924, "step": 2952 }, { "epoch": 1.0714739229024943, "grad_norm": 0.177734375, "learning_rate": 0.00013016766613633583, "loss": 0.9808, "step": 2954 }, { "epoch": 1.0721995464852607, "grad_norm": 0.181640625, "learning_rate": 0.00013011871251988741, "loss": 1.0503, "step": 2956 }, { "epoch": 1.0729251700680271, "grad_norm": 0.1328125, "learning_rate": 0.000130069758903439, "loss": 1.042, "step": 2958 }, { "epoch": 1.0736507936507937, "grad_norm": 0.154296875, "learning_rate": 0.00013002080528699058, "loss": 1.0484, "step": 2960 }, { "epoch": 1.0743764172335601, "grad_norm": 0.14453125, "learning_rate": 0.0001299718516705422, "loss": 0.925, "step": 2962 }, { "epoch": 1.0751020408163265, "grad_norm": 0.154296875, "learning_rate": 0.00012992289805409374, "loss": 1.0459, "step": 2964 }, { "epoch": 1.075827664399093, "grad_norm": 0.1845703125, "learning_rate": 0.00012987394443764533, "loss": 0.9794, "step": 2966 }, { "epoch": 1.0765532879818593, "grad_norm": 0.1884765625, "learning_rate": 0.00012982499082119694, "loss": 0.887, "step": 2968 }, { "epoch": 1.0772789115646257, "grad_norm": 0.15234375, "learning_rate": 0.00012977603720474852, "loss": 1.0116, "step": 2970 }, { "epoch": 1.0780045351473924, "grad_norm": 0.1650390625, "learning_rate": 0.0001297270835883001, "loss": 0.9688, "step": 2972 }, { "epoch": 1.0787301587301588, "grad_norm": 0.1591796875, "learning_rate": 0.00012967812997185168, "loss": 1.024, "step": 2974 }, { "epoch": 1.0794557823129252, "grad_norm": 0.154296875, "learning_rate": 0.00012962917635540326, "loss": 1.0757, "step": 2976 }, { "epoch": 1.0801814058956916, "grad_norm": 0.1728515625, "learning_rate": 0.00012958022273895485, "loss": 0.9665, "step": 2978 }, { "epoch": 1.080907029478458, "grad_norm": 0.18359375, "learning_rate": 0.00012953126912250643, "loss": 0.9317, "step": 2980 }, { "epoch": 1.0816326530612246, "grad_norm": 0.1748046875, "learning_rate": 0.00012948231550605804, "loss": 0.9634, "step": 2982 }, { "epoch": 1.082358276643991, "grad_norm": 0.154296875, "learning_rate": 0.0001294333618896096, "loss": 0.9887, "step": 2984 }, { "epoch": 1.0830839002267574, "grad_norm": 0.15625, "learning_rate": 0.00012938440827316118, "loss": 0.8876, "step": 2986 }, { "epoch": 1.0838095238095238, "grad_norm": 0.169921875, "learning_rate": 0.00012933545465671279, "loss": 0.9634, "step": 2988 }, { "epoch": 1.0845351473922902, "grad_norm": 0.15625, "learning_rate": 0.00012928650104026437, "loss": 0.9834, "step": 2990 }, { "epoch": 1.0852607709750566, "grad_norm": 0.154296875, "learning_rate": 0.00012923754742381592, "loss": 0.99, "step": 2992 }, { "epoch": 1.0859863945578232, "grad_norm": 0.142578125, "learning_rate": 0.00012918859380736753, "loss": 0.9375, "step": 2994 }, { "epoch": 1.0867120181405896, "grad_norm": 0.154296875, "learning_rate": 0.00012913964019091911, "loss": 0.9757, "step": 2996 }, { "epoch": 1.087437641723356, "grad_norm": 0.1494140625, "learning_rate": 0.0001290906865744707, "loss": 0.9934, "step": 2998 }, { "epoch": 1.0881632653061224, "grad_norm": 0.1591796875, "learning_rate": 0.00012904173295802228, "loss": 0.9745, "step": 3000 }, { "epoch": 1.0881632653061224, "eval_loss": 1.0154409408569336, "eval_runtime": 92.0958, "eval_samples_per_second": 19.545, "eval_steps_per_second": 0.619, "step": 3000 }, { "epoch": 1.0888888888888888, "grad_norm": 0.140625, "learning_rate": 0.00012899277934157386, "loss": 0.9854, "step": 3002 }, { "epoch": 1.0896145124716554, "grad_norm": 0.19140625, "learning_rate": 0.00012894382572512544, "loss": 0.9796, "step": 3004 }, { "epoch": 1.0903401360544218, "grad_norm": 0.166015625, "learning_rate": 0.00012889487210867703, "loss": 0.9706, "step": 3006 }, { "epoch": 1.0910657596371882, "grad_norm": 0.15625, "learning_rate": 0.00012884591849222864, "loss": 0.9691, "step": 3008 }, { "epoch": 1.0917913832199546, "grad_norm": 0.1728515625, "learning_rate": 0.00012879696487578022, "loss": 1.0203, "step": 3010 }, { "epoch": 1.092517006802721, "grad_norm": 0.1865234375, "learning_rate": 0.00012874801125933177, "loss": 0.9979, "step": 3012 }, { "epoch": 1.0932426303854874, "grad_norm": 0.166015625, "learning_rate": 0.00012869905764288338, "loss": 1.0656, "step": 3014 }, { "epoch": 1.093968253968254, "grad_norm": 0.1591796875, "learning_rate": 0.00012865010402643496, "loss": 0.9869, "step": 3016 }, { "epoch": 1.0946938775510204, "grad_norm": 0.162109375, "learning_rate": 0.00012860115040998655, "loss": 1.0094, "step": 3018 }, { "epoch": 1.0954195011337868, "grad_norm": 0.1748046875, "learning_rate": 0.00012855219679353813, "loss": 1.0243, "step": 3020 }, { "epoch": 1.0961451247165532, "grad_norm": 0.1767578125, "learning_rate": 0.0001285032431770897, "loss": 0.9527, "step": 3022 }, { "epoch": 1.0968707482993196, "grad_norm": 0.150390625, "learning_rate": 0.0001284542895606413, "loss": 1.0238, "step": 3024 }, { "epoch": 1.0975963718820863, "grad_norm": 0.1708984375, "learning_rate": 0.00012840533594419288, "loss": 0.9843, "step": 3026 }, { "epoch": 1.0983219954648527, "grad_norm": 0.15625, "learning_rate": 0.00012835638232774448, "loss": 0.9041, "step": 3028 }, { "epoch": 1.099047619047619, "grad_norm": 0.1533203125, "learning_rate": 0.00012830742871129607, "loss": 0.9152, "step": 3030 }, { "epoch": 1.0997732426303855, "grad_norm": 0.1669921875, "learning_rate": 0.00012825847509484762, "loss": 0.9771, "step": 3032 }, { "epoch": 1.1004988662131518, "grad_norm": 0.1572265625, "learning_rate": 0.00012820952147839923, "loss": 0.952, "step": 3034 }, { "epoch": 1.1012244897959185, "grad_norm": 0.17578125, "learning_rate": 0.00012816056786195081, "loss": 0.9888, "step": 3036 }, { "epoch": 1.1019501133786849, "grad_norm": 0.1640625, "learning_rate": 0.0001281116142455024, "loss": 0.9741, "step": 3038 }, { "epoch": 1.1026757369614513, "grad_norm": 0.1689453125, "learning_rate": 0.00012806266062905398, "loss": 0.9094, "step": 3040 }, { "epoch": 1.1034013605442177, "grad_norm": 0.1728515625, "learning_rate": 0.00012801370701260556, "loss": 0.9386, "step": 3042 }, { "epoch": 1.104126984126984, "grad_norm": 0.15234375, "learning_rate": 0.00012796475339615714, "loss": 0.9503, "step": 3044 }, { "epoch": 1.1048526077097505, "grad_norm": 0.1455078125, "learning_rate": 0.00012791579977970873, "loss": 0.9595, "step": 3046 }, { "epoch": 1.105578231292517, "grad_norm": 0.1513671875, "learning_rate": 0.00012786684616326033, "loss": 0.9771, "step": 3048 }, { "epoch": 1.1063038548752835, "grad_norm": 0.17578125, "learning_rate": 0.00012781789254681192, "loss": 1.054, "step": 3050 }, { "epoch": 1.1070294784580499, "grad_norm": 0.173828125, "learning_rate": 0.00012776893893036347, "loss": 0.9665, "step": 3052 }, { "epoch": 1.1077551020408163, "grad_norm": 0.138671875, "learning_rate": 0.00012771998531391508, "loss": 1.0094, "step": 3054 }, { "epoch": 1.1084807256235827, "grad_norm": 0.1650390625, "learning_rate": 0.00012767103169746666, "loss": 0.9462, "step": 3056 }, { "epoch": 1.1092063492063493, "grad_norm": 0.1572265625, "learning_rate": 0.00012762207808101825, "loss": 1.0243, "step": 3058 }, { "epoch": 1.1099319727891157, "grad_norm": 0.1513671875, "learning_rate": 0.00012757312446456983, "loss": 0.9708, "step": 3060 }, { "epoch": 1.110657596371882, "grad_norm": 0.15625, "learning_rate": 0.0001275241708481214, "loss": 0.9986, "step": 3062 }, { "epoch": 1.1113832199546485, "grad_norm": 0.1513671875, "learning_rate": 0.000127475217231673, "loss": 0.9911, "step": 3064 }, { "epoch": 1.112108843537415, "grad_norm": 0.1787109375, "learning_rate": 0.00012742626361522458, "loss": 1.0371, "step": 3066 }, { "epoch": 1.1128344671201813, "grad_norm": 0.14453125, "learning_rate": 0.00012737730999877618, "loss": 1.0138, "step": 3068 }, { "epoch": 1.113560090702948, "grad_norm": 0.154296875, "learning_rate": 0.00012732835638232774, "loss": 0.9498, "step": 3070 }, { "epoch": 1.1142857142857143, "grad_norm": 0.15625, "learning_rate": 0.00012727940276587932, "loss": 1.0003, "step": 3072 }, { "epoch": 1.1150113378684807, "grad_norm": 0.1396484375, "learning_rate": 0.00012723044914943093, "loss": 1.0018, "step": 3074 }, { "epoch": 1.1157369614512471, "grad_norm": 0.140625, "learning_rate": 0.00012718149553298251, "loss": 0.9182, "step": 3076 }, { "epoch": 1.1164625850340135, "grad_norm": 0.166015625, "learning_rate": 0.0001271325419165341, "loss": 0.9443, "step": 3078 }, { "epoch": 1.1171882086167801, "grad_norm": 0.171875, "learning_rate": 0.00012708358830008568, "loss": 1.0236, "step": 3080 }, { "epoch": 1.1179138321995465, "grad_norm": 0.162109375, "learning_rate": 0.00012703463468363726, "loss": 0.977, "step": 3082 }, { "epoch": 1.118639455782313, "grad_norm": 0.15234375, "learning_rate": 0.00012698568106718884, "loss": 0.9371, "step": 3084 }, { "epoch": 1.1193650793650793, "grad_norm": 0.1767578125, "learning_rate": 0.00012693672745074043, "loss": 0.9309, "step": 3086 }, { "epoch": 1.1200907029478457, "grad_norm": 0.1669921875, "learning_rate": 0.00012688777383429203, "loss": 0.9547, "step": 3088 }, { "epoch": 1.1208163265306124, "grad_norm": 0.23828125, "learning_rate": 0.0001268388202178436, "loss": 1.0593, "step": 3090 }, { "epoch": 1.1215419501133788, "grad_norm": 0.1748046875, "learning_rate": 0.00012678986660139517, "loss": 0.9916, "step": 3092 }, { "epoch": 1.1222675736961452, "grad_norm": 0.171875, "learning_rate": 0.00012674091298494678, "loss": 0.9708, "step": 3094 }, { "epoch": 1.1229931972789116, "grad_norm": 0.1904296875, "learning_rate": 0.00012669195936849836, "loss": 0.9498, "step": 3096 }, { "epoch": 1.123718820861678, "grad_norm": 0.1728515625, "learning_rate": 0.00012664300575204995, "loss": 0.9388, "step": 3098 }, { "epoch": 1.1244444444444444, "grad_norm": 0.1494140625, "learning_rate": 0.00012659405213560153, "loss": 0.9372, "step": 3100 }, { "epoch": 1.1244444444444444, "eval_loss": 1.0154296159744263, "eval_runtime": 89.182, "eval_samples_per_second": 20.183, "eval_steps_per_second": 0.639, "step": 3100 }, { "epoch": 1.125170068027211, "grad_norm": 0.19140625, "learning_rate": 0.0001265450985191531, "loss": 0.9911, "step": 3102 }, { "epoch": 1.1258956916099774, "grad_norm": 0.1533203125, "learning_rate": 0.0001264961449027047, "loss": 1.0097, "step": 3104 }, { "epoch": 1.1266213151927438, "grad_norm": 0.162109375, "learning_rate": 0.00012644719128625628, "loss": 0.9942, "step": 3106 }, { "epoch": 1.1273469387755102, "grad_norm": 0.162109375, "learning_rate": 0.00012639823766980788, "loss": 0.9891, "step": 3108 }, { "epoch": 1.1280725623582766, "grad_norm": 0.1689453125, "learning_rate": 0.00012634928405335944, "loss": 1.0132, "step": 3110 }, { "epoch": 1.1287981859410432, "grad_norm": 0.1689453125, "learning_rate": 0.00012630033043691102, "loss": 0.9039, "step": 3112 }, { "epoch": 1.1295238095238096, "grad_norm": 0.1767578125, "learning_rate": 0.00012625137682046263, "loss": 0.9398, "step": 3114 }, { "epoch": 1.130249433106576, "grad_norm": 0.1923828125, "learning_rate": 0.00012620242320401421, "loss": 0.9643, "step": 3116 }, { "epoch": 1.1309750566893424, "grad_norm": 0.1474609375, "learning_rate": 0.0001261534695875658, "loss": 0.9252, "step": 3118 }, { "epoch": 1.1317006802721088, "grad_norm": 0.17578125, "learning_rate": 0.00012610451597111738, "loss": 0.9849, "step": 3120 }, { "epoch": 1.1324263038548752, "grad_norm": 0.162109375, "learning_rate": 0.00012605556235466896, "loss": 1.004, "step": 3122 }, { "epoch": 1.1331519274376418, "grad_norm": 0.1845703125, "learning_rate": 0.00012600660873822054, "loss": 1.0351, "step": 3124 }, { "epoch": 1.1338775510204082, "grad_norm": 0.1533203125, "learning_rate": 0.00012595765512177213, "loss": 0.9665, "step": 3126 }, { "epoch": 1.1346031746031746, "grad_norm": 0.1552734375, "learning_rate": 0.00012590870150532373, "loss": 1.0028, "step": 3128 }, { "epoch": 1.135328798185941, "grad_norm": 0.1689453125, "learning_rate": 0.0001258597478888753, "loss": 0.9811, "step": 3130 }, { "epoch": 1.1360544217687074, "grad_norm": 0.1650390625, "learning_rate": 0.00012581079427242687, "loss": 1.01, "step": 3132 }, { "epoch": 1.136780045351474, "grad_norm": 0.15625, "learning_rate": 0.00012576184065597848, "loss": 0.9534, "step": 3134 }, { "epoch": 1.1375056689342404, "grad_norm": 0.1630859375, "learning_rate": 0.00012571288703953006, "loss": 0.9672, "step": 3136 }, { "epoch": 1.1382312925170068, "grad_norm": 0.150390625, "learning_rate": 0.00012566393342308162, "loss": 1.0314, "step": 3138 }, { "epoch": 1.1389569160997732, "grad_norm": 0.1904296875, "learning_rate": 0.00012561497980663323, "loss": 0.96, "step": 3140 }, { "epoch": 1.1396825396825396, "grad_norm": 0.1630859375, "learning_rate": 0.0001255660261901848, "loss": 0.9722, "step": 3142 }, { "epoch": 1.140408163265306, "grad_norm": 0.1455078125, "learning_rate": 0.0001255170725737364, "loss": 1.063, "step": 3144 }, { "epoch": 1.1411337868480727, "grad_norm": 0.1796875, "learning_rate": 0.00012546811895728798, "loss": 0.987, "step": 3146 }, { "epoch": 1.141859410430839, "grad_norm": 0.1552734375, "learning_rate": 0.00012541916534083956, "loss": 0.9693, "step": 3148 }, { "epoch": 1.1425850340136054, "grad_norm": 0.140625, "learning_rate": 0.00012537021172439114, "loss": 0.9687, "step": 3150 }, { "epoch": 1.1433106575963718, "grad_norm": 0.1533203125, "learning_rate": 0.00012532125810794272, "loss": 0.9822, "step": 3152 }, { "epoch": 1.1440362811791382, "grad_norm": 0.162109375, "learning_rate": 0.00012527230449149433, "loss": 0.9652, "step": 3154 }, { "epoch": 1.1447619047619049, "grad_norm": 0.158203125, "learning_rate": 0.0001252233508750459, "loss": 0.8838, "step": 3156 }, { "epoch": 1.1454875283446713, "grad_norm": 0.1611328125, "learning_rate": 0.00012517439725859747, "loss": 0.9707, "step": 3158 }, { "epoch": 1.1462131519274377, "grad_norm": 0.16015625, "learning_rate": 0.00012512544364214908, "loss": 1.0355, "step": 3160 }, { "epoch": 1.146938775510204, "grad_norm": 0.15625, "learning_rate": 0.00012507649002570066, "loss": 1.0814, "step": 3162 }, { "epoch": 1.1476643990929705, "grad_norm": 0.15234375, "learning_rate": 0.00012502753640925224, "loss": 1.005, "step": 3164 }, { "epoch": 1.1483900226757369, "grad_norm": 0.166015625, "learning_rate": 0.00012497858279280382, "loss": 0.9581, "step": 3166 }, { "epoch": 1.1491156462585035, "grad_norm": 0.1650390625, "learning_rate": 0.0001249296291763554, "loss": 1.0175, "step": 3168 }, { "epoch": 1.1498412698412699, "grad_norm": 0.16796875, "learning_rate": 0.000124880675559907, "loss": 0.9524, "step": 3170 }, { "epoch": 1.1505668934240363, "grad_norm": 0.158203125, "learning_rate": 0.00012483172194345857, "loss": 1.0104, "step": 3172 }, { "epoch": 1.1512925170068027, "grad_norm": 0.169921875, "learning_rate": 0.00012478276832701018, "loss": 0.9864, "step": 3174 }, { "epoch": 1.152018140589569, "grad_norm": 0.1572265625, "learning_rate": 0.00012473381471056176, "loss": 1.0231, "step": 3176 }, { "epoch": 1.1527437641723357, "grad_norm": 0.162109375, "learning_rate": 0.00012468486109411332, "loss": 0.9632, "step": 3178 }, { "epoch": 1.153469387755102, "grad_norm": 0.1806640625, "learning_rate": 0.00012463590747766493, "loss": 0.9549, "step": 3180 }, { "epoch": 1.1541950113378685, "grad_norm": 0.150390625, "learning_rate": 0.0001245869538612165, "loss": 1.0362, "step": 3182 }, { "epoch": 1.154920634920635, "grad_norm": 0.1591796875, "learning_rate": 0.0001245380002447681, "loss": 0.9991, "step": 3184 }, { "epoch": 1.1556462585034013, "grad_norm": 0.1396484375, "learning_rate": 0.00012448904662831967, "loss": 0.9353, "step": 3186 }, { "epoch": 1.1563718820861677, "grad_norm": 0.1494140625, "learning_rate": 0.00012444009301187126, "loss": 0.9651, "step": 3188 }, { "epoch": 1.1570975056689343, "grad_norm": 0.1484375, "learning_rate": 0.00012439113939542284, "loss": 0.9336, "step": 3190 }, { "epoch": 1.1578231292517007, "grad_norm": 0.166015625, "learning_rate": 0.00012434218577897442, "loss": 0.9326, "step": 3192 }, { "epoch": 1.1585487528344671, "grad_norm": 0.1767578125, "learning_rate": 0.00012429323216252603, "loss": 1.0304, "step": 3194 }, { "epoch": 1.1592743764172335, "grad_norm": 0.1591796875, "learning_rate": 0.0001242442785460776, "loss": 0.9833, "step": 3196 }, { "epoch": 1.16, "grad_norm": 0.1533203125, "learning_rate": 0.00012419532492962917, "loss": 1.0228, "step": 3198 }, { "epoch": 1.1607256235827665, "grad_norm": 0.16015625, "learning_rate": 0.00012414637131318078, "loss": 0.988, "step": 3200 }, { "epoch": 1.1607256235827665, "eval_loss": 1.014972448348999, "eval_runtime": 94.878, "eval_samples_per_second": 18.972, "eval_steps_per_second": 0.601, "step": 3200 }, { "epoch": 1.161451247165533, "grad_norm": 0.1552734375, "learning_rate": 0.00012409741769673236, "loss": 0.9443, "step": 3202 }, { "epoch": 1.1621768707482993, "grad_norm": 0.171875, "learning_rate": 0.00012404846408028394, "loss": 1.0144, "step": 3204 }, { "epoch": 1.1629024943310657, "grad_norm": 0.166015625, "learning_rate": 0.0001239995104638355, "loss": 0.8949, "step": 3206 }, { "epoch": 1.1636281179138321, "grad_norm": 0.1533203125, "learning_rate": 0.0001239505568473871, "loss": 0.9979, "step": 3208 }, { "epoch": 1.1643537414965985, "grad_norm": 0.173828125, "learning_rate": 0.0001239016032309387, "loss": 0.985, "step": 3210 }, { "epoch": 1.1650793650793652, "grad_norm": 0.1689453125, "learning_rate": 0.00012385264961449027, "loss": 1.018, "step": 3212 }, { "epoch": 1.1658049886621316, "grad_norm": 0.1904296875, "learning_rate": 0.00012380369599804188, "loss": 0.9736, "step": 3214 }, { "epoch": 1.166530612244898, "grad_norm": 0.1728515625, "learning_rate": 0.00012375474238159344, "loss": 0.9888, "step": 3216 }, { "epoch": 1.1672562358276644, "grad_norm": 0.166015625, "learning_rate": 0.00012370578876514502, "loss": 0.9882, "step": 3218 }, { "epoch": 1.1679818594104308, "grad_norm": 0.162109375, "learning_rate": 0.00012365683514869663, "loss": 1.0883, "step": 3220 }, { "epoch": 1.1687074829931974, "grad_norm": 0.185546875, "learning_rate": 0.0001236078815322482, "loss": 1.0099, "step": 3222 }, { "epoch": 1.1694331065759638, "grad_norm": 0.203125, "learning_rate": 0.0001235589279157998, "loss": 0.991, "step": 3224 }, { "epoch": 1.1701587301587302, "grad_norm": 0.1533203125, "learning_rate": 0.00012350997429935135, "loss": 0.9971, "step": 3226 }, { "epoch": 1.1708843537414966, "grad_norm": 0.181640625, "learning_rate": 0.00012346102068290296, "loss": 1.0944, "step": 3228 }, { "epoch": 1.171609977324263, "grad_norm": 0.1533203125, "learning_rate": 0.00012341206706645454, "loss": 1.0144, "step": 3230 }, { "epoch": 1.1723356009070294, "grad_norm": 0.158203125, "learning_rate": 0.00012336311345000612, "loss": 0.9596, "step": 3232 }, { "epoch": 1.173061224489796, "grad_norm": 0.16796875, "learning_rate": 0.00012331415983355773, "loss": 1.0307, "step": 3234 }, { "epoch": 1.1737868480725624, "grad_norm": 0.1787109375, "learning_rate": 0.00012326520621710929, "loss": 0.9639, "step": 3236 }, { "epoch": 1.1745124716553288, "grad_norm": 0.1591796875, "learning_rate": 0.00012321625260066087, "loss": 0.9618, "step": 3238 }, { "epoch": 1.1752380952380952, "grad_norm": 0.1796875, "learning_rate": 0.00012316729898421248, "loss": 0.9739, "step": 3240 }, { "epoch": 1.1759637188208616, "grad_norm": 0.173828125, "learning_rate": 0.00012311834536776406, "loss": 1.0316, "step": 3242 }, { "epoch": 1.1766893424036282, "grad_norm": 0.15625, "learning_rate": 0.00012306939175131564, "loss": 0.9901, "step": 3244 }, { "epoch": 1.1774149659863946, "grad_norm": 0.173828125, "learning_rate": 0.0001230204381348672, "loss": 1.0439, "step": 3246 }, { "epoch": 1.178140589569161, "grad_norm": 0.1552734375, "learning_rate": 0.0001229714845184188, "loss": 0.9763, "step": 3248 }, { "epoch": 1.1788662131519274, "grad_norm": 0.1767578125, "learning_rate": 0.0001229225309019704, "loss": 0.9723, "step": 3250 }, { "epoch": 1.1795918367346938, "grad_norm": 0.1552734375, "learning_rate": 0.00012287357728552197, "loss": 1.1122, "step": 3252 }, { "epoch": 1.1803174603174602, "grad_norm": 0.18359375, "learning_rate": 0.00012282462366907358, "loss": 1.0547, "step": 3254 }, { "epoch": 1.1810430839002268, "grad_norm": 0.142578125, "learning_rate": 0.00012277567005262514, "loss": 0.9226, "step": 3256 }, { "epoch": 1.1817687074829932, "grad_norm": 0.1396484375, "learning_rate": 0.00012272671643617672, "loss": 0.9439, "step": 3258 }, { "epoch": 1.1824943310657596, "grad_norm": 0.15234375, "learning_rate": 0.00012267776281972833, "loss": 1.0307, "step": 3260 }, { "epoch": 1.183219954648526, "grad_norm": 0.1669921875, "learning_rate": 0.0001226288092032799, "loss": 1.0502, "step": 3262 }, { "epoch": 1.1839455782312924, "grad_norm": 0.1923828125, "learning_rate": 0.0001225798555868315, "loss": 1.0492, "step": 3264 }, { "epoch": 1.184671201814059, "grad_norm": 0.1533203125, "learning_rate": 0.00012253090197038305, "loss": 0.962, "step": 3266 }, { "epoch": 1.1853968253968254, "grad_norm": 0.166015625, "learning_rate": 0.00012248194835393466, "loss": 0.9904, "step": 3268 }, { "epoch": 1.1861224489795918, "grad_norm": 0.146484375, "learning_rate": 0.00012243299473748624, "loss": 0.9677, "step": 3270 }, { "epoch": 1.1868480725623582, "grad_norm": 0.1611328125, "learning_rate": 0.00012238404112103782, "loss": 1.0684, "step": 3272 }, { "epoch": 1.1875736961451246, "grad_norm": 0.169921875, "learning_rate": 0.00012233508750458943, "loss": 1.0299, "step": 3274 }, { "epoch": 1.188299319727891, "grad_norm": 0.1572265625, "learning_rate": 0.00012228613388814099, "loss": 0.98, "step": 3276 }, { "epoch": 1.1890249433106577, "grad_norm": 0.1591796875, "learning_rate": 0.00012223718027169257, "loss": 1.0087, "step": 3278 }, { "epoch": 1.189750566893424, "grad_norm": 0.166015625, "learning_rate": 0.00012218822665524418, "loss": 1.0092, "step": 3280 }, { "epoch": 1.1904761904761905, "grad_norm": 0.162109375, "learning_rate": 0.00012213927303879576, "loss": 1.0301, "step": 3282 }, { "epoch": 1.1912018140589569, "grad_norm": 0.154296875, "learning_rate": 0.00012209031942234731, "loss": 1.0265, "step": 3284 }, { "epoch": 1.1919274376417233, "grad_norm": 0.25, "learning_rate": 0.00012204136580589891, "loss": 0.9792, "step": 3286 }, { "epoch": 1.1926530612244899, "grad_norm": 0.1640625, "learning_rate": 0.0001219924121894505, "loss": 0.9287, "step": 3288 }, { "epoch": 1.1933786848072563, "grad_norm": 0.1650390625, "learning_rate": 0.00012194345857300209, "loss": 0.9637, "step": 3290 }, { "epoch": 1.1941043083900227, "grad_norm": 0.1640625, "learning_rate": 0.00012189450495655368, "loss": 0.984, "step": 3292 }, { "epoch": 1.194829931972789, "grad_norm": 0.1708984375, "learning_rate": 0.00012184555134010524, "loss": 1.0349, "step": 3294 }, { "epoch": 1.1955555555555555, "grad_norm": 0.1357421875, "learning_rate": 0.00012179659772365684, "loss": 0.9786, "step": 3296 }, { "epoch": 1.1962811791383219, "grad_norm": 0.19921875, "learning_rate": 0.00012174764410720843, "loss": 0.9701, "step": 3298 }, { "epoch": 1.1970068027210885, "grad_norm": 0.1572265625, "learning_rate": 0.00012169869049076001, "loss": 1.0021, "step": 3300 }, { "epoch": 1.1970068027210885, "eval_loss": 1.014991044998169, "eval_runtime": 91.949, "eval_samples_per_second": 19.576, "eval_steps_per_second": 0.62, "step": 3300 }, { "epoch": 1.197732426303855, "grad_norm": 0.1533203125, "learning_rate": 0.00012164973687431161, "loss": 0.8863, "step": 3302 }, { "epoch": 1.1984580498866213, "grad_norm": 0.1650390625, "learning_rate": 0.00012160078325786316, "loss": 1.0103, "step": 3304 }, { "epoch": 1.1991836734693877, "grad_norm": 0.1640625, "learning_rate": 0.00012155182964141476, "loss": 0.9499, "step": 3306 }, { "epoch": 1.199909297052154, "grad_norm": 0.169921875, "learning_rate": 0.00012150287602496636, "loss": 1.0125, "step": 3308 }, { "epoch": 1.2006349206349207, "grad_norm": 0.16796875, "learning_rate": 0.00012145392240851794, "loss": 0.9703, "step": 3310 }, { "epoch": 1.2013605442176871, "grad_norm": 0.1767578125, "learning_rate": 0.00012140496879206953, "loss": 0.9431, "step": 3312 }, { "epoch": 1.2020861678004535, "grad_norm": 0.1474609375, "learning_rate": 0.00012135601517562109, "loss": 0.9832, "step": 3314 }, { "epoch": 1.20281179138322, "grad_norm": 0.1611328125, "learning_rate": 0.00012130706155917269, "loss": 0.9843, "step": 3316 }, { "epoch": 1.2035374149659863, "grad_norm": 0.1494140625, "learning_rate": 0.00012125810794272428, "loss": 0.9591, "step": 3318 }, { "epoch": 1.2042630385487527, "grad_norm": 0.150390625, "learning_rate": 0.00012120915432627586, "loss": 0.9599, "step": 3320 }, { "epoch": 1.2049886621315193, "grad_norm": 0.158203125, "learning_rate": 0.00012116020070982746, "loss": 1.0021, "step": 3322 }, { "epoch": 1.2057142857142857, "grad_norm": 0.166015625, "learning_rate": 0.00012111124709337901, "loss": 1.0136, "step": 3324 }, { "epoch": 1.2064399092970521, "grad_norm": 0.205078125, "learning_rate": 0.00012106229347693061, "loss": 0.9422, "step": 3326 }, { "epoch": 1.2071655328798185, "grad_norm": 0.1806640625, "learning_rate": 0.0001210133398604822, "loss": 1.0347, "step": 3328 }, { "epoch": 1.2078911564625852, "grad_norm": 0.1591796875, "learning_rate": 0.00012096438624403379, "loss": 1.0006, "step": 3330 }, { "epoch": 1.2086167800453516, "grad_norm": 0.158203125, "learning_rate": 0.00012091543262758538, "loss": 0.9784, "step": 3332 }, { "epoch": 1.209342403628118, "grad_norm": 0.166015625, "learning_rate": 0.00012086647901113694, "loss": 1.0127, "step": 3334 }, { "epoch": 1.2100680272108844, "grad_norm": 0.15234375, "learning_rate": 0.00012081752539468854, "loss": 1.0538, "step": 3336 }, { "epoch": 1.2107936507936508, "grad_norm": 0.1591796875, "learning_rate": 0.00012076857177824013, "loss": 0.9858, "step": 3338 }, { "epoch": 1.2115192743764172, "grad_norm": 0.1806640625, "learning_rate": 0.00012071961816179171, "loss": 0.978, "step": 3340 }, { "epoch": 1.2122448979591836, "grad_norm": 0.1669921875, "learning_rate": 0.00012067066454534331, "loss": 0.9754, "step": 3342 }, { "epoch": 1.2129705215419502, "grad_norm": 0.1728515625, "learning_rate": 0.00012062171092889486, "loss": 0.9934, "step": 3344 }, { "epoch": 1.2136961451247166, "grad_norm": 0.1689453125, "learning_rate": 0.00012057275731244646, "loss": 0.9599, "step": 3346 }, { "epoch": 1.214421768707483, "grad_norm": 0.1611328125, "learning_rate": 0.00012052380369599806, "loss": 1.0248, "step": 3348 }, { "epoch": 1.2151473922902494, "grad_norm": 0.1611328125, "learning_rate": 0.00012047485007954964, "loss": 1.0095, "step": 3350 }, { "epoch": 1.215873015873016, "grad_norm": 0.154296875, "learning_rate": 0.00012042589646310121, "loss": 0.9968, "step": 3352 }, { "epoch": 1.2165986394557824, "grad_norm": 0.16796875, "learning_rate": 0.00012037694284665279, "loss": 0.9093, "step": 3354 }, { "epoch": 1.2173242630385488, "grad_norm": 0.142578125, "learning_rate": 0.00012032798923020439, "loss": 0.9017, "step": 3356 }, { "epoch": 1.2180498866213152, "grad_norm": 0.1640625, "learning_rate": 0.00012027903561375598, "loss": 0.9775, "step": 3358 }, { "epoch": 1.2187755102040816, "grad_norm": 0.169921875, "learning_rate": 0.00012023008199730756, "loss": 1.0245, "step": 3360 }, { "epoch": 1.219501133786848, "grad_norm": 0.158203125, "learning_rate": 0.00012018112838085913, "loss": 1.0026, "step": 3362 }, { "epoch": 1.2202267573696144, "grad_norm": 0.1748046875, "learning_rate": 0.00012013217476441071, "loss": 0.991, "step": 3364 }, { "epoch": 1.220952380952381, "grad_norm": 0.1787109375, "learning_rate": 0.00012008322114796231, "loss": 0.9144, "step": 3366 }, { "epoch": 1.2216780045351474, "grad_norm": 0.1572265625, "learning_rate": 0.0001200342675315139, "loss": 0.9246, "step": 3368 }, { "epoch": 1.2224036281179138, "grad_norm": 0.1630859375, "learning_rate": 0.00011998531391506549, "loss": 0.991, "step": 3370 }, { "epoch": 1.2231292517006802, "grad_norm": 0.16015625, "learning_rate": 0.00011993636029861706, "loss": 1.0354, "step": 3372 }, { "epoch": 1.2238548752834468, "grad_norm": 0.1513671875, "learning_rate": 0.00011988740668216864, "loss": 1.0143, "step": 3374 }, { "epoch": 1.2245804988662132, "grad_norm": 0.259765625, "learning_rate": 0.00011983845306572024, "loss": 1.0311, "step": 3376 }, { "epoch": 1.2253061224489796, "grad_norm": 0.1572265625, "learning_rate": 0.00011978949944927183, "loss": 0.9852, "step": 3378 }, { "epoch": 1.226031746031746, "grad_norm": 0.18359375, "learning_rate": 0.00011974054583282341, "loss": 0.9433, "step": 3380 }, { "epoch": 1.2267573696145124, "grad_norm": 0.1552734375, "learning_rate": 0.00011969159221637498, "loss": 0.9716, "step": 3382 }, { "epoch": 1.2274829931972788, "grad_norm": 0.15234375, "learning_rate": 0.00011964263859992656, "loss": 0.9286, "step": 3384 }, { "epoch": 1.2282086167800454, "grad_norm": 0.142578125, "learning_rate": 0.00011959368498347816, "loss": 0.9936, "step": 3386 }, { "epoch": 1.2289342403628118, "grad_norm": 0.16015625, "learning_rate": 0.00011954473136702976, "loss": 0.9766, "step": 3388 }, { "epoch": 1.2296598639455782, "grad_norm": 0.1630859375, "learning_rate": 0.00011949577775058134, "loss": 0.9575, "step": 3390 }, { "epoch": 1.2303854875283446, "grad_norm": 0.1650390625, "learning_rate": 0.00011944682413413291, "loss": 0.9791, "step": 3392 }, { "epoch": 1.231111111111111, "grad_norm": 0.1806640625, "learning_rate": 0.00011939787051768449, "loss": 1.0045, "step": 3394 }, { "epoch": 1.2318367346938777, "grad_norm": 0.1689453125, "learning_rate": 0.00011934891690123609, "loss": 1.0042, "step": 3396 }, { "epoch": 1.232562358276644, "grad_norm": 0.1689453125, "learning_rate": 0.00011929996328478768, "loss": 0.9742, "step": 3398 }, { "epoch": 1.2332879818594105, "grad_norm": 0.1904296875, "learning_rate": 0.00011925100966833926, "loss": 0.9748, "step": 3400 }, { "epoch": 1.2332879818594105, "eval_loss": 1.0148086547851562, "eval_runtime": 94.1179, "eval_samples_per_second": 19.125, "eval_steps_per_second": 0.606, "step": 3400 }, { "epoch": 1.2340136054421769, "grad_norm": 0.150390625, "learning_rate": 0.00011920205605189083, "loss": 0.9734, "step": 3402 }, { "epoch": 1.2347392290249433, "grad_norm": 0.150390625, "learning_rate": 0.00011915310243544241, "loss": 0.9926, "step": 3404 }, { "epoch": 1.2354648526077097, "grad_norm": 0.162109375, "learning_rate": 0.00011910414881899401, "loss": 0.9998, "step": 3406 }, { "epoch": 1.2361904761904763, "grad_norm": 0.1572265625, "learning_rate": 0.0001190551952025456, "loss": 0.9564, "step": 3408 }, { "epoch": 1.2369160997732427, "grad_norm": 0.1572265625, "learning_rate": 0.00011900624158609719, "loss": 0.9322, "step": 3410 }, { "epoch": 1.237641723356009, "grad_norm": 0.1494140625, "learning_rate": 0.00011895728796964876, "loss": 0.9525, "step": 3412 }, { "epoch": 1.2383673469387755, "grad_norm": 0.1669921875, "learning_rate": 0.00011890833435320034, "loss": 1.0165, "step": 3414 }, { "epoch": 1.2390929705215419, "grad_norm": 0.15625, "learning_rate": 0.00011885938073675194, "loss": 0.9926, "step": 3416 }, { "epoch": 1.2398185941043085, "grad_norm": 0.1611328125, "learning_rate": 0.00011881042712030353, "loss": 0.9599, "step": 3418 }, { "epoch": 1.240544217687075, "grad_norm": 0.1669921875, "learning_rate": 0.00011876147350385511, "loss": 0.9723, "step": 3420 }, { "epoch": 1.2412698412698413, "grad_norm": 0.16796875, "learning_rate": 0.00011871251988740668, "loss": 1.0077, "step": 3422 }, { "epoch": 1.2419954648526077, "grad_norm": 0.1533203125, "learning_rate": 0.00011866356627095826, "loss": 0.9708, "step": 3424 }, { "epoch": 1.242721088435374, "grad_norm": 0.17578125, "learning_rate": 0.00011861461265450986, "loss": 1.0205, "step": 3426 }, { "epoch": 1.2434467120181405, "grad_norm": 0.166015625, "learning_rate": 0.00011856565903806146, "loss": 0.9795, "step": 3428 }, { "epoch": 1.2441723356009071, "grad_norm": 0.1552734375, "learning_rate": 0.00011851670542161301, "loss": 1.0165, "step": 3430 }, { "epoch": 1.2448979591836735, "grad_norm": 0.1708984375, "learning_rate": 0.00011846775180516461, "loss": 1.0622, "step": 3432 }, { "epoch": 1.24562358276644, "grad_norm": 0.169921875, "learning_rate": 0.00011841879818871619, "loss": 1.0099, "step": 3434 }, { "epoch": 1.2463492063492063, "grad_norm": 0.1455078125, "learning_rate": 0.00011836984457226779, "loss": 0.9006, "step": 3436 }, { "epoch": 1.2470748299319727, "grad_norm": 0.16015625, "learning_rate": 0.00011832089095581938, "loss": 1.0503, "step": 3438 }, { "epoch": 1.2478004535147393, "grad_norm": 0.158203125, "learning_rate": 0.00011827193733937094, "loss": 0.9561, "step": 3440 }, { "epoch": 1.2485260770975057, "grad_norm": 0.15234375, "learning_rate": 0.00011822298372292253, "loss": 0.9966, "step": 3442 }, { "epoch": 1.2492517006802721, "grad_norm": 0.1552734375, "learning_rate": 0.00011817403010647411, "loss": 1.0036, "step": 3444 }, { "epoch": 1.2499773242630385, "grad_norm": 0.1669921875, "learning_rate": 0.00011812507649002571, "loss": 1.0254, "step": 3446 }, { "epoch": 1.250702947845805, "grad_norm": 0.1396484375, "learning_rate": 0.0001180761228735773, "loss": 0.9544, "step": 3448 }, { "epoch": 1.2514285714285713, "grad_norm": 0.17578125, "learning_rate": 0.00011802716925712886, "loss": 0.9625, "step": 3450 }, { "epoch": 1.2521541950113377, "grad_norm": 0.142578125, "learning_rate": 0.00011797821564068046, "loss": 0.9375, "step": 3452 }, { "epoch": 1.2528798185941044, "grad_norm": 0.16796875, "learning_rate": 0.00011792926202423204, "loss": 1.0041, "step": 3454 }, { "epoch": 1.2536054421768708, "grad_norm": 0.1455078125, "learning_rate": 0.00011788030840778364, "loss": 0.9661, "step": 3456 }, { "epoch": 1.2543310657596372, "grad_norm": 0.1748046875, "learning_rate": 0.00011783135479133523, "loss": 1.0311, "step": 3458 }, { "epoch": 1.2550566893424036, "grad_norm": 0.146484375, "learning_rate": 0.00011778240117488679, "loss": 0.9824, "step": 3460 }, { "epoch": 1.2557823129251702, "grad_norm": 0.16796875, "learning_rate": 0.00011773344755843838, "loss": 0.9351, "step": 3462 }, { "epoch": 1.2565079365079366, "grad_norm": 0.181640625, "learning_rate": 0.00011768449394198996, "loss": 1.0692, "step": 3464 }, { "epoch": 1.257233560090703, "grad_norm": 0.16796875, "learning_rate": 0.00011763554032554156, "loss": 1.0367, "step": 3466 }, { "epoch": 1.2579591836734694, "grad_norm": 0.162109375, "learning_rate": 0.00011758658670909316, "loss": 0.9531, "step": 3468 }, { "epoch": 1.2586848072562358, "grad_norm": 0.16015625, "learning_rate": 0.00011753763309264471, "loss": 1.0415, "step": 3470 }, { "epoch": 1.2594104308390022, "grad_norm": 0.16015625, "learning_rate": 0.00011748867947619631, "loss": 0.9577, "step": 3472 }, { "epoch": 1.2601360544217686, "grad_norm": 0.1552734375, "learning_rate": 0.00011743972585974789, "loss": 1.0151, "step": 3474 }, { "epoch": 1.2608616780045352, "grad_norm": 0.158203125, "learning_rate": 0.00011739077224329948, "loss": 0.9575, "step": 3476 }, { "epoch": 1.2615873015873016, "grad_norm": 0.150390625, "learning_rate": 0.00011734181862685108, "loss": 0.9676, "step": 3478 }, { "epoch": 1.262312925170068, "grad_norm": 0.1611328125, "learning_rate": 0.00011729286501040264, "loss": 0.907, "step": 3480 }, { "epoch": 1.2630385487528344, "grad_norm": 0.1689453125, "learning_rate": 0.00011724391139395423, "loss": 0.9871, "step": 3482 }, { "epoch": 1.263764172335601, "grad_norm": 0.16015625, "learning_rate": 0.00011719495777750581, "loss": 0.9981, "step": 3484 }, { "epoch": 1.2644897959183674, "grad_norm": 0.1552734375, "learning_rate": 0.00011714600416105741, "loss": 0.993, "step": 3486 }, { "epoch": 1.2652154195011338, "grad_norm": 0.1787109375, "learning_rate": 0.000117097050544609, "loss": 1.0024, "step": 3488 }, { "epoch": 1.2659410430839002, "grad_norm": 0.1748046875, "learning_rate": 0.00011704809692816056, "loss": 0.9617, "step": 3490 }, { "epoch": 1.2666666666666666, "grad_norm": 0.1640625, "learning_rate": 0.00011699914331171216, "loss": 0.9186, "step": 3492 }, { "epoch": 1.267392290249433, "grad_norm": 0.162109375, "learning_rate": 0.00011695018969526374, "loss": 0.9795, "step": 3494 }, { "epoch": 1.2681179138321996, "grad_norm": 0.1669921875, "learning_rate": 0.00011690123607881533, "loss": 1.0009, "step": 3496 }, { "epoch": 1.268843537414966, "grad_norm": 0.171875, "learning_rate": 0.0001168522824623669, "loss": 1.001, "step": 3498 }, { "epoch": 1.2695691609977324, "grad_norm": 0.1689453125, "learning_rate": 0.00011680332884591849, "loss": 1.0089, "step": 3500 }, { "epoch": 1.2695691609977324, "eval_loss": 1.0145820379257202, "eval_runtime": 87.8413, "eval_samples_per_second": 20.491, "eval_steps_per_second": 0.649, "step": 3500 }, { "epoch": 1.2702947845804988, "grad_norm": 0.1767578125, "learning_rate": 0.00011675437522947008, "loss": 1.0022, "step": 3502 }, { "epoch": 1.2710204081632652, "grad_norm": 0.154296875, "learning_rate": 0.00011670542161302166, "loss": 1.0038, "step": 3504 }, { "epoch": 1.2717460317460318, "grad_norm": 0.158203125, "learning_rate": 0.00011665646799657326, "loss": 0.9343, "step": 3506 }, { "epoch": 1.2724716553287982, "grad_norm": 0.16015625, "learning_rate": 0.00011660751438012483, "loss": 1.0078, "step": 3508 }, { "epoch": 1.2731972789115646, "grad_norm": 0.154296875, "learning_rate": 0.00011655856076367641, "loss": 0.9831, "step": 3510 }, { "epoch": 1.273922902494331, "grad_norm": 0.1640625, "learning_rate": 0.000116509607147228, "loss": 1.0197, "step": 3512 }, { "epoch": 1.2746485260770974, "grad_norm": 0.162109375, "learning_rate": 0.00011646065353077959, "loss": 1.0065, "step": 3514 }, { "epoch": 1.2753741496598638, "grad_norm": 0.16015625, "learning_rate": 0.00011641169991433118, "loss": 0.9438, "step": 3516 }, { "epoch": 1.2760997732426305, "grad_norm": 0.1708984375, "learning_rate": 0.00011636274629788275, "loss": 1.0004, "step": 3518 }, { "epoch": 1.2768253968253969, "grad_norm": 0.1708984375, "learning_rate": 0.00011631379268143434, "loss": 0.9518, "step": 3520 }, { "epoch": 1.2775510204081633, "grad_norm": 0.1689453125, "learning_rate": 0.00011626483906498593, "loss": 0.9643, "step": 3522 }, { "epoch": 1.2782766439909297, "grad_norm": 0.1552734375, "learning_rate": 0.00011621588544853751, "loss": 0.977, "step": 3524 }, { "epoch": 1.2790022675736963, "grad_norm": 0.1708984375, "learning_rate": 0.00011616693183208911, "loss": 1.0163, "step": 3526 }, { "epoch": 1.2797278911564627, "grad_norm": 0.2109375, "learning_rate": 0.00011611797821564068, "loss": 0.9853, "step": 3528 }, { "epoch": 1.280453514739229, "grad_norm": 0.1611328125, "learning_rate": 0.00011606902459919226, "loss": 1.0327, "step": 3530 }, { "epoch": 1.2811791383219955, "grad_norm": 0.1708984375, "learning_rate": 0.00011602007098274386, "loss": 1.039, "step": 3532 }, { "epoch": 1.2819047619047619, "grad_norm": 0.1640625, "learning_rate": 0.00011597111736629544, "loss": 1.0176, "step": 3534 }, { "epoch": 1.2826303854875283, "grad_norm": 0.1533203125, "learning_rate": 0.00011592216374984703, "loss": 0.9863, "step": 3536 }, { "epoch": 1.2833560090702947, "grad_norm": 0.1640625, "learning_rate": 0.0001158732101333986, "loss": 0.9882, "step": 3538 }, { "epoch": 1.2840816326530613, "grad_norm": 0.1630859375, "learning_rate": 0.00011582425651695019, "loss": 1.0326, "step": 3540 }, { "epoch": 1.2848072562358277, "grad_norm": 0.171875, "learning_rate": 0.00011577530290050178, "loss": 0.9556, "step": 3542 }, { "epoch": 1.285532879818594, "grad_norm": 0.1484375, "learning_rate": 0.00011572634928405336, "loss": 0.9931, "step": 3544 }, { "epoch": 1.2862585034013605, "grad_norm": 0.16796875, "learning_rate": 0.00011567739566760496, "loss": 0.9452, "step": 3546 }, { "epoch": 1.2869841269841271, "grad_norm": 0.1669921875, "learning_rate": 0.00011562844205115653, "loss": 0.9842, "step": 3548 }, { "epoch": 1.2877097505668935, "grad_norm": 0.197265625, "learning_rate": 0.00011557948843470811, "loss": 1.0206, "step": 3550 }, { "epoch": 1.28843537414966, "grad_norm": 0.134765625, "learning_rate": 0.0001155305348182597, "loss": 0.909, "step": 3552 }, { "epoch": 1.2891609977324263, "grad_norm": 0.1640625, "learning_rate": 0.00011548158120181129, "loss": 1.0186, "step": 3554 }, { "epoch": 1.2898866213151927, "grad_norm": 0.1396484375, "learning_rate": 0.00011543262758536288, "loss": 1.0197, "step": 3556 }, { "epoch": 1.2906122448979591, "grad_norm": 0.1669921875, "learning_rate": 0.00011538367396891445, "loss": 0.9857, "step": 3558 }, { "epoch": 1.2913378684807255, "grad_norm": 0.15625, "learning_rate": 0.00011533472035246604, "loss": 0.9722, "step": 3560 }, { "epoch": 1.2920634920634921, "grad_norm": 0.1650390625, "learning_rate": 0.00011528576673601763, "loss": 1.0129, "step": 3562 }, { "epoch": 1.2927891156462585, "grad_norm": 0.1650390625, "learning_rate": 0.00011523681311956921, "loss": 0.9774, "step": 3564 }, { "epoch": 1.293514739229025, "grad_norm": 0.1767578125, "learning_rate": 0.00011518785950312081, "loss": 1.0214, "step": 3566 }, { "epoch": 1.2942403628117913, "grad_norm": 0.166015625, "learning_rate": 0.00011513890588667238, "loss": 0.9524, "step": 3568 }, { "epoch": 1.294965986394558, "grad_norm": 0.1611328125, "learning_rate": 0.00011508995227022396, "loss": 0.9879, "step": 3570 }, { "epoch": 1.2956916099773244, "grad_norm": 0.1552734375, "learning_rate": 0.00011504099865377556, "loss": 0.946, "step": 3572 }, { "epoch": 1.2964172335600908, "grad_norm": 0.17578125, "learning_rate": 0.00011499204503732714, "loss": 1.0078, "step": 3574 }, { "epoch": 1.2971428571428572, "grad_norm": 0.162109375, "learning_rate": 0.00011494309142087871, "loss": 1.0155, "step": 3576 }, { "epoch": 1.2978684807256236, "grad_norm": 0.1435546875, "learning_rate": 0.0001148941378044303, "loss": 0.9912, "step": 3578 }, { "epoch": 1.29859410430839, "grad_norm": 0.1689453125, "learning_rate": 0.00011484518418798189, "loss": 0.9606, "step": 3580 }, { "epoch": 1.2993197278911564, "grad_norm": 0.1650390625, "learning_rate": 0.00011479623057153348, "loss": 0.9986, "step": 3582 }, { "epoch": 1.300045351473923, "grad_norm": 0.1796875, "learning_rate": 0.00011474727695508506, "loss": 0.9684, "step": 3584 }, { "epoch": 1.3007709750566894, "grad_norm": 0.1689453125, "learning_rate": 0.00011469832333863663, "loss": 0.9675, "step": 3586 }, { "epoch": 1.3014965986394558, "grad_norm": 0.1533203125, "learning_rate": 0.00011464936972218823, "loss": 0.971, "step": 3588 }, { "epoch": 1.3022222222222222, "grad_norm": 0.1591796875, "learning_rate": 0.00011460041610573981, "loss": 0.9238, "step": 3590 }, { "epoch": 1.3029478458049888, "grad_norm": 0.1630859375, "learning_rate": 0.0001145514624892914, "loss": 0.9196, "step": 3592 }, { "epoch": 1.3036734693877552, "grad_norm": 0.1552734375, "learning_rate": 0.00011450250887284299, "loss": 0.9861, "step": 3594 }, { "epoch": 1.3043990929705216, "grad_norm": 0.185546875, "learning_rate": 0.00011445355525639456, "loss": 1.0305, "step": 3596 }, { "epoch": 1.305124716553288, "grad_norm": 0.1513671875, "learning_rate": 0.00011440460163994615, "loss": 0.9368, "step": 3598 }, { "epoch": 1.3058503401360544, "grad_norm": 0.15625, "learning_rate": 0.00011435564802349774, "loss": 0.9788, "step": 3600 }, { "epoch": 1.3058503401360544, "eval_loss": 1.014317274093628, "eval_runtime": 87.2865, "eval_samples_per_second": 20.622, "eval_steps_per_second": 0.653, "step": 3600 }, { "epoch": 1.3065759637188208, "grad_norm": 0.15234375, "learning_rate": 0.00011430669440704933, "loss": 0.9431, "step": 3602 }, { "epoch": 1.3073015873015872, "grad_norm": 0.154296875, "learning_rate": 0.00011425774079060091, "loss": 0.9195, "step": 3604 }, { "epoch": 1.3080272108843538, "grad_norm": 0.1806640625, "learning_rate": 0.00011420878717415248, "loss": 0.9597, "step": 3606 }, { "epoch": 1.3087528344671202, "grad_norm": 0.1533203125, "learning_rate": 0.00011415983355770408, "loss": 0.9325, "step": 3608 }, { "epoch": 1.3094784580498866, "grad_norm": 0.1689453125, "learning_rate": 0.00011411087994125566, "loss": 0.9972, "step": 3610 }, { "epoch": 1.310204081632653, "grad_norm": 0.1591796875, "learning_rate": 0.00011406192632480726, "loss": 1.0023, "step": 3612 }, { "epoch": 1.3109297052154196, "grad_norm": 0.1767578125, "learning_rate": 0.00011401297270835884, "loss": 0.9872, "step": 3614 }, { "epoch": 1.311655328798186, "grad_norm": 0.150390625, "learning_rate": 0.00011396401909191041, "loss": 1.0391, "step": 3616 }, { "epoch": 1.3123809523809524, "grad_norm": 0.1572265625, "learning_rate": 0.000113915065475462, "loss": 0.9912, "step": 3618 }, { "epoch": 1.3131065759637188, "grad_norm": 0.1640625, "learning_rate": 0.00011386611185901359, "loss": 1.0563, "step": 3620 }, { "epoch": 1.3138321995464852, "grad_norm": 0.1640625, "learning_rate": 0.00011381715824256518, "loss": 0.9985, "step": 3622 }, { "epoch": 1.3145578231292516, "grad_norm": 0.1591796875, "learning_rate": 0.00011376820462611676, "loss": 0.9683, "step": 3624 }, { "epoch": 1.315283446712018, "grad_norm": 0.17578125, "learning_rate": 0.00011371925100966833, "loss": 0.9682, "step": 3626 }, { "epoch": 1.3160090702947846, "grad_norm": 0.16015625, "learning_rate": 0.00011367029739321993, "loss": 1.0611, "step": 3628 }, { "epoch": 1.316734693877551, "grad_norm": 0.166015625, "learning_rate": 0.00011362134377677151, "loss": 0.9994, "step": 3630 }, { "epoch": 1.3174603174603174, "grad_norm": 0.140625, "learning_rate": 0.0001135723901603231, "loss": 0.9085, "step": 3632 }, { "epoch": 1.3181859410430838, "grad_norm": 0.16796875, "learning_rate": 0.00011352343654387469, "loss": 0.9891, "step": 3634 }, { "epoch": 1.3189115646258505, "grad_norm": 0.1591796875, "learning_rate": 0.00011347448292742626, "loss": 1.0091, "step": 3636 }, { "epoch": 1.3196371882086169, "grad_norm": 0.1650390625, "learning_rate": 0.00011342552931097785, "loss": 0.9623, "step": 3638 }, { "epoch": 1.3203628117913833, "grad_norm": 0.1650390625, "learning_rate": 0.00011337657569452944, "loss": 1.0065, "step": 3640 }, { "epoch": 1.3210884353741497, "grad_norm": 0.1533203125, "learning_rate": 0.00011332762207808103, "loss": 0.9271, "step": 3642 }, { "epoch": 1.321814058956916, "grad_norm": 0.17578125, "learning_rate": 0.0001132786684616326, "loss": 0.9958, "step": 3644 }, { "epoch": 1.3225396825396825, "grad_norm": 0.15625, "learning_rate": 0.00011322971484518418, "loss": 0.9128, "step": 3646 }, { "epoch": 1.3232653061224489, "grad_norm": 0.158203125, "learning_rate": 0.00011318076122873578, "loss": 0.9164, "step": 3648 }, { "epoch": 1.3239909297052155, "grad_norm": 0.1708984375, "learning_rate": 0.00011313180761228736, "loss": 1.0168, "step": 3650 }, { "epoch": 1.3247165532879819, "grad_norm": 0.162109375, "learning_rate": 0.00011308285399583896, "loss": 0.9683, "step": 3652 }, { "epoch": 1.3254421768707483, "grad_norm": 0.1689453125, "learning_rate": 0.00011303390037939052, "loss": 0.971, "step": 3654 }, { "epoch": 1.3261678004535147, "grad_norm": 0.14453125, "learning_rate": 0.00011298494676294211, "loss": 0.9244, "step": 3656 }, { "epoch": 1.3268934240362813, "grad_norm": 0.2060546875, "learning_rate": 0.0001129359931464937, "loss": 0.9584, "step": 3658 }, { "epoch": 1.3276190476190477, "grad_norm": 0.16015625, "learning_rate": 0.00011288703953004529, "loss": 0.882, "step": 3660 }, { "epoch": 1.328344671201814, "grad_norm": 0.1708984375, "learning_rate": 0.00011283808591359688, "loss": 0.9848, "step": 3662 }, { "epoch": 1.3290702947845805, "grad_norm": 0.1650390625, "learning_rate": 0.00011278913229714845, "loss": 1.1064, "step": 3664 }, { "epoch": 1.329795918367347, "grad_norm": 0.1533203125, "learning_rate": 0.00011274017868070003, "loss": 0.9595, "step": 3666 }, { "epoch": 1.3305215419501133, "grad_norm": 0.1455078125, "learning_rate": 0.00011269122506425163, "loss": 0.9298, "step": 3668 }, { "epoch": 1.3312471655328797, "grad_norm": 0.185546875, "learning_rate": 0.00011264227144780321, "loss": 1.0159, "step": 3670 }, { "epoch": 1.3319727891156463, "grad_norm": 0.14453125, "learning_rate": 0.0001125933178313548, "loss": 0.9367, "step": 3672 }, { "epoch": 1.3326984126984127, "grad_norm": 0.166015625, "learning_rate": 0.00011254436421490637, "loss": 0.9847, "step": 3674 }, { "epoch": 1.3334240362811791, "grad_norm": 0.1376953125, "learning_rate": 0.00011249541059845796, "loss": 0.9506, "step": 3676 }, { "epoch": 1.3341496598639455, "grad_norm": 0.154296875, "learning_rate": 0.00011244645698200955, "loss": 1.0063, "step": 3678 }, { "epoch": 1.3348752834467121, "grad_norm": 0.1376953125, "learning_rate": 0.00011239750336556114, "loss": 0.9438, "step": 3680 }, { "epoch": 1.3356009070294785, "grad_norm": 0.1767578125, "learning_rate": 0.00011234854974911273, "loss": 0.968, "step": 3682 }, { "epoch": 1.336326530612245, "grad_norm": 0.1650390625, "learning_rate": 0.0001122995961326643, "loss": 0.9975, "step": 3684 }, { "epoch": 1.3370521541950113, "grad_norm": 0.171875, "learning_rate": 0.00011225064251621588, "loss": 1.0121, "step": 3686 }, { "epoch": 1.3377777777777777, "grad_norm": 0.15234375, "learning_rate": 0.00011220168889976748, "loss": 0.9627, "step": 3688 }, { "epoch": 1.3385034013605441, "grad_norm": 0.1572265625, "learning_rate": 0.00011215273528331906, "loss": 1.0203, "step": 3690 }, { "epoch": 1.3392290249433105, "grad_norm": 0.185546875, "learning_rate": 0.00011210378166687066, "loss": 1.0594, "step": 3692 }, { "epoch": 1.3399546485260772, "grad_norm": 0.1513671875, "learning_rate": 0.00011205482805042222, "loss": 0.9768, "step": 3694 }, { "epoch": 1.3406802721088436, "grad_norm": 0.1484375, "learning_rate": 0.00011200587443397381, "loss": 1.027, "step": 3696 }, { "epoch": 1.34140589569161, "grad_norm": 0.162109375, "learning_rate": 0.0001119569208175254, "loss": 1.0098, "step": 3698 }, { "epoch": 1.3421315192743764, "grad_norm": 0.1572265625, "learning_rate": 0.00011190796720107698, "loss": 0.975, "step": 3700 }, { "epoch": 1.3421315192743764, "eval_loss": 1.0139893293380737, "eval_runtime": 87.2842, "eval_samples_per_second": 20.622, "eval_steps_per_second": 0.653, "step": 3700 }, { "epoch": 1.342857142857143, "grad_norm": 0.1787109375, "learning_rate": 0.00011185901358462858, "loss": 0.9606, "step": 3702 }, { "epoch": 1.3435827664399094, "grad_norm": 0.1826171875, "learning_rate": 0.00011181005996818015, "loss": 1.0558, "step": 3704 }, { "epoch": 1.3443083900226758, "grad_norm": 0.1640625, "learning_rate": 0.00011176110635173173, "loss": 0.9632, "step": 3706 }, { "epoch": 1.3450340136054422, "grad_norm": 0.1591796875, "learning_rate": 0.00011171215273528333, "loss": 1.0016, "step": 3708 }, { "epoch": 1.3457596371882086, "grad_norm": 0.158203125, "learning_rate": 0.00011166319911883491, "loss": 0.9078, "step": 3710 }, { "epoch": 1.346485260770975, "grad_norm": 0.1669921875, "learning_rate": 0.0001116142455023865, "loss": 1.0287, "step": 3712 }, { "epoch": 1.3472108843537414, "grad_norm": 0.15625, "learning_rate": 0.00011156529188593807, "loss": 0.9694, "step": 3714 }, { "epoch": 1.347936507936508, "grad_norm": 0.2001953125, "learning_rate": 0.00011151633826948966, "loss": 0.9519, "step": 3716 }, { "epoch": 1.3486621315192744, "grad_norm": 0.158203125, "learning_rate": 0.00011146738465304125, "loss": 0.9624, "step": 3718 }, { "epoch": 1.3493877551020408, "grad_norm": 0.154296875, "learning_rate": 0.00011141843103659283, "loss": 0.9302, "step": 3720 }, { "epoch": 1.3501133786848072, "grad_norm": 0.150390625, "learning_rate": 0.0001113694774201444, "loss": 0.9119, "step": 3722 }, { "epoch": 1.3508390022675738, "grad_norm": 0.1455078125, "learning_rate": 0.000111320523803696, "loss": 0.9269, "step": 3724 }, { "epoch": 1.3515646258503402, "grad_norm": 0.1767578125, "learning_rate": 0.00011127157018724758, "loss": 1.0261, "step": 3726 }, { "epoch": 1.3522902494331066, "grad_norm": 0.1591796875, "learning_rate": 0.00011122261657079918, "loss": 1.0123, "step": 3728 }, { "epoch": 1.353015873015873, "grad_norm": 0.1689453125, "learning_rate": 0.00011117366295435076, "loss": 1.0042, "step": 3730 }, { "epoch": 1.3537414965986394, "grad_norm": 0.169921875, "learning_rate": 0.00011112470933790233, "loss": 0.9707, "step": 3732 }, { "epoch": 1.3544671201814058, "grad_norm": 0.1806640625, "learning_rate": 0.00011107575572145392, "loss": 0.9504, "step": 3734 }, { "epoch": 1.3551927437641722, "grad_norm": 0.1796875, "learning_rate": 0.0001110268021050055, "loss": 0.9999, "step": 3736 }, { "epoch": 1.3559183673469388, "grad_norm": 0.189453125, "learning_rate": 0.0001109778484885571, "loss": 1.0107, "step": 3738 }, { "epoch": 1.3566439909297052, "grad_norm": 0.16015625, "learning_rate": 0.00011092889487210868, "loss": 0.9823, "step": 3740 }, { "epoch": 1.3573696145124716, "grad_norm": 0.1689453125, "learning_rate": 0.00011087994125566025, "loss": 0.9654, "step": 3742 }, { "epoch": 1.358095238095238, "grad_norm": 0.2099609375, "learning_rate": 0.00011083098763921185, "loss": 0.983, "step": 3744 }, { "epoch": 1.3588208616780046, "grad_norm": 0.158203125, "learning_rate": 0.00011078203402276343, "loss": 1.0468, "step": 3746 }, { "epoch": 1.359546485260771, "grad_norm": 0.1669921875, "learning_rate": 0.00011073308040631503, "loss": 0.9342, "step": 3748 }, { "epoch": 1.3602721088435374, "grad_norm": 0.15625, "learning_rate": 0.00011068412678986661, "loss": 0.9798, "step": 3750 }, { "epoch": 1.3609977324263038, "grad_norm": 0.1787109375, "learning_rate": 0.00011063517317341818, "loss": 0.9705, "step": 3752 }, { "epoch": 1.3617233560090702, "grad_norm": 0.1884765625, "learning_rate": 0.00011058621955696977, "loss": 1.0327, "step": 3754 }, { "epoch": 1.3624489795918366, "grad_norm": 0.1708984375, "learning_rate": 0.00011053726594052136, "loss": 1.0191, "step": 3756 }, { "epoch": 1.363174603174603, "grad_norm": 0.1689453125, "learning_rate": 0.00011048831232407295, "loss": 0.9705, "step": 3758 }, { "epoch": 1.3639002267573697, "grad_norm": 0.1611328125, "learning_rate": 0.00011043935870762453, "loss": 0.9572, "step": 3760 }, { "epoch": 1.364625850340136, "grad_norm": 0.15234375, "learning_rate": 0.0001103904050911761, "loss": 0.9237, "step": 3762 }, { "epoch": 1.3653514739229025, "grad_norm": 0.169921875, "learning_rate": 0.0001103414514747277, "loss": 0.8971, "step": 3764 }, { "epoch": 1.3660770975056689, "grad_norm": 0.154296875, "learning_rate": 0.00011029249785827928, "loss": 0.9106, "step": 3766 }, { "epoch": 1.3668027210884355, "grad_norm": 0.216796875, "learning_rate": 0.00011024354424183088, "loss": 1.0726, "step": 3768 }, { "epoch": 1.3675283446712019, "grad_norm": 0.1474609375, "learning_rate": 0.00011019459062538246, "loss": 0.946, "step": 3770 }, { "epoch": 1.3682539682539683, "grad_norm": 0.1513671875, "learning_rate": 0.00011014563700893403, "loss": 0.936, "step": 3772 }, { "epoch": 1.3689795918367347, "grad_norm": 0.166015625, "learning_rate": 0.00011009668339248562, "loss": 0.9972, "step": 3774 }, { "epoch": 1.369705215419501, "grad_norm": 0.171875, "learning_rate": 0.0001100477297760372, "loss": 1.0203, "step": 3776 }, { "epoch": 1.3704308390022675, "grad_norm": 0.1826171875, "learning_rate": 0.0001099987761595888, "loss": 1.016, "step": 3778 }, { "epoch": 1.3711564625850339, "grad_norm": 0.18359375, "learning_rate": 0.00010994982254314038, "loss": 0.9996, "step": 3780 }, { "epoch": 1.3718820861678005, "grad_norm": 0.16796875, "learning_rate": 0.00010990086892669195, "loss": 0.9701, "step": 3782 }, { "epoch": 1.372607709750567, "grad_norm": 0.1630859375, "learning_rate": 0.00010985191531024355, "loss": 0.9935, "step": 3784 }, { "epoch": 1.3733333333333333, "grad_norm": 0.1396484375, "learning_rate": 0.00010980296169379513, "loss": 0.9978, "step": 3786 }, { "epoch": 1.3740589569160997, "grad_norm": 0.16015625, "learning_rate": 0.00010975400807734673, "loss": 0.9602, "step": 3788 }, { "epoch": 1.3747845804988663, "grad_norm": 0.150390625, "learning_rate": 0.0001097050544608983, "loss": 0.9894, "step": 3790 }, { "epoch": 1.3755102040816327, "grad_norm": 0.154296875, "learning_rate": 0.00010965610084444988, "loss": 0.9851, "step": 3792 }, { "epoch": 1.3762358276643991, "grad_norm": 0.1669921875, "learning_rate": 0.00010960714722800147, "loss": 1.0069, "step": 3794 }, { "epoch": 1.3769614512471655, "grad_norm": 0.158203125, "learning_rate": 0.00010955819361155306, "loss": 0.9355, "step": 3796 }, { "epoch": 1.377687074829932, "grad_norm": 0.1884765625, "learning_rate": 0.00010950923999510465, "loss": 1.0524, "step": 3798 }, { "epoch": 1.3784126984126983, "grad_norm": 0.169921875, "learning_rate": 0.00010946028637865622, "loss": 0.9241, "step": 3800 }, { "epoch": 1.3784126984126983, "eval_loss": 1.0139381885528564, "eval_runtime": 88.7285, "eval_samples_per_second": 20.287, "eval_steps_per_second": 0.642, "step": 3800 }, { "epoch": 1.379138321995465, "grad_norm": 0.1494140625, "learning_rate": 0.0001094113327622078, "loss": 0.9646, "step": 3802 }, { "epoch": 1.3798639455782313, "grad_norm": 0.169921875, "learning_rate": 0.0001093623791457594, "loss": 0.9912, "step": 3804 }, { "epoch": 1.3805895691609977, "grad_norm": 0.166015625, "learning_rate": 0.00010931342552931098, "loss": 1.1027, "step": 3806 }, { "epoch": 1.3813151927437641, "grad_norm": 0.1689453125, "learning_rate": 0.00010926447191286258, "loss": 1.0653, "step": 3808 }, { "epoch": 1.3820408163265305, "grad_norm": 0.1484375, "learning_rate": 0.00010921551829641415, "loss": 0.8741, "step": 3810 }, { "epoch": 1.3827664399092972, "grad_norm": 0.16015625, "learning_rate": 0.00010916656467996573, "loss": 1.0081, "step": 3812 }, { "epoch": 1.3834920634920636, "grad_norm": 0.1689453125, "learning_rate": 0.00010911761106351732, "loss": 1.0412, "step": 3814 }, { "epoch": 1.38421768707483, "grad_norm": 0.158203125, "learning_rate": 0.0001090686574470689, "loss": 1.0026, "step": 3816 }, { "epoch": 1.3849433106575963, "grad_norm": 0.177734375, "learning_rate": 0.0001090197038306205, "loss": 1.0484, "step": 3818 }, { "epoch": 1.3856689342403627, "grad_norm": 0.1572265625, "learning_rate": 0.00010897075021417207, "loss": 1.0069, "step": 3820 }, { "epoch": 1.3863945578231291, "grad_norm": 0.1650390625, "learning_rate": 0.00010892179659772365, "loss": 0.9699, "step": 3822 }, { "epoch": 1.3871201814058958, "grad_norm": 0.1572265625, "learning_rate": 0.00010887284298127525, "loss": 0.9712, "step": 3824 }, { "epoch": 1.3878458049886622, "grad_norm": 0.154296875, "learning_rate": 0.00010882388936482683, "loss": 0.9634, "step": 3826 }, { "epoch": 1.3885714285714286, "grad_norm": 0.1416015625, "learning_rate": 0.00010877493574837843, "loss": 0.9306, "step": 3828 }, { "epoch": 1.389297052154195, "grad_norm": 0.18359375, "learning_rate": 0.00010872598213193, "loss": 0.9489, "step": 3830 }, { "epoch": 1.3900226757369616, "grad_norm": 0.15625, "learning_rate": 0.00010867702851548158, "loss": 1.0097, "step": 3832 }, { "epoch": 1.390748299319728, "grad_norm": 0.1513671875, "learning_rate": 0.00010862807489903317, "loss": 0.9603, "step": 3834 }, { "epoch": 1.3914739229024944, "grad_norm": 0.1513671875, "learning_rate": 0.00010857912128258476, "loss": 0.944, "step": 3836 }, { "epoch": 1.3921995464852608, "grad_norm": 0.1650390625, "learning_rate": 0.00010853016766613635, "loss": 0.9558, "step": 3838 }, { "epoch": 1.3929251700680272, "grad_norm": 0.181640625, "learning_rate": 0.00010848121404968792, "loss": 0.98, "step": 3840 }, { "epoch": 1.3936507936507936, "grad_norm": 0.158203125, "learning_rate": 0.0001084322604332395, "loss": 1.0133, "step": 3842 }, { "epoch": 1.39437641723356, "grad_norm": 0.1708984375, "learning_rate": 0.0001083833068167911, "loss": 1.0192, "step": 3844 }, { "epoch": 1.3951020408163266, "grad_norm": 0.1650390625, "learning_rate": 0.00010833435320034268, "loss": 1.0125, "step": 3846 }, { "epoch": 1.395827664399093, "grad_norm": 0.158203125, "learning_rate": 0.00010828539958389428, "loss": 0.9468, "step": 3848 }, { "epoch": 1.3965532879818594, "grad_norm": 0.1884765625, "learning_rate": 0.00010823644596744585, "loss": 0.9093, "step": 3850 }, { "epoch": 1.3972789115646258, "grad_norm": 0.171875, "learning_rate": 0.00010818749235099743, "loss": 0.9601, "step": 3852 }, { "epoch": 1.3980045351473924, "grad_norm": 0.1513671875, "learning_rate": 0.00010813853873454902, "loss": 0.9524, "step": 3854 }, { "epoch": 1.3987301587301588, "grad_norm": 0.1533203125, "learning_rate": 0.0001080895851181006, "loss": 0.9915, "step": 3856 }, { "epoch": 1.3994557823129252, "grad_norm": 0.162109375, "learning_rate": 0.0001080406315016522, "loss": 0.9875, "step": 3858 }, { "epoch": 1.4001814058956916, "grad_norm": 0.1689453125, "learning_rate": 0.00010799167788520377, "loss": 0.9353, "step": 3860 }, { "epoch": 1.400907029478458, "grad_norm": 0.1796875, "learning_rate": 0.00010794272426875535, "loss": 0.974, "step": 3862 }, { "epoch": 1.4016326530612244, "grad_norm": 0.16796875, "learning_rate": 0.00010789377065230695, "loss": 0.9781, "step": 3864 }, { "epoch": 1.4023582766439908, "grad_norm": 0.181640625, "learning_rate": 0.00010784481703585853, "loss": 0.9633, "step": 3866 }, { "epoch": 1.4030839002267574, "grad_norm": 0.1455078125, "learning_rate": 0.0001077958634194101, "loss": 0.909, "step": 3868 }, { "epoch": 1.4038095238095238, "grad_norm": 0.1806640625, "learning_rate": 0.0001077469098029617, "loss": 1.0195, "step": 3870 }, { "epoch": 1.4045351473922902, "grad_norm": 0.1748046875, "learning_rate": 0.00010769795618651328, "loss": 1.0009, "step": 3872 }, { "epoch": 1.4052607709750566, "grad_norm": 0.1630859375, "learning_rate": 0.00010764900257006487, "loss": 0.941, "step": 3874 }, { "epoch": 1.4059863945578233, "grad_norm": 0.146484375, "learning_rate": 0.00010760004895361646, "loss": 0.8787, "step": 3876 }, { "epoch": 1.4067120181405897, "grad_norm": 0.15234375, "learning_rate": 0.00010755109533716802, "loss": 1.0049, "step": 3878 }, { "epoch": 1.407437641723356, "grad_norm": 0.1669921875, "learning_rate": 0.00010750214172071962, "loss": 1.0413, "step": 3880 }, { "epoch": 1.4081632653061225, "grad_norm": 0.1748046875, "learning_rate": 0.0001074531881042712, "loss": 1.0002, "step": 3882 }, { "epoch": 1.4088888888888889, "grad_norm": 0.19140625, "learning_rate": 0.0001074042344878228, "loss": 1.0016, "step": 3884 }, { "epoch": 1.4096145124716553, "grad_norm": 0.1884765625, "learning_rate": 0.00010735528087137438, "loss": 0.9948, "step": 3886 }, { "epoch": 1.4103401360544217, "grad_norm": 0.173828125, "learning_rate": 0.00010730632725492595, "loss": 0.9512, "step": 3888 }, { "epoch": 1.4110657596371883, "grad_norm": 0.16796875, "learning_rate": 0.00010725737363847755, "loss": 0.9985, "step": 3890 }, { "epoch": 1.4117913832199547, "grad_norm": 0.166015625, "learning_rate": 0.00010720842002202913, "loss": 0.9329, "step": 3892 }, { "epoch": 1.412517006802721, "grad_norm": 0.1650390625, "learning_rate": 0.00010715946640558072, "loss": 0.9697, "step": 3894 }, { "epoch": 1.4132426303854875, "grad_norm": 0.1591796875, "learning_rate": 0.0001071105127891323, "loss": 0.9691, "step": 3896 }, { "epoch": 1.413968253968254, "grad_norm": 0.1474609375, "learning_rate": 0.00010706155917268387, "loss": 0.9706, "step": 3898 }, { "epoch": 1.4146938775510205, "grad_norm": 0.171875, "learning_rate": 0.00010701260555623547, "loss": 1.0023, "step": 3900 }, { "epoch": 1.4146938775510205, "eval_loss": 1.013636589050293, "eval_runtime": 92.7582, "eval_samples_per_second": 19.405, "eval_steps_per_second": 0.615, "step": 3900 }, { "epoch": 1.415419501133787, "grad_norm": 0.154296875, "learning_rate": 0.00010696365193978705, "loss": 0.9634, "step": 3902 }, { "epoch": 1.4161451247165533, "grad_norm": 0.1611328125, "learning_rate": 0.00010691469832333865, "loss": 0.9808, "step": 3904 }, { "epoch": 1.4168707482993197, "grad_norm": 0.169921875, "learning_rate": 0.00010686574470689023, "loss": 0.9637, "step": 3906 }, { "epoch": 1.417596371882086, "grad_norm": 0.1474609375, "learning_rate": 0.0001068167910904418, "loss": 0.9661, "step": 3908 }, { "epoch": 1.4183219954648525, "grad_norm": 0.1650390625, "learning_rate": 0.0001067678374739934, "loss": 0.9844, "step": 3910 }, { "epoch": 1.4190476190476191, "grad_norm": 0.150390625, "learning_rate": 0.00010671888385754498, "loss": 0.9415, "step": 3912 }, { "epoch": 1.4197732426303855, "grad_norm": 0.1552734375, "learning_rate": 0.00010666993024109657, "loss": 1.0491, "step": 3914 }, { "epoch": 1.420498866213152, "grad_norm": 0.1845703125, "learning_rate": 0.00010662097662464816, "loss": 1.007, "step": 3916 }, { "epoch": 1.4212244897959183, "grad_norm": 0.154296875, "learning_rate": 0.00010657202300819972, "loss": 0.9802, "step": 3918 }, { "epoch": 1.421950113378685, "grad_norm": 0.1611328125, "learning_rate": 0.00010652306939175132, "loss": 1.0043, "step": 3920 }, { "epoch": 1.4226757369614513, "grad_norm": 0.1611328125, "learning_rate": 0.0001064741157753029, "loss": 1.0034, "step": 3922 }, { "epoch": 1.4234013605442177, "grad_norm": 0.1806640625, "learning_rate": 0.0001064251621588545, "loss": 0.9387, "step": 3924 }, { "epoch": 1.4241269841269841, "grad_norm": 0.1650390625, "learning_rate": 0.00010637620854240608, "loss": 0.9047, "step": 3926 }, { "epoch": 1.4248526077097505, "grad_norm": 0.1552734375, "learning_rate": 0.00010632725492595765, "loss": 0.9024, "step": 3928 }, { "epoch": 1.425578231292517, "grad_norm": 0.1552734375, "learning_rate": 0.00010627830130950925, "loss": 1.024, "step": 3930 }, { "epoch": 1.4263038548752833, "grad_norm": 0.169921875, "learning_rate": 0.00010622934769306083, "loss": 0.9421, "step": 3932 }, { "epoch": 1.42702947845805, "grad_norm": 0.1640625, "learning_rate": 0.00010618039407661242, "loss": 1.0215, "step": 3934 }, { "epoch": 1.4277551020408163, "grad_norm": 0.1650390625, "learning_rate": 0.00010613144046016399, "loss": 0.9815, "step": 3936 }, { "epoch": 1.4284807256235827, "grad_norm": 0.1494140625, "learning_rate": 0.00010608248684371557, "loss": 0.9606, "step": 3938 }, { "epoch": 1.4292063492063491, "grad_norm": 0.158203125, "learning_rate": 0.00010603353322726717, "loss": 0.9918, "step": 3940 }, { "epoch": 1.4299319727891158, "grad_norm": 0.1630859375, "learning_rate": 0.00010598457961081875, "loss": 0.9695, "step": 3942 }, { "epoch": 1.4306575963718822, "grad_norm": 0.177734375, "learning_rate": 0.00010593562599437035, "loss": 0.9732, "step": 3944 }, { "epoch": 1.4313832199546486, "grad_norm": 0.1787109375, "learning_rate": 0.00010588667237792192, "loss": 1.0474, "step": 3946 }, { "epoch": 1.432108843537415, "grad_norm": 0.1728515625, "learning_rate": 0.0001058377187614735, "loss": 1.0052, "step": 3948 }, { "epoch": 1.4328344671201814, "grad_norm": 0.1533203125, "learning_rate": 0.0001057887651450251, "loss": 0.9909, "step": 3950 }, { "epoch": 1.4335600907029478, "grad_norm": 0.2109375, "learning_rate": 0.00010573981152857668, "loss": 0.9774, "step": 3952 }, { "epoch": 1.4342857142857142, "grad_norm": 0.1787109375, "learning_rate": 0.00010569085791212827, "loss": 0.9736, "step": 3954 }, { "epoch": 1.4350113378684808, "grad_norm": 0.1689453125, "learning_rate": 0.00010564190429567984, "loss": 1.0008, "step": 3956 }, { "epoch": 1.4357369614512472, "grad_norm": 0.1669921875, "learning_rate": 0.00010559295067923142, "loss": 0.9877, "step": 3958 }, { "epoch": 1.4364625850340136, "grad_norm": 0.171875, "learning_rate": 0.00010554399706278302, "loss": 0.9768, "step": 3960 }, { "epoch": 1.43718820861678, "grad_norm": 0.1572265625, "learning_rate": 0.0001054950434463346, "loss": 0.9943, "step": 3962 }, { "epoch": 1.4379138321995466, "grad_norm": 0.1865234375, "learning_rate": 0.0001054460898298862, "loss": 1.0235, "step": 3964 }, { "epoch": 1.438639455782313, "grad_norm": 0.1767578125, "learning_rate": 0.00010539713621343777, "loss": 1.0468, "step": 3966 }, { "epoch": 1.4393650793650794, "grad_norm": 0.16015625, "learning_rate": 0.00010534818259698935, "loss": 0.9501, "step": 3968 }, { "epoch": 1.4400907029478458, "grad_norm": 0.1572265625, "learning_rate": 0.00010529922898054095, "loss": 1.0001, "step": 3970 }, { "epoch": 1.4408163265306122, "grad_norm": 0.1865234375, "learning_rate": 0.00010525027536409253, "loss": 0.9847, "step": 3972 }, { "epoch": 1.4415419501133786, "grad_norm": 0.1552734375, "learning_rate": 0.00010520132174764412, "loss": 0.9209, "step": 3974 }, { "epoch": 1.442267573696145, "grad_norm": 0.1748046875, "learning_rate": 0.00010515236813119569, "loss": 0.969, "step": 3976 }, { "epoch": 1.4429931972789116, "grad_norm": 0.140625, "learning_rate": 0.00010510341451474727, "loss": 0.8542, "step": 3978 }, { "epoch": 1.443718820861678, "grad_norm": 0.1650390625, "learning_rate": 0.00010505446089829887, "loss": 0.9088, "step": 3980 }, { "epoch": 1.4444444444444444, "grad_norm": 0.1875, "learning_rate": 0.00010500550728185045, "loss": 0.9864, "step": 3982 }, { "epoch": 1.4451700680272108, "grad_norm": 0.1806640625, "learning_rate": 0.00010495655366540205, "loss": 1.0233, "step": 3984 }, { "epoch": 1.4458956916099774, "grad_norm": 0.15234375, "learning_rate": 0.00010490760004895362, "loss": 0.9392, "step": 3986 }, { "epoch": 1.4466213151927438, "grad_norm": 0.1748046875, "learning_rate": 0.0001048586464325052, "loss": 0.9399, "step": 3988 }, { "epoch": 1.4473469387755102, "grad_norm": 0.171875, "learning_rate": 0.0001048096928160568, "loss": 1.0332, "step": 3990 }, { "epoch": 1.4480725623582766, "grad_norm": 0.2060546875, "learning_rate": 0.00010476073919960838, "loss": 1.0093, "step": 3992 }, { "epoch": 1.448798185941043, "grad_norm": 0.1806640625, "learning_rate": 0.00010471178558315997, "loss": 0.9698, "step": 3994 }, { "epoch": 1.4495238095238094, "grad_norm": 0.162109375, "learning_rate": 0.00010466283196671154, "loss": 0.9719, "step": 3996 }, { "epoch": 1.4502494331065758, "grad_norm": 0.1787109375, "learning_rate": 0.00010461387835026312, "loss": 1.0843, "step": 3998 }, { "epoch": 1.4509750566893425, "grad_norm": 0.1748046875, "learning_rate": 0.00010456492473381472, "loss": 0.9405, "step": 4000 }, { "epoch": 1.4509750566893425, "eval_loss": 1.0138306617736816, "eval_runtime": 90.2991, "eval_samples_per_second": 19.934, "eval_steps_per_second": 0.631, "step": 4000 }, { "epoch": 1.4517006802721089, "grad_norm": 0.16015625, "learning_rate": 0.0001045159711173663, "loss": 0.9971, "step": 4002 }, { "epoch": 1.4524263038548753, "grad_norm": 0.1552734375, "learning_rate": 0.0001044670175009179, "loss": 0.9311, "step": 4004 }, { "epoch": 1.4531519274376417, "grad_norm": 0.1513671875, "learning_rate": 0.00010441806388446947, "loss": 0.9054, "step": 4006 }, { "epoch": 1.4538775510204083, "grad_norm": 0.16015625, "learning_rate": 0.00010436911026802105, "loss": 1.0426, "step": 4008 }, { "epoch": 1.4546031746031747, "grad_norm": 0.1728515625, "learning_rate": 0.00010432015665157264, "loss": 0.9125, "step": 4010 }, { "epoch": 1.455328798185941, "grad_norm": 0.16796875, "learning_rate": 0.00010427120303512423, "loss": 0.9421, "step": 4012 }, { "epoch": 1.4560544217687075, "grad_norm": 0.1591796875, "learning_rate": 0.0001042222494186758, "loss": 0.9698, "step": 4014 }, { "epoch": 1.4567800453514739, "grad_norm": 0.1689453125, "learning_rate": 0.00010417329580222739, "loss": 1.0262, "step": 4016 }, { "epoch": 1.4575056689342403, "grad_norm": 0.166015625, "learning_rate": 0.00010412434218577897, "loss": 0.9472, "step": 4018 }, { "epoch": 1.4582312925170067, "grad_norm": 0.142578125, "learning_rate": 0.00010407538856933057, "loss": 0.981, "step": 4020 }, { "epoch": 1.4589569160997733, "grad_norm": 0.1591796875, "learning_rate": 0.00010402643495288215, "loss": 0.9012, "step": 4022 }, { "epoch": 1.4596825396825397, "grad_norm": 0.158203125, "learning_rate": 0.00010397748133643372, "loss": 0.9633, "step": 4024 }, { "epoch": 1.460408163265306, "grad_norm": 0.1669921875, "learning_rate": 0.00010392852771998532, "loss": 0.9586, "step": 4026 }, { "epoch": 1.4611337868480725, "grad_norm": 0.193359375, "learning_rate": 0.0001038795741035369, "loss": 1.0336, "step": 4028 }, { "epoch": 1.4618594104308391, "grad_norm": 0.1669921875, "learning_rate": 0.0001038306204870885, "loss": 0.9896, "step": 4030 }, { "epoch": 1.4625850340136055, "grad_norm": 0.185546875, "learning_rate": 0.00010378166687064008, "loss": 1.0079, "step": 4032 }, { "epoch": 1.463310657596372, "grad_norm": 0.166015625, "learning_rate": 0.00010373271325419165, "loss": 0.9626, "step": 4034 }, { "epoch": 1.4640362811791383, "grad_norm": 0.1689453125, "learning_rate": 0.00010368375963774324, "loss": 0.9859, "step": 4036 }, { "epoch": 1.4647619047619047, "grad_norm": 0.1484375, "learning_rate": 0.00010363480602129482, "loss": 0.9595, "step": 4038 }, { "epoch": 1.465487528344671, "grad_norm": 0.1650390625, "learning_rate": 0.00010358585240484642, "loss": 0.9691, "step": 4040 }, { "epoch": 1.4662131519274375, "grad_norm": 0.15234375, "learning_rate": 0.000103536898788398, "loss": 0.983, "step": 4042 }, { "epoch": 1.4669387755102041, "grad_norm": 0.171875, "learning_rate": 0.00010348794517194957, "loss": 1.0172, "step": 4044 }, { "epoch": 1.4676643990929705, "grad_norm": 0.1728515625, "learning_rate": 0.00010343899155550117, "loss": 0.9945, "step": 4046 }, { "epoch": 1.468390022675737, "grad_norm": 0.1708984375, "learning_rate": 0.00010339003793905275, "loss": 0.9569, "step": 4048 }, { "epoch": 1.4691156462585033, "grad_norm": 0.1591796875, "learning_rate": 0.00010334108432260434, "loss": 0.9445, "step": 4050 }, { "epoch": 1.46984126984127, "grad_norm": 0.18359375, "learning_rate": 0.00010329213070615593, "loss": 1.0193, "step": 4052 }, { "epoch": 1.4705668934240363, "grad_norm": 0.1806640625, "learning_rate": 0.0001032431770897075, "loss": 1.0227, "step": 4054 }, { "epoch": 1.4712925170068027, "grad_norm": 0.171875, "learning_rate": 0.00010319422347325909, "loss": 0.9067, "step": 4056 }, { "epoch": 1.4720181405895691, "grad_norm": 0.1630859375, "learning_rate": 0.00010314526985681067, "loss": 0.9907, "step": 4058 }, { "epoch": 1.4727437641723355, "grad_norm": 0.1689453125, "learning_rate": 0.00010309631624036227, "loss": 0.963, "step": 4060 }, { "epoch": 1.473469387755102, "grad_norm": 0.15234375, "learning_rate": 0.00010304736262391385, "loss": 0.9769, "step": 4062 }, { "epoch": 1.4741950113378683, "grad_norm": 0.1689453125, "learning_rate": 0.00010299840900746542, "loss": 0.9689, "step": 4064 }, { "epoch": 1.474920634920635, "grad_norm": 0.1796875, "learning_rate": 0.00010294945539101702, "loss": 0.9922, "step": 4066 }, { "epoch": 1.4756462585034014, "grad_norm": 0.1572265625, "learning_rate": 0.0001029005017745686, "loss": 0.9429, "step": 4068 }, { "epoch": 1.4763718820861678, "grad_norm": 0.1884765625, "learning_rate": 0.0001028515481581202, "loss": 1.0386, "step": 4070 }, { "epoch": 1.4770975056689342, "grad_norm": 0.1630859375, "learning_rate": 0.00010280259454167178, "loss": 0.9666, "step": 4072 }, { "epoch": 1.4778231292517008, "grad_norm": 0.146484375, "learning_rate": 0.00010275364092522335, "loss": 1.0141, "step": 4074 }, { "epoch": 1.4785487528344672, "grad_norm": 0.1591796875, "learning_rate": 0.00010270468730877494, "loss": 0.9448, "step": 4076 }, { "epoch": 1.4792743764172336, "grad_norm": 0.1484375, "learning_rate": 0.00010265573369232652, "loss": 1.0185, "step": 4078 }, { "epoch": 1.48, "grad_norm": 0.1591796875, "learning_rate": 0.00010260678007587812, "loss": 0.9499, "step": 4080 }, { "epoch": 1.4807256235827664, "grad_norm": 0.1708984375, "learning_rate": 0.00010255782645942969, "loss": 0.9646, "step": 4082 }, { "epoch": 1.4814512471655328, "grad_norm": 0.1708984375, "learning_rate": 0.00010250887284298127, "loss": 1.0299, "step": 4084 }, { "epoch": 1.4821768707482994, "grad_norm": 0.1611328125, "learning_rate": 0.00010245991922653287, "loss": 0.9877, "step": 4086 }, { "epoch": 1.4829024943310658, "grad_norm": 0.173828125, "learning_rate": 0.00010241096561008445, "loss": 0.9352, "step": 4088 }, { "epoch": 1.4836281179138322, "grad_norm": 0.158203125, "learning_rate": 0.00010236201199363604, "loss": 0.9672, "step": 4090 }, { "epoch": 1.4843537414965986, "grad_norm": 0.1767578125, "learning_rate": 0.00010231305837718761, "loss": 0.9205, "step": 4092 }, { "epoch": 1.485079365079365, "grad_norm": 0.1640625, "learning_rate": 0.0001022641047607392, "loss": 1.0447, "step": 4094 }, { "epoch": 1.4858049886621316, "grad_norm": 0.1611328125, "learning_rate": 0.00010221515114429079, "loss": 0.9848, "step": 4096 }, { "epoch": 1.486530612244898, "grad_norm": 0.1875, "learning_rate": 0.00010216619752784237, "loss": 0.9774, "step": 4098 }, { "epoch": 1.4872562358276644, "grad_norm": 0.1708984375, "learning_rate": 0.00010211724391139397, "loss": 0.9378, "step": 4100 }, { "epoch": 1.4872562358276644, "eval_loss": 1.0134267807006836, "eval_runtime": 90.0252, "eval_samples_per_second": 19.994, "eval_steps_per_second": 0.633, "step": 4100 }, { "epoch": 1.4879818594104308, "grad_norm": 0.16796875, "learning_rate": 0.00010206829029494554, "loss": 0.9782, "step": 4102 }, { "epoch": 1.4887074829931972, "grad_norm": 0.1708984375, "learning_rate": 0.00010201933667849712, "loss": 1.0532, "step": 4104 }, { "epoch": 1.4894331065759636, "grad_norm": 0.1513671875, "learning_rate": 0.00010197038306204872, "loss": 0.9129, "step": 4106 }, { "epoch": 1.4901587301587302, "grad_norm": 0.1494140625, "learning_rate": 0.0001019214294456003, "loss": 0.9277, "step": 4108 }, { "epoch": 1.4908843537414966, "grad_norm": 0.17578125, "learning_rate": 0.0001018724758291519, "loss": 0.9731, "step": 4110 }, { "epoch": 1.491609977324263, "grad_norm": 0.158203125, "learning_rate": 0.00010182352221270346, "loss": 0.9863, "step": 4112 }, { "epoch": 1.4923356009070294, "grad_norm": 0.15625, "learning_rate": 0.00010177456859625505, "loss": 1.1322, "step": 4114 }, { "epoch": 1.4930612244897958, "grad_norm": 0.181640625, "learning_rate": 0.00010172561497980664, "loss": 0.9763, "step": 4116 }, { "epoch": 1.4937868480725625, "grad_norm": 0.173828125, "learning_rate": 0.00010167666136335822, "loss": 0.978, "step": 4118 }, { "epoch": 1.4945124716553289, "grad_norm": 0.1748046875, "learning_rate": 0.00010162770774690982, "loss": 1.0198, "step": 4120 }, { "epoch": 1.4952380952380953, "grad_norm": 0.162109375, "learning_rate": 0.00010157875413046139, "loss": 0.8959, "step": 4122 }, { "epoch": 1.4959637188208617, "grad_norm": 0.173828125, "learning_rate": 0.00010152980051401297, "loss": 0.9886, "step": 4124 }, { "epoch": 1.496689342403628, "grad_norm": 0.1513671875, "learning_rate": 0.00010148084689756457, "loss": 0.9965, "step": 4126 }, { "epoch": 1.4974149659863945, "grad_norm": 0.1689453125, "learning_rate": 0.00010143189328111615, "loss": 0.961, "step": 4128 }, { "epoch": 1.498140589569161, "grad_norm": 0.1455078125, "learning_rate": 0.00010138293966466774, "loss": 0.9631, "step": 4130 }, { "epoch": 1.4988662131519275, "grad_norm": 0.208984375, "learning_rate": 0.00010133398604821931, "loss": 1.0086, "step": 4132 }, { "epoch": 1.4995918367346939, "grad_norm": 0.154296875, "learning_rate": 0.0001012850324317709, "loss": 0.9727, "step": 4134 }, { "epoch": 1.5003174603174603, "grad_norm": 0.150390625, "learning_rate": 0.00010123607881532249, "loss": 0.9614, "step": 4136 }, { "epoch": 1.501043083900227, "grad_norm": 0.1708984375, "learning_rate": 0.00010118712519887407, "loss": 0.9524, "step": 4138 }, { "epoch": 1.5017687074829933, "grad_norm": 0.171875, "learning_rate": 0.00010113817158242567, "loss": 0.9623, "step": 4140 }, { "epoch": 1.5024943310657597, "grad_norm": 0.1728515625, "learning_rate": 0.00010108921796597724, "loss": 0.9599, "step": 4142 }, { "epoch": 1.503219954648526, "grad_norm": 0.162109375, "learning_rate": 0.00010104026434952882, "loss": 0.9384, "step": 4144 }, { "epoch": 1.5039455782312925, "grad_norm": 0.1591796875, "learning_rate": 0.00010099131073308042, "loss": 0.9673, "step": 4146 }, { "epoch": 1.504671201814059, "grad_norm": 0.16015625, "learning_rate": 0.000100942357116632, "loss": 0.9375, "step": 4148 }, { "epoch": 1.5053968253968253, "grad_norm": 0.181640625, "learning_rate": 0.0001008934035001836, "loss": 0.884, "step": 4150 }, { "epoch": 1.5061224489795917, "grad_norm": 0.1513671875, "learning_rate": 0.00010084444988373516, "loss": 0.9254, "step": 4152 }, { "epoch": 1.5068480725623583, "grad_norm": 0.203125, "learning_rate": 0.00010079549626728675, "loss": 1.0317, "step": 4154 }, { "epoch": 1.5075736961451247, "grad_norm": 0.17578125, "learning_rate": 0.00010074654265083834, "loss": 0.9853, "step": 4156 }, { "epoch": 1.508299319727891, "grad_norm": 0.166015625, "learning_rate": 0.00010069758903438992, "loss": 0.9551, "step": 4158 }, { "epoch": 1.5090249433106577, "grad_norm": 0.1640625, "learning_rate": 0.00010064863541794149, "loss": 1.0439, "step": 4160 }, { "epoch": 1.5097505668934241, "grad_norm": 0.1875, "learning_rate": 0.00010059968180149309, "loss": 0.9087, "step": 4162 }, { "epoch": 1.5104761904761905, "grad_norm": 0.15234375, "learning_rate": 0.00010055072818504467, "loss": 0.9663, "step": 4164 }, { "epoch": 1.511201814058957, "grad_norm": 0.162109375, "learning_rate": 0.00010050177456859627, "loss": 0.9691, "step": 4166 }, { "epoch": 1.5119274376417233, "grad_norm": 0.171875, "learning_rate": 0.00010045282095214785, "loss": 1.0104, "step": 4168 }, { "epoch": 1.5126530612244897, "grad_norm": 0.169921875, "learning_rate": 0.00010040386733569942, "loss": 1.022, "step": 4170 }, { "epoch": 1.5133786848072561, "grad_norm": 0.1396484375, "learning_rate": 0.00010035491371925101, "loss": 1.0086, "step": 4172 }, { "epoch": 1.5141043083900225, "grad_norm": 0.166015625, "learning_rate": 0.0001003059601028026, "loss": 0.9908, "step": 4174 }, { "epoch": 1.5148299319727891, "grad_norm": 0.22265625, "learning_rate": 0.00010025700648635419, "loss": 0.9026, "step": 4176 }, { "epoch": 1.5155555555555555, "grad_norm": 0.1630859375, "learning_rate": 0.00010020805286990577, "loss": 1.0208, "step": 4178 }, { "epoch": 1.516281179138322, "grad_norm": 0.1591796875, "learning_rate": 0.00010015909925345734, "loss": 0.9172, "step": 4180 }, { "epoch": 1.5170068027210886, "grad_norm": 0.1708984375, "learning_rate": 0.00010011014563700894, "loss": 0.9348, "step": 4182 }, { "epoch": 1.517732426303855, "grad_norm": 0.1591796875, "learning_rate": 0.00010006119202056052, "loss": 0.9899, "step": 4184 }, { "epoch": 1.5184580498866214, "grad_norm": 0.1953125, "learning_rate": 0.00010001223840411212, "loss": 1.0367, "step": 4186 }, { "epoch": 1.5191836734693878, "grad_norm": 0.1669921875, "learning_rate": 9.996328478766368e-05, "loss": 0.9138, "step": 4188 }, { "epoch": 1.5199092970521542, "grad_norm": 0.1689453125, "learning_rate": 9.991433117121528e-05, "loss": 1.0288, "step": 4190 }, { "epoch": 1.5206349206349206, "grad_norm": 0.158203125, "learning_rate": 9.986537755476686e-05, "loss": 0.9677, "step": 4192 }, { "epoch": 1.521360544217687, "grad_norm": 0.1962890625, "learning_rate": 9.981642393831845e-05, "loss": 1.0642, "step": 4194 }, { "epoch": 1.5220861678004534, "grad_norm": 0.16796875, "learning_rate": 9.976747032187004e-05, "loss": 1.0481, "step": 4196 }, { "epoch": 1.52281179138322, "grad_norm": 0.1669921875, "learning_rate": 9.971851670542161e-05, "loss": 1.0118, "step": 4198 }, { "epoch": 1.5235374149659864, "grad_norm": 0.1650390625, "learning_rate": 9.96695630889732e-05, "loss": 0.9331, "step": 4200 }, { "epoch": 1.5235374149659864, "eval_loss": 1.013152003288269, "eval_runtime": 89.7624, "eval_samples_per_second": 20.053, "eval_steps_per_second": 0.635, "step": 4200 }, { "epoch": 1.5242630385487528, "grad_norm": 0.16796875, "learning_rate": 9.962060947252479e-05, "loss": 1.0175, "step": 4202 }, { "epoch": 1.5249886621315194, "grad_norm": 0.16796875, "learning_rate": 9.957165585607637e-05, "loss": 1.0108, "step": 4204 }, { "epoch": 1.5257142857142858, "grad_norm": 0.1572265625, "learning_rate": 9.952270223962795e-05, "loss": 1.0153, "step": 4206 }, { "epoch": 1.5264399092970522, "grad_norm": 0.1767578125, "learning_rate": 9.947374862317953e-05, "loss": 1.0924, "step": 4208 }, { "epoch": 1.5271655328798186, "grad_norm": 0.162109375, "learning_rate": 9.942479500673113e-05, "loss": 0.9743, "step": 4210 }, { "epoch": 1.527891156462585, "grad_norm": 0.177734375, "learning_rate": 9.937584139028271e-05, "loss": 0.9311, "step": 4212 }, { "epoch": 1.5286167800453514, "grad_norm": 0.158203125, "learning_rate": 9.93268877738343e-05, "loss": 0.9674, "step": 4214 }, { "epoch": 1.5293424036281178, "grad_norm": 0.1611328125, "learning_rate": 9.927793415738588e-05, "loss": 0.9447, "step": 4216 }, { "epoch": 1.5300680272108842, "grad_norm": 0.1533203125, "learning_rate": 9.922898054093746e-05, "loss": 1.0299, "step": 4218 }, { "epoch": 1.5307936507936508, "grad_norm": 0.166015625, "learning_rate": 9.918002692448906e-05, "loss": 0.954, "step": 4220 }, { "epoch": 1.5315192743764172, "grad_norm": 0.171875, "learning_rate": 9.913107330804064e-05, "loss": 1.012, "step": 4222 }, { "epoch": 1.5322448979591838, "grad_norm": 0.1748046875, "learning_rate": 9.908211969159222e-05, "loss": 1.0334, "step": 4224 }, { "epoch": 1.5329705215419502, "grad_norm": 0.1474609375, "learning_rate": 9.90331660751438e-05, "loss": 0.9732, "step": 4226 }, { "epoch": 1.5336961451247166, "grad_norm": 0.15234375, "learning_rate": 9.898421245869538e-05, "loss": 0.9949, "step": 4228 }, { "epoch": 1.534421768707483, "grad_norm": 0.1806640625, "learning_rate": 9.893525884224698e-05, "loss": 1.0049, "step": 4230 }, { "epoch": 1.5351473922902494, "grad_norm": 0.158203125, "learning_rate": 9.888630522579856e-05, "loss": 0.9727, "step": 4232 }, { "epoch": 1.5358730158730158, "grad_norm": 0.1630859375, "learning_rate": 9.883735160935014e-05, "loss": 0.9889, "step": 4234 }, { "epoch": 1.5365986394557822, "grad_norm": 0.1669921875, "learning_rate": 9.878839799290173e-05, "loss": 0.9699, "step": 4236 }, { "epoch": 1.5373242630385486, "grad_norm": 0.1650390625, "learning_rate": 9.873944437645331e-05, "loss": 0.9659, "step": 4238 }, { "epoch": 1.538049886621315, "grad_norm": 0.1533203125, "learning_rate": 9.86904907600049e-05, "loss": 0.9404, "step": 4240 }, { "epoch": 1.5387755102040817, "grad_norm": 0.17578125, "learning_rate": 9.864153714355649e-05, "loss": 0.9822, "step": 4242 }, { "epoch": 1.539501133786848, "grad_norm": 0.154296875, "learning_rate": 9.859258352710807e-05, "loss": 0.9127, "step": 4244 }, { "epoch": 1.5402267573696147, "grad_norm": 0.1630859375, "learning_rate": 9.854362991065965e-05, "loss": 1.0608, "step": 4246 }, { "epoch": 1.540952380952381, "grad_norm": 0.1689453125, "learning_rate": 9.849467629421123e-05, "loss": 1.0184, "step": 4248 }, { "epoch": 1.5416780045351475, "grad_norm": 0.1767578125, "learning_rate": 9.844572267776282e-05, "loss": 1.0275, "step": 4250 }, { "epoch": 1.5424036281179139, "grad_norm": 0.1630859375, "learning_rate": 9.839676906131441e-05, "loss": 0.9631, "step": 4252 }, { "epoch": 1.5431292517006803, "grad_norm": 0.1611328125, "learning_rate": 9.8347815444866e-05, "loss": 0.9986, "step": 4254 }, { "epoch": 1.5438548752834467, "grad_norm": 0.16015625, "learning_rate": 9.829886182841758e-05, "loss": 0.9216, "step": 4256 }, { "epoch": 1.544580498866213, "grad_norm": 0.1669921875, "learning_rate": 9.824990821196916e-05, "loss": 0.939, "step": 4258 }, { "epoch": 1.5453061224489795, "grad_norm": 0.1630859375, "learning_rate": 9.820095459552074e-05, "loss": 1.049, "step": 4260 }, { "epoch": 1.5460317460317459, "grad_norm": 0.1435546875, "learning_rate": 9.815200097907234e-05, "loss": 0.9569, "step": 4262 }, { "epoch": 1.5467573696145125, "grad_norm": 0.17578125, "learning_rate": 9.810304736262392e-05, "loss": 1.0119, "step": 4264 }, { "epoch": 1.5474829931972789, "grad_norm": 0.1875, "learning_rate": 9.80540937461755e-05, "loss": 1.1073, "step": 4266 }, { "epoch": 1.5482086167800455, "grad_norm": 0.1494140625, "learning_rate": 9.800514012972708e-05, "loss": 0.9696, "step": 4268 }, { "epoch": 1.548934240362812, "grad_norm": 0.154296875, "learning_rate": 9.795618651327867e-05, "loss": 0.8951, "step": 4270 }, { "epoch": 1.5496598639455783, "grad_norm": 0.193359375, "learning_rate": 9.790723289683026e-05, "loss": 0.94, "step": 4272 }, { "epoch": 1.5503854875283447, "grad_norm": 0.1708984375, "learning_rate": 9.785827928038184e-05, "loss": 1.0004, "step": 4274 }, { "epoch": 1.551111111111111, "grad_norm": 0.1591796875, "learning_rate": 9.780932566393343e-05, "loss": 0.9606, "step": 4276 }, { "epoch": 1.5518367346938775, "grad_norm": 0.162109375, "learning_rate": 9.776037204748501e-05, "loss": 1.0272, "step": 4278 }, { "epoch": 1.552562358276644, "grad_norm": 0.1767578125, "learning_rate": 9.771141843103659e-05, "loss": 0.9806, "step": 4280 }, { "epoch": 1.5532879818594103, "grad_norm": 0.1484375, "learning_rate": 9.766246481458819e-05, "loss": 0.8835, "step": 4282 }, { "epoch": 1.5540136054421767, "grad_norm": 0.16015625, "learning_rate": 9.761351119813976e-05, "loss": 0.9617, "step": 4284 }, { "epoch": 1.5547392290249433, "grad_norm": 0.154296875, "learning_rate": 9.756455758169135e-05, "loss": 1.047, "step": 4286 }, { "epoch": 1.5554648526077097, "grad_norm": 0.1591796875, "learning_rate": 9.751560396524293e-05, "loss": 0.9978, "step": 4288 }, { "epoch": 1.5561904761904763, "grad_norm": 0.1689453125, "learning_rate": 9.746665034879452e-05, "loss": 0.9766, "step": 4290 }, { "epoch": 1.5569160997732427, "grad_norm": 0.173828125, "learning_rate": 9.741769673234611e-05, "loss": 1.0446, "step": 4292 }, { "epoch": 1.5576417233560091, "grad_norm": 0.177734375, "learning_rate": 9.736874311589768e-05, "loss": 1.0301, "step": 4294 }, { "epoch": 1.5583673469387755, "grad_norm": 0.1767578125, "learning_rate": 9.731978949944928e-05, "loss": 1.0035, "step": 4296 }, { "epoch": 1.559092970521542, "grad_norm": 0.1962890625, "learning_rate": 9.727083588300086e-05, "loss": 0.9809, "step": 4298 }, { "epoch": 1.5598185941043083, "grad_norm": 0.1728515625, "learning_rate": 9.722188226655244e-05, "loss": 0.9959, "step": 4300 }, { "epoch": 1.5598185941043083, "eval_loss": 1.0130585432052612, "eval_runtime": 87.5971, "eval_samples_per_second": 20.549, "eval_steps_per_second": 0.651, "step": 4300 }, { "epoch": 1.5605442176870747, "grad_norm": 0.15234375, "learning_rate": 9.717292865010404e-05, "loss": 1.0011, "step": 4302 }, { "epoch": 1.5612698412698411, "grad_norm": 0.1767578125, "learning_rate": 9.71239750336556e-05, "loss": 0.934, "step": 4304 }, { "epoch": 1.5619954648526078, "grad_norm": 0.14453125, "learning_rate": 9.70750214172072e-05, "loss": 0.9783, "step": 4306 }, { "epoch": 1.5627210884353742, "grad_norm": 0.1796875, "learning_rate": 9.702606780075878e-05, "loss": 0.995, "step": 4308 }, { "epoch": 1.5634467120181406, "grad_norm": 0.162109375, "learning_rate": 9.697711418431037e-05, "loss": 0.932, "step": 4310 }, { "epoch": 1.5641723356009072, "grad_norm": 0.154296875, "learning_rate": 9.692816056786196e-05, "loss": 1.0029, "step": 4312 }, { "epoch": 1.5648979591836736, "grad_norm": 0.1708984375, "learning_rate": 9.687920695141353e-05, "loss": 0.9971, "step": 4314 }, { "epoch": 1.56562358276644, "grad_norm": 0.1845703125, "learning_rate": 9.683025333496513e-05, "loss": 0.9928, "step": 4316 }, { "epoch": 1.5663492063492064, "grad_norm": 0.16015625, "learning_rate": 9.678129971851671e-05, "loss": 0.9279, "step": 4318 }, { "epoch": 1.5670748299319728, "grad_norm": 0.17578125, "learning_rate": 9.673234610206829e-05, "loss": 0.937, "step": 4320 }, { "epoch": 1.5678004535147392, "grad_norm": 0.1611328125, "learning_rate": 9.668339248561989e-05, "loss": 1.1021, "step": 4322 }, { "epoch": 1.5685260770975056, "grad_norm": 0.1513671875, "learning_rate": 9.663443886917146e-05, "loss": 0.938, "step": 4324 }, { "epoch": 1.569251700680272, "grad_norm": 0.177734375, "learning_rate": 9.658548525272305e-05, "loss": 0.992, "step": 4326 }, { "epoch": 1.5699773242630386, "grad_norm": 0.1787109375, "learning_rate": 9.653653163627463e-05, "loss": 1.0522, "step": 4328 }, { "epoch": 1.570702947845805, "grad_norm": 0.16015625, "learning_rate": 9.648757801982622e-05, "loss": 1.0242, "step": 4330 }, { "epoch": 1.5714285714285714, "grad_norm": 0.154296875, "learning_rate": 9.643862440337781e-05, "loss": 0.9797, "step": 4332 }, { "epoch": 1.572154195011338, "grad_norm": 0.185546875, "learning_rate": 9.638967078692938e-05, "loss": 1.0288, "step": 4334 }, { "epoch": 1.5728798185941044, "grad_norm": 0.162109375, "learning_rate": 9.634071717048098e-05, "loss": 0.9567, "step": 4336 }, { "epoch": 1.5736054421768708, "grad_norm": 0.1591796875, "learning_rate": 9.629176355403256e-05, "loss": 1.0486, "step": 4338 }, { "epoch": 1.5743310657596372, "grad_norm": 0.1611328125, "learning_rate": 9.624280993758414e-05, "loss": 0.9999, "step": 4340 }, { "epoch": 1.5750566893424036, "grad_norm": 0.1904296875, "learning_rate": 9.619385632113574e-05, "loss": 0.9742, "step": 4342 }, { "epoch": 1.57578231292517, "grad_norm": 0.154296875, "learning_rate": 9.61449027046873e-05, "loss": 0.9984, "step": 4344 }, { "epoch": 1.5765079365079364, "grad_norm": 0.15625, "learning_rate": 9.60959490882389e-05, "loss": 0.987, "step": 4346 }, { "epoch": 1.5772335600907028, "grad_norm": 0.158203125, "learning_rate": 9.604699547179048e-05, "loss": 1.0147, "step": 4348 }, { "epoch": 1.5779591836734694, "grad_norm": 0.169921875, "learning_rate": 9.599804185534207e-05, "loss": 0.9514, "step": 4350 }, { "epoch": 1.5786848072562358, "grad_norm": 0.150390625, "learning_rate": 9.594908823889365e-05, "loss": 0.8686, "step": 4352 }, { "epoch": 1.5794104308390022, "grad_norm": 0.173828125, "learning_rate": 9.590013462244523e-05, "loss": 0.9818, "step": 4354 }, { "epoch": 1.5801360544217689, "grad_norm": 0.1572265625, "learning_rate": 9.585118100599683e-05, "loss": 0.9206, "step": 4356 }, { "epoch": 1.5808616780045353, "grad_norm": 0.1630859375, "learning_rate": 9.580222738954841e-05, "loss": 1.0003, "step": 4358 }, { "epoch": 1.5815873015873017, "grad_norm": 0.1533203125, "learning_rate": 9.575327377309999e-05, "loss": 0.9899, "step": 4360 }, { "epoch": 1.582312925170068, "grad_norm": 0.1650390625, "learning_rate": 9.570432015665157e-05, "loss": 0.9102, "step": 4362 }, { "epoch": 1.5830385487528345, "grad_norm": 0.1484375, "learning_rate": 9.565536654020316e-05, "loss": 0.9594, "step": 4364 }, { "epoch": 1.5837641723356009, "grad_norm": 0.21875, "learning_rate": 9.560641292375475e-05, "loss": 0.9382, "step": 4366 }, { "epoch": 1.5844897959183672, "grad_norm": 0.158203125, "learning_rate": 9.555745930730633e-05, "loss": 1.0231, "step": 4368 }, { "epoch": 1.5852154195011336, "grad_norm": 0.1689453125, "learning_rate": 9.550850569085792e-05, "loss": 0.9838, "step": 4370 }, { "epoch": 1.5859410430839003, "grad_norm": 0.17578125, "learning_rate": 9.54595520744095e-05, "loss": 1.0085, "step": 4372 }, { "epoch": 1.5866666666666667, "grad_norm": 0.1640625, "learning_rate": 9.541059845796108e-05, "loss": 0.9167, "step": 4374 }, { "epoch": 1.587392290249433, "grad_norm": 0.181640625, "learning_rate": 9.536164484151268e-05, "loss": 0.9559, "step": 4376 }, { "epoch": 1.5881179138321997, "grad_norm": 0.1689453125, "learning_rate": 9.531269122506426e-05, "loss": 0.9368, "step": 4378 }, { "epoch": 1.588843537414966, "grad_norm": 0.1767578125, "learning_rate": 9.526373760861584e-05, "loss": 0.9922, "step": 4380 }, { "epoch": 1.5895691609977325, "grad_norm": 0.1640625, "learning_rate": 9.521478399216742e-05, "loss": 0.9523, "step": 4382 }, { "epoch": 1.5902947845804989, "grad_norm": 0.17578125, "learning_rate": 9.5165830375719e-05, "loss": 0.9917, "step": 4384 }, { "epoch": 1.5910204081632653, "grad_norm": 0.1787109375, "learning_rate": 9.51168767592706e-05, "loss": 0.9945, "step": 4386 }, { "epoch": 1.5917460317460317, "grad_norm": 0.205078125, "learning_rate": 9.506792314282218e-05, "loss": 1.0831, "step": 4388 }, { "epoch": 1.592471655328798, "grad_norm": 0.1982421875, "learning_rate": 9.501896952637377e-05, "loss": 1.0329, "step": 4390 }, { "epoch": 1.5931972789115645, "grad_norm": 0.171875, "learning_rate": 9.497001590992535e-05, "loss": 0.8841, "step": 4392 }, { "epoch": 1.593922902494331, "grad_norm": 0.1650390625, "learning_rate": 9.492106229347693e-05, "loss": 0.8609, "step": 4394 }, { "epoch": 1.5946485260770975, "grad_norm": 0.16796875, "learning_rate": 9.487210867702851e-05, "loss": 0.9959, "step": 4396 }, { "epoch": 1.595374149659864, "grad_norm": 0.1669921875, "learning_rate": 9.482315506058011e-05, "loss": 0.9974, "step": 4398 }, { "epoch": 1.5960997732426305, "grad_norm": 0.1728515625, "learning_rate": 9.477420144413169e-05, "loss": 0.9548, "step": 4400 }, { "epoch": 1.5960997732426305, "eval_loss": 1.0129448175430298, "eval_runtime": 87.6312, "eval_samples_per_second": 20.541, "eval_steps_per_second": 0.65, "step": 4400 }, { "epoch": 1.596825396825397, "grad_norm": 0.166015625, "learning_rate": 9.472524782768327e-05, "loss": 1.0032, "step": 4402 }, { "epoch": 1.5975510204081633, "grad_norm": 0.1787109375, "learning_rate": 9.467629421123486e-05, "loss": 0.9867, "step": 4404 }, { "epoch": 1.5982766439909297, "grad_norm": 0.162109375, "learning_rate": 9.462734059478644e-05, "loss": 1.0104, "step": 4406 }, { "epoch": 1.5990022675736961, "grad_norm": 0.177734375, "learning_rate": 9.457838697833803e-05, "loss": 0.881, "step": 4408 }, { "epoch": 1.5997278911564625, "grad_norm": 0.1650390625, "learning_rate": 9.452943336188962e-05, "loss": 0.9992, "step": 4410 }, { "epoch": 1.600453514739229, "grad_norm": 0.2236328125, "learning_rate": 9.44804797454412e-05, "loss": 0.9666, "step": 4412 }, { "epoch": 1.6011791383219953, "grad_norm": 0.1748046875, "learning_rate": 9.443152612899278e-05, "loss": 1.0162, "step": 4414 }, { "epoch": 1.601904761904762, "grad_norm": 0.1494140625, "learning_rate": 9.438257251254436e-05, "loss": 0.8902, "step": 4416 }, { "epoch": 1.6026303854875283, "grad_norm": 0.1650390625, "learning_rate": 9.433361889609596e-05, "loss": 0.9424, "step": 4418 }, { "epoch": 1.6033560090702947, "grad_norm": 0.1767578125, "learning_rate": 9.428466527964754e-05, "loss": 0.9927, "step": 4420 }, { "epoch": 1.6040816326530614, "grad_norm": 0.177734375, "learning_rate": 9.423571166319912e-05, "loss": 0.9972, "step": 4422 }, { "epoch": 1.6048072562358278, "grad_norm": 0.19921875, "learning_rate": 9.41867580467507e-05, "loss": 0.938, "step": 4424 }, { "epoch": 1.6055328798185942, "grad_norm": 0.1552734375, "learning_rate": 9.413780443030229e-05, "loss": 0.8708, "step": 4426 }, { "epoch": 1.6062585034013606, "grad_norm": 0.1796875, "learning_rate": 9.408885081385388e-05, "loss": 1.0519, "step": 4428 }, { "epoch": 1.606984126984127, "grad_norm": 0.1748046875, "learning_rate": 9.403989719740545e-05, "loss": 0.9837, "step": 4430 }, { "epoch": 1.6077097505668934, "grad_norm": 0.1611328125, "learning_rate": 9.399094358095705e-05, "loss": 1.0007, "step": 4432 }, { "epoch": 1.6084353741496598, "grad_norm": 0.1767578125, "learning_rate": 9.394198996450863e-05, "loss": 0.9906, "step": 4434 }, { "epoch": 1.6091609977324262, "grad_norm": 0.1533203125, "learning_rate": 9.389303634806021e-05, "loss": 0.9736, "step": 4436 }, { "epoch": 1.6098866213151928, "grad_norm": 0.193359375, "learning_rate": 9.384408273161181e-05, "loss": 1.0087, "step": 4438 }, { "epoch": 1.6106122448979592, "grad_norm": 0.1455078125, "learning_rate": 9.379512911516338e-05, "loss": 0.9447, "step": 4440 }, { "epoch": 1.6113378684807256, "grad_norm": 0.1708984375, "learning_rate": 9.374617549871497e-05, "loss": 1.0428, "step": 4442 }, { "epoch": 1.6120634920634922, "grad_norm": 0.154296875, "learning_rate": 9.369722188226656e-05, "loss": 0.9664, "step": 4444 }, { "epoch": 1.6127891156462586, "grad_norm": 0.173828125, "learning_rate": 9.364826826581814e-05, "loss": 0.9886, "step": 4446 }, { "epoch": 1.613514739229025, "grad_norm": 0.19921875, "learning_rate": 9.359931464936973e-05, "loss": 1.0257, "step": 4448 }, { "epoch": 1.6142403628117914, "grad_norm": 0.15234375, "learning_rate": 9.35503610329213e-05, "loss": 0.9166, "step": 4450 }, { "epoch": 1.6149659863945578, "grad_norm": 0.201171875, "learning_rate": 9.35014074164729e-05, "loss": 1.0876, "step": 4452 }, { "epoch": 1.6156916099773242, "grad_norm": 0.1533203125, "learning_rate": 9.345245380002448e-05, "loss": 0.9517, "step": 4454 }, { "epoch": 1.6164172335600906, "grad_norm": 0.154296875, "learning_rate": 9.340350018357606e-05, "loss": 0.9019, "step": 4456 }, { "epoch": 1.617142857142857, "grad_norm": 0.1533203125, "learning_rate": 9.335454656712766e-05, "loss": 0.9497, "step": 4458 }, { "epoch": 1.6178684807256236, "grad_norm": 0.1572265625, "learning_rate": 9.330559295067923e-05, "loss": 0.9503, "step": 4460 }, { "epoch": 1.61859410430839, "grad_norm": 0.1689453125, "learning_rate": 9.325663933423082e-05, "loss": 0.9898, "step": 4462 }, { "epoch": 1.6193197278911564, "grad_norm": 0.15234375, "learning_rate": 9.32076857177824e-05, "loss": 0.9641, "step": 4464 }, { "epoch": 1.620045351473923, "grad_norm": 0.1865234375, "learning_rate": 9.315873210133399e-05, "loss": 0.963, "step": 4466 }, { "epoch": 1.6207709750566894, "grad_norm": 0.1826171875, "learning_rate": 9.310977848488558e-05, "loss": 0.9416, "step": 4468 }, { "epoch": 1.6214965986394558, "grad_norm": 0.162109375, "learning_rate": 9.306082486843715e-05, "loss": 1.0424, "step": 4470 }, { "epoch": 1.6222222222222222, "grad_norm": 0.1474609375, "learning_rate": 9.301187125198875e-05, "loss": 0.9659, "step": 4472 }, { "epoch": 1.6229478458049886, "grad_norm": 0.1513671875, "learning_rate": 9.296291763554033e-05, "loss": 0.8367, "step": 4474 }, { "epoch": 1.623673469387755, "grad_norm": 0.1455078125, "learning_rate": 9.291396401909191e-05, "loss": 0.9668, "step": 4476 }, { "epoch": 1.6243990929705214, "grad_norm": 0.1416015625, "learning_rate": 9.286501040264351e-05, "loss": 0.9655, "step": 4478 }, { "epoch": 1.6251247165532878, "grad_norm": 0.189453125, "learning_rate": 9.281605678619508e-05, "loss": 0.9554, "step": 4480 }, { "epoch": 1.6258503401360545, "grad_norm": 0.158203125, "learning_rate": 9.276710316974667e-05, "loss": 0.9243, "step": 4482 }, { "epoch": 1.6265759637188208, "grad_norm": 0.1572265625, "learning_rate": 9.271814955329826e-05, "loss": 0.9596, "step": 4484 }, { "epoch": 1.6273015873015872, "grad_norm": 0.15625, "learning_rate": 9.266919593684984e-05, "loss": 0.9765, "step": 4486 }, { "epoch": 1.6280272108843539, "grad_norm": 0.16015625, "learning_rate": 9.262024232040143e-05, "loss": 0.9482, "step": 4488 }, { "epoch": 1.6287528344671203, "grad_norm": 0.1689453125, "learning_rate": 9.2571288703953e-05, "loss": 0.9621, "step": 4490 }, { "epoch": 1.6294784580498867, "grad_norm": 0.18359375, "learning_rate": 9.25223350875046e-05, "loss": 1.0181, "step": 4492 }, { "epoch": 1.630204081632653, "grad_norm": 0.1669921875, "learning_rate": 9.247338147105618e-05, "loss": 0.9446, "step": 4494 }, { "epoch": 1.6309297052154195, "grad_norm": 0.162109375, "learning_rate": 9.242442785460776e-05, "loss": 0.9495, "step": 4496 }, { "epoch": 1.6316553287981859, "grad_norm": 0.169921875, "learning_rate": 9.237547423815934e-05, "loss": 1.0283, "step": 4498 }, { "epoch": 1.6323809523809523, "grad_norm": 0.1826171875, "learning_rate": 9.232652062171093e-05, "loss": 1.0349, "step": 4500 }, { "epoch": 1.6323809523809523, "eval_loss": 1.0125901699066162, "eval_runtime": 86.0097, "eval_samples_per_second": 20.928, "eval_steps_per_second": 0.663, "step": 4500 }, { "epoch": 1.6331065759637187, "grad_norm": 0.169921875, "learning_rate": 9.227756700526252e-05, "loss": 1.0119, "step": 4502 }, { "epoch": 1.6338321995464853, "grad_norm": 0.162109375, "learning_rate": 9.22286133888141e-05, "loss": 0.9775, "step": 4504 }, { "epoch": 1.6345578231292517, "grad_norm": 0.177734375, "learning_rate": 9.217965977236569e-05, "loss": 0.9544, "step": 4506 }, { "epoch": 1.6352834467120183, "grad_norm": 0.16796875, "learning_rate": 9.213070615591727e-05, "loss": 0.941, "step": 4508 }, { "epoch": 1.6360090702947847, "grad_norm": 0.1669921875, "learning_rate": 9.208175253946885e-05, "loss": 0.9603, "step": 4510 }, { "epoch": 1.636734693877551, "grad_norm": 0.154296875, "learning_rate": 9.203279892302045e-05, "loss": 0.9505, "step": 4512 }, { "epoch": 1.6374603174603175, "grad_norm": 0.1689453125, "learning_rate": 9.198384530657203e-05, "loss": 1.0291, "step": 4514 }, { "epoch": 1.638185941043084, "grad_norm": 0.1923828125, "learning_rate": 9.193489169012361e-05, "loss": 0.8945, "step": 4516 }, { "epoch": 1.6389115646258503, "grad_norm": 0.166015625, "learning_rate": 9.18859380736752e-05, "loss": 1.0053, "step": 4518 }, { "epoch": 1.6396371882086167, "grad_norm": 0.17578125, "learning_rate": 9.183698445722678e-05, "loss": 0.9705, "step": 4520 }, { "epoch": 1.640362811791383, "grad_norm": 0.1611328125, "learning_rate": 9.178803084077837e-05, "loss": 0.9802, "step": 4522 }, { "epoch": 1.6410884353741495, "grad_norm": 0.1611328125, "learning_rate": 9.173907722432996e-05, "loss": 0.9674, "step": 4524 }, { "epoch": 1.6418140589569161, "grad_norm": 0.1708984375, "learning_rate": 9.169012360788154e-05, "loss": 1.0555, "step": 4526 }, { "epoch": 1.6425396825396825, "grad_norm": 0.1689453125, "learning_rate": 9.164116999143312e-05, "loss": 0.9737, "step": 4528 }, { "epoch": 1.6432653061224491, "grad_norm": 0.1845703125, "learning_rate": 9.15922163749847e-05, "loss": 0.9983, "step": 4530 }, { "epoch": 1.6439909297052155, "grad_norm": 0.1552734375, "learning_rate": 9.15432627585363e-05, "loss": 0.9307, "step": 4532 }, { "epoch": 1.644716553287982, "grad_norm": 0.1552734375, "learning_rate": 9.149430914208788e-05, "loss": 0.9447, "step": 4534 }, { "epoch": 1.6454421768707483, "grad_norm": 0.15234375, "learning_rate": 9.144535552563946e-05, "loss": 0.9611, "step": 4536 }, { "epoch": 1.6461678004535147, "grad_norm": 0.1787109375, "learning_rate": 9.139640190919104e-05, "loss": 1.0478, "step": 4538 }, { "epoch": 1.6468934240362811, "grad_norm": 0.17578125, "learning_rate": 9.134744829274263e-05, "loss": 1.002, "step": 4540 }, { "epoch": 1.6476190476190475, "grad_norm": 0.1533203125, "learning_rate": 9.129849467629421e-05, "loss": 0.9249, "step": 4542 }, { "epoch": 1.648344671201814, "grad_norm": 0.1826171875, "learning_rate": 9.12495410598458e-05, "loss": 0.9697, "step": 4544 }, { "epoch": 1.6490702947845803, "grad_norm": 0.1728515625, "learning_rate": 9.120058744339739e-05, "loss": 0.9693, "step": 4546 }, { "epoch": 1.649795918367347, "grad_norm": 0.1806640625, "learning_rate": 9.115163382694897e-05, "loss": 0.9989, "step": 4548 }, { "epoch": 1.6505215419501134, "grad_norm": 0.1650390625, "learning_rate": 9.110268021050055e-05, "loss": 1.0239, "step": 4550 }, { "epoch": 1.65124716553288, "grad_norm": 0.1533203125, "learning_rate": 9.105372659405213e-05, "loss": 0.9383, "step": 4552 }, { "epoch": 1.6519727891156464, "grad_norm": 0.1796875, "learning_rate": 9.100477297760373e-05, "loss": 0.9475, "step": 4554 }, { "epoch": 1.6526984126984128, "grad_norm": 0.1474609375, "learning_rate": 9.095581936115531e-05, "loss": 0.9904, "step": 4556 }, { "epoch": 1.6534240362811792, "grad_norm": 0.16796875, "learning_rate": 9.09068657447069e-05, "loss": 0.9174, "step": 4558 }, { "epoch": 1.6541496598639456, "grad_norm": 0.166015625, "learning_rate": 9.085791212825848e-05, "loss": 1.0161, "step": 4560 }, { "epoch": 1.654875283446712, "grad_norm": 0.16015625, "learning_rate": 9.080895851181006e-05, "loss": 1.1105, "step": 4562 }, { "epoch": 1.6556009070294784, "grad_norm": 0.1669921875, "learning_rate": 9.076000489536165e-05, "loss": 1.0257, "step": 4564 }, { "epoch": 1.6563265306122448, "grad_norm": 0.1513671875, "learning_rate": 9.071105127891324e-05, "loss": 0.9142, "step": 4566 }, { "epoch": 1.6570521541950112, "grad_norm": 0.1826171875, "learning_rate": 9.066209766246482e-05, "loss": 0.9849, "step": 4568 }, { "epoch": 1.6577777777777778, "grad_norm": 0.1787109375, "learning_rate": 9.06131440460164e-05, "loss": 0.9778, "step": 4570 }, { "epoch": 1.6585034013605442, "grad_norm": 0.1787109375, "learning_rate": 9.056419042956798e-05, "loss": 0.9673, "step": 4572 }, { "epoch": 1.6592290249433108, "grad_norm": 0.16015625, "learning_rate": 9.051523681311958e-05, "loss": 1.0083, "step": 4574 }, { "epoch": 1.6599546485260772, "grad_norm": 0.1689453125, "learning_rate": 9.046628319667115e-05, "loss": 0.9852, "step": 4576 }, { "epoch": 1.6606802721088436, "grad_norm": 0.19921875, "learning_rate": 9.041732958022274e-05, "loss": 0.9384, "step": 4578 }, { "epoch": 1.66140589569161, "grad_norm": 0.166015625, "learning_rate": 9.036837596377433e-05, "loss": 1.0178, "step": 4580 }, { "epoch": 1.6621315192743764, "grad_norm": 0.166015625, "learning_rate": 9.031942234732591e-05, "loss": 0.9881, "step": 4582 }, { "epoch": 1.6628571428571428, "grad_norm": 0.1845703125, "learning_rate": 9.02704687308775e-05, "loss": 1.0091, "step": 4584 }, { "epoch": 1.6635827664399092, "grad_norm": 0.1767578125, "learning_rate": 9.022151511442907e-05, "loss": 1.0278, "step": 4586 }, { "epoch": 1.6643083900226756, "grad_norm": 0.1640625, "learning_rate": 9.017256149798067e-05, "loss": 0.9655, "step": 4588 }, { "epoch": 1.6650340136054422, "grad_norm": 0.15625, "learning_rate": 9.012360788153225e-05, "loss": 0.9821, "step": 4590 }, { "epoch": 1.6657596371882086, "grad_norm": 0.1572265625, "learning_rate": 9.007465426508383e-05, "loss": 1.0376, "step": 4592 }, { "epoch": 1.666485260770975, "grad_norm": 0.150390625, "learning_rate": 9.002570064863543e-05, "loss": 0.9686, "step": 4594 }, { "epoch": 1.6672108843537417, "grad_norm": 0.1787109375, "learning_rate": 8.9976747032187e-05, "loss": 1.0581, "step": 4596 }, { "epoch": 1.667936507936508, "grad_norm": 0.201171875, "learning_rate": 8.99277934157386e-05, "loss": 0.9976, "step": 4598 }, { "epoch": 1.6686621315192744, "grad_norm": 0.1806640625, "learning_rate": 8.987883979929018e-05, "loss": 0.9585, "step": 4600 }, { "epoch": 1.6686621315192744, "eval_loss": 1.0123869180679321, "eval_runtime": 90.9583, "eval_samples_per_second": 19.789, "eval_steps_per_second": 0.627, "step": 4600 }, { "epoch": 1.6693877551020408, "grad_norm": 0.15234375, "learning_rate": 8.982988618284176e-05, "loss": 0.9106, "step": 4602 }, { "epoch": 1.6701133786848072, "grad_norm": 0.1611328125, "learning_rate": 8.978093256639335e-05, "loss": 0.9744, "step": 4604 }, { "epoch": 1.6708390022675736, "grad_norm": 0.1875, "learning_rate": 8.973197894994492e-05, "loss": 0.9738, "step": 4606 }, { "epoch": 1.67156462585034, "grad_norm": 0.158203125, "learning_rate": 8.968302533349652e-05, "loss": 0.9827, "step": 4608 }, { "epoch": 1.6722902494331064, "grad_norm": 0.220703125, "learning_rate": 8.96340717170481e-05, "loss": 1.0603, "step": 4610 }, { "epoch": 1.673015873015873, "grad_norm": 0.162109375, "learning_rate": 8.958511810059968e-05, "loss": 0.9742, "step": 4612 }, { "epoch": 1.6737414965986395, "grad_norm": 0.1591796875, "learning_rate": 8.953616448415128e-05, "loss": 0.9745, "step": 4614 }, { "epoch": 1.6744671201814059, "grad_norm": 0.1806640625, "learning_rate": 8.948721086770285e-05, "loss": 1.039, "step": 4616 }, { "epoch": 1.6751927437641725, "grad_norm": 0.1748046875, "learning_rate": 8.943825725125444e-05, "loss": 0.917, "step": 4618 }, { "epoch": 1.6759183673469389, "grad_norm": 0.1728515625, "learning_rate": 8.938930363480603e-05, "loss": 1.0145, "step": 4620 }, { "epoch": 1.6766439909297053, "grad_norm": 0.1552734375, "learning_rate": 8.934035001835761e-05, "loss": 0.8931, "step": 4622 }, { "epoch": 1.6773696145124717, "grad_norm": 0.169921875, "learning_rate": 8.92913964019092e-05, "loss": 0.9811, "step": 4624 }, { "epoch": 1.678095238095238, "grad_norm": 0.185546875, "learning_rate": 8.924244278546077e-05, "loss": 0.9463, "step": 4626 }, { "epoch": 1.6788208616780045, "grad_norm": 0.1552734375, "learning_rate": 8.919348916901237e-05, "loss": 0.9064, "step": 4628 }, { "epoch": 1.6795464852607709, "grad_norm": 0.177734375, "learning_rate": 8.914453555256395e-05, "loss": 1.0477, "step": 4630 }, { "epoch": 1.6802721088435373, "grad_norm": 0.201171875, "learning_rate": 8.909558193611553e-05, "loss": 1.0279, "step": 4632 }, { "epoch": 1.680997732426304, "grad_norm": 0.1640625, "learning_rate": 8.904662831966713e-05, "loss": 0.9726, "step": 4634 }, { "epoch": 1.6817233560090703, "grad_norm": 0.1484375, "learning_rate": 8.89976747032187e-05, "loss": 0.9378, "step": 4636 }, { "epoch": 1.6824489795918367, "grad_norm": 0.177734375, "learning_rate": 8.89487210867703e-05, "loss": 0.9877, "step": 4638 }, { "epoch": 1.6831746031746033, "grad_norm": 0.1494140625, "learning_rate": 8.889976747032188e-05, "loss": 0.9537, "step": 4640 }, { "epoch": 1.6839002267573697, "grad_norm": 0.181640625, "learning_rate": 8.885081385387346e-05, "loss": 0.9996, "step": 4642 }, { "epoch": 1.6846258503401361, "grad_norm": 0.1611328125, "learning_rate": 8.880186023742504e-05, "loss": 0.9682, "step": 4644 }, { "epoch": 1.6853514739229025, "grad_norm": 0.1533203125, "learning_rate": 8.875290662097662e-05, "loss": 0.995, "step": 4646 }, { "epoch": 1.686077097505669, "grad_norm": 0.1455078125, "learning_rate": 8.870395300452822e-05, "loss": 0.9594, "step": 4648 }, { "epoch": 1.6868027210884353, "grad_norm": 0.16796875, "learning_rate": 8.86549993880798e-05, "loss": 0.9323, "step": 4650 }, { "epoch": 1.6875283446712017, "grad_norm": 0.171875, "learning_rate": 8.860604577163138e-05, "loss": 1.0264, "step": 4652 }, { "epoch": 1.6882539682539681, "grad_norm": 0.1552734375, "learning_rate": 8.855709215518297e-05, "loss": 0.9587, "step": 4654 }, { "epoch": 1.6889795918367347, "grad_norm": 0.16796875, "learning_rate": 8.850813853873455e-05, "loss": 0.9402, "step": 4656 }, { "epoch": 1.6897052154195011, "grad_norm": 0.1708984375, "learning_rate": 8.845918492228614e-05, "loss": 0.9673, "step": 4658 }, { "epoch": 1.6904308390022675, "grad_norm": 0.1611328125, "learning_rate": 8.841023130583773e-05, "loss": 0.9555, "step": 4660 }, { "epoch": 1.6911564625850342, "grad_norm": 0.1650390625, "learning_rate": 8.836127768938931e-05, "loss": 0.9361, "step": 4662 }, { "epoch": 1.6918820861678006, "grad_norm": 0.1533203125, "learning_rate": 8.831232407294089e-05, "loss": 0.9679, "step": 4664 }, { "epoch": 1.692607709750567, "grad_norm": 0.166015625, "learning_rate": 8.826337045649247e-05, "loss": 0.9455, "step": 4666 }, { "epoch": 1.6933333333333334, "grad_norm": 0.1669921875, "learning_rate": 8.821441684004407e-05, "loss": 0.9773, "step": 4668 }, { "epoch": 1.6940589569160998, "grad_norm": 0.1591796875, "learning_rate": 8.816546322359565e-05, "loss": 0.95, "step": 4670 }, { "epoch": 1.6947845804988662, "grad_norm": 0.150390625, "learning_rate": 8.811650960714723e-05, "loss": 0.9737, "step": 4672 }, { "epoch": 1.6955102040816326, "grad_norm": 0.1708984375, "learning_rate": 8.806755599069882e-05, "loss": 1.0247, "step": 4674 }, { "epoch": 1.696235827664399, "grad_norm": 0.1474609375, "learning_rate": 8.80186023742504e-05, "loss": 0.9881, "step": 4676 }, { "epoch": 1.6969614512471656, "grad_norm": 0.173828125, "learning_rate": 8.7969648757802e-05, "loss": 1.0309, "step": 4678 }, { "epoch": 1.697687074829932, "grad_norm": 0.1689453125, "learning_rate": 8.792069514135358e-05, "loss": 0.9641, "step": 4680 }, { "epoch": 1.6984126984126984, "grad_norm": 0.1640625, "learning_rate": 8.787174152490516e-05, "loss": 0.9768, "step": 4682 }, { "epoch": 1.699138321995465, "grad_norm": 0.1650390625, "learning_rate": 8.782278790845674e-05, "loss": 0.9526, "step": 4684 }, { "epoch": 1.6998639455782314, "grad_norm": 0.1611328125, "learning_rate": 8.777383429200832e-05, "loss": 1.0107, "step": 4686 }, { "epoch": 1.7005895691609978, "grad_norm": 0.17578125, "learning_rate": 8.77248806755599e-05, "loss": 1.0372, "step": 4688 }, { "epoch": 1.7013151927437642, "grad_norm": 0.16015625, "learning_rate": 8.76759270591115e-05, "loss": 0.949, "step": 4690 }, { "epoch": 1.7020408163265306, "grad_norm": 0.1552734375, "learning_rate": 8.762697344266308e-05, "loss": 0.9846, "step": 4692 }, { "epoch": 1.702766439909297, "grad_norm": 0.1904296875, "learning_rate": 8.757801982621467e-05, "loss": 0.9488, "step": 4694 }, { "epoch": 1.7034920634920634, "grad_norm": 0.1455078125, "learning_rate": 8.752906620976625e-05, "loss": 0.9439, "step": 4696 }, { "epoch": 1.7042176870748298, "grad_norm": 0.1669921875, "learning_rate": 8.748011259331783e-05, "loss": 0.9623, "step": 4698 }, { "epoch": 1.7049433106575964, "grad_norm": 0.1728515625, "learning_rate": 8.743115897686943e-05, "loss": 1.0633, "step": 4700 }, { "epoch": 1.7049433106575964, "eval_loss": 1.0122075080871582, "eval_runtime": 84.9355, "eval_samples_per_second": 21.193, "eval_steps_per_second": 0.671, "step": 4700 }, { "epoch": 1.7056689342403628, "grad_norm": 0.1591796875, "learning_rate": 8.738220536042101e-05, "loss": 0.9102, "step": 4702 }, { "epoch": 1.7063945578231292, "grad_norm": 0.1552734375, "learning_rate": 8.733325174397259e-05, "loss": 0.9428, "step": 4704 }, { "epoch": 1.7071201814058958, "grad_norm": 0.15625, "learning_rate": 8.728429812752417e-05, "loss": 0.9601, "step": 4706 }, { "epoch": 1.7078458049886622, "grad_norm": 0.16015625, "learning_rate": 8.723534451107576e-05, "loss": 0.9997, "step": 4708 }, { "epoch": 1.7085714285714286, "grad_norm": 0.1806640625, "learning_rate": 8.718639089462735e-05, "loss": 0.9719, "step": 4710 }, { "epoch": 1.709297052154195, "grad_norm": 0.162109375, "learning_rate": 8.713743727817893e-05, "loss": 0.938, "step": 4712 }, { "epoch": 1.7100226757369614, "grad_norm": 0.16796875, "learning_rate": 8.708848366173052e-05, "loss": 0.9769, "step": 4714 }, { "epoch": 1.7107482993197278, "grad_norm": 0.18359375, "learning_rate": 8.70395300452821e-05, "loss": 0.9675, "step": 4716 }, { "epoch": 1.7114739229024942, "grad_norm": 0.1865234375, "learning_rate": 8.699057642883368e-05, "loss": 0.9881, "step": 4718 }, { "epoch": 1.7121995464852606, "grad_norm": 0.1640625, "learning_rate": 8.694162281238528e-05, "loss": 1.0473, "step": 4720 }, { "epoch": 1.7129251700680272, "grad_norm": 0.15234375, "learning_rate": 8.689266919593684e-05, "loss": 0.9488, "step": 4722 }, { "epoch": 1.7136507936507936, "grad_norm": 0.1796875, "learning_rate": 8.684371557948844e-05, "loss": 1.0057, "step": 4724 }, { "epoch": 1.71437641723356, "grad_norm": 0.1689453125, "learning_rate": 8.679476196304002e-05, "loss": 0.9249, "step": 4726 }, { "epoch": 1.7151020408163267, "grad_norm": 0.1943359375, "learning_rate": 8.67458083465916e-05, "loss": 0.9699, "step": 4728 }, { "epoch": 1.715827664399093, "grad_norm": 0.1787109375, "learning_rate": 8.66968547301432e-05, "loss": 1.001, "step": 4730 }, { "epoch": 1.7165532879818595, "grad_norm": 0.1708984375, "learning_rate": 8.664790111369477e-05, "loss": 0.901, "step": 4732 }, { "epoch": 1.7172789115646259, "grad_norm": 0.173828125, "learning_rate": 8.659894749724637e-05, "loss": 1.026, "step": 4734 }, { "epoch": 1.7180045351473923, "grad_norm": 0.173828125, "learning_rate": 8.654999388079795e-05, "loss": 0.9686, "step": 4736 }, { "epoch": 1.7187301587301587, "grad_norm": 0.203125, "learning_rate": 8.650104026434953e-05, "loss": 1.0217, "step": 4738 }, { "epoch": 1.719455782312925, "grad_norm": 0.1669921875, "learning_rate": 8.645208664790113e-05, "loss": 1.0212, "step": 4740 }, { "epoch": 1.7201814058956915, "grad_norm": 0.162109375, "learning_rate": 8.64031330314527e-05, "loss": 0.9317, "step": 4742 }, { "epoch": 1.720907029478458, "grad_norm": 0.158203125, "learning_rate": 8.635417941500429e-05, "loss": 0.9294, "step": 4744 }, { "epoch": 1.7216326530612245, "grad_norm": 0.1474609375, "learning_rate": 8.630522579855587e-05, "loss": 0.9838, "step": 4746 }, { "epoch": 1.7223582766439909, "grad_norm": 0.1796875, "learning_rate": 8.625627218210746e-05, "loss": 0.9376, "step": 4748 }, { "epoch": 1.7230839002267575, "grad_norm": 0.177734375, "learning_rate": 8.620731856565905e-05, "loss": 1.0101, "step": 4750 }, { "epoch": 1.723809523809524, "grad_norm": 0.16796875, "learning_rate": 8.615836494921062e-05, "loss": 0.9394, "step": 4752 }, { "epoch": 1.7245351473922903, "grad_norm": 0.1689453125, "learning_rate": 8.610941133276222e-05, "loss": 1.0197, "step": 4754 }, { "epoch": 1.7252607709750567, "grad_norm": 0.166015625, "learning_rate": 8.60604577163138e-05, "loss": 1.0194, "step": 4756 }, { "epoch": 1.725986394557823, "grad_norm": 0.1865234375, "learning_rate": 8.601150409986538e-05, "loss": 1.0242, "step": 4758 }, { "epoch": 1.7267120181405895, "grad_norm": 0.185546875, "learning_rate": 8.596255048341698e-05, "loss": 0.9163, "step": 4760 }, { "epoch": 1.727437641723356, "grad_norm": 0.1826171875, "learning_rate": 8.591359686696854e-05, "loss": 0.9777, "step": 4762 }, { "epoch": 1.7281632653061223, "grad_norm": 0.162109375, "learning_rate": 8.586464325052014e-05, "loss": 0.9872, "step": 4764 }, { "epoch": 1.728888888888889, "grad_norm": 0.162109375, "learning_rate": 8.581568963407172e-05, "loss": 1.059, "step": 4766 }, { "epoch": 1.7296145124716553, "grad_norm": 0.1845703125, "learning_rate": 8.57667360176233e-05, "loss": 1.0383, "step": 4768 }, { "epoch": 1.7303401360544217, "grad_norm": 0.158203125, "learning_rate": 8.57177824011749e-05, "loss": 0.9363, "step": 4770 }, { "epoch": 1.7310657596371883, "grad_norm": 0.154296875, "learning_rate": 8.566882878472647e-05, "loss": 0.8923, "step": 4772 }, { "epoch": 1.7317913832199547, "grad_norm": 0.1640625, "learning_rate": 8.561987516827807e-05, "loss": 0.9878, "step": 4774 }, { "epoch": 1.7325170068027211, "grad_norm": 0.1875, "learning_rate": 8.557092155182965e-05, "loss": 1.0398, "step": 4776 }, { "epoch": 1.7332426303854875, "grad_norm": 0.19140625, "learning_rate": 8.552196793538123e-05, "loss": 0.9416, "step": 4778 }, { "epoch": 1.733968253968254, "grad_norm": 0.1953125, "learning_rate": 8.547301431893283e-05, "loss": 1.051, "step": 4780 }, { "epoch": 1.7346938775510203, "grad_norm": 0.212890625, "learning_rate": 8.54240607024844e-05, "loss": 1.0494, "step": 4782 }, { "epoch": 1.7354195011337867, "grad_norm": 0.1748046875, "learning_rate": 8.537510708603599e-05, "loss": 1.0535, "step": 4784 }, { "epoch": 1.7361451247165531, "grad_norm": 0.173828125, "learning_rate": 8.532615346958757e-05, "loss": 1.0028, "step": 4786 }, { "epoch": 1.7368707482993198, "grad_norm": 0.189453125, "learning_rate": 8.527719985313915e-05, "loss": 1.0047, "step": 4788 }, { "epoch": 1.7375963718820862, "grad_norm": 0.166015625, "learning_rate": 8.522824623669074e-05, "loss": 0.9975, "step": 4790 }, { "epoch": 1.7383219954648526, "grad_norm": 0.1650390625, "learning_rate": 8.517929262024232e-05, "loss": 1.0143, "step": 4792 }, { "epoch": 1.7390476190476192, "grad_norm": 0.1826171875, "learning_rate": 8.513033900379392e-05, "loss": 0.948, "step": 4794 }, { "epoch": 1.7397732426303856, "grad_norm": 0.1923828125, "learning_rate": 8.50813853873455e-05, "loss": 0.9424, "step": 4796 }, { "epoch": 1.740498866213152, "grad_norm": 0.1962890625, "learning_rate": 8.503243177089708e-05, "loss": 0.9775, "step": 4798 }, { "epoch": 1.7412244897959184, "grad_norm": 0.19140625, "learning_rate": 8.498347815444866e-05, "loss": 0.9664, "step": 4800 }, { "epoch": 1.7412244897959184, "eval_loss": 1.012211561203003, "eval_runtime": 102.4092, "eval_samples_per_second": 17.577, "eval_steps_per_second": 0.557, "step": 4800 }, { "epoch": 1.7419501133786848, "grad_norm": 0.18359375, "learning_rate": 8.493452453800024e-05, "loss": 0.9895, "step": 4802 }, { "epoch": 1.7426757369614512, "grad_norm": 0.1484375, "learning_rate": 8.488557092155184e-05, "loss": 0.995, "step": 4804 }, { "epoch": 1.7434013605442176, "grad_norm": 0.1669921875, "learning_rate": 8.483661730510342e-05, "loss": 0.9932, "step": 4806 }, { "epoch": 1.744126984126984, "grad_norm": 0.1484375, "learning_rate": 8.4787663688655e-05, "loss": 1.005, "step": 4808 }, { "epoch": 1.7448526077097506, "grad_norm": 0.1728515625, "learning_rate": 8.473871007220659e-05, "loss": 0.9672, "step": 4810 }, { "epoch": 1.745578231292517, "grad_norm": 0.1435546875, "learning_rate": 8.468975645575817e-05, "loss": 0.9399, "step": 4812 }, { "epoch": 1.7463038548752836, "grad_norm": 0.15625, "learning_rate": 8.464080283930977e-05, "loss": 0.9642, "step": 4814 }, { "epoch": 1.74702947845805, "grad_norm": 0.1748046875, "learning_rate": 8.459184922286135e-05, "loss": 0.9698, "step": 4816 }, { "epoch": 1.7477551020408164, "grad_norm": 0.1611328125, "learning_rate": 8.454289560641293e-05, "loss": 0.981, "step": 4818 }, { "epoch": 1.7484807256235828, "grad_norm": 0.169921875, "learning_rate": 8.449394198996451e-05, "loss": 0.9702, "step": 4820 }, { "epoch": 1.7492063492063492, "grad_norm": 0.150390625, "learning_rate": 8.44449883735161e-05, "loss": 0.9315, "step": 4822 }, { "epoch": 1.7499319727891156, "grad_norm": 0.1669921875, "learning_rate": 8.439603475706769e-05, "loss": 0.9465, "step": 4824 }, { "epoch": 1.750657596371882, "grad_norm": 0.1572265625, "learning_rate": 8.434708114061927e-05, "loss": 1.013, "step": 4826 }, { "epoch": 1.7513832199546484, "grad_norm": 0.1806640625, "learning_rate": 8.429812752417085e-05, "loss": 0.9432, "step": 4828 }, { "epoch": 1.7521088435374148, "grad_norm": 0.181640625, "learning_rate": 8.424917390772244e-05, "loss": 1.0176, "step": 4830 }, { "epoch": 1.7528344671201814, "grad_norm": 0.171875, "learning_rate": 8.420022029127402e-05, "loss": 1.0079, "step": 4832 }, { "epoch": 1.7535600907029478, "grad_norm": 0.1474609375, "learning_rate": 8.41512666748256e-05, "loss": 0.9487, "step": 4834 }, { "epoch": 1.7542857142857144, "grad_norm": 0.1611328125, "learning_rate": 8.41023130583772e-05, "loss": 0.945, "step": 4836 }, { "epoch": 1.7550113378684808, "grad_norm": 0.166015625, "learning_rate": 8.405335944192878e-05, "loss": 0.9486, "step": 4838 }, { "epoch": 1.7557369614512472, "grad_norm": 0.1796875, "learning_rate": 8.400440582548036e-05, "loss": 0.9777, "step": 4840 }, { "epoch": 1.7564625850340136, "grad_norm": 0.16015625, "learning_rate": 8.395545220903194e-05, "loss": 0.9026, "step": 4842 }, { "epoch": 1.75718820861678, "grad_norm": 0.1669921875, "learning_rate": 8.390649859258353e-05, "loss": 1.082, "step": 4844 }, { "epoch": 1.7579138321995464, "grad_norm": 0.1669921875, "learning_rate": 8.385754497613512e-05, "loss": 0.9822, "step": 4846 }, { "epoch": 1.7586394557823128, "grad_norm": 0.162109375, "learning_rate": 8.38085913596867e-05, "loss": 0.9299, "step": 4848 }, { "epoch": 1.7593650793650792, "grad_norm": 0.1640625, "learning_rate": 8.375963774323829e-05, "loss": 0.9228, "step": 4850 }, { "epoch": 1.7600907029478456, "grad_norm": 0.1767578125, "learning_rate": 8.371068412678987e-05, "loss": 0.985, "step": 4852 }, { "epoch": 1.7608163265306123, "grad_norm": 0.181640625, "learning_rate": 8.366173051034145e-05, "loss": 0.9734, "step": 4854 }, { "epoch": 1.7615419501133787, "grad_norm": 0.1640625, "learning_rate": 8.361277689389305e-05, "loss": 0.9293, "step": 4856 }, { "epoch": 1.7622675736961453, "grad_norm": 0.158203125, "learning_rate": 8.356382327744463e-05, "loss": 0.99, "step": 4858 }, { "epoch": 1.7629931972789117, "grad_norm": 0.1640625, "learning_rate": 8.351486966099621e-05, "loss": 0.9639, "step": 4860 }, { "epoch": 1.763718820861678, "grad_norm": 0.177734375, "learning_rate": 8.34659160445478e-05, "loss": 1.0291, "step": 4862 }, { "epoch": 1.7644444444444445, "grad_norm": 0.1611328125, "learning_rate": 8.341696242809938e-05, "loss": 0.9175, "step": 4864 }, { "epoch": 1.7651700680272109, "grad_norm": 0.17578125, "learning_rate": 8.336800881165097e-05, "loss": 1.0067, "step": 4866 }, { "epoch": 1.7658956916099773, "grad_norm": 0.1669921875, "learning_rate": 8.331905519520254e-05, "loss": 0.9996, "step": 4868 }, { "epoch": 1.7666213151927437, "grad_norm": 0.16796875, "learning_rate": 8.327010157875414e-05, "loss": 0.9372, "step": 4870 }, { "epoch": 1.76734693877551, "grad_norm": 0.1572265625, "learning_rate": 8.322114796230572e-05, "loss": 0.9316, "step": 4872 }, { "epoch": 1.7680725623582765, "grad_norm": 0.15234375, "learning_rate": 8.31721943458573e-05, "loss": 1.0023, "step": 4874 }, { "epoch": 1.768798185941043, "grad_norm": 0.177734375, "learning_rate": 8.31232407294089e-05, "loss": 1.0662, "step": 4876 }, { "epoch": 1.7695238095238095, "grad_norm": 0.1796875, "learning_rate": 8.307428711296047e-05, "loss": 0.9817, "step": 4878 }, { "epoch": 1.7702494331065761, "grad_norm": 0.162109375, "learning_rate": 8.302533349651206e-05, "loss": 0.9366, "step": 4880 }, { "epoch": 1.7709750566893425, "grad_norm": 0.1826171875, "learning_rate": 8.297637988006364e-05, "loss": 0.956, "step": 4882 }, { "epoch": 1.771700680272109, "grad_norm": 0.140625, "learning_rate": 8.292742626361523e-05, "loss": 0.9326, "step": 4884 }, { "epoch": 1.7724263038548753, "grad_norm": 0.1630859375, "learning_rate": 8.287847264716682e-05, "loss": 0.9756, "step": 4886 }, { "epoch": 1.7731519274376417, "grad_norm": 0.173828125, "learning_rate": 8.282951903071839e-05, "loss": 1.0435, "step": 4888 }, { "epoch": 1.7738775510204081, "grad_norm": 0.1689453125, "learning_rate": 8.278056541426999e-05, "loss": 1.0294, "step": 4890 }, { "epoch": 1.7746031746031745, "grad_norm": 0.154296875, "learning_rate": 8.273161179782157e-05, "loss": 0.9823, "step": 4892 }, { "epoch": 1.775328798185941, "grad_norm": 0.162109375, "learning_rate": 8.268265818137315e-05, "loss": 0.9292, "step": 4894 }, { "epoch": 1.7760544217687075, "grad_norm": 0.1728515625, "learning_rate": 8.263370456492475e-05, "loss": 0.9869, "step": 4896 }, { "epoch": 1.776780045351474, "grad_norm": 0.16015625, "learning_rate": 8.258475094847632e-05, "loss": 0.9646, "step": 4898 }, { "epoch": 1.7775056689342403, "grad_norm": 0.1875, "learning_rate": 8.253579733202791e-05, "loss": 0.9385, "step": 4900 }, { "epoch": 1.7775056689342403, "eval_loss": 1.0119349956512451, "eval_runtime": 86.664, "eval_samples_per_second": 20.77, "eval_steps_per_second": 0.658, "step": 4900 }, { "epoch": 1.778231292517007, "grad_norm": 0.1513671875, "learning_rate": 8.248684371557948e-05, "loss": 0.9462, "step": 4902 }, { "epoch": 1.7789569160997734, "grad_norm": 0.1533203125, "learning_rate": 8.243789009913108e-05, "loss": 1.0036, "step": 4904 }, { "epoch": 1.7796825396825398, "grad_norm": 0.166015625, "learning_rate": 8.238893648268267e-05, "loss": 0.9715, "step": 4906 }, { "epoch": 1.7804081632653062, "grad_norm": 0.1591796875, "learning_rate": 8.233998286623424e-05, "loss": 1.0437, "step": 4908 }, { "epoch": 1.7811337868480726, "grad_norm": 0.1728515625, "learning_rate": 8.229102924978584e-05, "loss": 0.9423, "step": 4910 }, { "epoch": 1.781859410430839, "grad_norm": 0.1630859375, "learning_rate": 8.22420756333374e-05, "loss": 0.9602, "step": 4912 }, { "epoch": 1.7825850340136054, "grad_norm": 0.16015625, "learning_rate": 8.2193122016889e-05, "loss": 0.9675, "step": 4914 }, { "epoch": 1.7833106575963718, "grad_norm": 0.1748046875, "learning_rate": 8.21441684004406e-05, "loss": 0.9956, "step": 4916 }, { "epoch": 1.7840362811791384, "grad_norm": 0.1396484375, "learning_rate": 8.209521478399217e-05, "loss": 0.9582, "step": 4918 }, { "epoch": 1.7847619047619048, "grad_norm": 0.1708984375, "learning_rate": 8.204626116754376e-05, "loss": 0.9657, "step": 4920 }, { "epoch": 1.7854875283446712, "grad_norm": 0.169921875, "learning_rate": 8.199730755109533e-05, "loss": 0.9727, "step": 4922 }, { "epoch": 1.7862131519274378, "grad_norm": 0.185546875, "learning_rate": 8.194835393464693e-05, "loss": 0.9675, "step": 4924 }, { "epoch": 1.7869387755102042, "grad_norm": 0.1806640625, "learning_rate": 8.189940031819852e-05, "loss": 0.9393, "step": 4926 }, { "epoch": 1.7876643990929706, "grad_norm": 0.16015625, "learning_rate": 8.185044670175009e-05, "loss": 0.9847, "step": 4928 }, { "epoch": 1.788390022675737, "grad_norm": 0.1728515625, "learning_rate": 8.180149308530169e-05, "loss": 0.9224, "step": 4930 }, { "epoch": 1.7891156462585034, "grad_norm": 0.162109375, "learning_rate": 8.175253946885327e-05, "loss": 1.0141, "step": 4932 }, { "epoch": 1.7898412698412698, "grad_norm": 0.173828125, "learning_rate": 8.170358585240485e-05, "loss": 1.0221, "step": 4934 }, { "epoch": 1.7905668934240362, "grad_norm": 0.1630859375, "learning_rate": 8.165463223595643e-05, "loss": 0.9043, "step": 4936 }, { "epoch": 1.7912925170068026, "grad_norm": 0.173828125, "learning_rate": 8.160567861950802e-05, "loss": 0.9836, "step": 4938 }, { "epoch": 1.7920181405895692, "grad_norm": 0.1640625, "learning_rate": 8.155672500305961e-05, "loss": 0.9173, "step": 4940 }, { "epoch": 1.7927437641723356, "grad_norm": 0.16796875, "learning_rate": 8.15077713866112e-05, "loss": 0.9854, "step": 4942 }, { "epoch": 1.793469387755102, "grad_norm": 0.1748046875, "learning_rate": 8.145881777016278e-05, "loss": 0.9927, "step": 4944 }, { "epoch": 1.7941950113378686, "grad_norm": 0.16796875, "learning_rate": 8.140986415371436e-05, "loss": 0.9674, "step": 4946 }, { "epoch": 1.794920634920635, "grad_norm": 0.1572265625, "learning_rate": 8.136091053726594e-05, "loss": 0.9844, "step": 4948 }, { "epoch": 1.7956462585034014, "grad_norm": 0.1630859375, "learning_rate": 8.131195692081754e-05, "loss": 0.8923, "step": 4950 }, { "epoch": 1.7963718820861678, "grad_norm": 0.154296875, "learning_rate": 8.126300330436912e-05, "loss": 0.9434, "step": 4952 }, { "epoch": 1.7970975056689342, "grad_norm": 0.185546875, "learning_rate": 8.12140496879207e-05, "loss": 1.0141, "step": 4954 }, { "epoch": 1.7978231292517006, "grad_norm": 0.1689453125, "learning_rate": 8.116509607147228e-05, "loss": 0.9718, "step": 4956 }, { "epoch": 1.798548752834467, "grad_norm": 0.1884765625, "learning_rate": 8.111614245502387e-05, "loss": 1.1004, "step": 4958 }, { "epoch": 1.7992743764172334, "grad_norm": 0.169921875, "learning_rate": 8.106718883857546e-05, "loss": 0.9479, "step": 4960 }, { "epoch": 1.8, "grad_norm": 0.154296875, "learning_rate": 8.101823522212704e-05, "loss": 0.9463, "step": 4962 }, { "epoch": 1.8007256235827664, "grad_norm": 0.171875, "learning_rate": 8.096928160567863e-05, "loss": 0.9659, "step": 4964 }, { "epoch": 1.8014512471655328, "grad_norm": 0.1611328125, "learning_rate": 8.092032798923021e-05, "loss": 1.0237, "step": 4966 }, { "epoch": 1.8021768707482995, "grad_norm": 0.17578125, "learning_rate": 8.087137437278179e-05, "loss": 1.0192, "step": 4968 }, { "epoch": 1.8029024943310659, "grad_norm": 0.1640625, "learning_rate": 8.082242075633339e-05, "loss": 0.9587, "step": 4970 }, { "epoch": 1.8036281179138323, "grad_norm": 0.16796875, "learning_rate": 8.077346713988497e-05, "loss": 1.0048, "step": 4972 }, { "epoch": 1.8043537414965987, "grad_norm": 0.18359375, "learning_rate": 8.072451352343655e-05, "loss": 0.9607, "step": 4974 }, { "epoch": 1.805079365079365, "grad_norm": 0.1640625, "learning_rate": 8.067555990698813e-05, "loss": 1.0168, "step": 4976 }, { "epoch": 1.8058049886621315, "grad_norm": 0.1728515625, "learning_rate": 8.062660629053972e-05, "loss": 0.9538, "step": 4978 }, { "epoch": 1.8065306122448979, "grad_norm": 0.1796875, "learning_rate": 8.05776526740913e-05, "loss": 0.9468, "step": 4980 }, { "epoch": 1.8072562358276643, "grad_norm": 0.1728515625, "learning_rate": 8.05286990576429e-05, "loss": 0.995, "step": 4982 }, { "epoch": 1.8079818594104309, "grad_norm": 0.1787109375, "learning_rate": 8.047974544119448e-05, "loss": 0.954, "step": 4984 }, { "epoch": 1.8087074829931973, "grad_norm": 0.1484375, "learning_rate": 8.043079182474606e-05, "loss": 0.957, "step": 4986 }, { "epoch": 1.8094331065759637, "grad_norm": 0.1806640625, "learning_rate": 8.038183820829764e-05, "loss": 0.9804, "step": 4988 }, { "epoch": 1.8101587301587303, "grad_norm": 0.1904296875, "learning_rate": 8.033288459184922e-05, "loss": 0.9968, "step": 4990 }, { "epoch": 1.8108843537414967, "grad_norm": 0.1640625, "learning_rate": 8.028393097540082e-05, "loss": 0.9741, "step": 4992 }, { "epoch": 1.811609977324263, "grad_norm": 0.1982421875, "learning_rate": 8.02349773589524e-05, "loss": 0.9552, "step": 4994 }, { "epoch": 1.8123356009070295, "grad_norm": 0.16015625, "learning_rate": 8.018602374250398e-05, "loss": 1.0086, "step": 4996 }, { "epoch": 1.813061224489796, "grad_norm": 0.1572265625, "learning_rate": 8.013707012605557e-05, "loss": 0.9693, "step": 4998 }, { "epoch": 1.8137868480725623, "grad_norm": 0.1767578125, "learning_rate": 8.008811650960715e-05, "loss": 1.0028, "step": 5000 }, { "epoch": 1.8137868480725623, "eval_loss": 1.0117406845092773, "eval_runtime": 82.6702, "eval_samples_per_second": 21.773, "eval_steps_per_second": 0.689, "step": 5000 }, { "epoch": 1.8145124716553287, "grad_norm": 0.15234375, "learning_rate": 8.003916289315874e-05, "loss": 0.968, "step": 5002 }, { "epoch": 1.815238095238095, "grad_norm": 0.1474609375, "learning_rate": 7.999020927671033e-05, "loss": 1.0029, "step": 5004 }, { "epoch": 1.8159637188208617, "grad_norm": 0.1826171875, "learning_rate": 7.994125566026191e-05, "loss": 0.9718, "step": 5006 }, { "epoch": 1.8166893424036281, "grad_norm": 0.1552734375, "learning_rate": 7.989230204381349e-05, "loss": 0.9443, "step": 5008 }, { "epoch": 1.8174149659863945, "grad_norm": 0.1962890625, "learning_rate": 7.984334842736507e-05, "loss": 1.054, "step": 5010 }, { "epoch": 1.8181405895691611, "grad_norm": 0.171875, "learning_rate": 7.979439481091667e-05, "loss": 1.0275, "step": 5012 }, { "epoch": 1.8188662131519275, "grad_norm": 0.181640625, "learning_rate": 7.974544119446824e-05, "loss": 1.0257, "step": 5014 }, { "epoch": 1.819591836734694, "grad_norm": 0.16015625, "learning_rate": 7.969648757801983e-05, "loss": 0.9781, "step": 5016 }, { "epoch": 1.8203174603174603, "grad_norm": 0.1787109375, "learning_rate": 7.964753396157142e-05, "loss": 0.9828, "step": 5018 }, { "epoch": 1.8210430839002267, "grad_norm": 0.1845703125, "learning_rate": 7.9598580345123e-05, "loss": 0.9198, "step": 5020 }, { "epoch": 1.8217687074829931, "grad_norm": 0.173828125, "learning_rate": 7.95496267286746e-05, "loss": 0.9939, "step": 5022 }, { "epoch": 1.8224943310657595, "grad_norm": 0.1650390625, "learning_rate": 7.950067311222616e-05, "loss": 0.9844, "step": 5024 }, { "epoch": 1.823219954648526, "grad_norm": 0.1875, "learning_rate": 7.945171949577776e-05, "loss": 0.9858, "step": 5026 }, { "epoch": 1.8239455782312926, "grad_norm": 0.1826171875, "learning_rate": 7.940276587932934e-05, "loss": 0.9681, "step": 5028 }, { "epoch": 1.824671201814059, "grad_norm": 0.1650390625, "learning_rate": 7.935381226288092e-05, "loss": 1.0324, "step": 5030 }, { "epoch": 1.8253968253968254, "grad_norm": 0.216796875, "learning_rate": 7.930485864643252e-05, "loss": 1.0411, "step": 5032 }, { "epoch": 1.826122448979592, "grad_norm": 0.2080078125, "learning_rate": 7.925590502998409e-05, "loss": 0.9887, "step": 5034 }, { "epoch": 1.8268480725623584, "grad_norm": 0.15625, "learning_rate": 7.920695141353568e-05, "loss": 0.985, "step": 5036 }, { "epoch": 1.8275736961451248, "grad_norm": 0.1923828125, "learning_rate": 7.915799779708727e-05, "loss": 0.9819, "step": 5038 }, { "epoch": 1.8282993197278912, "grad_norm": 0.1640625, "learning_rate": 7.910904418063885e-05, "loss": 0.9861, "step": 5040 }, { "epoch": 1.8290249433106576, "grad_norm": 0.255859375, "learning_rate": 7.906009056419044e-05, "loss": 1.0396, "step": 5042 }, { "epoch": 1.829750566893424, "grad_norm": 0.1826171875, "learning_rate": 7.901113694774201e-05, "loss": 1.0312, "step": 5044 }, { "epoch": 1.8304761904761904, "grad_norm": 0.1630859375, "learning_rate": 7.896218333129361e-05, "loss": 1.0328, "step": 5046 }, { "epoch": 1.8312018140589568, "grad_norm": 0.169921875, "learning_rate": 7.891322971484518e-05, "loss": 0.9466, "step": 5048 }, { "epoch": 1.8319274376417234, "grad_norm": 0.1796875, "learning_rate": 7.886427609839677e-05, "loss": 0.9832, "step": 5050 }, { "epoch": 1.8326530612244898, "grad_norm": 0.2060546875, "learning_rate": 7.881532248194837e-05, "loss": 0.9828, "step": 5052 }, { "epoch": 1.8333786848072562, "grad_norm": 0.15234375, "learning_rate": 7.876636886549994e-05, "loss": 0.9619, "step": 5054 }, { "epoch": 1.8341043083900228, "grad_norm": 0.15625, "learning_rate": 7.871741524905153e-05, "loss": 0.8759, "step": 5056 }, { "epoch": 1.8348299319727892, "grad_norm": 0.1572265625, "learning_rate": 7.86684616326031e-05, "loss": 1.0589, "step": 5058 }, { "epoch": 1.8355555555555556, "grad_norm": 0.1767578125, "learning_rate": 7.86195080161547e-05, "loss": 1.0069, "step": 5060 }, { "epoch": 1.836281179138322, "grad_norm": 0.17578125, "learning_rate": 7.857055439970629e-05, "loss": 0.9726, "step": 5062 }, { "epoch": 1.8370068027210884, "grad_norm": 0.173828125, "learning_rate": 7.852160078325786e-05, "loss": 0.9907, "step": 5064 }, { "epoch": 1.8377324263038548, "grad_norm": 0.1611328125, "learning_rate": 7.847264716680946e-05, "loss": 0.9354, "step": 5066 }, { "epoch": 1.8384580498866212, "grad_norm": 0.1591796875, "learning_rate": 7.842369355036103e-05, "loss": 0.9583, "step": 5068 }, { "epoch": 1.8391836734693876, "grad_norm": 0.15234375, "learning_rate": 7.837473993391262e-05, "loss": 0.9899, "step": 5070 }, { "epoch": 1.8399092970521542, "grad_norm": 0.1630859375, "learning_rate": 7.832578631746422e-05, "loss": 1.0061, "step": 5072 }, { "epoch": 1.8406349206349206, "grad_norm": 0.1396484375, "learning_rate": 7.827683270101579e-05, "loss": 0.8678, "step": 5074 }, { "epoch": 1.841360544217687, "grad_norm": 0.1806640625, "learning_rate": 7.822787908456738e-05, "loss": 0.9876, "step": 5076 }, { "epoch": 1.8420861678004536, "grad_norm": 0.158203125, "learning_rate": 7.817892546811895e-05, "loss": 0.9555, "step": 5078 }, { "epoch": 1.84281179138322, "grad_norm": 0.169921875, "learning_rate": 7.812997185167055e-05, "loss": 1.0482, "step": 5080 }, { "epoch": 1.8435374149659864, "grad_norm": 0.1708984375, "learning_rate": 7.808101823522213e-05, "loss": 0.9854, "step": 5082 }, { "epoch": 1.8442630385487528, "grad_norm": 0.173828125, "learning_rate": 7.803206461877371e-05, "loss": 0.9896, "step": 5084 }, { "epoch": 1.8449886621315192, "grad_norm": 0.1650390625, "learning_rate": 7.798311100232531e-05, "loss": 0.9083, "step": 5086 }, { "epoch": 1.8457142857142856, "grad_norm": 0.1875, "learning_rate": 7.793415738587688e-05, "loss": 0.9821, "step": 5088 }, { "epoch": 1.846439909297052, "grad_norm": 0.1708984375, "learning_rate": 7.788520376942847e-05, "loss": 1.0252, "step": 5090 }, { "epoch": 1.8471655328798184, "grad_norm": 0.1640625, "learning_rate": 7.783625015298005e-05, "loss": 1.0168, "step": 5092 }, { "epoch": 1.847891156462585, "grad_norm": 0.18359375, "learning_rate": 7.778729653653164e-05, "loss": 1.0525, "step": 5094 }, { "epoch": 1.8486167800453515, "grad_norm": 0.16796875, "learning_rate": 7.773834292008323e-05, "loss": 0.946, "step": 5096 }, { "epoch": 1.8493424036281179, "grad_norm": 0.2001953125, "learning_rate": 7.76893893036348e-05, "loss": 0.9013, "step": 5098 }, { "epoch": 1.8500680272108845, "grad_norm": 0.1875, "learning_rate": 7.76404356871864e-05, "loss": 0.9918, "step": 5100 }, { "epoch": 1.8500680272108845, "eval_loss": 1.0116716623306274, "eval_runtime": 85.8661, "eval_samples_per_second": 20.963, "eval_steps_per_second": 0.664, "step": 5100 }, { "epoch": 1.8507936507936509, "grad_norm": 0.1513671875, "learning_rate": 7.759148207073798e-05, "loss": 0.988, "step": 5102 }, { "epoch": 1.8515192743764173, "grad_norm": 0.1904296875, "learning_rate": 7.754252845428956e-05, "loss": 0.9003, "step": 5104 }, { "epoch": 1.8522448979591837, "grad_norm": 0.1494140625, "learning_rate": 7.749357483784116e-05, "loss": 0.9411, "step": 5106 }, { "epoch": 1.85297052154195, "grad_norm": 0.16015625, "learning_rate": 7.744462122139273e-05, "loss": 0.9319, "step": 5108 }, { "epoch": 1.8536961451247165, "grad_norm": 0.189453125, "learning_rate": 7.739566760494432e-05, "loss": 1.046, "step": 5110 }, { "epoch": 1.8544217687074829, "grad_norm": 0.14453125, "learning_rate": 7.73467139884959e-05, "loss": 0.8999, "step": 5112 }, { "epoch": 1.8551473922902493, "grad_norm": 0.169921875, "learning_rate": 7.729776037204749e-05, "loss": 1.0543, "step": 5114 }, { "epoch": 1.855873015873016, "grad_norm": 0.1689453125, "learning_rate": 7.724880675559908e-05, "loss": 1.0118, "step": 5116 }, { "epoch": 1.8565986394557823, "grad_norm": 0.1650390625, "learning_rate": 7.719985313915065e-05, "loss": 1.0171, "step": 5118 }, { "epoch": 1.857324263038549, "grad_norm": 0.158203125, "learning_rate": 7.715089952270225e-05, "loss": 0.9521, "step": 5120 }, { "epoch": 1.8580498866213153, "grad_norm": 0.154296875, "learning_rate": 7.710194590625383e-05, "loss": 0.9663, "step": 5122 }, { "epoch": 1.8587755102040817, "grad_norm": 0.1826171875, "learning_rate": 7.705299228980541e-05, "loss": 0.9721, "step": 5124 }, { "epoch": 1.8595011337868481, "grad_norm": 0.17578125, "learning_rate": 7.7004038673357e-05, "loss": 1.0055, "step": 5126 }, { "epoch": 1.8602267573696145, "grad_norm": 0.1650390625, "learning_rate": 7.695508505690858e-05, "loss": 0.9677, "step": 5128 }, { "epoch": 1.860952380952381, "grad_norm": 0.16796875, "learning_rate": 7.690613144046017e-05, "loss": 0.9718, "step": 5130 }, { "epoch": 1.8616780045351473, "grad_norm": 0.1865234375, "learning_rate": 7.685717782401175e-05, "loss": 0.9995, "step": 5132 }, { "epoch": 1.8624036281179137, "grad_norm": 0.154296875, "learning_rate": 7.680822420756334e-05, "loss": 0.9169, "step": 5134 }, { "epoch": 1.8631292517006801, "grad_norm": 0.166015625, "learning_rate": 7.675927059111492e-05, "loss": 0.9085, "step": 5136 }, { "epoch": 1.8638548752834467, "grad_norm": 0.1796875, "learning_rate": 7.67103169746665e-05, "loss": 0.9904, "step": 5138 }, { "epoch": 1.8645804988662131, "grad_norm": 0.1708984375, "learning_rate": 7.66613633582181e-05, "loss": 1.0268, "step": 5140 }, { "epoch": 1.8653061224489798, "grad_norm": 0.1533203125, "learning_rate": 7.661240974176968e-05, "loss": 1.0152, "step": 5142 }, { "epoch": 1.8660317460317462, "grad_norm": 0.1708984375, "learning_rate": 7.656345612532126e-05, "loss": 1.0148, "step": 5144 }, { "epoch": 1.8667573696145126, "grad_norm": 0.16015625, "learning_rate": 7.651450250887284e-05, "loss": 0.9883, "step": 5146 }, { "epoch": 1.867482993197279, "grad_norm": 0.1669921875, "learning_rate": 7.646554889242443e-05, "loss": 1.0816, "step": 5148 }, { "epoch": 1.8682086167800454, "grad_norm": 0.17578125, "learning_rate": 7.641659527597602e-05, "loss": 0.9106, "step": 5150 }, { "epoch": 1.8689342403628117, "grad_norm": 0.177734375, "learning_rate": 7.63676416595276e-05, "loss": 0.9509, "step": 5152 }, { "epoch": 1.8696598639455781, "grad_norm": 0.1689453125, "learning_rate": 7.631868804307919e-05, "loss": 0.8867, "step": 5154 }, { "epoch": 1.8703854875283445, "grad_norm": 0.1953125, "learning_rate": 7.626973442663077e-05, "loss": 1.0658, "step": 5156 }, { "epoch": 1.871111111111111, "grad_norm": 0.1396484375, "learning_rate": 7.622078081018235e-05, "loss": 0.9254, "step": 5158 }, { "epoch": 1.8718367346938776, "grad_norm": 0.15234375, "learning_rate": 7.617182719373393e-05, "loss": 0.9928, "step": 5160 }, { "epoch": 1.872562358276644, "grad_norm": 0.16796875, "learning_rate": 7.612287357728553e-05, "loss": 0.9523, "step": 5162 }, { "epoch": 1.8732879818594106, "grad_norm": 0.1826171875, "learning_rate": 7.607391996083711e-05, "loss": 1.0135, "step": 5164 }, { "epoch": 1.874013605442177, "grad_norm": 0.1708984375, "learning_rate": 7.60249663443887e-05, "loss": 1.0446, "step": 5166 }, { "epoch": 1.8747392290249434, "grad_norm": 0.1787109375, "learning_rate": 7.597601272794028e-05, "loss": 0.9577, "step": 5168 }, { "epoch": 1.8754648526077098, "grad_norm": 0.16796875, "learning_rate": 7.592705911149186e-05, "loss": 1.0376, "step": 5170 }, { "epoch": 1.8761904761904762, "grad_norm": 0.15625, "learning_rate": 7.587810549504345e-05, "loss": 0.9016, "step": 5172 }, { "epoch": 1.8769160997732426, "grad_norm": 0.1689453125, "learning_rate": 7.582915187859504e-05, "loss": 0.9314, "step": 5174 }, { "epoch": 1.877641723356009, "grad_norm": 0.1728515625, "learning_rate": 7.578019826214662e-05, "loss": 0.9535, "step": 5176 }, { "epoch": 1.8783673469387754, "grad_norm": 0.166015625, "learning_rate": 7.57312446456982e-05, "loss": 1.0041, "step": 5178 }, { "epoch": 1.879092970521542, "grad_norm": 0.1611328125, "learning_rate": 7.568229102924978e-05, "loss": 0.9118, "step": 5180 }, { "epoch": 1.8798185941043084, "grad_norm": 0.1650390625, "learning_rate": 7.563333741280138e-05, "loss": 0.9734, "step": 5182 }, { "epoch": 1.8805442176870748, "grad_norm": 0.162109375, "learning_rate": 7.558438379635296e-05, "loss": 0.941, "step": 5184 }, { "epoch": 1.8812698412698414, "grad_norm": 0.173828125, "learning_rate": 7.553543017990454e-05, "loss": 1.0394, "step": 5186 }, { "epoch": 1.8819954648526078, "grad_norm": 0.154296875, "learning_rate": 7.548647656345613e-05, "loss": 0.9472, "step": 5188 }, { "epoch": 1.8827210884353742, "grad_norm": 0.1787109375, "learning_rate": 7.543752294700771e-05, "loss": 0.9045, "step": 5190 }, { "epoch": 1.8834467120181406, "grad_norm": 0.18359375, "learning_rate": 7.53885693305593e-05, "loss": 1.0148, "step": 5192 }, { "epoch": 1.884172335600907, "grad_norm": 0.1572265625, "learning_rate": 7.533961571411087e-05, "loss": 0.9824, "step": 5194 }, { "epoch": 1.8848979591836734, "grad_norm": 0.1689453125, "learning_rate": 7.529066209766247e-05, "loss": 1.0019, "step": 5196 }, { "epoch": 1.8856235827664398, "grad_norm": 0.216796875, "learning_rate": 7.524170848121405e-05, "loss": 1.0268, "step": 5198 }, { "epoch": 1.8863492063492062, "grad_norm": 0.166015625, "learning_rate": 7.519275486476563e-05, "loss": 1.0249, "step": 5200 }, { "epoch": 1.8863492063492062, "eval_loss": 1.011323094367981, "eval_runtime": 85.9575, "eval_samples_per_second": 20.941, "eval_steps_per_second": 0.663, "step": 5200 }, { "epoch": 1.8870748299319728, "grad_norm": 0.181640625, "learning_rate": 7.514380124831723e-05, "loss": 0.9658, "step": 5202 }, { "epoch": 1.8878004535147392, "grad_norm": 0.16796875, "learning_rate": 7.50948476318688e-05, "loss": 1.0172, "step": 5204 }, { "epoch": 1.8885260770975056, "grad_norm": 0.1875, "learning_rate": 7.50458940154204e-05, "loss": 0.9279, "step": 5206 }, { "epoch": 1.8892517006802723, "grad_norm": 0.162109375, "learning_rate": 7.499694039897198e-05, "loss": 0.9781, "step": 5208 }, { "epoch": 1.8899773242630387, "grad_norm": 0.1845703125, "learning_rate": 7.494798678252356e-05, "loss": 1.0788, "step": 5210 }, { "epoch": 1.890702947845805, "grad_norm": 0.13671875, "learning_rate": 7.489903316607515e-05, "loss": 0.9595, "step": 5212 }, { "epoch": 1.8914285714285715, "grad_norm": 0.158203125, "learning_rate": 7.485007954962672e-05, "loss": 0.9927, "step": 5214 }, { "epoch": 1.8921541950113379, "grad_norm": 0.1748046875, "learning_rate": 7.480112593317832e-05, "loss": 1.0254, "step": 5216 }, { "epoch": 1.8928798185941043, "grad_norm": 0.1923828125, "learning_rate": 7.47521723167299e-05, "loss": 1.0419, "step": 5218 }, { "epoch": 1.8936054421768707, "grad_norm": 0.205078125, "learning_rate": 7.470321870028148e-05, "loss": 0.9921, "step": 5220 }, { "epoch": 1.894331065759637, "grad_norm": 0.17578125, "learning_rate": 7.465426508383308e-05, "loss": 0.9195, "step": 5222 }, { "epoch": 1.8950566893424037, "grad_norm": 0.1748046875, "learning_rate": 7.460531146738465e-05, "loss": 1.0073, "step": 5224 }, { "epoch": 1.89578231292517, "grad_norm": 0.19921875, "learning_rate": 7.455635785093624e-05, "loss": 1.0088, "step": 5226 }, { "epoch": 1.8965079365079365, "grad_norm": 0.1689453125, "learning_rate": 7.450740423448783e-05, "loss": 0.938, "step": 5228 }, { "epoch": 1.897233560090703, "grad_norm": 0.146484375, "learning_rate": 7.445845061803941e-05, "loss": 0.9745, "step": 5230 }, { "epoch": 1.8979591836734695, "grad_norm": 0.16796875, "learning_rate": 7.4409497001591e-05, "loss": 0.9514, "step": 5232 }, { "epoch": 1.898684807256236, "grad_norm": 0.173828125, "learning_rate": 7.436054338514257e-05, "loss": 0.9912, "step": 5234 }, { "epoch": 1.8994104308390023, "grad_norm": 0.1640625, "learning_rate": 7.431158976869417e-05, "loss": 0.9794, "step": 5236 }, { "epoch": 1.9001360544217687, "grad_norm": 0.2099609375, "learning_rate": 7.426263615224575e-05, "loss": 1.1444, "step": 5238 }, { "epoch": 1.900861678004535, "grad_norm": 0.1591796875, "learning_rate": 7.421368253579733e-05, "loss": 0.9846, "step": 5240 }, { "epoch": 1.9015873015873015, "grad_norm": 0.18359375, "learning_rate": 7.416472891934893e-05, "loss": 0.9898, "step": 5242 }, { "epoch": 1.902312925170068, "grad_norm": 0.1455078125, "learning_rate": 7.41157753029005e-05, "loss": 0.9763, "step": 5244 }, { "epoch": 1.9030385487528345, "grad_norm": 0.171875, "learning_rate": 7.40668216864521e-05, "loss": 0.9776, "step": 5246 }, { "epoch": 1.903764172335601, "grad_norm": 0.17578125, "learning_rate": 7.401786807000368e-05, "loss": 0.9866, "step": 5248 }, { "epoch": 1.9044897959183673, "grad_norm": 0.166015625, "learning_rate": 7.396891445355526e-05, "loss": 0.9892, "step": 5250 }, { "epoch": 1.905215419501134, "grad_norm": 0.15625, "learning_rate": 7.391996083710685e-05, "loss": 0.9734, "step": 5252 }, { "epoch": 1.9059410430839003, "grad_norm": 0.1953125, "learning_rate": 7.387100722065842e-05, "loss": 1.0002, "step": 5254 }, { "epoch": 1.9066666666666667, "grad_norm": 0.1533203125, "learning_rate": 7.382205360421002e-05, "loss": 0.9108, "step": 5256 }, { "epoch": 1.9073922902494331, "grad_norm": 0.1806640625, "learning_rate": 7.37730999877616e-05, "loss": 0.9652, "step": 5258 }, { "epoch": 1.9081179138321995, "grad_norm": 0.2041015625, "learning_rate": 7.372414637131318e-05, "loss": 0.9549, "step": 5260 }, { "epoch": 1.908843537414966, "grad_norm": 0.16796875, "learning_rate": 7.367519275486478e-05, "loss": 1.0341, "step": 5262 }, { "epoch": 1.9095691609977323, "grad_norm": 0.1767578125, "learning_rate": 7.362623913841635e-05, "loss": 0.9873, "step": 5264 }, { "epoch": 1.9102947845804987, "grad_norm": 0.173828125, "learning_rate": 7.357728552196794e-05, "loss": 0.9916, "step": 5266 }, { "epoch": 1.9110204081632653, "grad_norm": 0.162109375, "learning_rate": 7.352833190551953e-05, "loss": 0.96, "step": 5268 }, { "epoch": 1.9117460317460317, "grad_norm": 0.1708984375, "learning_rate": 7.347937828907111e-05, "loss": 0.9318, "step": 5270 }, { "epoch": 1.9124716553287981, "grad_norm": 0.1611328125, "learning_rate": 7.343042467262269e-05, "loss": 0.9651, "step": 5272 }, { "epoch": 1.9131972789115648, "grad_norm": 0.1669921875, "learning_rate": 7.338147105617427e-05, "loss": 0.9995, "step": 5274 }, { "epoch": 1.9139229024943312, "grad_norm": 0.1494140625, "learning_rate": 7.333251743972587e-05, "loss": 1.0009, "step": 5276 }, { "epoch": 1.9146485260770976, "grad_norm": 0.171875, "learning_rate": 7.328356382327745e-05, "loss": 1.0127, "step": 5278 }, { "epoch": 1.915374149659864, "grad_norm": 0.1416015625, "learning_rate": 7.323461020682903e-05, "loss": 0.9979, "step": 5280 }, { "epoch": 1.9160997732426304, "grad_norm": 0.2001953125, "learning_rate": 7.318565659038062e-05, "loss": 1.057, "step": 5282 }, { "epoch": 1.9168253968253968, "grad_norm": 0.16015625, "learning_rate": 7.31367029739322e-05, "loss": 1.0084, "step": 5284 }, { "epoch": 1.9175510204081632, "grad_norm": 0.1787109375, "learning_rate": 7.308774935748379e-05, "loss": 1.0373, "step": 5286 }, { "epoch": 1.9182766439909296, "grad_norm": 0.1923828125, "learning_rate": 7.303879574103538e-05, "loss": 1.0019, "step": 5288 }, { "epoch": 1.9190022675736962, "grad_norm": 0.18359375, "learning_rate": 7.298984212458696e-05, "loss": 0.9535, "step": 5290 }, { "epoch": 1.9197278911564626, "grad_norm": 0.169921875, "learning_rate": 7.294088850813854e-05, "loss": 0.9318, "step": 5292 }, { "epoch": 1.920453514739229, "grad_norm": 0.1484375, "learning_rate": 7.289193489169012e-05, "loss": 0.9384, "step": 5294 }, { "epoch": 1.9211791383219956, "grad_norm": 0.177734375, "learning_rate": 7.284298127524172e-05, "loss": 0.9661, "step": 5296 }, { "epoch": 1.921904761904762, "grad_norm": 0.166015625, "learning_rate": 7.27940276587933e-05, "loss": 0.9253, "step": 5298 }, { "epoch": 1.9226303854875284, "grad_norm": 0.1513671875, "learning_rate": 7.274507404234488e-05, "loss": 0.9973, "step": 5300 }, { "epoch": 1.9226303854875284, "eval_loss": 1.0112332105636597, "eval_runtime": 86.071, "eval_samples_per_second": 20.913, "eval_steps_per_second": 0.662, "step": 5300 }, { "epoch": 1.9233560090702948, "grad_norm": 0.1796875, "learning_rate": 7.269612042589646e-05, "loss": 1.0532, "step": 5302 }, { "epoch": 1.9240816326530612, "grad_norm": 0.1669921875, "learning_rate": 7.264716680944805e-05, "loss": 0.9237, "step": 5304 }, { "epoch": 1.9248072562358276, "grad_norm": 0.15625, "learning_rate": 7.259821319299963e-05, "loss": 0.9588, "step": 5306 }, { "epoch": 1.925532879818594, "grad_norm": 0.16796875, "learning_rate": 7.254925957655123e-05, "loss": 1.0311, "step": 5308 }, { "epoch": 1.9262585034013604, "grad_norm": 0.15625, "learning_rate": 7.250030596010281e-05, "loss": 0.901, "step": 5310 }, { "epoch": 1.926984126984127, "grad_norm": 0.18359375, "learning_rate": 7.245135234365439e-05, "loss": 1.0205, "step": 5312 }, { "epoch": 1.9277097505668934, "grad_norm": 0.1533203125, "learning_rate": 7.240239872720597e-05, "loss": 0.958, "step": 5314 }, { "epoch": 1.9284353741496598, "grad_norm": 0.17578125, "learning_rate": 7.235344511075755e-05, "loss": 0.996, "step": 5316 }, { "epoch": 1.9291609977324264, "grad_norm": 0.162109375, "learning_rate": 7.230449149430915e-05, "loss": 0.9822, "step": 5318 }, { "epoch": 1.9298866213151928, "grad_norm": 0.1484375, "learning_rate": 7.225553787786073e-05, "loss": 0.9584, "step": 5320 }, { "epoch": 1.9306122448979592, "grad_norm": 0.1611328125, "learning_rate": 7.220658426141231e-05, "loss": 0.9161, "step": 5322 }, { "epoch": 1.9313378684807256, "grad_norm": 0.158203125, "learning_rate": 7.21576306449639e-05, "loss": 0.9766, "step": 5324 }, { "epoch": 1.932063492063492, "grad_norm": 0.166015625, "learning_rate": 7.210867702851548e-05, "loss": 1.0074, "step": 5326 }, { "epoch": 1.9327891156462584, "grad_norm": 0.166015625, "learning_rate": 7.205972341206708e-05, "loss": 0.9183, "step": 5328 }, { "epoch": 1.9335147392290248, "grad_norm": 0.2021484375, "learning_rate": 7.201076979561866e-05, "loss": 1.01, "step": 5330 }, { "epoch": 1.9342403628117912, "grad_norm": 0.18359375, "learning_rate": 7.196181617917024e-05, "loss": 0.969, "step": 5332 }, { "epoch": 1.9349659863945579, "grad_norm": 0.177734375, "learning_rate": 7.191286256272182e-05, "loss": 1.0358, "step": 5334 }, { "epoch": 1.9356916099773243, "grad_norm": 0.1865234375, "learning_rate": 7.18639089462734e-05, "loss": 1.0049, "step": 5336 }, { "epoch": 1.9364172335600907, "grad_norm": 0.1611328125, "learning_rate": 7.1814955329825e-05, "loss": 0.9693, "step": 5338 }, { "epoch": 1.9371428571428573, "grad_norm": 0.14453125, "learning_rate": 7.176600171337657e-05, "loss": 0.9089, "step": 5340 }, { "epoch": 1.9378684807256237, "grad_norm": 0.18359375, "learning_rate": 7.171704809692816e-05, "loss": 0.9832, "step": 5342 }, { "epoch": 1.93859410430839, "grad_norm": 0.162109375, "learning_rate": 7.166809448047975e-05, "loss": 0.9502, "step": 5344 }, { "epoch": 1.9393197278911565, "grad_norm": 0.15625, "learning_rate": 7.161914086403133e-05, "loss": 0.9835, "step": 5346 }, { "epoch": 1.9400453514739229, "grad_norm": 0.1689453125, "learning_rate": 7.157018724758293e-05, "loss": 0.9826, "step": 5348 }, { "epoch": 1.9407709750566893, "grad_norm": 0.166015625, "learning_rate": 7.15212336311345e-05, "loss": 0.9625, "step": 5350 }, { "epoch": 1.9414965986394557, "grad_norm": 0.1728515625, "learning_rate": 7.147228001468609e-05, "loss": 0.9491, "step": 5352 }, { "epoch": 1.942222222222222, "grad_norm": 0.1845703125, "learning_rate": 7.142332639823767e-05, "loss": 1.0431, "step": 5354 }, { "epoch": 1.9429478458049887, "grad_norm": 0.15625, "learning_rate": 7.137437278178925e-05, "loss": 1.0504, "step": 5356 }, { "epoch": 1.943673469387755, "grad_norm": 0.1865234375, "learning_rate": 7.132541916534085e-05, "loss": 1.0378, "step": 5358 }, { "epoch": 1.9443990929705215, "grad_norm": 0.1572265625, "learning_rate": 7.127646554889242e-05, "loss": 0.949, "step": 5360 }, { "epoch": 1.9451247165532881, "grad_norm": 0.1640625, "learning_rate": 7.122751193244401e-05, "loss": 0.9973, "step": 5362 }, { "epoch": 1.9458503401360545, "grad_norm": 0.1630859375, "learning_rate": 7.11785583159956e-05, "loss": 0.996, "step": 5364 }, { "epoch": 1.946575963718821, "grad_norm": 0.171875, "learning_rate": 7.112960469954718e-05, "loss": 1.0218, "step": 5366 }, { "epoch": 1.9473015873015873, "grad_norm": 0.1669921875, "learning_rate": 7.108065108309878e-05, "loss": 1.0081, "step": 5368 }, { "epoch": 1.9480272108843537, "grad_norm": 0.1611328125, "learning_rate": 7.103169746665034e-05, "loss": 0.9438, "step": 5370 }, { "epoch": 1.94875283446712, "grad_norm": 0.2109375, "learning_rate": 7.098274385020194e-05, "loss": 0.9731, "step": 5372 }, { "epoch": 1.9494784580498865, "grad_norm": 0.1611328125, "learning_rate": 7.093379023375352e-05, "loss": 0.9133, "step": 5374 }, { "epoch": 1.950204081632653, "grad_norm": 0.1767578125, "learning_rate": 7.08848366173051e-05, "loss": 0.946, "step": 5376 }, { "epoch": 1.9509297052154195, "grad_norm": 0.19140625, "learning_rate": 7.08358830008567e-05, "loss": 0.9632, "step": 5378 }, { "epoch": 1.951655328798186, "grad_norm": 0.1650390625, "learning_rate": 7.078692938440827e-05, "loss": 0.9542, "step": 5380 }, { "epoch": 1.9523809523809523, "grad_norm": 0.1455078125, "learning_rate": 7.073797576795986e-05, "loss": 0.9328, "step": 5382 }, { "epoch": 1.953106575963719, "grad_norm": 0.1728515625, "learning_rate": 7.068902215151145e-05, "loss": 0.9566, "step": 5384 }, { "epoch": 1.9538321995464853, "grad_norm": 0.1865234375, "learning_rate": 7.064006853506303e-05, "loss": 0.9863, "step": 5386 }, { "epoch": 1.9545578231292517, "grad_norm": 0.1494140625, "learning_rate": 7.059111491861462e-05, "loss": 0.9328, "step": 5388 }, { "epoch": 1.9552834467120181, "grad_norm": 0.169921875, "learning_rate": 7.05421613021662e-05, "loss": 0.9978, "step": 5390 }, { "epoch": 1.9560090702947845, "grad_norm": 0.177734375, "learning_rate": 7.049320768571779e-05, "loss": 1.0063, "step": 5392 }, { "epoch": 1.956734693877551, "grad_norm": 0.1669921875, "learning_rate": 7.044425406926937e-05, "loss": 0.9898, "step": 5394 }, { "epoch": 1.9574603174603173, "grad_norm": 0.166015625, "learning_rate": 7.039530045282095e-05, "loss": 0.9487, "step": 5396 }, { "epoch": 1.9581859410430837, "grad_norm": 0.162109375, "learning_rate": 7.034634683637255e-05, "loss": 0.9949, "step": 5398 }, { "epoch": 1.9589115646258504, "grad_norm": 0.1494140625, "learning_rate": 7.029739321992412e-05, "loss": 0.9803, "step": 5400 }, { "epoch": 1.9589115646258504, "eval_loss": 1.011096715927124, "eval_runtime": 82.3524, "eval_samples_per_second": 21.857, "eval_steps_per_second": 0.692, "step": 5400 }, { "epoch": 1.9596371882086168, "grad_norm": 0.1298828125, "learning_rate": 7.024843960347571e-05, "loss": 0.9387, "step": 5402 }, { "epoch": 1.9603628117913834, "grad_norm": 0.1630859375, "learning_rate": 7.01994859870273e-05, "loss": 0.9136, "step": 5404 }, { "epoch": 1.9610884353741498, "grad_norm": 0.1748046875, "learning_rate": 7.015053237057888e-05, "loss": 0.9976, "step": 5406 }, { "epoch": 1.9618140589569162, "grad_norm": 0.150390625, "learning_rate": 7.010157875413047e-05, "loss": 0.9514, "step": 5408 }, { "epoch": 1.9625396825396826, "grad_norm": 0.1591796875, "learning_rate": 7.005262513768204e-05, "loss": 0.9502, "step": 5410 }, { "epoch": 1.963265306122449, "grad_norm": 0.16015625, "learning_rate": 7.000367152123364e-05, "loss": 0.9255, "step": 5412 }, { "epoch": 1.9639909297052154, "grad_norm": 0.1748046875, "learning_rate": 6.995471790478522e-05, "loss": 0.9789, "step": 5414 }, { "epoch": 1.9647165532879818, "grad_norm": 0.15234375, "learning_rate": 6.99057642883368e-05, "loss": 0.9001, "step": 5416 }, { "epoch": 1.9654421768707482, "grad_norm": 0.150390625, "learning_rate": 6.985681067188839e-05, "loss": 1.0297, "step": 5418 }, { "epoch": 1.9661678004535146, "grad_norm": 0.1748046875, "learning_rate": 6.980785705543997e-05, "loss": 0.9494, "step": 5420 }, { "epoch": 1.9668934240362812, "grad_norm": 0.1591796875, "learning_rate": 6.975890343899156e-05, "loss": 0.9232, "step": 5422 }, { "epoch": 1.9676190476190476, "grad_norm": 0.158203125, "learning_rate": 6.970994982254315e-05, "loss": 0.9913, "step": 5424 }, { "epoch": 1.9683446712018142, "grad_norm": 0.1494140625, "learning_rate": 6.966099620609473e-05, "loss": 0.9597, "step": 5426 }, { "epoch": 1.9690702947845806, "grad_norm": 0.1650390625, "learning_rate": 6.961204258964631e-05, "loss": 0.9634, "step": 5428 }, { "epoch": 1.969795918367347, "grad_norm": 0.1943359375, "learning_rate": 6.95630889731979e-05, "loss": 1.027, "step": 5430 }, { "epoch": 1.9705215419501134, "grad_norm": 0.17578125, "learning_rate": 6.951413535674949e-05, "loss": 0.886, "step": 5432 }, { "epoch": 1.9712471655328798, "grad_norm": 0.1650390625, "learning_rate": 6.946518174030107e-05, "loss": 1.0277, "step": 5434 }, { "epoch": 1.9719727891156462, "grad_norm": 0.166015625, "learning_rate": 6.941622812385265e-05, "loss": 1.018, "step": 5436 }, { "epoch": 1.9726984126984126, "grad_norm": 0.17578125, "learning_rate": 6.936727450740424e-05, "loss": 0.9642, "step": 5438 }, { "epoch": 1.973424036281179, "grad_norm": 0.1748046875, "learning_rate": 6.931832089095582e-05, "loss": 0.9943, "step": 5440 }, { "epoch": 1.9741496598639454, "grad_norm": 0.166015625, "learning_rate": 6.926936727450741e-05, "loss": 0.9534, "step": 5442 }, { "epoch": 1.974875283446712, "grad_norm": 0.1845703125, "learning_rate": 6.9220413658059e-05, "loss": 0.9578, "step": 5444 }, { "epoch": 1.9756009070294784, "grad_norm": 0.1787109375, "learning_rate": 6.917146004161058e-05, "loss": 0.9278, "step": 5446 }, { "epoch": 1.976326530612245, "grad_norm": 0.1962890625, "learning_rate": 6.912250642516216e-05, "loss": 1.0084, "step": 5448 }, { "epoch": 1.9770521541950115, "grad_norm": 0.171875, "learning_rate": 6.907355280871374e-05, "loss": 0.9462, "step": 5450 }, { "epoch": 1.9777777777777779, "grad_norm": 0.1611328125, "learning_rate": 6.902459919226533e-05, "loss": 0.9877, "step": 5452 }, { "epoch": 1.9785034013605443, "grad_norm": 0.169921875, "learning_rate": 6.897564557581692e-05, "loss": 0.9353, "step": 5454 }, { "epoch": 1.9792290249433107, "grad_norm": 0.15625, "learning_rate": 6.89266919593685e-05, "loss": 1.0031, "step": 5456 }, { "epoch": 1.979954648526077, "grad_norm": 0.171875, "learning_rate": 6.887773834292009e-05, "loss": 0.9874, "step": 5458 }, { "epoch": 1.9806802721088435, "grad_norm": 0.1767578125, "learning_rate": 6.882878472647167e-05, "loss": 0.9689, "step": 5460 }, { "epoch": 1.9814058956916099, "grad_norm": 0.1630859375, "learning_rate": 6.877983111002325e-05, "loss": 0.9945, "step": 5462 }, { "epoch": 1.9821315192743763, "grad_norm": 0.1748046875, "learning_rate": 6.873087749357485e-05, "loss": 0.9695, "step": 5464 }, { "epoch": 1.9828571428571429, "grad_norm": 0.1591796875, "learning_rate": 6.868192387712643e-05, "loss": 0.9294, "step": 5466 }, { "epoch": 1.9835827664399093, "grad_norm": 0.16015625, "learning_rate": 6.863297026067801e-05, "loss": 1.0133, "step": 5468 }, { "epoch": 1.984308390022676, "grad_norm": 0.171875, "learning_rate": 6.85840166442296e-05, "loss": 0.9074, "step": 5470 }, { "epoch": 1.9850340136054423, "grad_norm": 0.1982421875, "learning_rate": 6.853506302778118e-05, "loss": 0.9792, "step": 5472 }, { "epoch": 1.9857596371882087, "grad_norm": 0.1845703125, "learning_rate": 6.848610941133277e-05, "loss": 1.0607, "step": 5474 }, { "epoch": 1.986485260770975, "grad_norm": 0.1572265625, "learning_rate": 6.843715579488435e-05, "loss": 0.9649, "step": 5476 }, { "epoch": 1.9872108843537415, "grad_norm": 0.1650390625, "learning_rate": 6.838820217843594e-05, "loss": 0.9928, "step": 5478 }, { "epoch": 1.987936507936508, "grad_norm": 0.2119140625, "learning_rate": 6.833924856198752e-05, "loss": 0.9035, "step": 5480 }, { "epoch": 1.9886621315192743, "grad_norm": 0.1875, "learning_rate": 6.82902949455391e-05, "loss": 1.0789, "step": 5482 }, { "epoch": 1.9893877551020407, "grad_norm": 0.181640625, "learning_rate": 6.82413413290907e-05, "loss": 0.9973, "step": 5484 }, { "epoch": 1.9901133786848073, "grad_norm": 0.1669921875, "learning_rate": 6.819238771264227e-05, "loss": 1.0186, "step": 5486 }, { "epoch": 1.9908390022675737, "grad_norm": 0.1630859375, "learning_rate": 6.814343409619386e-05, "loss": 0.9575, "step": 5488 }, { "epoch": 1.99156462585034, "grad_norm": 0.1513671875, "learning_rate": 6.809448047974544e-05, "loss": 0.9181, "step": 5490 }, { "epoch": 1.9922902494331067, "grad_norm": 0.1826171875, "learning_rate": 6.804552686329703e-05, "loss": 0.9937, "step": 5492 }, { "epoch": 1.9930158730158731, "grad_norm": 0.1923828125, "learning_rate": 6.799657324684862e-05, "loss": 0.9643, "step": 5494 }, { "epoch": 1.9937414965986395, "grad_norm": 0.154296875, "learning_rate": 6.794761963040019e-05, "loss": 0.9727, "step": 5496 }, { "epoch": 1.994467120181406, "grad_norm": 0.1552734375, "learning_rate": 6.789866601395179e-05, "loss": 0.9577, "step": 5498 }, { "epoch": 1.9951927437641723, "grad_norm": 0.1611328125, "learning_rate": 6.784971239750337e-05, "loss": 0.9759, "step": 5500 }, { "epoch": 1.9951927437641723, "eval_loss": 1.0108295679092407, "eval_runtime": 88.2979, "eval_samples_per_second": 20.386, "eval_steps_per_second": 0.646, "step": 5500 }, { "epoch": 1.9959183673469387, "grad_norm": 0.1748046875, "learning_rate": 6.780075878105495e-05, "loss": 1.0464, "step": 5502 }, { "epoch": 1.9966439909297051, "grad_norm": 0.177734375, "learning_rate": 6.775180516460655e-05, "loss": 1.0008, "step": 5504 }, { "epoch": 1.9973696145124715, "grad_norm": 0.19140625, "learning_rate": 6.770285154815812e-05, "loss": 0.9546, "step": 5506 }, { "epoch": 1.9980952380952381, "grad_norm": 0.1611328125, "learning_rate": 6.765389793170971e-05, "loss": 0.9799, "step": 5508 }, { "epoch": 1.9988208616780045, "grad_norm": 0.1650390625, "learning_rate": 6.760494431526129e-05, "loss": 1.0274, "step": 5510 }, { "epoch": 1.999546485260771, "grad_norm": 0.16015625, "learning_rate": 6.755599069881288e-05, "loss": 0.9454, "step": 5512 }, { "epoch": 2.0, "grad_norm": 0.3046875, "learning_rate": 6.750703708236447e-05, "loss": 0.9132, "step": 5514 }, { "epoch": 2.0007256235827664, "grad_norm": 0.19140625, "learning_rate": 6.745808346591604e-05, "loss": 0.9337, "step": 5516 }, { "epoch": 2.001451247165533, "grad_norm": 0.150390625, "learning_rate": 6.740912984946764e-05, "loss": 0.9514, "step": 5518 }, { "epoch": 2.002176870748299, "grad_norm": 0.1806640625, "learning_rate": 6.736017623301922e-05, "loss": 1.0014, "step": 5520 }, { "epoch": 2.0029024943310656, "grad_norm": 0.1796875, "learning_rate": 6.73112226165708e-05, "loss": 0.9509, "step": 5522 }, { "epoch": 2.003628117913832, "grad_norm": 0.154296875, "learning_rate": 6.72622690001224e-05, "loss": 1.0404, "step": 5524 }, { "epoch": 2.004353741496599, "grad_norm": 0.16796875, "learning_rate": 6.721331538367396e-05, "loss": 0.9281, "step": 5526 }, { "epoch": 2.0050793650793652, "grad_norm": 0.16796875, "learning_rate": 6.716436176722556e-05, "loss": 0.9993, "step": 5528 }, { "epoch": 2.0058049886621316, "grad_norm": 0.162109375, "learning_rate": 6.711540815077714e-05, "loss": 1.0137, "step": 5530 }, { "epoch": 2.006530612244898, "grad_norm": 0.1708984375, "learning_rate": 6.706645453432873e-05, "loss": 0.9299, "step": 5532 }, { "epoch": 2.0072562358276644, "grad_norm": 0.1689453125, "learning_rate": 6.701750091788032e-05, "loss": 0.9612, "step": 5534 }, { "epoch": 2.007981859410431, "grad_norm": 0.171875, "learning_rate": 6.696854730143189e-05, "loss": 1.0255, "step": 5536 }, { "epoch": 2.0087074829931972, "grad_norm": 0.158203125, "learning_rate": 6.691959368498349e-05, "loss": 0.9482, "step": 5538 }, { "epoch": 2.0094331065759636, "grad_norm": 0.1630859375, "learning_rate": 6.687064006853507e-05, "loss": 0.9571, "step": 5540 }, { "epoch": 2.01015873015873, "grad_norm": 0.1572265625, "learning_rate": 6.682168645208665e-05, "loss": 0.9913, "step": 5542 }, { "epoch": 2.0108843537414964, "grad_norm": 0.169921875, "learning_rate": 6.677273283563825e-05, "loss": 0.9695, "step": 5544 }, { "epoch": 2.011609977324263, "grad_norm": 0.1591796875, "learning_rate": 6.672377921918981e-05, "loss": 0.9805, "step": 5546 }, { "epoch": 2.0123356009070297, "grad_norm": 0.1689453125, "learning_rate": 6.667482560274141e-05, "loss": 0.9468, "step": 5548 }, { "epoch": 2.013061224489796, "grad_norm": 0.1484375, "learning_rate": 6.662587198629299e-05, "loss": 0.9305, "step": 5550 }, { "epoch": 2.0137868480725625, "grad_norm": 0.1494140625, "learning_rate": 6.657691836984458e-05, "loss": 0.956, "step": 5552 }, { "epoch": 2.014512471655329, "grad_norm": 0.1689453125, "learning_rate": 6.652796475339617e-05, "loss": 0.9536, "step": 5554 }, { "epoch": 2.0152380952380953, "grad_norm": 0.173828125, "learning_rate": 6.647901113694774e-05, "loss": 0.9555, "step": 5556 }, { "epoch": 2.0159637188208617, "grad_norm": 0.150390625, "learning_rate": 6.643005752049934e-05, "loss": 0.9548, "step": 5558 }, { "epoch": 2.016689342403628, "grad_norm": 0.1689453125, "learning_rate": 6.638110390405092e-05, "loss": 1.0376, "step": 5560 }, { "epoch": 2.0174149659863945, "grad_norm": 0.173828125, "learning_rate": 6.63321502876025e-05, "loss": 0.9717, "step": 5562 }, { "epoch": 2.018140589569161, "grad_norm": 0.1865234375, "learning_rate": 6.628319667115408e-05, "loss": 1.0095, "step": 5564 }, { "epoch": 2.0188662131519273, "grad_norm": 0.1533203125, "learning_rate": 6.623424305470566e-05, "loss": 0.9583, "step": 5566 }, { "epoch": 2.0195918367346937, "grad_norm": 0.1640625, "learning_rate": 6.618528943825726e-05, "loss": 1.0386, "step": 5568 }, { "epoch": 2.0203174603174605, "grad_norm": 0.18359375, "learning_rate": 6.613633582180884e-05, "loss": 1.0481, "step": 5570 }, { "epoch": 2.021043083900227, "grad_norm": 0.15625, "learning_rate": 6.608738220536043e-05, "loss": 0.9812, "step": 5572 }, { "epoch": 2.0217687074829933, "grad_norm": 0.16796875, "learning_rate": 6.603842858891201e-05, "loss": 0.9529, "step": 5574 }, { "epoch": 2.0224943310657597, "grad_norm": 0.1611328125, "learning_rate": 6.598947497246359e-05, "loss": 0.9702, "step": 5576 }, { "epoch": 2.023219954648526, "grad_norm": 0.1796875, "learning_rate": 6.594052135601519e-05, "loss": 1.0463, "step": 5578 }, { "epoch": 2.0239455782312925, "grad_norm": 0.1552734375, "learning_rate": 6.589156773956677e-05, "loss": 0.9735, "step": 5580 }, { "epoch": 2.024671201814059, "grad_norm": 0.1591796875, "learning_rate": 6.584261412311835e-05, "loss": 0.9631, "step": 5582 }, { "epoch": 2.0253968253968253, "grad_norm": 0.1494140625, "learning_rate": 6.579366050666993e-05, "loss": 1.0417, "step": 5584 }, { "epoch": 2.0261224489795917, "grad_norm": 0.171875, "learning_rate": 6.574470689022151e-05, "loss": 0.9679, "step": 5586 }, { "epoch": 2.026848072562358, "grad_norm": 0.15625, "learning_rate": 6.569575327377311e-05, "loss": 0.9433, "step": 5588 }, { "epoch": 2.0275736961451245, "grad_norm": 0.1650390625, "learning_rate": 6.564679965732469e-05, "loss": 1.1191, "step": 5590 }, { "epoch": 2.0282993197278913, "grad_norm": 0.166015625, "learning_rate": 6.559784604087628e-05, "loss": 1.0056, "step": 5592 }, { "epoch": 2.0290249433106577, "grad_norm": 0.140625, "learning_rate": 6.554889242442786e-05, "loss": 0.8881, "step": 5594 }, { "epoch": 2.029750566893424, "grad_norm": 0.15234375, "learning_rate": 6.549993880797944e-05, "loss": 1.005, "step": 5596 }, { "epoch": 2.0304761904761905, "grad_norm": 0.16015625, "learning_rate": 6.545098519153102e-05, "loss": 0.9251, "step": 5598 }, { "epoch": 2.031201814058957, "grad_norm": 0.1630859375, "learning_rate": 6.540203157508262e-05, "loss": 0.9907, "step": 5600 }, { "epoch": 2.031201814058957, "eval_loss": 1.0105645656585693, "eval_runtime": 88.0197, "eval_samples_per_second": 20.45, "eval_steps_per_second": 0.648, "step": 5600 }, { "epoch": 2.0319274376417233, "grad_norm": 0.1669921875, "learning_rate": 6.53530779586342e-05, "loss": 1.0097, "step": 5602 }, { "epoch": 2.0326530612244897, "grad_norm": 0.1806640625, "learning_rate": 6.530412434218578e-05, "loss": 0.9781, "step": 5604 }, { "epoch": 2.033378684807256, "grad_norm": 0.1552734375, "learning_rate": 6.525517072573736e-05, "loss": 0.9432, "step": 5606 }, { "epoch": 2.0341043083900225, "grad_norm": 0.1875, "learning_rate": 6.520621710928895e-05, "loss": 0.9518, "step": 5608 }, { "epoch": 2.034829931972789, "grad_norm": 0.1650390625, "learning_rate": 6.515726349284054e-05, "loss": 0.9617, "step": 5610 }, { "epoch": 2.0355555555555553, "grad_norm": 0.181640625, "learning_rate": 6.510830987639212e-05, "loss": 0.954, "step": 5612 }, { "epoch": 2.036281179138322, "grad_norm": 0.173828125, "learning_rate": 6.505935625994371e-05, "loss": 1.0424, "step": 5614 }, { "epoch": 2.0370068027210886, "grad_norm": 0.251953125, "learning_rate": 6.501040264349529e-05, "loss": 1.0404, "step": 5616 }, { "epoch": 2.037732426303855, "grad_norm": 0.181640625, "learning_rate": 6.496144902704687e-05, "loss": 0.989, "step": 5618 }, { "epoch": 2.0384580498866214, "grad_norm": 0.1787109375, "learning_rate": 6.491249541059847e-05, "loss": 0.9792, "step": 5620 }, { "epoch": 2.039183673469388, "grad_norm": 0.169921875, "learning_rate": 6.486354179415005e-05, "loss": 1.0176, "step": 5622 }, { "epoch": 2.039909297052154, "grad_norm": 0.18359375, "learning_rate": 6.481458817770163e-05, "loss": 0.9985, "step": 5624 }, { "epoch": 2.0406349206349206, "grad_norm": 0.1689453125, "learning_rate": 6.476563456125321e-05, "loss": 0.9782, "step": 5626 }, { "epoch": 2.041360544217687, "grad_norm": 0.158203125, "learning_rate": 6.47166809448048e-05, "loss": 0.9238, "step": 5628 }, { "epoch": 2.0420861678004534, "grad_norm": 0.162109375, "learning_rate": 6.466772732835639e-05, "loss": 1.0303, "step": 5630 }, { "epoch": 2.0428117913832198, "grad_norm": 0.1767578125, "learning_rate": 6.461877371190796e-05, "loss": 1.0011, "step": 5632 }, { "epoch": 2.043537414965986, "grad_norm": 0.162109375, "learning_rate": 6.456982009545956e-05, "loss": 0.981, "step": 5634 }, { "epoch": 2.044263038548753, "grad_norm": 0.19140625, "learning_rate": 6.452086647901114e-05, "loss": 0.9464, "step": 5636 }, { "epoch": 2.0449886621315194, "grad_norm": 0.1669921875, "learning_rate": 6.447191286256272e-05, "loss": 1.0066, "step": 5638 }, { "epoch": 2.045714285714286, "grad_norm": 0.16015625, "learning_rate": 6.442295924611432e-05, "loss": 0.9959, "step": 5640 }, { "epoch": 2.046439909297052, "grad_norm": 0.16796875, "learning_rate": 6.437400562966589e-05, "loss": 0.942, "step": 5642 }, { "epoch": 2.0471655328798186, "grad_norm": 0.1630859375, "learning_rate": 6.432505201321748e-05, "loss": 0.9856, "step": 5644 }, { "epoch": 2.047891156462585, "grad_norm": 0.1962890625, "learning_rate": 6.427609839676906e-05, "loss": 0.9828, "step": 5646 }, { "epoch": 2.0486167800453514, "grad_norm": 0.1728515625, "learning_rate": 6.422714478032065e-05, "loss": 0.9923, "step": 5648 }, { "epoch": 2.049342403628118, "grad_norm": 0.1748046875, "learning_rate": 6.417819116387224e-05, "loss": 1.0303, "step": 5650 }, { "epoch": 2.050068027210884, "grad_norm": 0.1689453125, "learning_rate": 6.412923754742381e-05, "loss": 1.0702, "step": 5652 }, { "epoch": 2.0507936507936506, "grad_norm": 0.21875, "learning_rate": 6.408028393097541e-05, "loss": 1.0109, "step": 5654 }, { "epoch": 2.0515192743764175, "grad_norm": 0.1826171875, "learning_rate": 6.403133031452699e-05, "loss": 0.9457, "step": 5656 }, { "epoch": 2.052244897959184, "grad_norm": 0.1630859375, "learning_rate": 6.398237669807857e-05, "loss": 1.0746, "step": 5658 }, { "epoch": 2.0529705215419503, "grad_norm": 0.16796875, "learning_rate": 6.393342308163017e-05, "loss": 0.9641, "step": 5660 }, { "epoch": 2.0536961451247167, "grad_norm": 0.1669921875, "learning_rate": 6.388446946518174e-05, "loss": 1.0007, "step": 5662 }, { "epoch": 2.054421768707483, "grad_norm": 0.203125, "learning_rate": 6.383551584873333e-05, "loss": 0.9972, "step": 5664 }, { "epoch": 2.0551473922902495, "grad_norm": 0.1787109375, "learning_rate": 6.378656223228491e-05, "loss": 1.0813, "step": 5666 }, { "epoch": 2.055873015873016, "grad_norm": 0.154296875, "learning_rate": 6.37376086158365e-05, "loss": 0.9023, "step": 5668 }, { "epoch": 2.0565986394557823, "grad_norm": 0.154296875, "learning_rate": 6.368865499938809e-05, "loss": 0.9559, "step": 5670 }, { "epoch": 2.0573242630385487, "grad_norm": 0.150390625, "learning_rate": 6.363970138293966e-05, "loss": 0.9024, "step": 5672 }, { "epoch": 2.058049886621315, "grad_norm": 0.1494140625, "learning_rate": 6.359074776649126e-05, "loss": 0.9186, "step": 5674 }, { "epoch": 2.0587755102040814, "grad_norm": 0.1767578125, "learning_rate": 6.354179415004284e-05, "loss": 1.0022, "step": 5676 }, { "epoch": 2.0595011337868483, "grad_norm": 0.1865234375, "learning_rate": 6.349284053359442e-05, "loss": 1.0076, "step": 5678 }, { "epoch": 2.0602267573696147, "grad_norm": 0.166015625, "learning_rate": 6.344388691714602e-05, "loss": 0.9731, "step": 5680 }, { "epoch": 2.060952380952381, "grad_norm": 0.166015625, "learning_rate": 6.339493330069759e-05, "loss": 0.9956, "step": 5682 }, { "epoch": 2.0616780045351475, "grad_norm": 0.18359375, "learning_rate": 6.334597968424918e-05, "loss": 0.918, "step": 5684 }, { "epoch": 2.062403628117914, "grad_norm": 0.1689453125, "learning_rate": 6.329702606780076e-05, "loss": 0.9229, "step": 5686 }, { "epoch": 2.0631292517006803, "grad_norm": 0.15234375, "learning_rate": 6.324807245135235e-05, "loss": 0.9616, "step": 5688 }, { "epoch": 2.0638548752834467, "grad_norm": 0.177734375, "learning_rate": 6.319911883490394e-05, "loss": 1.0174, "step": 5690 }, { "epoch": 2.064580498866213, "grad_norm": 0.1533203125, "learning_rate": 6.315016521845551e-05, "loss": 1.0449, "step": 5692 }, { "epoch": 2.0653061224489795, "grad_norm": 0.1787109375, "learning_rate": 6.310121160200711e-05, "loss": 0.9916, "step": 5694 }, { "epoch": 2.066031746031746, "grad_norm": 0.166015625, "learning_rate": 6.305225798555869e-05, "loss": 0.9486, "step": 5696 }, { "epoch": 2.0667573696145123, "grad_norm": 0.1494140625, "learning_rate": 6.300330436911027e-05, "loss": 0.9569, "step": 5698 }, { "epoch": 2.067482993197279, "grad_norm": 0.1640625, "learning_rate": 6.295435075266187e-05, "loss": 0.9016, "step": 5700 }, { "epoch": 2.067482993197279, "eval_loss": 1.0106786489486694, "eval_runtime": 89.0366, "eval_samples_per_second": 20.216, "eval_steps_per_second": 0.64, "step": 5700 }, { "epoch": 2.0682086167800455, "grad_norm": 0.1640625, "learning_rate": 6.290539713621344e-05, "loss": 0.9435, "step": 5702 }, { "epoch": 2.068934240362812, "grad_norm": 0.1904296875, "learning_rate": 6.285644351976503e-05, "loss": 1.0447, "step": 5704 }, { "epoch": 2.0696598639455783, "grad_norm": 0.1689453125, "learning_rate": 6.280748990331661e-05, "loss": 0.9554, "step": 5706 }, { "epoch": 2.0703854875283447, "grad_norm": 0.162109375, "learning_rate": 6.27585362868682e-05, "loss": 1.03, "step": 5708 }, { "epoch": 2.071111111111111, "grad_norm": 0.18359375, "learning_rate": 6.270958267041978e-05, "loss": 0.9301, "step": 5710 }, { "epoch": 2.0718367346938775, "grad_norm": 0.1787109375, "learning_rate": 6.266062905397136e-05, "loss": 1.0046, "step": 5712 }, { "epoch": 2.072562358276644, "grad_norm": 0.185546875, "learning_rate": 6.261167543752296e-05, "loss": 1.0349, "step": 5714 }, { "epoch": 2.0732879818594103, "grad_norm": 0.171875, "learning_rate": 6.256272182107454e-05, "loss": 1.0612, "step": 5716 }, { "epoch": 2.0740136054421767, "grad_norm": 0.1591796875, "learning_rate": 6.251376820462612e-05, "loss": 0.9338, "step": 5718 }, { "epoch": 2.074739229024943, "grad_norm": 0.1826171875, "learning_rate": 6.24648145881777e-05, "loss": 1.0275, "step": 5720 }, { "epoch": 2.07546485260771, "grad_norm": 0.1552734375, "learning_rate": 6.241586097172929e-05, "loss": 0.9196, "step": 5722 }, { "epoch": 2.0761904761904764, "grad_norm": 0.150390625, "learning_rate": 6.236690735528088e-05, "loss": 0.9807, "step": 5724 }, { "epoch": 2.0769160997732428, "grad_norm": 0.171875, "learning_rate": 6.231795373883246e-05, "loss": 0.9343, "step": 5726 }, { "epoch": 2.077641723356009, "grad_norm": 0.181640625, "learning_rate": 6.226900012238405e-05, "loss": 0.9618, "step": 5728 }, { "epoch": 2.0783673469387756, "grad_norm": 0.18359375, "learning_rate": 6.222004650593563e-05, "loss": 0.9586, "step": 5730 }, { "epoch": 2.079092970521542, "grad_norm": 0.1640625, "learning_rate": 6.217109288948721e-05, "loss": 1.0001, "step": 5732 }, { "epoch": 2.0798185941043084, "grad_norm": 0.18359375, "learning_rate": 6.21221392730388e-05, "loss": 1.0526, "step": 5734 }, { "epoch": 2.0805442176870748, "grad_norm": 0.1806640625, "learning_rate": 6.207318565659039e-05, "loss": 0.9504, "step": 5736 }, { "epoch": 2.081269841269841, "grad_norm": 0.150390625, "learning_rate": 6.202423204014197e-05, "loss": 0.923, "step": 5738 }, { "epoch": 2.0819954648526076, "grad_norm": 0.1845703125, "learning_rate": 6.197527842369355e-05, "loss": 0.9957, "step": 5740 }, { "epoch": 2.082721088435374, "grad_norm": 0.1728515625, "learning_rate": 6.192632480724514e-05, "loss": 0.8884, "step": 5742 }, { "epoch": 2.083446712018141, "grad_norm": 0.162109375, "learning_rate": 6.187737119079672e-05, "loss": 0.9332, "step": 5744 }, { "epoch": 2.084172335600907, "grad_norm": 0.1904296875, "learning_rate": 6.182841757434831e-05, "loss": 0.9743, "step": 5746 }, { "epoch": 2.0848979591836736, "grad_norm": 0.1669921875, "learning_rate": 6.17794639578999e-05, "loss": 0.9578, "step": 5748 }, { "epoch": 2.08562358276644, "grad_norm": 0.1865234375, "learning_rate": 6.173051034145148e-05, "loss": 0.9793, "step": 5750 }, { "epoch": 2.0863492063492064, "grad_norm": 0.162109375, "learning_rate": 6.168155672500306e-05, "loss": 0.9264, "step": 5752 }, { "epoch": 2.087074829931973, "grad_norm": 0.171875, "learning_rate": 6.163260310855464e-05, "loss": 0.92, "step": 5754 }, { "epoch": 2.087800453514739, "grad_norm": 0.1806640625, "learning_rate": 6.158364949210624e-05, "loss": 1.0329, "step": 5756 }, { "epoch": 2.0885260770975056, "grad_norm": 0.1796875, "learning_rate": 6.153469587565782e-05, "loss": 0.9942, "step": 5758 }, { "epoch": 2.089251700680272, "grad_norm": 0.1494140625, "learning_rate": 6.14857422592094e-05, "loss": 0.9291, "step": 5760 }, { "epoch": 2.0899773242630384, "grad_norm": 0.150390625, "learning_rate": 6.143678864276099e-05, "loss": 0.9504, "step": 5762 }, { "epoch": 2.090702947845805, "grad_norm": 0.1728515625, "learning_rate": 6.138783502631257e-05, "loss": 0.9671, "step": 5764 }, { "epoch": 2.0914285714285716, "grad_norm": 0.20703125, "learning_rate": 6.133888140986416e-05, "loss": 1.0002, "step": 5766 }, { "epoch": 2.092154195011338, "grad_norm": 0.1591796875, "learning_rate": 6.128992779341575e-05, "loss": 0.9597, "step": 5768 }, { "epoch": 2.0928798185941044, "grad_norm": 0.1884765625, "learning_rate": 6.124097417696733e-05, "loss": 1.0265, "step": 5770 }, { "epoch": 2.093605442176871, "grad_norm": 0.1796875, "learning_rate": 6.119202056051891e-05, "loss": 1.035, "step": 5772 }, { "epoch": 2.0943310657596372, "grad_norm": 0.166015625, "learning_rate": 6.114306694407049e-05, "loss": 0.9861, "step": 5774 }, { "epoch": 2.0950566893424036, "grad_norm": 0.173828125, "learning_rate": 6.109411332762209e-05, "loss": 1.011, "step": 5776 }, { "epoch": 2.09578231292517, "grad_norm": 0.162109375, "learning_rate": 6.104515971117366e-05, "loss": 0.9615, "step": 5778 }, { "epoch": 2.0965079365079364, "grad_norm": 0.1650390625, "learning_rate": 6.099620609472525e-05, "loss": 1.0444, "step": 5780 }, { "epoch": 2.097233560090703, "grad_norm": 0.17578125, "learning_rate": 6.094725247827684e-05, "loss": 0.9458, "step": 5782 }, { "epoch": 2.0979591836734692, "grad_norm": 0.1533203125, "learning_rate": 6.089829886182842e-05, "loss": 0.9296, "step": 5784 }, { "epoch": 2.0986848072562356, "grad_norm": 0.1630859375, "learning_rate": 6.084934524538001e-05, "loss": 0.8853, "step": 5786 }, { "epoch": 2.0994104308390025, "grad_norm": 0.2177734375, "learning_rate": 6.080039162893158e-05, "loss": 0.9348, "step": 5788 }, { "epoch": 2.100136054421769, "grad_norm": 0.1611328125, "learning_rate": 6.075143801248318e-05, "loss": 0.934, "step": 5790 }, { "epoch": 2.1008616780045353, "grad_norm": 0.1875, "learning_rate": 6.070248439603477e-05, "loss": 0.9812, "step": 5792 }, { "epoch": 2.1015873015873017, "grad_norm": 0.1806640625, "learning_rate": 6.065353077958634e-05, "loss": 0.9687, "step": 5794 }, { "epoch": 2.102312925170068, "grad_norm": 0.150390625, "learning_rate": 6.060457716313793e-05, "loss": 0.9097, "step": 5796 }, { "epoch": 2.1030385487528345, "grad_norm": 0.166015625, "learning_rate": 6.055562354668951e-05, "loss": 0.9202, "step": 5798 }, { "epoch": 2.103764172335601, "grad_norm": 0.17578125, "learning_rate": 6.05066699302411e-05, "loss": 0.9264, "step": 5800 }, { "epoch": 2.103764172335601, "eval_loss": 1.010640025138855, "eval_runtime": 86.5952, "eval_samples_per_second": 20.786, "eval_steps_per_second": 0.658, "step": 5800 }, { "epoch": 2.1044897959183673, "grad_norm": 0.1689453125, "learning_rate": 6.045771631379269e-05, "loss": 0.9798, "step": 5802 }, { "epoch": 2.1052154195011337, "grad_norm": 0.1640625, "learning_rate": 6.040876269734427e-05, "loss": 0.953, "step": 5804 }, { "epoch": 2.1059410430839, "grad_norm": 0.169921875, "learning_rate": 6.035980908089586e-05, "loss": 0.9906, "step": 5806 }, { "epoch": 2.1066666666666665, "grad_norm": 0.197265625, "learning_rate": 6.031085546444743e-05, "loss": 0.9762, "step": 5808 }, { "epoch": 2.1073922902494333, "grad_norm": 0.166015625, "learning_rate": 6.026190184799903e-05, "loss": 0.9851, "step": 5810 }, { "epoch": 2.1081179138321997, "grad_norm": 0.171875, "learning_rate": 6.0212948231550604e-05, "loss": 0.9887, "step": 5812 }, { "epoch": 2.108843537414966, "grad_norm": 0.169921875, "learning_rate": 6.016399461510219e-05, "loss": 0.9786, "step": 5814 }, { "epoch": 2.1095691609977325, "grad_norm": 0.1650390625, "learning_rate": 6.011504099865378e-05, "loss": 0.952, "step": 5816 }, { "epoch": 2.110294784580499, "grad_norm": 0.1728515625, "learning_rate": 6.006608738220536e-05, "loss": 1.0122, "step": 5818 }, { "epoch": 2.1110204081632653, "grad_norm": 0.15625, "learning_rate": 6.001713376575695e-05, "loss": 0.9715, "step": 5820 }, { "epoch": 2.1117460317460317, "grad_norm": 0.1806640625, "learning_rate": 5.996818014930853e-05, "loss": 1.005, "step": 5822 }, { "epoch": 2.112471655328798, "grad_norm": 0.1875, "learning_rate": 5.991922653286012e-05, "loss": 1.021, "step": 5824 }, { "epoch": 2.1131972789115645, "grad_norm": 0.1591796875, "learning_rate": 5.987027291641171e-05, "loss": 0.9739, "step": 5826 }, { "epoch": 2.113922902494331, "grad_norm": 0.173828125, "learning_rate": 5.982131929996328e-05, "loss": 0.9251, "step": 5828 }, { "epoch": 2.1146485260770973, "grad_norm": 0.1640625, "learning_rate": 5.977236568351488e-05, "loss": 1.0101, "step": 5830 }, { "epoch": 2.115374149659864, "grad_norm": 0.1728515625, "learning_rate": 5.9723412067066454e-05, "loss": 0.9652, "step": 5832 }, { "epoch": 2.1160997732426305, "grad_norm": 0.16015625, "learning_rate": 5.967445845061804e-05, "loss": 0.9139, "step": 5834 }, { "epoch": 2.116825396825397, "grad_norm": 0.158203125, "learning_rate": 5.962550483416963e-05, "loss": 1.0025, "step": 5836 }, { "epoch": 2.1175510204081633, "grad_norm": 0.1708984375, "learning_rate": 5.957655121772121e-05, "loss": 0.973, "step": 5838 }, { "epoch": 2.1182766439909297, "grad_norm": 0.1484375, "learning_rate": 5.95275976012728e-05, "loss": 0.918, "step": 5840 }, { "epoch": 2.119002267573696, "grad_norm": 0.1572265625, "learning_rate": 5.947864398482438e-05, "loss": 0.9193, "step": 5842 }, { "epoch": 2.1197278911564625, "grad_norm": 0.177734375, "learning_rate": 5.942969036837597e-05, "loss": 0.949, "step": 5844 }, { "epoch": 2.120453514739229, "grad_norm": 0.1796875, "learning_rate": 5.938073675192756e-05, "loss": 0.9686, "step": 5846 }, { "epoch": 2.1211791383219953, "grad_norm": 0.1796875, "learning_rate": 5.933178313547913e-05, "loss": 1.0197, "step": 5848 }, { "epoch": 2.1219047619047617, "grad_norm": 0.1708984375, "learning_rate": 5.928282951903073e-05, "loss": 0.9537, "step": 5850 }, { "epoch": 2.122630385487528, "grad_norm": 0.1669921875, "learning_rate": 5.9233875902582304e-05, "loss": 0.9571, "step": 5852 }, { "epoch": 2.123356009070295, "grad_norm": 0.1748046875, "learning_rate": 5.918492228613389e-05, "loss": 0.9315, "step": 5854 }, { "epoch": 2.1240816326530614, "grad_norm": 0.1689453125, "learning_rate": 5.913596866968547e-05, "loss": 0.9558, "step": 5856 }, { "epoch": 2.124807256235828, "grad_norm": 0.16015625, "learning_rate": 5.908701505323706e-05, "loss": 0.9256, "step": 5858 }, { "epoch": 2.125532879818594, "grad_norm": 0.21875, "learning_rate": 5.903806143678865e-05, "loss": 0.9814, "step": 5860 }, { "epoch": 2.1262585034013606, "grad_norm": 0.1982421875, "learning_rate": 5.898910782034023e-05, "loss": 1.0241, "step": 5862 }, { "epoch": 2.126984126984127, "grad_norm": 0.16796875, "learning_rate": 5.894015420389182e-05, "loss": 0.9556, "step": 5864 }, { "epoch": 2.1277097505668934, "grad_norm": 0.1748046875, "learning_rate": 5.889120058744339e-05, "loss": 1.0065, "step": 5866 }, { "epoch": 2.1284353741496598, "grad_norm": 0.1708984375, "learning_rate": 5.884224697099498e-05, "loss": 0.9229, "step": 5868 }, { "epoch": 2.129160997732426, "grad_norm": 0.1591796875, "learning_rate": 5.879329335454658e-05, "loss": 0.8913, "step": 5870 }, { "epoch": 2.1298866213151926, "grad_norm": 0.193359375, "learning_rate": 5.8744339738098153e-05, "loss": 0.9706, "step": 5872 }, { "epoch": 2.130612244897959, "grad_norm": 0.1796875, "learning_rate": 5.869538612164974e-05, "loss": 0.9607, "step": 5874 }, { "epoch": 2.131337868480726, "grad_norm": 0.1728515625, "learning_rate": 5.864643250520132e-05, "loss": 0.9314, "step": 5876 }, { "epoch": 2.132063492063492, "grad_norm": 0.173828125, "learning_rate": 5.859747888875291e-05, "loss": 0.97, "step": 5878 }, { "epoch": 2.1327891156462586, "grad_norm": 0.1650390625, "learning_rate": 5.85485252723045e-05, "loss": 0.9887, "step": 5880 }, { "epoch": 2.133514739229025, "grad_norm": 0.162109375, "learning_rate": 5.849957165585608e-05, "loss": 1.013, "step": 5882 }, { "epoch": 2.1342403628117914, "grad_norm": 0.15625, "learning_rate": 5.845061803940767e-05, "loss": 0.9536, "step": 5884 }, { "epoch": 2.134965986394558, "grad_norm": 0.166015625, "learning_rate": 5.840166442295924e-05, "loss": 0.9715, "step": 5886 }, { "epoch": 2.135691609977324, "grad_norm": 0.15234375, "learning_rate": 5.835271080651083e-05, "loss": 1.0258, "step": 5888 }, { "epoch": 2.1364172335600906, "grad_norm": 0.1689453125, "learning_rate": 5.8303757190062414e-05, "loss": 0.9495, "step": 5890 }, { "epoch": 2.137142857142857, "grad_norm": 0.166015625, "learning_rate": 5.8254803573614e-05, "loss": 0.9514, "step": 5892 }, { "epoch": 2.1378684807256234, "grad_norm": 0.181640625, "learning_rate": 5.820584995716559e-05, "loss": 1.001, "step": 5894 }, { "epoch": 2.13859410430839, "grad_norm": 0.177734375, "learning_rate": 5.815689634071717e-05, "loss": 0.9775, "step": 5896 }, { "epoch": 2.1393197278911567, "grad_norm": 0.1826171875, "learning_rate": 5.810794272426876e-05, "loss": 0.9368, "step": 5898 }, { "epoch": 2.140045351473923, "grad_norm": 0.1650390625, "learning_rate": 5.805898910782034e-05, "loss": 0.9397, "step": 5900 }, { "epoch": 2.140045351473923, "eval_loss": 1.0105234384536743, "eval_runtime": 86.09, "eval_samples_per_second": 20.908, "eval_steps_per_second": 0.662, "step": 5900 }, { "epoch": 2.1407709750566895, "grad_norm": 0.1787109375, "learning_rate": 5.801003549137193e-05, "loss": 1.0924, "step": 5902 }, { "epoch": 2.141496598639456, "grad_norm": 0.1796875, "learning_rate": 5.796108187492352e-05, "loss": 0.9779, "step": 5904 }, { "epoch": 2.1422222222222222, "grad_norm": 0.169921875, "learning_rate": 5.791212825847509e-05, "loss": 0.9686, "step": 5906 }, { "epoch": 2.1429478458049886, "grad_norm": 0.1728515625, "learning_rate": 5.786317464202668e-05, "loss": 0.9359, "step": 5908 }, { "epoch": 2.143673469387755, "grad_norm": 0.185546875, "learning_rate": 5.7814221025578264e-05, "loss": 0.982, "step": 5910 }, { "epoch": 2.1443990929705214, "grad_norm": 0.1787109375, "learning_rate": 5.776526740912985e-05, "loss": 0.887, "step": 5912 }, { "epoch": 2.145124716553288, "grad_norm": 0.1611328125, "learning_rate": 5.771631379268144e-05, "loss": 0.9527, "step": 5914 }, { "epoch": 2.1458503401360542, "grad_norm": 0.1875, "learning_rate": 5.766736017623302e-05, "loss": 0.9901, "step": 5916 }, { "epoch": 2.1465759637188206, "grad_norm": 0.169921875, "learning_rate": 5.761840655978461e-05, "loss": 1.0351, "step": 5918 }, { "epoch": 2.1473015873015875, "grad_norm": 0.16796875, "learning_rate": 5.756945294333619e-05, "loss": 1.0367, "step": 5920 }, { "epoch": 2.148027210884354, "grad_norm": 0.1796875, "learning_rate": 5.752049932688778e-05, "loss": 0.9757, "step": 5922 }, { "epoch": 2.1487528344671203, "grad_norm": 0.1708984375, "learning_rate": 5.7471545710439354e-05, "loss": 0.9131, "step": 5924 }, { "epoch": 2.1494784580498867, "grad_norm": 0.1650390625, "learning_rate": 5.742259209399094e-05, "loss": 0.9754, "step": 5926 }, { "epoch": 2.150204081632653, "grad_norm": 0.1640625, "learning_rate": 5.737363847754253e-05, "loss": 1.0091, "step": 5928 }, { "epoch": 2.1509297052154195, "grad_norm": 0.197265625, "learning_rate": 5.7324684861094114e-05, "loss": 1.0215, "step": 5930 }, { "epoch": 2.151655328798186, "grad_norm": 0.1826171875, "learning_rate": 5.72757312446457e-05, "loss": 0.9583, "step": 5932 }, { "epoch": 2.1523809523809523, "grad_norm": 0.1796875, "learning_rate": 5.722677762819728e-05, "loss": 1.0407, "step": 5934 }, { "epoch": 2.1531065759637187, "grad_norm": 0.1484375, "learning_rate": 5.717782401174887e-05, "loss": 0.8784, "step": 5936 }, { "epoch": 2.153832199546485, "grad_norm": 0.1962890625, "learning_rate": 5.712887039530046e-05, "loss": 0.9976, "step": 5938 }, { "epoch": 2.1545578231292515, "grad_norm": 0.1923828125, "learning_rate": 5.707991677885204e-05, "loss": 0.9972, "step": 5940 }, { "epoch": 2.1552834467120183, "grad_norm": 0.1650390625, "learning_rate": 5.703096316240363e-05, "loss": 0.9872, "step": 5942 }, { "epoch": 2.1560090702947847, "grad_norm": 0.169921875, "learning_rate": 5.6982009545955204e-05, "loss": 0.9176, "step": 5944 }, { "epoch": 2.156734693877551, "grad_norm": 0.1591796875, "learning_rate": 5.693305592950679e-05, "loss": 0.9286, "step": 5946 }, { "epoch": 2.1574603174603175, "grad_norm": 0.16015625, "learning_rate": 5.688410231305838e-05, "loss": 0.9204, "step": 5948 }, { "epoch": 2.158185941043084, "grad_norm": 0.1591796875, "learning_rate": 5.6835148696609964e-05, "loss": 0.9486, "step": 5950 }, { "epoch": 2.1589115646258503, "grad_norm": 0.158203125, "learning_rate": 5.678619508016155e-05, "loss": 1.0603, "step": 5952 }, { "epoch": 2.1596371882086167, "grad_norm": 0.1748046875, "learning_rate": 5.673724146371313e-05, "loss": 0.9545, "step": 5954 }, { "epoch": 2.160362811791383, "grad_norm": 0.1884765625, "learning_rate": 5.668828784726472e-05, "loss": 1.0123, "step": 5956 }, { "epoch": 2.1610884353741495, "grad_norm": 0.150390625, "learning_rate": 5.66393342308163e-05, "loss": 0.9411, "step": 5958 }, { "epoch": 2.161814058956916, "grad_norm": 0.15625, "learning_rate": 5.659038061436789e-05, "loss": 0.9927, "step": 5960 }, { "epoch": 2.1625396825396823, "grad_norm": 0.166015625, "learning_rate": 5.654142699791948e-05, "loss": 0.9126, "step": 5962 }, { "epoch": 2.163265306122449, "grad_norm": 0.193359375, "learning_rate": 5.6492473381471054e-05, "loss": 0.9676, "step": 5964 }, { "epoch": 2.1639909297052156, "grad_norm": 0.1630859375, "learning_rate": 5.644351976502264e-05, "loss": 0.9483, "step": 5966 }, { "epoch": 2.164716553287982, "grad_norm": 0.1474609375, "learning_rate": 5.6394566148574225e-05, "loss": 0.9706, "step": 5968 }, { "epoch": 2.1654421768707484, "grad_norm": 0.162109375, "learning_rate": 5.6345612532125814e-05, "loss": 0.978, "step": 5970 }, { "epoch": 2.1661678004535148, "grad_norm": 0.19140625, "learning_rate": 5.62966589156774e-05, "loss": 0.9953, "step": 5972 }, { "epoch": 2.166893424036281, "grad_norm": 0.1650390625, "learning_rate": 5.624770529922898e-05, "loss": 0.9638, "step": 5974 }, { "epoch": 2.1676190476190476, "grad_norm": 0.1787109375, "learning_rate": 5.619875168278057e-05, "loss": 1.0009, "step": 5976 }, { "epoch": 2.168344671201814, "grad_norm": 0.1875, "learning_rate": 5.614979806633215e-05, "loss": 1.0743, "step": 5978 }, { "epoch": 2.1690702947845804, "grad_norm": 0.1640625, "learning_rate": 5.610084444988374e-05, "loss": 0.9951, "step": 5980 }, { "epoch": 2.1697959183673468, "grad_norm": 0.1826171875, "learning_rate": 5.605189083343533e-05, "loss": 0.9734, "step": 5982 }, { "epoch": 2.170521541950113, "grad_norm": 0.18359375, "learning_rate": 5.6002937216986903e-05, "loss": 1.0389, "step": 5984 }, { "epoch": 2.17124716553288, "grad_norm": 0.1845703125, "learning_rate": 5.595398360053849e-05, "loss": 1.0822, "step": 5986 }, { "epoch": 2.1719727891156464, "grad_norm": 0.1767578125, "learning_rate": 5.5905029984090075e-05, "loss": 0.9547, "step": 5988 }, { "epoch": 2.172698412698413, "grad_norm": 0.15625, "learning_rate": 5.5856076367641664e-05, "loss": 0.9614, "step": 5990 }, { "epoch": 2.173424036281179, "grad_norm": 0.177734375, "learning_rate": 5.580712275119325e-05, "loss": 0.9657, "step": 5992 }, { "epoch": 2.1741496598639456, "grad_norm": 0.177734375, "learning_rate": 5.575816913474483e-05, "loss": 0.9888, "step": 5994 }, { "epoch": 2.174875283446712, "grad_norm": 0.1572265625, "learning_rate": 5.570921551829642e-05, "loss": 0.9204, "step": 5996 }, { "epoch": 2.1756009070294784, "grad_norm": 0.1748046875, "learning_rate": 5.5660261901848e-05, "loss": 1.0059, "step": 5998 }, { "epoch": 2.176326530612245, "grad_norm": 0.1640625, "learning_rate": 5.561130828539959e-05, "loss": 1.0289, "step": 6000 }, { "epoch": 2.176326530612245, "eval_loss": 1.0105350017547607, "eval_runtime": 89.1474, "eval_samples_per_second": 20.191, "eval_steps_per_second": 0.639, "step": 6000 } ], "logging_steps": 2, "max_steps": 8271, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 100, "stateful_callbacks": { "EarlyStoppingCallback": { "args": { "early_stopping_patience": 3, "early_stopping_threshold": 0.0 }, "attributes": { "early_stopping_patience_counter": 1 } }, "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.0187451388598944e+19, "train_batch_size": 2, "trial_name": null, "trial_params": null }