{ "best_metric": null, "best_model_checkpoint": null, "epoch": 4.0, "eval_steps": 798, "global_step": 12760, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00031347962382445143, "grad_norm": 4.473039150238037, "learning_rate": 5e-10, "loss": 8.3121, "step": 1 }, { "epoch": 0.00031347962382445143, "eval_loss": 28.89470672607422, "eval_runtime": 20.3862, "eval_samples_per_second": 131.805, "eval_steps_per_second": 8.241, "step": 1 }, { "epoch": 0.0006269592476489029, "grad_norm": 3.657515287399292, "learning_rate": 1e-09, "loss": 7.0439, "step": 2 }, { "epoch": 0.0009404388714733542, "grad_norm": 2.806065559387207, "learning_rate": 1.5e-09, "loss": 6.6257, "step": 3 }, { "epoch": 0.0012539184952978057, "grad_norm": 2.810598611831665, "learning_rate": 2e-09, "loss": 6.5132, "step": 4 }, { "epoch": 0.001567398119122257, "grad_norm": 4.099343299865723, "learning_rate": 2.5e-09, "loss": 8.1055, "step": 5 }, { "epoch": 0.0018808777429467085, "grad_norm": 3.389047384262085, "learning_rate": 3e-09, "loss": 8.7388, "step": 6 }, { "epoch": 0.00219435736677116, "grad_norm": 3.4108972549438477, "learning_rate": 3.5000000000000003e-09, "loss": 7.8426, "step": 7 }, { "epoch": 0.0025078369905956114, "grad_norm": 3.3495447635650635, "learning_rate": 4e-09, "loss": 8.2811, "step": 8 }, { "epoch": 0.0028213166144200625, "grad_norm": 3.485499143600464, "learning_rate": 4.500000000000001e-09, "loss": 7.3836, "step": 9 }, { "epoch": 0.003134796238244514, "grad_norm": 4.053262233734131, "learning_rate": 5e-09, "loss": 9.1597, "step": 10 }, { "epoch": 0.0034482758620689655, "grad_norm": 4.419150352478027, "learning_rate": 5.5000000000000004e-09, "loss": 12.4238, "step": 11 }, { "epoch": 0.003761755485893417, "grad_norm": 3.277268409729004, "learning_rate": 6e-09, "loss": 8.4173, "step": 12 }, { "epoch": 0.0040752351097178684, "grad_norm": 3.8386521339416504, "learning_rate": 6.5e-09, "loss": 8.2201, "step": 13 }, { "epoch": 0.00438871473354232, "grad_norm": 4.788142681121826, "learning_rate": 7.000000000000001e-09, "loss": 11.7976, "step": 14 }, { "epoch": 0.004702194357366771, "grad_norm": 3.0357067584991455, "learning_rate": 7.500000000000001e-09, "loss": 8.5286, "step": 15 }, { "epoch": 0.005015673981191223, "grad_norm": 3.4486613273620605, "learning_rate": 8e-09, "loss": 7.839, "step": 16 }, { "epoch": 0.005329153605015674, "grad_norm": 4.829311370849609, "learning_rate": 8.5e-09, "loss": 9.6776, "step": 17 }, { "epoch": 0.005642633228840125, "grad_norm": 4.219234466552734, "learning_rate": 9.000000000000001e-09, "loss": 7.8809, "step": 18 }, { "epoch": 0.0059561128526645765, "grad_norm": 3.634462833404541, "learning_rate": 9.5e-09, "loss": 9.879, "step": 19 }, { "epoch": 0.006269592476489028, "grad_norm": 3.924381732940674, "learning_rate": 1e-08, "loss": 9.0212, "step": 20 }, { "epoch": 0.0065830721003134795, "grad_norm": 3.562445640563965, "learning_rate": 1.05e-08, "loss": 8.593, "step": 21 }, { "epoch": 0.006896551724137931, "grad_norm": 2.989351749420166, "learning_rate": 1.1000000000000001e-08, "loss": 6.3834, "step": 22 }, { "epoch": 0.007210031347962382, "grad_norm": 3.755312919616699, "learning_rate": 1.15e-08, "loss": 9.2207, "step": 23 }, { "epoch": 0.007523510971786834, "grad_norm": 3.771742582321167, "learning_rate": 1.2e-08, "loss": 9.2504, "step": 24 }, { "epoch": 0.007836990595611285, "grad_norm": 2.664695978164673, "learning_rate": 1.2500000000000001e-08, "loss": 7.0286, "step": 25 }, { "epoch": 0.008150470219435737, "grad_norm": 2.916489601135254, "learning_rate": 1.3e-08, "loss": 7.5708, "step": 26 }, { "epoch": 0.008463949843260187, "grad_norm": 3.4173834323883057, "learning_rate": 1.3500000000000002e-08, "loss": 7.0245, "step": 27 }, { "epoch": 0.00877742946708464, "grad_norm": 4.134421348571777, "learning_rate": 1.4000000000000001e-08, "loss": 8.218, "step": 28 }, { "epoch": 0.00909090909090909, "grad_norm": 2.478123188018799, "learning_rate": 1.45e-08, "loss": 6.6974, "step": 29 }, { "epoch": 0.009404388714733543, "grad_norm": 4.880210876464844, "learning_rate": 1.5000000000000002e-08, "loss": 13.0694, "step": 30 }, { "epoch": 0.009717868338557993, "grad_norm": 3.532501220703125, "learning_rate": 1.55e-08, "loss": 6.8898, "step": 31 }, { "epoch": 0.010031347962382446, "grad_norm": 3.8485097885131836, "learning_rate": 1.6e-08, "loss": 10.3871, "step": 32 }, { "epoch": 0.010344827586206896, "grad_norm": 4.742996692657471, "learning_rate": 1.6500000000000002e-08, "loss": 10.987, "step": 33 }, { "epoch": 0.010658307210031349, "grad_norm": 3.595039129257202, "learning_rate": 1.7e-08, "loss": 9.6221, "step": 34 }, { "epoch": 0.0109717868338558, "grad_norm": 3.5569448471069336, "learning_rate": 1.75e-08, "loss": 6.5841, "step": 35 }, { "epoch": 0.01128526645768025, "grad_norm": 4.763426303863525, "learning_rate": 1.8000000000000002e-08, "loss": 7.6671, "step": 36 }, { "epoch": 0.011598746081504702, "grad_norm": 3.6167099475860596, "learning_rate": 1.8500000000000004e-08, "loss": 10.4234, "step": 37 }, { "epoch": 0.011912225705329153, "grad_norm": 4.075372219085693, "learning_rate": 1.9e-08, "loss": 11.068, "step": 38 }, { "epoch": 0.012225705329153605, "grad_norm": 6.240333080291748, "learning_rate": 1.95e-08, "loss": 17.4708, "step": 39 }, { "epoch": 0.012539184952978056, "grad_norm": 3.7709567546844482, "learning_rate": 2e-08, "loss": 8.2594, "step": 40 }, { "epoch": 0.012852664576802508, "grad_norm": 2.885645866394043, "learning_rate": 2.0500000000000005e-08, "loss": 6.7495, "step": 41 }, { "epoch": 0.013166144200626959, "grad_norm": 3.3022165298461914, "learning_rate": 2.1e-08, "loss": 7.0996, "step": 42 }, { "epoch": 0.013479623824451411, "grad_norm": 3.8264548778533936, "learning_rate": 2.15e-08, "loss": 8.166, "step": 43 }, { "epoch": 0.013793103448275862, "grad_norm": 4.580362796783447, "learning_rate": 2.2000000000000002e-08, "loss": 14.0935, "step": 44 }, { "epoch": 0.014106583072100314, "grad_norm": 4.131213665008545, "learning_rate": 2.25e-08, "loss": 12.5754, "step": 45 }, { "epoch": 0.014420062695924765, "grad_norm": 3.5261685848236084, "learning_rate": 2.3e-08, "loss": 7.9894, "step": 46 }, { "epoch": 0.014733542319749215, "grad_norm": 3.0584185123443604, "learning_rate": 2.3500000000000002e-08, "loss": 6.4833, "step": 47 }, { "epoch": 0.015047021943573668, "grad_norm": 3.39809513092041, "learning_rate": 2.4e-08, "loss": 8.1705, "step": 48 }, { "epoch": 0.015360501567398118, "grad_norm": 4.006636619567871, "learning_rate": 2.45e-08, "loss": 8.5717, "step": 49 }, { "epoch": 0.01567398119122257, "grad_norm": 5.995611190795898, "learning_rate": 2.5000000000000002e-08, "loss": 8.4349, "step": 50 }, { "epoch": 0.01598746081504702, "grad_norm": 3.6572229862213135, "learning_rate": 2.5500000000000003e-08, "loss": 7.9315, "step": 51 }, { "epoch": 0.016300940438871474, "grad_norm": 3.6359832286834717, "learning_rate": 2.6e-08, "loss": 8.9688, "step": 52 }, { "epoch": 0.016614420062695926, "grad_norm": 2.8677127361297607, "learning_rate": 2.6500000000000002e-08, "loss": 6.3906, "step": 53 }, { "epoch": 0.016927899686520375, "grad_norm": 2.7862868309020996, "learning_rate": 2.7000000000000004e-08, "loss": 7.7946, "step": 54 }, { "epoch": 0.017241379310344827, "grad_norm": 3.8618948459625244, "learning_rate": 2.75e-08, "loss": 7.8249, "step": 55 }, { "epoch": 0.01755485893416928, "grad_norm": 3.185887336730957, "learning_rate": 2.8000000000000003e-08, "loss": 7.2935, "step": 56 }, { "epoch": 0.017868338557993732, "grad_norm": 2.7501015663146973, "learning_rate": 2.8500000000000004e-08, "loss": 6.9817, "step": 57 }, { "epoch": 0.01818181818181818, "grad_norm": 3.1869773864746094, "learning_rate": 2.9e-08, "loss": 7.383, "step": 58 }, { "epoch": 0.018495297805642633, "grad_norm": 5.212038993835449, "learning_rate": 2.9500000000000003e-08, "loss": 14.1096, "step": 59 }, { "epoch": 0.018808777429467086, "grad_norm": 3.6909945011138916, "learning_rate": 3.0000000000000004e-08, "loss": 9.0418, "step": 60 }, { "epoch": 0.019122257053291535, "grad_norm": 4.0268120765686035, "learning_rate": 3.05e-08, "loss": 11.0808, "step": 61 }, { "epoch": 0.019435736677115987, "grad_norm": 4.155067443847656, "learning_rate": 3.1e-08, "loss": 10.4174, "step": 62 }, { "epoch": 0.01974921630094044, "grad_norm": 4.077535629272461, "learning_rate": 3.1500000000000004e-08, "loss": 8.0036, "step": 63 }, { "epoch": 0.02006269592476489, "grad_norm": 3.3879809379577637, "learning_rate": 3.2e-08, "loss": 7.2178, "step": 64 }, { "epoch": 0.02037617554858934, "grad_norm": 4.345617771148682, "learning_rate": 3.25e-08, "loss": 11.3773, "step": 65 }, { "epoch": 0.020689655172413793, "grad_norm": 3.596099853515625, "learning_rate": 3.3000000000000004e-08, "loss": 7.7855, "step": 66 }, { "epoch": 0.021003134796238245, "grad_norm": 2.798985004425049, "learning_rate": 3.35e-08, "loss": 6.8671, "step": 67 }, { "epoch": 0.021316614420062698, "grad_norm": 3.921792507171631, "learning_rate": 3.4e-08, "loss": 11.5327, "step": 68 }, { "epoch": 0.021630094043887146, "grad_norm": 3.9658515453338623, "learning_rate": 3.4500000000000005e-08, "loss": 7.4312, "step": 69 }, { "epoch": 0.0219435736677116, "grad_norm": 3.8455910682678223, "learning_rate": 3.5e-08, "loss": 11.7665, "step": 70 }, { "epoch": 0.02225705329153605, "grad_norm": 3.4687187671661377, "learning_rate": 3.550000000000001e-08, "loss": 9.1295, "step": 71 }, { "epoch": 0.0225705329153605, "grad_norm": 3.2690117359161377, "learning_rate": 3.6000000000000005e-08, "loss": 8.421, "step": 72 }, { "epoch": 0.022884012539184952, "grad_norm": 4.495233535766602, "learning_rate": 3.65e-08, "loss": 8.9433, "step": 73 }, { "epoch": 0.023197492163009405, "grad_norm": 3.4736380577087402, "learning_rate": 3.700000000000001e-08, "loss": 7.5205, "step": 74 }, { "epoch": 0.023510971786833857, "grad_norm": 3.021930456161499, "learning_rate": 3.7500000000000005e-08, "loss": 6.0582, "step": 75 }, { "epoch": 0.023824451410658306, "grad_norm": 3.8124265670776367, "learning_rate": 3.8e-08, "loss": 9.6513, "step": 76 }, { "epoch": 0.02413793103448276, "grad_norm": 2.6047370433807373, "learning_rate": 3.850000000000001e-08, "loss": 6.0856, "step": 77 }, { "epoch": 0.02445141065830721, "grad_norm": 4.412445545196533, "learning_rate": 3.9e-08, "loss": 9.3832, "step": 78 }, { "epoch": 0.024764890282131663, "grad_norm": 4.0916547775268555, "learning_rate": 3.950000000000001e-08, "loss": 10.2509, "step": 79 }, { "epoch": 0.025078369905956112, "grad_norm": 3.16266131401062, "learning_rate": 4e-08, "loss": 8.3115, "step": 80 }, { "epoch": 0.025391849529780564, "grad_norm": 3.2230985164642334, "learning_rate": 4.05e-08, "loss": 8.4255, "step": 81 }, { "epoch": 0.025705329153605017, "grad_norm": 4.53424072265625, "learning_rate": 4.100000000000001e-08, "loss": 9.7329, "step": 82 }, { "epoch": 0.026018808777429465, "grad_norm": 2.889455795288086, "learning_rate": 4.15e-08, "loss": 7.6362, "step": 83 }, { "epoch": 0.026332288401253918, "grad_norm": 4.101494789123535, "learning_rate": 4.2e-08, "loss": 9.3998, "step": 84 }, { "epoch": 0.02664576802507837, "grad_norm": 3.166928291320801, "learning_rate": 4.2500000000000003e-08, "loss": 8.9123, "step": 85 }, { "epoch": 0.026959247648902823, "grad_norm": 4.550684452056885, "learning_rate": 4.3e-08, "loss": 10.9083, "step": 86 }, { "epoch": 0.02727272727272727, "grad_norm": 3.8120524883270264, "learning_rate": 4.35e-08, "loss": 11.0573, "step": 87 }, { "epoch": 0.027586206896551724, "grad_norm": 3.2697463035583496, "learning_rate": 4.4000000000000004e-08, "loss": 6.9374, "step": 88 }, { "epoch": 0.027899686520376176, "grad_norm": 4.755700588226318, "learning_rate": 4.45e-08, "loss": 9.1937, "step": 89 }, { "epoch": 0.02821316614420063, "grad_norm": 2.45084547996521, "learning_rate": 4.5e-08, "loss": 8.1863, "step": 90 }, { "epoch": 0.028526645768025077, "grad_norm": 4.488142490386963, "learning_rate": 4.5500000000000004e-08, "loss": 10.0932, "step": 91 }, { "epoch": 0.02884012539184953, "grad_norm": 2.933330774307251, "learning_rate": 4.6e-08, "loss": 6.199, "step": 92 }, { "epoch": 0.029153605015673982, "grad_norm": 3.5675182342529297, "learning_rate": 4.65e-08, "loss": 11.0667, "step": 93 }, { "epoch": 0.02946708463949843, "grad_norm": 3.300973653793335, "learning_rate": 4.7000000000000004e-08, "loss": 8.7929, "step": 94 }, { "epoch": 0.029780564263322883, "grad_norm": 4.447729587554932, "learning_rate": 4.75e-08, "loss": 13.4648, "step": 95 }, { "epoch": 0.030094043887147336, "grad_norm": 3.820695638656616, "learning_rate": 4.8e-08, "loss": 8.4362, "step": 96 }, { "epoch": 0.030407523510971788, "grad_norm": 3.2529191970825195, "learning_rate": 4.8500000000000004e-08, "loss": 7.8372, "step": 97 }, { "epoch": 0.030721003134796237, "grad_norm": 5.665775775909424, "learning_rate": 4.9e-08, "loss": 13.5665, "step": 98 }, { "epoch": 0.03103448275862069, "grad_norm": 3.9593966007232666, "learning_rate": 4.9500000000000006e-08, "loss": 8.0411, "step": 99 }, { "epoch": 0.03134796238244514, "grad_norm": 3.611478567123413, "learning_rate": 5.0000000000000004e-08, "loss": 7.8446, "step": 100 }, { "epoch": 0.031661442006269594, "grad_norm": 3.8838236331939697, "learning_rate": 5.05e-08, "loss": 8.1889, "step": 101 }, { "epoch": 0.03197492163009404, "grad_norm": 3.930360794067383, "learning_rate": 5.100000000000001e-08, "loss": 11.6122, "step": 102 }, { "epoch": 0.0322884012539185, "grad_norm": 3.1364479064941406, "learning_rate": 5.1500000000000005e-08, "loss": 7.8931, "step": 103 }, { "epoch": 0.03260188087774295, "grad_norm": 4.427796840667725, "learning_rate": 5.2e-08, "loss": 10.2633, "step": 104 }, { "epoch": 0.032915360501567396, "grad_norm": 4.472553730010986, "learning_rate": 5.250000000000001e-08, "loss": 9.7383, "step": 105 }, { "epoch": 0.03322884012539185, "grad_norm": 3.7383227348327637, "learning_rate": 5.3000000000000005e-08, "loss": 10.8243, "step": 106 }, { "epoch": 0.0335423197492163, "grad_norm": 4.171073913574219, "learning_rate": 5.35e-08, "loss": 9.0764, "step": 107 }, { "epoch": 0.03385579937304075, "grad_norm": 3.8834450244903564, "learning_rate": 5.400000000000001e-08, "loss": 11.7289, "step": 108 }, { "epoch": 0.034169278996865206, "grad_norm": 5.5527753829956055, "learning_rate": 5.4500000000000005e-08, "loss": 13.2347, "step": 109 }, { "epoch": 0.034482758620689655, "grad_norm": 3.0170059204101562, "learning_rate": 5.5e-08, "loss": 8.0964, "step": 110 }, { "epoch": 0.034796238244514104, "grad_norm": 4.954318046569824, "learning_rate": 5.550000000000001e-08, "loss": 9.1267, "step": 111 }, { "epoch": 0.03510971786833856, "grad_norm": 5.236219882965088, "learning_rate": 5.6000000000000005e-08, "loss": 13.9724, "step": 112 }, { "epoch": 0.03542319749216301, "grad_norm": 3.203052520751953, "learning_rate": 5.65e-08, "loss": 7.3345, "step": 113 }, { "epoch": 0.035736677115987464, "grad_norm": 3.730330228805542, "learning_rate": 5.700000000000001e-08, "loss": 11.3994, "step": 114 }, { "epoch": 0.03605015673981191, "grad_norm": 3.654137372970581, "learning_rate": 5.7500000000000005e-08, "loss": 7.6924, "step": 115 }, { "epoch": 0.03636363636363636, "grad_norm": 3.058237314224243, "learning_rate": 5.8e-08, "loss": 6.7775, "step": 116 }, { "epoch": 0.03667711598746082, "grad_norm": 3.9933547973632812, "learning_rate": 5.850000000000001e-08, "loss": 9.6667, "step": 117 }, { "epoch": 0.03699059561128527, "grad_norm": 4.147707462310791, "learning_rate": 5.9000000000000006e-08, "loss": 7.6471, "step": 118 }, { "epoch": 0.037304075235109715, "grad_norm": 4.296032428741455, "learning_rate": 5.950000000000001e-08, "loss": 13.4849, "step": 119 }, { "epoch": 0.03761755485893417, "grad_norm": 3.6536829471588135, "learning_rate": 6.000000000000001e-08, "loss": 10.2448, "step": 120 }, { "epoch": 0.03793103448275862, "grad_norm": 3.77724552154541, "learning_rate": 6.05e-08, "loss": 8.568, "step": 121 }, { "epoch": 0.03824451410658307, "grad_norm": 3.5914738178253174, "learning_rate": 6.1e-08, "loss": 7.9479, "step": 122 }, { "epoch": 0.038557993730407525, "grad_norm": 8.78665828704834, "learning_rate": 6.15e-08, "loss": 7.3475, "step": 123 }, { "epoch": 0.038871473354231974, "grad_norm": 3.715463399887085, "learning_rate": 6.2e-08, "loss": 8.8613, "step": 124 }, { "epoch": 0.03918495297805643, "grad_norm": 3.4635043144226074, "learning_rate": 6.250000000000001e-08, "loss": 7.8405, "step": 125 }, { "epoch": 0.03949843260188088, "grad_norm": 3.9361000061035156, "learning_rate": 6.300000000000001e-08, "loss": 7.8298, "step": 126 }, { "epoch": 0.03981191222570533, "grad_norm": 3.872082233428955, "learning_rate": 6.35e-08, "loss": 11.5712, "step": 127 }, { "epoch": 0.04012539184952978, "grad_norm": 3.6292521953582764, "learning_rate": 6.4e-08, "loss": 8.9746, "step": 128 }, { "epoch": 0.04043887147335423, "grad_norm": 3.303443193435669, "learning_rate": 6.45e-08, "loss": 7.3576, "step": 129 }, { "epoch": 0.04075235109717868, "grad_norm": 2.838657855987549, "learning_rate": 6.5e-08, "loss": 8.1253, "step": 130 }, { "epoch": 0.04106583072100314, "grad_norm": 3.875380039215088, "learning_rate": 6.550000000000001e-08, "loss": 9.9537, "step": 131 }, { "epoch": 0.041379310344827586, "grad_norm": 3.1287903785705566, "learning_rate": 6.600000000000001e-08, "loss": 8.871, "step": 132 }, { "epoch": 0.041692789968652035, "grad_norm": 3.414733648300171, "learning_rate": 6.65e-08, "loss": 6.58, "step": 133 }, { "epoch": 0.04200626959247649, "grad_norm": 3.1217563152313232, "learning_rate": 6.7e-08, "loss": 7.591, "step": 134 }, { "epoch": 0.04231974921630094, "grad_norm": 4.4469733238220215, "learning_rate": 6.75e-08, "loss": 9.0155, "step": 135 }, { "epoch": 0.042633228840125395, "grad_norm": 3.728848695755005, "learning_rate": 6.8e-08, "loss": 8.98, "step": 136 }, { "epoch": 0.042946708463949844, "grad_norm": 4.085875988006592, "learning_rate": 6.850000000000001e-08, "loss": 9.7549, "step": 137 }, { "epoch": 0.04326018808777429, "grad_norm": 2.9348063468933105, "learning_rate": 6.900000000000001e-08, "loss": 7.0279, "step": 138 }, { "epoch": 0.04357366771159875, "grad_norm": 3.2348380088806152, "learning_rate": 6.950000000000001e-08, "loss": 8.1917, "step": 139 }, { "epoch": 0.0438871473354232, "grad_norm": 3.731046438217163, "learning_rate": 7e-08, "loss": 8.1675, "step": 140 }, { "epoch": 0.044200626959247646, "grad_norm": 2.7796294689178467, "learning_rate": 7.05e-08, "loss": 7.1764, "step": 141 }, { "epoch": 0.0445141065830721, "grad_norm": 3.467015027999878, "learning_rate": 7.100000000000001e-08, "loss": 7.7669, "step": 142 }, { "epoch": 0.04482758620689655, "grad_norm": 4.046057224273682, "learning_rate": 7.150000000000001e-08, "loss": 11.0822, "step": 143 }, { "epoch": 0.045141065830721, "grad_norm": 4.335888385772705, "learning_rate": 7.200000000000001e-08, "loss": 9.0507, "step": 144 }, { "epoch": 0.045454545454545456, "grad_norm": 3.208477258682251, "learning_rate": 7.250000000000001e-08, "loss": 8.5856, "step": 145 }, { "epoch": 0.045768025078369905, "grad_norm": 4.974661350250244, "learning_rate": 7.3e-08, "loss": 13.4032, "step": 146 }, { "epoch": 0.04608150470219436, "grad_norm": 5.451123237609863, "learning_rate": 7.35e-08, "loss": 11.6684, "step": 147 }, { "epoch": 0.04639498432601881, "grad_norm": 5.630722999572754, "learning_rate": 7.400000000000001e-08, "loss": 14.3619, "step": 148 }, { "epoch": 0.04670846394984326, "grad_norm": 3.298954725265503, "learning_rate": 7.450000000000001e-08, "loss": 7.9049, "step": 149 }, { "epoch": 0.047021943573667714, "grad_norm": 6.209390640258789, "learning_rate": 7.500000000000001e-08, "loss": 12.7281, "step": 150 }, { "epoch": 0.04733542319749216, "grad_norm": 4.096133232116699, "learning_rate": 7.550000000000001e-08, "loss": 10.6689, "step": 151 }, { "epoch": 0.04764890282131661, "grad_norm": 3.8314168453216553, "learning_rate": 7.6e-08, "loss": 9.3038, "step": 152 }, { "epoch": 0.04796238244514107, "grad_norm": 4.284617900848389, "learning_rate": 7.65e-08, "loss": 11.3222, "step": 153 }, { "epoch": 0.04827586206896552, "grad_norm": 3.8246536254882812, "learning_rate": 7.700000000000001e-08, "loss": 9.8774, "step": 154 }, { "epoch": 0.048589341692789965, "grad_norm": 3.011291265487671, "learning_rate": 7.750000000000001e-08, "loss": 6.99, "step": 155 }, { "epoch": 0.04890282131661442, "grad_norm": 3.5155868530273438, "learning_rate": 7.8e-08, "loss": 10.0299, "step": 156 }, { "epoch": 0.04921630094043887, "grad_norm": 3.2512242794036865, "learning_rate": 7.85e-08, "loss": 7.9809, "step": 157 }, { "epoch": 0.049529780564263326, "grad_norm": 3.8495590686798096, "learning_rate": 7.900000000000002e-08, "loss": 8.4488, "step": 158 }, { "epoch": 0.049843260188087775, "grad_norm": 4.043474197387695, "learning_rate": 7.950000000000002e-08, "loss": 7.3039, "step": 159 }, { "epoch": 0.050156739811912224, "grad_norm": 3.638864755630493, "learning_rate": 8e-08, "loss": 12.0005, "step": 160 }, { "epoch": 0.05047021943573668, "grad_norm": 3.9693477153778076, "learning_rate": 8.05e-08, "loss": 5.8005, "step": 161 }, { "epoch": 0.05078369905956113, "grad_norm": 4.173606872558594, "learning_rate": 8.1e-08, "loss": 14.1363, "step": 162 }, { "epoch": 0.05109717868338558, "grad_norm": 3.1908795833587646, "learning_rate": 8.15e-08, "loss": 8.1196, "step": 163 }, { "epoch": 0.05141065830721003, "grad_norm": 2.548922061920166, "learning_rate": 8.200000000000002e-08, "loss": 6.7172, "step": 164 }, { "epoch": 0.05172413793103448, "grad_norm": 3.2565834522247314, "learning_rate": 8.25e-08, "loss": 7.6421, "step": 165 }, { "epoch": 0.05203761755485893, "grad_norm": 4.065040111541748, "learning_rate": 8.3e-08, "loss": 9.3676, "step": 166 }, { "epoch": 0.05235109717868339, "grad_norm": 3.38653826713562, "learning_rate": 8.35e-08, "loss": 8.983, "step": 167 }, { "epoch": 0.052664576802507836, "grad_norm": 3.8572115898132324, "learning_rate": 8.4e-08, "loss": 8.1122, "step": 168 }, { "epoch": 0.05297805642633229, "grad_norm": 3.2368323802948, "learning_rate": 8.45e-08, "loss": 8.4592, "step": 169 }, { "epoch": 0.05329153605015674, "grad_norm": 3.422175884246826, "learning_rate": 8.500000000000001e-08, "loss": 9.0227, "step": 170 }, { "epoch": 0.05360501567398119, "grad_norm": 4.226418972015381, "learning_rate": 8.55e-08, "loss": 10.3479, "step": 171 }, { "epoch": 0.053918495297805645, "grad_norm": 4.698319435119629, "learning_rate": 8.6e-08, "loss": 9.7526, "step": 172 }, { "epoch": 0.054231974921630094, "grad_norm": 3.448711395263672, "learning_rate": 8.65e-08, "loss": 9.1493, "step": 173 }, { "epoch": 0.05454545454545454, "grad_norm": 3.3972206115722656, "learning_rate": 8.7e-08, "loss": 6.7097, "step": 174 }, { "epoch": 0.054858934169279, "grad_norm": 3.6650757789611816, "learning_rate": 8.750000000000001e-08, "loss": 10.019, "step": 175 }, { "epoch": 0.05517241379310345, "grad_norm": 3.743401288986206, "learning_rate": 8.800000000000001e-08, "loss": 9.1552, "step": 176 }, { "epoch": 0.055485893416927896, "grad_norm": 3.6061248779296875, "learning_rate": 8.85e-08, "loss": 8.0651, "step": 177 }, { "epoch": 0.05579937304075235, "grad_norm": 3.2070910930633545, "learning_rate": 8.9e-08, "loss": 8.7052, "step": 178 }, { "epoch": 0.0561128526645768, "grad_norm": 4.579248428344727, "learning_rate": 8.95e-08, "loss": 8.3114, "step": 179 }, { "epoch": 0.05642633228840126, "grad_norm": 4.0575690269470215, "learning_rate": 9e-08, "loss": 8.8177, "step": 180 }, { "epoch": 0.056739811912225706, "grad_norm": 3.667318820953369, "learning_rate": 9.050000000000001e-08, "loss": 9.8964, "step": 181 }, { "epoch": 0.057053291536050155, "grad_norm": 4.42782735824585, "learning_rate": 9.100000000000001e-08, "loss": 10.0656, "step": 182 }, { "epoch": 0.05736677115987461, "grad_norm": 3.8123972415924072, "learning_rate": 9.15e-08, "loss": 10.2998, "step": 183 }, { "epoch": 0.05768025078369906, "grad_norm": 4.3586745262146, "learning_rate": 9.2e-08, "loss": 8.8101, "step": 184 }, { "epoch": 0.05799373040752351, "grad_norm": 3.785413980484009, "learning_rate": 9.25e-08, "loss": 9.7424, "step": 185 }, { "epoch": 0.058307210031347964, "grad_norm": 3.5556790828704834, "learning_rate": 9.3e-08, "loss": 10.6945, "step": 186 }, { "epoch": 0.05862068965517241, "grad_norm": 4.004053592681885, "learning_rate": 9.350000000000001e-08, "loss": 9.8506, "step": 187 }, { "epoch": 0.05893416927899686, "grad_norm": 5.22344970703125, "learning_rate": 9.400000000000001e-08, "loss": 12.796, "step": 188 }, { "epoch": 0.05924764890282132, "grad_norm": 3.7727713584899902, "learning_rate": 9.45e-08, "loss": 7.7093, "step": 189 }, { "epoch": 0.05956112852664577, "grad_norm": 3.474813461303711, "learning_rate": 9.5e-08, "loss": 8.4177, "step": 190 }, { "epoch": 0.05987460815047022, "grad_norm": 3.4020941257476807, "learning_rate": 9.55e-08, "loss": 10.3753, "step": 191 }, { "epoch": 0.06018808777429467, "grad_norm": 4.846778392791748, "learning_rate": 9.6e-08, "loss": 9.5903, "step": 192 }, { "epoch": 0.06050156739811912, "grad_norm": 4.567539691925049, "learning_rate": 9.650000000000001e-08, "loss": 9.7159, "step": 193 }, { "epoch": 0.060815047021943576, "grad_norm": 3.0759778022766113, "learning_rate": 9.700000000000001e-08, "loss": 7.0748, "step": 194 }, { "epoch": 0.061128526645768025, "grad_norm": 3.4420666694641113, "learning_rate": 9.75e-08, "loss": 7.8556, "step": 195 }, { "epoch": 0.061442006269592474, "grad_norm": 3.7455813884735107, "learning_rate": 9.8e-08, "loss": 7.4662, "step": 196 }, { "epoch": 0.06175548589341693, "grad_norm": 3.7451171875, "learning_rate": 9.85e-08, "loss": 8.2137, "step": 197 }, { "epoch": 0.06206896551724138, "grad_norm": 3.513235330581665, "learning_rate": 9.900000000000001e-08, "loss": 8.2557, "step": 198 }, { "epoch": 0.06238244514106583, "grad_norm": 4.297177314758301, "learning_rate": 9.950000000000001e-08, "loss": 10.0288, "step": 199 }, { "epoch": 0.06269592476489028, "grad_norm": 4.298759460449219, "learning_rate": 1.0000000000000001e-07, "loss": 10.5449, "step": 200 }, { "epoch": 0.06300940438871473, "grad_norm": 3.4075138568878174, "learning_rate": 1.005e-07, "loss": 7.7592, "step": 201 }, { "epoch": 0.06332288401253919, "grad_norm": 3.19218373298645, "learning_rate": 1.01e-07, "loss": 7.7692, "step": 202 }, { "epoch": 0.06363636363636363, "grad_norm": 4.200982093811035, "learning_rate": 1.015e-07, "loss": 8.7216, "step": 203 }, { "epoch": 0.06394984326018809, "grad_norm": 3.4024147987365723, "learning_rate": 1.0200000000000001e-07, "loss": 7.9235, "step": 204 }, { "epoch": 0.06426332288401254, "grad_norm": 3.751147747039795, "learning_rate": 1.0250000000000001e-07, "loss": 10.3086, "step": 205 }, { "epoch": 0.064576802507837, "grad_norm": 3.3769607543945312, "learning_rate": 1.0300000000000001e-07, "loss": 7.629, "step": 206 }, { "epoch": 0.06489028213166144, "grad_norm": 3.3052303791046143, "learning_rate": 1.0350000000000001e-07, "loss": 9.5893, "step": 207 }, { "epoch": 0.0652037617554859, "grad_norm": 4.127976894378662, "learning_rate": 1.04e-07, "loss": 12.7904, "step": 208 }, { "epoch": 0.06551724137931035, "grad_norm": 3.8568077087402344, "learning_rate": 1.045e-07, "loss": 9.6994, "step": 209 }, { "epoch": 0.06583072100313479, "grad_norm": 3.236760377883911, "learning_rate": 1.0500000000000001e-07, "loss": 8.0482, "step": 210 }, { "epoch": 0.06614420062695925, "grad_norm": 4.277583122253418, "learning_rate": 1.0550000000000001e-07, "loss": 8.1819, "step": 211 }, { "epoch": 0.0664576802507837, "grad_norm": 3.3786604404449463, "learning_rate": 1.0600000000000001e-07, "loss": 7.0545, "step": 212 }, { "epoch": 0.06677115987460815, "grad_norm": 4.007237911224365, "learning_rate": 1.0650000000000001e-07, "loss": 9.9794, "step": 213 }, { "epoch": 0.0670846394984326, "grad_norm": 4.616070747375488, "learning_rate": 1.07e-07, "loss": 11.1093, "step": 214 }, { "epoch": 0.06739811912225706, "grad_norm": 3.0655412673950195, "learning_rate": 1.075e-07, "loss": 9.5085, "step": 215 }, { "epoch": 0.0677115987460815, "grad_norm": 3.05678391456604, "learning_rate": 1.0800000000000001e-07, "loss": 6.3335, "step": 216 }, { "epoch": 0.06802507836990596, "grad_norm": 2.941502094268799, "learning_rate": 1.0850000000000001e-07, "loss": 7.2431, "step": 217 }, { "epoch": 0.06833855799373041, "grad_norm": 3.810181140899658, "learning_rate": 1.0900000000000001e-07, "loss": 9.8534, "step": 218 }, { "epoch": 0.06865203761755485, "grad_norm": 3.007319450378418, "learning_rate": 1.0950000000000001e-07, "loss": 9.2206, "step": 219 }, { "epoch": 0.06896551724137931, "grad_norm": 3.6502251625061035, "learning_rate": 1.1e-07, "loss": 7.9423, "step": 220 }, { "epoch": 0.06927899686520377, "grad_norm": 3.4484660625457764, "learning_rate": 1.1050000000000002e-07, "loss": 8.6399, "step": 221 }, { "epoch": 0.06959247648902821, "grad_norm": 3.3999218940734863, "learning_rate": 1.1100000000000001e-07, "loss": 8.1017, "step": 222 }, { "epoch": 0.06990595611285266, "grad_norm": 3.0518319606781006, "learning_rate": 1.1150000000000001e-07, "loss": 7.0595, "step": 223 }, { "epoch": 0.07021943573667712, "grad_norm": 2.940213918685913, "learning_rate": 1.1200000000000001e-07, "loss": 7.2927, "step": 224 }, { "epoch": 0.07053291536050156, "grad_norm": 4.038904190063477, "learning_rate": 1.1250000000000001e-07, "loss": 9.1408, "step": 225 }, { "epoch": 0.07084639498432602, "grad_norm": 2.5958571434020996, "learning_rate": 1.13e-07, "loss": 6.325, "step": 226 }, { "epoch": 0.07115987460815047, "grad_norm": 3.0129003524780273, "learning_rate": 1.1350000000000002e-07, "loss": 6.963, "step": 227 }, { "epoch": 0.07147335423197493, "grad_norm": 2.9724435806274414, "learning_rate": 1.1400000000000001e-07, "loss": 8.6995, "step": 228 }, { "epoch": 0.07178683385579937, "grad_norm": 3.7590036392211914, "learning_rate": 1.1450000000000001e-07, "loss": 7.9243, "step": 229 }, { "epoch": 0.07210031347962383, "grad_norm": 3.5400218963623047, "learning_rate": 1.1500000000000001e-07, "loss": 6.981, "step": 230 }, { "epoch": 0.07241379310344828, "grad_norm": 2.9733407497406006, "learning_rate": 1.1550000000000001e-07, "loss": 7.1957, "step": 231 }, { "epoch": 0.07272727272727272, "grad_norm": 3.4962241649627686, "learning_rate": 1.16e-07, "loss": 10.1103, "step": 232 }, { "epoch": 0.07304075235109718, "grad_norm": 3.234700918197632, "learning_rate": 1.1650000000000002e-07, "loss": 7.5018, "step": 233 }, { "epoch": 0.07335423197492164, "grad_norm": 3.2388176918029785, "learning_rate": 1.1700000000000002e-07, "loss": 9.0859, "step": 234 }, { "epoch": 0.07366771159874608, "grad_norm": 3.4157941341400146, "learning_rate": 1.1750000000000001e-07, "loss": 7.8399, "step": 235 }, { "epoch": 0.07398119122257053, "grad_norm": 3.5833077430725098, "learning_rate": 1.1800000000000001e-07, "loss": 7.5087, "step": 236 }, { "epoch": 0.07429467084639499, "grad_norm": 4.769215106964111, "learning_rate": 1.1850000000000001e-07, "loss": 8.7621, "step": 237 }, { "epoch": 0.07460815047021943, "grad_norm": 3.941755771636963, "learning_rate": 1.1900000000000002e-07, "loss": 13.078, "step": 238 }, { "epoch": 0.07492163009404389, "grad_norm": 3.798787832260132, "learning_rate": 1.195e-07, "loss": 9.7572, "step": 239 }, { "epoch": 0.07523510971786834, "grad_norm": 4.713004112243652, "learning_rate": 1.2000000000000002e-07, "loss": 13.3091, "step": 240 }, { "epoch": 0.07554858934169278, "grad_norm": 3.408093214035034, "learning_rate": 1.205e-07, "loss": 7.7499, "step": 241 }, { "epoch": 0.07586206896551724, "grad_norm": 4.062272071838379, "learning_rate": 1.21e-07, "loss": 9.7988, "step": 242 }, { "epoch": 0.0761755485893417, "grad_norm": 4.664130210876465, "learning_rate": 1.215e-07, "loss": 9.4779, "step": 243 }, { "epoch": 0.07648902821316614, "grad_norm": 3.5159950256347656, "learning_rate": 1.22e-07, "loss": 8.8292, "step": 244 }, { "epoch": 0.0768025078369906, "grad_norm": 3.664834499359131, "learning_rate": 1.2250000000000002e-07, "loss": 8.0994, "step": 245 }, { "epoch": 0.07711598746081505, "grad_norm": 4.765824794769287, "learning_rate": 1.23e-07, "loss": 8.5952, "step": 246 }, { "epoch": 0.07742946708463949, "grad_norm": 3.9765498638153076, "learning_rate": 1.2350000000000001e-07, "loss": 9.6514, "step": 247 }, { "epoch": 0.07774294670846395, "grad_norm": 3.6185436248779297, "learning_rate": 1.24e-07, "loss": 8.74, "step": 248 }, { "epoch": 0.0780564263322884, "grad_norm": 3.6925275325775146, "learning_rate": 1.245e-07, "loss": 7.0158, "step": 249 }, { "epoch": 0.07836990595611286, "grad_norm": 3.878559112548828, "learning_rate": 1.2500000000000002e-07, "loss": 8.7004, "step": 250 }, { "epoch": 0.0786833855799373, "grad_norm": 4.704588413238525, "learning_rate": 1.255e-07, "loss": 9.9536, "step": 251 }, { "epoch": 0.07899686520376176, "grad_norm": 4.152799129486084, "learning_rate": 1.2600000000000002e-07, "loss": 9.6934, "step": 252 }, { "epoch": 0.07931034482758621, "grad_norm": 3.1007423400878906, "learning_rate": 1.265e-07, "loss": 7.1812, "step": 253 }, { "epoch": 0.07962382445141065, "grad_norm": 3.872610092163086, "learning_rate": 1.27e-07, "loss": 8.3354, "step": 254 }, { "epoch": 0.07993730407523511, "grad_norm": 3.334554433822632, "learning_rate": 1.275e-07, "loss": 9.6324, "step": 255 }, { "epoch": 0.08025078369905957, "grad_norm": 3.413045883178711, "learning_rate": 1.28e-07, "loss": 8.7391, "step": 256 }, { "epoch": 0.08056426332288401, "grad_norm": 3.656658887863159, "learning_rate": 1.2850000000000002e-07, "loss": 10.2712, "step": 257 }, { "epoch": 0.08087774294670846, "grad_norm": 3.1734461784362793, "learning_rate": 1.29e-07, "loss": 7.0908, "step": 258 }, { "epoch": 0.08119122257053292, "grad_norm": 5.781704425811768, "learning_rate": 1.2950000000000001e-07, "loss": 8.8973, "step": 259 }, { "epoch": 0.08150470219435736, "grad_norm": 3.3216211795806885, "learning_rate": 1.3e-07, "loss": 7.9809, "step": 260 }, { "epoch": 0.08181818181818182, "grad_norm": 3.4130654335021973, "learning_rate": 1.305e-07, "loss": 10.6107, "step": 261 }, { "epoch": 0.08213166144200627, "grad_norm": 3.198608160018921, "learning_rate": 1.3100000000000002e-07, "loss": 7.9311, "step": 262 }, { "epoch": 0.08244514106583072, "grad_norm": 3.2649126052856445, "learning_rate": 1.315e-07, "loss": 6.9358, "step": 263 }, { "epoch": 0.08275862068965517, "grad_norm": 3.805680513381958, "learning_rate": 1.3200000000000002e-07, "loss": 8.7701, "step": 264 }, { "epoch": 0.08307210031347963, "grad_norm": 3.8234152793884277, "learning_rate": 1.325e-07, "loss": 7.1507, "step": 265 }, { "epoch": 0.08338557993730407, "grad_norm": 4.14029598236084, "learning_rate": 1.33e-07, "loss": 10.002, "step": 266 }, { "epoch": 0.08369905956112852, "grad_norm": 3.70548677444458, "learning_rate": 1.3350000000000002e-07, "loss": 7.0396, "step": 267 }, { "epoch": 0.08401253918495298, "grad_norm": 4.227056503295898, "learning_rate": 1.34e-07, "loss": 10.8817, "step": 268 }, { "epoch": 0.08432601880877742, "grad_norm": 3.6237289905548096, "learning_rate": 1.3450000000000002e-07, "loss": 8.0917, "step": 269 }, { "epoch": 0.08463949843260188, "grad_norm": 2.858159303665161, "learning_rate": 1.35e-07, "loss": 6.4492, "step": 270 }, { "epoch": 0.08495297805642633, "grad_norm": 3.7545058727264404, "learning_rate": 1.3550000000000002e-07, "loss": 7.2091, "step": 271 }, { "epoch": 0.08526645768025079, "grad_norm": 3.4545750617980957, "learning_rate": 1.36e-07, "loss": 6.7025, "step": 272 }, { "epoch": 0.08557993730407523, "grad_norm": 2.947838068008423, "learning_rate": 1.365e-07, "loss": 6.9652, "step": 273 }, { "epoch": 0.08589341692789969, "grad_norm": 4.7777605056762695, "learning_rate": 1.3700000000000002e-07, "loss": 10.056, "step": 274 }, { "epoch": 0.08620689655172414, "grad_norm": 2.519544839859009, "learning_rate": 1.375e-07, "loss": 6.0347, "step": 275 }, { "epoch": 0.08652037617554859, "grad_norm": 4.664981842041016, "learning_rate": 1.3800000000000002e-07, "loss": 10.88, "step": 276 }, { "epoch": 0.08683385579937304, "grad_norm": 4.797813892364502, "learning_rate": 1.385e-07, "loss": 10.9775, "step": 277 }, { "epoch": 0.0871473354231975, "grad_norm": 3.1882078647613525, "learning_rate": 1.3900000000000001e-07, "loss": 7.7702, "step": 278 }, { "epoch": 0.08746081504702194, "grad_norm": 4.286925792694092, "learning_rate": 1.3950000000000002e-07, "loss": 10.9283, "step": 279 }, { "epoch": 0.0877742946708464, "grad_norm": 3.3588335514068604, "learning_rate": 1.4e-07, "loss": 7.7342, "step": 280 }, { "epoch": 0.08808777429467085, "grad_norm": 2.775568962097168, "learning_rate": 1.4050000000000002e-07, "loss": 6.9301, "step": 281 }, { "epoch": 0.08840125391849529, "grad_norm": 4.123598575592041, "learning_rate": 1.41e-07, "loss": 10.0182, "step": 282 }, { "epoch": 0.08871473354231975, "grad_norm": 4.854518413543701, "learning_rate": 1.4150000000000002e-07, "loss": 12.1932, "step": 283 }, { "epoch": 0.0890282131661442, "grad_norm": 2.8263025283813477, "learning_rate": 1.4200000000000003e-07, "loss": 7.8395, "step": 284 }, { "epoch": 0.08934169278996865, "grad_norm": 3.2582967281341553, "learning_rate": 1.425e-07, "loss": 7.4587, "step": 285 }, { "epoch": 0.0896551724137931, "grad_norm": 3.1975438594818115, "learning_rate": 1.4300000000000002e-07, "loss": 8.037, "step": 286 }, { "epoch": 0.08996865203761756, "grad_norm": 3.0459468364715576, "learning_rate": 1.435e-07, "loss": 7.1786, "step": 287 }, { "epoch": 0.090282131661442, "grad_norm": 4.034944534301758, "learning_rate": 1.4400000000000002e-07, "loss": 8.2916, "step": 288 }, { "epoch": 0.09059561128526646, "grad_norm": 5.0129499435424805, "learning_rate": 1.445e-07, "loss": 11.0094, "step": 289 }, { "epoch": 0.09090909090909091, "grad_norm": 4.1220903396606445, "learning_rate": 1.4500000000000001e-07, "loss": 8.5799, "step": 290 }, { "epoch": 0.09122257053291537, "grad_norm": 3.6785085201263428, "learning_rate": 1.4550000000000003e-07, "loss": 9.5831, "step": 291 }, { "epoch": 0.09153605015673981, "grad_norm": 3.44356369972229, "learning_rate": 1.46e-07, "loss": 7.9535, "step": 292 }, { "epoch": 0.09184952978056427, "grad_norm": 4.070002555847168, "learning_rate": 1.4650000000000002e-07, "loss": 11.4279, "step": 293 }, { "epoch": 0.09216300940438872, "grad_norm": 3.0748281478881836, "learning_rate": 1.47e-07, "loss": 7.3921, "step": 294 }, { "epoch": 0.09247648902821316, "grad_norm": 2.973388671875, "learning_rate": 1.4750000000000002e-07, "loss": 7.6754, "step": 295 }, { "epoch": 0.09278996865203762, "grad_norm": 3.4642834663391113, "learning_rate": 1.4800000000000003e-07, "loss": 8.1933, "step": 296 }, { "epoch": 0.09310344827586207, "grad_norm": 4.090353012084961, "learning_rate": 1.485e-07, "loss": 7.926, "step": 297 }, { "epoch": 0.09341692789968652, "grad_norm": 3.8677382469177246, "learning_rate": 1.4900000000000002e-07, "loss": 7.5799, "step": 298 }, { "epoch": 0.09373040752351097, "grad_norm": 3.7805559635162354, "learning_rate": 1.495e-07, "loss": 9.6573, "step": 299 }, { "epoch": 0.09404388714733543, "grad_norm": 3.8729629516601562, "learning_rate": 1.5000000000000002e-07, "loss": 10.3939, "step": 300 }, { "epoch": 0.09435736677115987, "grad_norm": 4.191401481628418, "learning_rate": 1.505e-07, "loss": 11.4678, "step": 301 }, { "epoch": 0.09467084639498433, "grad_norm": 3.4422829151153564, "learning_rate": 1.5100000000000002e-07, "loss": 7.9499, "step": 302 }, { "epoch": 0.09498432601880878, "grad_norm": 3.8708252906799316, "learning_rate": 1.5150000000000003e-07, "loss": 9.9911, "step": 303 }, { "epoch": 0.09529780564263322, "grad_norm": 3.080570936203003, "learning_rate": 1.52e-07, "loss": 7.4803, "step": 304 }, { "epoch": 0.09561128526645768, "grad_norm": 3.9624032974243164, "learning_rate": 1.5250000000000002e-07, "loss": 10.5125, "step": 305 }, { "epoch": 0.09592476489028214, "grad_norm": 4.132358551025391, "learning_rate": 1.53e-07, "loss": 10.1554, "step": 306 }, { "epoch": 0.09623824451410658, "grad_norm": 3.146574020385742, "learning_rate": 1.5350000000000002e-07, "loss": 6.7796, "step": 307 }, { "epoch": 0.09655172413793103, "grad_norm": 3.1497974395751953, "learning_rate": 1.5400000000000003e-07, "loss": 8.9689, "step": 308 }, { "epoch": 0.09686520376175549, "grad_norm": 4.384578227996826, "learning_rate": 1.5450000000000001e-07, "loss": 12.2934, "step": 309 }, { "epoch": 0.09717868338557993, "grad_norm": 3.5674219131469727, "learning_rate": 1.5500000000000002e-07, "loss": 10.6616, "step": 310 }, { "epoch": 0.09749216300940439, "grad_norm": 4.463435649871826, "learning_rate": 1.555e-07, "loss": 7.4585, "step": 311 }, { "epoch": 0.09780564263322884, "grad_norm": 4.1657609939575195, "learning_rate": 1.56e-07, "loss": 8.5832, "step": 312 }, { "epoch": 0.0981191222570533, "grad_norm": 5.202960968017578, "learning_rate": 1.5650000000000003e-07, "loss": 12.9104, "step": 313 }, { "epoch": 0.09843260188087774, "grad_norm": 4.485203742980957, "learning_rate": 1.57e-07, "loss": 7.8665, "step": 314 }, { "epoch": 0.0987460815047022, "grad_norm": 3.092864751815796, "learning_rate": 1.575e-07, "loss": 8.4677, "step": 315 }, { "epoch": 0.09905956112852665, "grad_norm": 3.2465710639953613, "learning_rate": 1.5800000000000004e-07, "loss": 7.498, "step": 316 }, { "epoch": 0.0993730407523511, "grad_norm": 5.711244106292725, "learning_rate": 1.585e-07, "loss": 12.3313, "step": 317 }, { "epoch": 0.09968652037617555, "grad_norm": 3.7532050609588623, "learning_rate": 1.5900000000000003e-07, "loss": 10.8737, "step": 318 }, { "epoch": 0.1, "grad_norm": 4.172240257263184, "learning_rate": 1.595e-07, "loss": 8.8672, "step": 319 }, { "epoch": 0.10031347962382445, "grad_norm": 3.577533721923828, "learning_rate": 1.6e-07, "loss": 11.197, "step": 320 }, { "epoch": 0.1006269592476489, "grad_norm": 3.8616108894348145, "learning_rate": 1.605e-07, "loss": 9.6396, "step": 321 }, { "epoch": 0.10094043887147336, "grad_norm": 5.926366329193115, "learning_rate": 1.61e-07, "loss": 12.75, "step": 322 }, { "epoch": 0.1012539184952978, "grad_norm": 3.613895893096924, "learning_rate": 1.6150000000000004e-07, "loss": 7.4106, "step": 323 }, { "epoch": 0.10156739811912226, "grad_norm": 3.902071952819824, "learning_rate": 1.62e-07, "loss": 11.4143, "step": 324 }, { "epoch": 0.10188087774294671, "grad_norm": 2.8985440731048584, "learning_rate": 1.625e-07, "loss": 7.3166, "step": 325 }, { "epoch": 0.10219435736677115, "grad_norm": 3.545724391937256, "learning_rate": 1.63e-07, "loss": 7.8454, "step": 326 }, { "epoch": 0.10250783699059561, "grad_norm": 3.259948253631592, "learning_rate": 1.635e-07, "loss": 8.4287, "step": 327 }, { "epoch": 0.10282131661442007, "grad_norm": 3.207242012023926, "learning_rate": 1.6400000000000004e-07, "loss": 7.234, "step": 328 }, { "epoch": 0.10313479623824451, "grad_norm": 3.2320971488952637, "learning_rate": 1.645e-07, "loss": 7.4103, "step": 329 }, { "epoch": 0.10344827586206896, "grad_norm": 4.266717910766602, "learning_rate": 1.65e-07, "loss": 10.7557, "step": 330 }, { "epoch": 0.10376175548589342, "grad_norm": 4.703216075897217, "learning_rate": 1.655e-07, "loss": 13.6921, "step": 331 }, { "epoch": 0.10407523510971786, "grad_norm": 2.98171329498291, "learning_rate": 1.66e-07, "loss": 6.4217, "step": 332 }, { "epoch": 0.10438871473354232, "grad_norm": 4.266834735870361, "learning_rate": 1.6650000000000004e-07, "loss": 8.7814, "step": 333 }, { "epoch": 0.10470219435736677, "grad_norm": 3.13301420211792, "learning_rate": 1.67e-07, "loss": 7.5465, "step": 334 }, { "epoch": 0.10501567398119123, "grad_norm": 3.93803071975708, "learning_rate": 1.675e-07, "loss": 9.7471, "step": 335 }, { "epoch": 0.10532915360501567, "grad_norm": 3.124204397201538, "learning_rate": 1.68e-07, "loss": 8.2321, "step": 336 }, { "epoch": 0.10564263322884013, "grad_norm": 3.1418778896331787, "learning_rate": 1.685e-07, "loss": 8.8142, "step": 337 }, { "epoch": 0.10595611285266458, "grad_norm": 3.865973472595215, "learning_rate": 1.69e-07, "loss": 10.8355, "step": 338 }, { "epoch": 0.10626959247648902, "grad_norm": 3.5527279376983643, "learning_rate": 1.695e-07, "loss": 7.6143, "step": 339 }, { "epoch": 0.10658307210031348, "grad_norm": 3.353538751602173, "learning_rate": 1.7000000000000001e-07, "loss": 9.4709, "step": 340 }, { "epoch": 0.10689655172413794, "grad_norm": 3.7409756183624268, "learning_rate": 1.705e-07, "loss": 11.7136, "step": 341 }, { "epoch": 0.10721003134796238, "grad_norm": 4.518954277038574, "learning_rate": 1.71e-07, "loss": 8.8255, "step": 342 }, { "epoch": 0.10752351097178683, "grad_norm": 4.619668960571289, "learning_rate": 1.715e-07, "loss": 12.3065, "step": 343 }, { "epoch": 0.10783699059561129, "grad_norm": 3.260843276977539, "learning_rate": 1.72e-07, "loss": 9.3232, "step": 344 }, { "epoch": 0.10815047021943573, "grad_norm": 3.9291486740112305, "learning_rate": 1.7250000000000002e-07, "loss": 9.9018, "step": 345 }, { "epoch": 0.10846394984326019, "grad_norm": 4.316411972045898, "learning_rate": 1.73e-07, "loss": 8.0697, "step": 346 }, { "epoch": 0.10877742946708464, "grad_norm": 3.2413835525512695, "learning_rate": 1.735e-07, "loss": 8.5543, "step": 347 }, { "epoch": 0.10909090909090909, "grad_norm": 2.881052017211914, "learning_rate": 1.74e-07, "loss": 6.481, "step": 348 }, { "epoch": 0.10940438871473354, "grad_norm": 3.0879931449890137, "learning_rate": 1.745e-07, "loss": 9.835, "step": 349 }, { "epoch": 0.109717868338558, "grad_norm": 3.5517656803131104, "learning_rate": 1.7500000000000002e-07, "loss": 8.4705, "step": 350 }, { "epoch": 0.11003134796238244, "grad_norm": 3.764478921890259, "learning_rate": 1.755e-07, "loss": 9.458, "step": 351 }, { "epoch": 0.1103448275862069, "grad_norm": 5.203456878662109, "learning_rate": 1.7600000000000001e-07, "loss": 11.9955, "step": 352 }, { "epoch": 0.11065830721003135, "grad_norm": 4.887335777282715, "learning_rate": 1.765e-07, "loss": 11.2742, "step": 353 }, { "epoch": 0.11097178683385579, "grad_norm": 3.6964869499206543, "learning_rate": 1.77e-07, "loss": 8.1842, "step": 354 }, { "epoch": 0.11128526645768025, "grad_norm": 3.0974836349487305, "learning_rate": 1.775e-07, "loss": 7.7249, "step": 355 }, { "epoch": 0.1115987460815047, "grad_norm": 3.3595480918884277, "learning_rate": 1.78e-07, "loss": 7.27, "step": 356 }, { "epoch": 0.11191222570532916, "grad_norm": 3.602126359939575, "learning_rate": 1.7850000000000002e-07, "loss": 10.0521, "step": 357 }, { "epoch": 0.1122257053291536, "grad_norm": 3.0182366371154785, "learning_rate": 1.79e-07, "loss": 8.1576, "step": 358 }, { "epoch": 0.11253918495297806, "grad_norm": 7.666667461395264, "learning_rate": 1.795e-07, "loss": 17.8031, "step": 359 }, { "epoch": 0.11285266457680251, "grad_norm": 3.1000211238861084, "learning_rate": 1.8e-07, "loss": 8.0788, "step": 360 }, { "epoch": 0.11316614420062696, "grad_norm": 3.830988883972168, "learning_rate": 1.805e-07, "loss": 10.3451, "step": 361 }, { "epoch": 0.11347962382445141, "grad_norm": 3.502091407775879, "learning_rate": 1.8100000000000002e-07, "loss": 7.3287, "step": 362 }, { "epoch": 0.11379310344827587, "grad_norm": 3.8298943042755127, "learning_rate": 1.815e-07, "loss": 7.3115, "step": 363 }, { "epoch": 0.11410658307210031, "grad_norm": 3.5254364013671875, "learning_rate": 1.8200000000000002e-07, "loss": 10.1726, "step": 364 }, { "epoch": 0.11442006269592477, "grad_norm": 2.8689956665039062, "learning_rate": 1.825e-07, "loss": 5.9041, "step": 365 }, { "epoch": 0.11473354231974922, "grad_norm": 3.9380292892456055, "learning_rate": 1.83e-07, "loss": 8.7651, "step": 366 }, { "epoch": 0.11504702194357366, "grad_norm": 4.940433502197266, "learning_rate": 1.8350000000000002e-07, "loss": 13.3309, "step": 367 }, { "epoch": 0.11536050156739812, "grad_norm": 3.1874752044677734, "learning_rate": 1.84e-07, "loss": 7.9231, "step": 368 }, { "epoch": 0.11567398119122257, "grad_norm": 3.361844778060913, "learning_rate": 1.8450000000000002e-07, "loss": 8.2631, "step": 369 }, { "epoch": 0.11598746081504702, "grad_norm": 4.501951217651367, "learning_rate": 1.85e-07, "loss": 11.7963, "step": 370 }, { "epoch": 0.11630094043887147, "grad_norm": 3.1025617122650146, "learning_rate": 1.8550000000000001e-07, "loss": 7.3833, "step": 371 }, { "epoch": 0.11661442006269593, "grad_norm": 3.30342960357666, "learning_rate": 1.86e-07, "loss": 8.7881, "step": 372 }, { "epoch": 0.11692789968652037, "grad_norm": 3.470710515975952, "learning_rate": 1.865e-07, "loss": 7.7015, "step": 373 }, { "epoch": 0.11724137931034483, "grad_norm": 3.2457809448242188, "learning_rate": 1.8700000000000002e-07, "loss": 8.8045, "step": 374 }, { "epoch": 0.11755485893416928, "grad_norm": 3.0047662258148193, "learning_rate": 1.875e-07, "loss": 8.3349, "step": 375 }, { "epoch": 0.11786833855799372, "grad_norm": 3.3185179233551025, "learning_rate": 1.8800000000000002e-07, "loss": 6.7955, "step": 376 }, { "epoch": 0.11818181818181818, "grad_norm": 3.7260663509368896, "learning_rate": 1.885e-07, "loss": 8.8171, "step": 377 }, { "epoch": 0.11849529780564264, "grad_norm": 3.9109082221984863, "learning_rate": 1.89e-07, "loss": 9.8559, "step": 378 }, { "epoch": 0.11880877742946709, "grad_norm": 3.2869834899902344, "learning_rate": 1.8950000000000002e-07, "loss": 11.0736, "step": 379 }, { "epoch": 0.11912225705329153, "grad_norm": 4.131401538848877, "learning_rate": 1.9e-07, "loss": 11.9681, "step": 380 }, { "epoch": 0.11943573667711599, "grad_norm": 3.241637945175171, "learning_rate": 1.9050000000000002e-07, "loss": 9.0305, "step": 381 }, { "epoch": 0.11974921630094044, "grad_norm": 3.635667324066162, "learning_rate": 1.91e-07, "loss": 9.6321, "step": 382 }, { "epoch": 0.12006269592476489, "grad_norm": 2.901669979095459, "learning_rate": 1.9150000000000001e-07, "loss": 7.8995, "step": 383 }, { "epoch": 0.12037617554858934, "grad_norm": 3.7463796138763428, "learning_rate": 1.92e-07, "loss": 7.9248, "step": 384 }, { "epoch": 0.1206896551724138, "grad_norm": 3.811286211013794, "learning_rate": 1.925e-07, "loss": 7.6093, "step": 385 }, { "epoch": 0.12100313479623824, "grad_norm": 3.3971803188323975, "learning_rate": 1.9300000000000002e-07, "loss": 7.9361, "step": 386 }, { "epoch": 0.1213166144200627, "grad_norm": 3.552212953567505, "learning_rate": 1.935e-07, "loss": 9.3616, "step": 387 }, { "epoch": 0.12163009404388715, "grad_norm": 3.006312608718872, "learning_rate": 1.9400000000000002e-07, "loss": 8.1449, "step": 388 }, { "epoch": 0.1219435736677116, "grad_norm": 3.3143630027770996, "learning_rate": 1.945e-07, "loss": 8.2126, "step": 389 }, { "epoch": 0.12225705329153605, "grad_norm": 3.8813610076904297, "learning_rate": 1.95e-07, "loss": 10.1397, "step": 390 }, { "epoch": 0.1225705329153605, "grad_norm": 2.9321842193603516, "learning_rate": 1.9550000000000002e-07, "loss": 6.3588, "step": 391 }, { "epoch": 0.12288401253918495, "grad_norm": 3.752800941467285, "learning_rate": 1.96e-07, "loss": 8.3612, "step": 392 }, { "epoch": 0.1231974921630094, "grad_norm": 3.759476661682129, "learning_rate": 1.9650000000000002e-07, "loss": 8.3461, "step": 393 }, { "epoch": 0.12351097178683386, "grad_norm": 2.937373161315918, "learning_rate": 1.97e-07, "loss": 6.9189, "step": 394 }, { "epoch": 0.1238244514106583, "grad_norm": 3.327111005783081, "learning_rate": 1.9750000000000001e-07, "loss": 9.0745, "step": 395 }, { "epoch": 0.12413793103448276, "grad_norm": 2.807039976119995, "learning_rate": 1.9800000000000003e-07, "loss": 6.5768, "step": 396 }, { "epoch": 0.12445141065830721, "grad_norm": 4.174190044403076, "learning_rate": 1.985e-07, "loss": 14.798, "step": 397 }, { "epoch": 0.12476489028213165, "grad_norm": 3.8052163124084473, "learning_rate": 1.9900000000000002e-07, "loss": 7.4751, "step": 398 }, { "epoch": 0.1250783699059561, "grad_norm": 3.2749545574188232, "learning_rate": 1.995e-07, "loss": 6.8837, "step": 399 }, { "epoch": 0.12539184952978055, "grad_norm": 5.2682695388793945, "learning_rate": 2.0000000000000002e-07, "loss": 17.3949, "step": 400 }, { "epoch": 0.12570532915360502, "grad_norm": 3.6819639205932617, "learning_rate": 2.005e-07, "loss": 8.8539, "step": 401 }, { "epoch": 0.12601880877742946, "grad_norm": 3.480807304382324, "learning_rate": 2.01e-07, "loss": 7.9894, "step": 402 }, { "epoch": 0.1263322884012539, "grad_norm": 4.27021598815918, "learning_rate": 2.0150000000000002e-07, "loss": 10.9579, "step": 403 }, { "epoch": 0.12664576802507838, "grad_norm": 3.2545857429504395, "learning_rate": 2.02e-07, "loss": 7.1311, "step": 404 }, { "epoch": 0.12695924764890282, "grad_norm": 3.5351126194000244, "learning_rate": 2.0250000000000002e-07, "loss": 6.8761, "step": 405 }, { "epoch": 0.12727272727272726, "grad_norm": 3.6219284534454346, "learning_rate": 2.03e-07, "loss": 8.3192, "step": 406 }, { "epoch": 0.12758620689655173, "grad_norm": 3.1589441299438477, "learning_rate": 2.0350000000000002e-07, "loss": 7.9498, "step": 407 }, { "epoch": 0.12789968652037617, "grad_norm": 3.862013101577759, "learning_rate": 2.0400000000000003e-07, "loss": 12.2729, "step": 408 }, { "epoch": 0.1282131661442006, "grad_norm": 3.1884796619415283, "learning_rate": 2.045e-07, "loss": 6.5746, "step": 409 }, { "epoch": 0.12852664576802508, "grad_norm": 2.733370065689087, "learning_rate": 2.0500000000000002e-07, "loss": 6.8392, "step": 410 }, { "epoch": 0.12884012539184952, "grad_norm": 4.183924198150635, "learning_rate": 2.055e-07, "loss": 9.7881, "step": 411 }, { "epoch": 0.129153605015674, "grad_norm": 4.186039924621582, "learning_rate": 2.0600000000000002e-07, "loss": 7.9172, "step": 412 }, { "epoch": 0.12946708463949844, "grad_norm": 4.129965305328369, "learning_rate": 2.0650000000000003e-07, "loss": 8.3458, "step": 413 }, { "epoch": 0.12978056426332288, "grad_norm": 3.3639910221099854, "learning_rate": 2.0700000000000001e-07, "loss": 7.8052, "step": 414 }, { "epoch": 0.13009404388714735, "grad_norm": 3.8352103233337402, "learning_rate": 2.0750000000000003e-07, "loss": 10.0176, "step": 415 }, { "epoch": 0.1304075235109718, "grad_norm": 3.672574281692505, "learning_rate": 2.08e-07, "loss": 10.4668, "step": 416 }, { "epoch": 0.13072100313479623, "grad_norm": 2.9237465858459473, "learning_rate": 2.0850000000000002e-07, "loss": 5.7591, "step": 417 }, { "epoch": 0.1310344827586207, "grad_norm": 4.392944812774658, "learning_rate": 2.09e-07, "loss": 10.9199, "step": 418 }, { "epoch": 0.13134796238244514, "grad_norm": 4.115564823150635, "learning_rate": 2.0950000000000002e-07, "loss": 7.3922, "step": 419 }, { "epoch": 0.13166144200626959, "grad_norm": 3.255119562149048, "learning_rate": 2.1000000000000003e-07, "loss": 7.2558, "step": 420 }, { "epoch": 0.13197492163009406, "grad_norm": 2.9578206539154053, "learning_rate": 2.105e-07, "loss": 8.0258, "step": 421 }, { "epoch": 0.1322884012539185, "grad_norm": 3.8864405155181885, "learning_rate": 2.1100000000000002e-07, "loss": 9.2595, "step": 422 }, { "epoch": 0.13260188087774294, "grad_norm": 3.2227253913879395, "learning_rate": 2.115e-07, "loss": 7.2088, "step": 423 }, { "epoch": 0.1329153605015674, "grad_norm": 3.5899605751037598, "learning_rate": 2.1200000000000002e-07, "loss": 8.3448, "step": 424 }, { "epoch": 0.13322884012539185, "grad_norm": 3.725222110748291, "learning_rate": 2.1250000000000003e-07, "loss": 8.4943, "step": 425 }, { "epoch": 0.1335423197492163, "grad_norm": 4.701845645904541, "learning_rate": 2.1300000000000001e-07, "loss": 12.9665, "step": 426 }, { "epoch": 0.13385579937304076, "grad_norm": 3.239241123199463, "learning_rate": 2.1350000000000003e-07, "loss": 7.3724, "step": 427 }, { "epoch": 0.1341692789968652, "grad_norm": 2.888547420501709, "learning_rate": 2.14e-07, "loss": 7.2136, "step": 428 }, { "epoch": 0.13448275862068965, "grad_norm": 3.1289238929748535, "learning_rate": 2.1450000000000002e-07, "loss": 8.2064, "step": 429 }, { "epoch": 0.13479623824451412, "grad_norm": 3.992865800857544, "learning_rate": 2.15e-07, "loss": 8.4046, "step": 430 }, { "epoch": 0.13510971786833856, "grad_norm": 3.2021284103393555, "learning_rate": 2.1550000000000002e-07, "loss": 12.1033, "step": 431 }, { "epoch": 0.135423197492163, "grad_norm": 3.114281177520752, "learning_rate": 2.1600000000000003e-07, "loss": 7.0616, "step": 432 }, { "epoch": 0.13573667711598747, "grad_norm": 3.175734043121338, "learning_rate": 2.165e-07, "loss": 7.1838, "step": 433 }, { "epoch": 0.1360501567398119, "grad_norm": 3.4493837356567383, "learning_rate": 2.1700000000000002e-07, "loss": 7.9487, "step": 434 }, { "epoch": 0.13636363636363635, "grad_norm": 3.04085373878479, "learning_rate": 2.175e-07, "loss": 7.6261, "step": 435 }, { "epoch": 0.13667711598746082, "grad_norm": 4.71550989151001, "learning_rate": 2.1800000000000002e-07, "loss": 13.7869, "step": 436 }, { "epoch": 0.13699059561128527, "grad_norm": 3.4197592735290527, "learning_rate": 2.1850000000000003e-07, "loss": 7.7423, "step": 437 }, { "epoch": 0.1373040752351097, "grad_norm": 3.3364362716674805, "learning_rate": 2.1900000000000002e-07, "loss": 8.9017, "step": 438 }, { "epoch": 0.13761755485893418, "grad_norm": 3.6165454387664795, "learning_rate": 2.1950000000000003e-07, "loss": 7.4735, "step": 439 }, { "epoch": 0.13793103448275862, "grad_norm": 5.154653072357178, "learning_rate": 2.2e-07, "loss": 14.2502, "step": 440 }, { "epoch": 0.13824451410658306, "grad_norm": 3.3679981231689453, "learning_rate": 2.2050000000000002e-07, "loss": 7.7449, "step": 441 }, { "epoch": 0.13855799373040753, "grad_norm": 3.918351888656616, "learning_rate": 2.2100000000000003e-07, "loss": 10.1961, "step": 442 }, { "epoch": 0.13887147335423197, "grad_norm": 3.199449300765991, "learning_rate": 2.2150000000000002e-07, "loss": 7.5792, "step": 443 }, { "epoch": 0.13918495297805641, "grad_norm": 3.322082757949829, "learning_rate": 2.2200000000000003e-07, "loss": 7.9956, "step": 444 }, { "epoch": 0.13949843260188088, "grad_norm": 3.1004014015197754, "learning_rate": 2.2250000000000001e-07, "loss": 7.0897, "step": 445 }, { "epoch": 0.13981191222570533, "grad_norm": 3.9262382984161377, "learning_rate": 2.2300000000000002e-07, "loss": 9.4931, "step": 446 }, { "epoch": 0.14012539184952977, "grad_norm": 3.103273868560791, "learning_rate": 2.235e-07, "loss": 6.7228, "step": 447 }, { "epoch": 0.14043887147335424, "grad_norm": 3.1082265377044678, "learning_rate": 2.2400000000000002e-07, "loss": 8.3279, "step": 448 }, { "epoch": 0.14075235109717868, "grad_norm": 3.3036298751831055, "learning_rate": 2.2450000000000003e-07, "loss": 7.875, "step": 449 }, { "epoch": 0.14106583072100312, "grad_norm": 3.8810064792633057, "learning_rate": 2.2500000000000002e-07, "loss": 8.3544, "step": 450 }, { "epoch": 0.1413793103448276, "grad_norm": 4.209482192993164, "learning_rate": 2.2550000000000003e-07, "loss": 9.1295, "step": 451 }, { "epoch": 0.14169278996865203, "grad_norm": 3.419205665588379, "learning_rate": 2.26e-07, "loss": 6.9125, "step": 452 }, { "epoch": 0.1420062695924765, "grad_norm": 3.985367774963379, "learning_rate": 2.2650000000000002e-07, "loss": 11.5769, "step": 453 }, { "epoch": 0.14231974921630094, "grad_norm": 3.542846202850342, "learning_rate": 2.2700000000000003e-07, "loss": 8.6733, "step": 454 }, { "epoch": 0.1426332288401254, "grad_norm": 3.4622037410736084, "learning_rate": 2.2750000000000002e-07, "loss": 9.0828, "step": 455 }, { "epoch": 0.14294670846394986, "grad_norm": 3.1978418827056885, "learning_rate": 2.2800000000000003e-07, "loss": 6.9524, "step": 456 }, { "epoch": 0.1432601880877743, "grad_norm": 3.3328914642333984, "learning_rate": 2.2850000000000001e-07, "loss": 7.4014, "step": 457 }, { "epoch": 0.14357366771159874, "grad_norm": 3.5065202713012695, "learning_rate": 2.2900000000000003e-07, "loss": 9.0877, "step": 458 }, { "epoch": 0.1438871473354232, "grad_norm": 6.033443927764893, "learning_rate": 2.2950000000000004e-07, "loss": 15.9271, "step": 459 }, { "epoch": 0.14420062695924765, "grad_norm": 4.368375301361084, "learning_rate": 2.3000000000000002e-07, "loss": 11.6588, "step": 460 }, { "epoch": 0.1445141065830721, "grad_norm": 3.0983519554138184, "learning_rate": 2.3050000000000003e-07, "loss": 6.6645, "step": 461 }, { "epoch": 0.14482758620689656, "grad_norm": 3.8745956420898438, "learning_rate": 2.3100000000000002e-07, "loss": 10.457, "step": 462 }, { "epoch": 0.145141065830721, "grad_norm": 3.9520504474639893, "learning_rate": 2.3150000000000003e-07, "loss": 10.4561, "step": 463 }, { "epoch": 0.14545454545454545, "grad_norm": 8.04711627960205, "learning_rate": 2.32e-07, "loss": 19.9331, "step": 464 }, { "epoch": 0.14576802507836992, "grad_norm": 4.129817485809326, "learning_rate": 2.3250000000000002e-07, "loss": 8.8334, "step": 465 }, { "epoch": 0.14608150470219436, "grad_norm": 4.049325466156006, "learning_rate": 2.3300000000000003e-07, "loss": 8.772, "step": 466 }, { "epoch": 0.1463949843260188, "grad_norm": 4.55631160736084, "learning_rate": 2.3350000000000002e-07, "loss": 13.48, "step": 467 }, { "epoch": 0.14670846394984327, "grad_norm": 3.488584518432617, "learning_rate": 2.3400000000000003e-07, "loss": 10.4798, "step": 468 }, { "epoch": 0.1470219435736677, "grad_norm": 5.771771430969238, "learning_rate": 2.3450000000000002e-07, "loss": 11.7797, "step": 469 }, { "epoch": 0.14733542319749215, "grad_norm": 3.116539716720581, "learning_rate": 2.3500000000000003e-07, "loss": 6.873, "step": 470 }, { "epoch": 0.14764890282131662, "grad_norm": 2.576658248901367, "learning_rate": 2.3550000000000004e-07, "loss": 7.21, "step": 471 }, { "epoch": 0.14796238244514107, "grad_norm": 3.1541895866394043, "learning_rate": 2.3600000000000002e-07, "loss": 7.4172, "step": 472 }, { "epoch": 0.1482758620689655, "grad_norm": 4.149017810821533, "learning_rate": 2.3650000000000003e-07, "loss": 13.571, "step": 473 }, { "epoch": 0.14858934169278998, "grad_norm": 4.210467338562012, "learning_rate": 2.3700000000000002e-07, "loss": 11.6499, "step": 474 }, { "epoch": 0.14890282131661442, "grad_norm": 5.193135738372803, "learning_rate": 2.3750000000000003e-07, "loss": 17.2001, "step": 475 }, { "epoch": 0.14921630094043886, "grad_norm": 3.6134843826293945, "learning_rate": 2.3800000000000004e-07, "loss": 6.4639, "step": 476 }, { "epoch": 0.14952978056426333, "grad_norm": 5.032776355743408, "learning_rate": 2.385e-07, "loss": 13.8185, "step": 477 }, { "epoch": 0.14984326018808777, "grad_norm": 4.616049766540527, "learning_rate": 2.39e-07, "loss": 10.5861, "step": 478 }, { "epoch": 0.15015673981191222, "grad_norm": 3.1708860397338867, "learning_rate": 2.395e-07, "loss": 7.8613, "step": 479 }, { "epoch": 0.15047021943573669, "grad_norm": 4.432284355163574, "learning_rate": 2.4000000000000003e-07, "loss": 10.7543, "step": 480 }, { "epoch": 0.15078369905956113, "grad_norm": 3.439668655395508, "learning_rate": 2.405e-07, "loss": 7.4266, "step": 481 }, { "epoch": 0.15109717868338557, "grad_norm": 4.03348445892334, "learning_rate": 2.41e-07, "loss": 11.5268, "step": 482 }, { "epoch": 0.15141065830721004, "grad_norm": 3.4068636894226074, "learning_rate": 2.4150000000000004e-07, "loss": 7.1183, "step": 483 }, { "epoch": 0.15172413793103448, "grad_norm": 4.846498966217041, "learning_rate": 2.42e-07, "loss": 11.0191, "step": 484 }, { "epoch": 0.15203761755485892, "grad_norm": 5.300605773925781, "learning_rate": 2.425e-07, "loss": 9.6552, "step": 485 }, { "epoch": 0.1523510971786834, "grad_norm": 4.646756649017334, "learning_rate": 2.43e-07, "loss": 13.7772, "step": 486 }, { "epoch": 0.15266457680250783, "grad_norm": 3.7247793674468994, "learning_rate": 2.4350000000000003e-07, "loss": 8.6148, "step": 487 }, { "epoch": 0.15297805642633228, "grad_norm": 3.396324634552002, "learning_rate": 2.44e-07, "loss": 11.5777, "step": 488 }, { "epoch": 0.15329153605015675, "grad_norm": 3.7958171367645264, "learning_rate": 2.445e-07, "loss": 8.7151, "step": 489 }, { "epoch": 0.1536050156739812, "grad_norm": 3.312392473220825, "learning_rate": 2.4500000000000004e-07, "loss": 8.2649, "step": 490 }, { "epoch": 0.15391849529780563, "grad_norm": 2.4669861793518066, "learning_rate": 2.455e-07, "loss": 6.4529, "step": 491 }, { "epoch": 0.1542319749216301, "grad_norm": 5.108162879943848, "learning_rate": 2.46e-07, "loss": 14.7079, "step": 492 }, { "epoch": 0.15454545454545454, "grad_norm": 3.430584192276001, "learning_rate": 2.465e-07, "loss": 8.4653, "step": 493 }, { "epoch": 0.15485893416927898, "grad_norm": 3.0151519775390625, "learning_rate": 2.4700000000000003e-07, "loss": 6.9225, "step": 494 }, { "epoch": 0.15517241379310345, "grad_norm": 4.027904510498047, "learning_rate": 2.475e-07, "loss": 8.5251, "step": 495 }, { "epoch": 0.1554858934169279, "grad_norm": 3.500556707382202, "learning_rate": 2.48e-07, "loss": 9.9441, "step": 496 }, { "epoch": 0.15579937304075236, "grad_norm": 3.032302141189575, "learning_rate": 2.4850000000000003e-07, "loss": 6.9624, "step": 497 }, { "epoch": 0.1561128526645768, "grad_norm": 4.127997875213623, "learning_rate": 2.49e-07, "loss": 9.5575, "step": 498 }, { "epoch": 0.15642633228840125, "grad_norm": 3.3242340087890625, "learning_rate": 2.495e-07, "loss": 7.8737, "step": 499 }, { "epoch": 0.15673981191222572, "grad_norm": 3.514852285385132, "learning_rate": 2.5000000000000004e-07, "loss": 7.5601, "step": 500 }, { "epoch": 0.15705329153605016, "grad_norm": 3.9590742588043213, "learning_rate": 2.505e-07, "loss": 8.2909, "step": 501 }, { "epoch": 0.1573667711598746, "grad_norm": 3.0089352130889893, "learning_rate": 2.51e-07, "loss": 8.9848, "step": 502 }, { "epoch": 0.15768025078369907, "grad_norm": 4.294312477111816, "learning_rate": 2.515e-07, "loss": 10.3314, "step": 503 }, { "epoch": 0.1579937304075235, "grad_norm": 4.103808879852295, "learning_rate": 2.5200000000000003e-07, "loss": 9.262, "step": 504 }, { "epoch": 0.15830721003134796, "grad_norm": 3.4118666648864746, "learning_rate": 2.525e-07, "loss": 7.6306, "step": 505 }, { "epoch": 0.15862068965517243, "grad_norm": 3.144343137741089, "learning_rate": 2.53e-07, "loss": 9.7107, "step": 506 }, { "epoch": 0.15893416927899687, "grad_norm": 3.416949987411499, "learning_rate": 2.5350000000000004e-07, "loss": 7.1229, "step": 507 }, { "epoch": 0.1592476489028213, "grad_norm": 3.30198335647583, "learning_rate": 2.54e-07, "loss": 8.9435, "step": 508 }, { "epoch": 0.15956112852664578, "grad_norm": 3.7316057682037354, "learning_rate": 2.545e-07, "loss": 8.5075, "step": 509 }, { "epoch": 0.15987460815047022, "grad_norm": 3.468623638153076, "learning_rate": 2.55e-07, "loss": 8.5576, "step": 510 }, { "epoch": 0.16018808777429466, "grad_norm": 3.579921245574951, "learning_rate": 2.5550000000000003e-07, "loss": 8.5891, "step": 511 }, { "epoch": 0.16050156739811913, "grad_norm": 2.719144105911255, "learning_rate": 2.56e-07, "loss": 7.0097, "step": 512 }, { "epoch": 0.16081504702194357, "grad_norm": 3.503176689147949, "learning_rate": 2.565e-07, "loss": 9.2646, "step": 513 }, { "epoch": 0.16112852664576802, "grad_norm": 3.4942426681518555, "learning_rate": 2.5700000000000004e-07, "loss": 8.3605, "step": 514 }, { "epoch": 0.1614420062695925, "grad_norm": 3.6798880100250244, "learning_rate": 2.575e-07, "loss": 7.8945, "step": 515 }, { "epoch": 0.16175548589341693, "grad_norm": 2.953244924545288, "learning_rate": 2.58e-07, "loss": 6.319, "step": 516 }, { "epoch": 0.16206896551724137, "grad_norm": 4.936113357543945, "learning_rate": 2.5850000000000004e-07, "loss": 13.3042, "step": 517 }, { "epoch": 0.16238244514106584, "grad_norm": 4.359250545501709, "learning_rate": 2.5900000000000003e-07, "loss": 8.4494, "step": 518 }, { "epoch": 0.16269592476489028, "grad_norm": 3.1021082401275635, "learning_rate": 2.595e-07, "loss": 8.4413, "step": 519 }, { "epoch": 0.16300940438871472, "grad_norm": 3.71295166015625, "learning_rate": 2.6e-07, "loss": 9.4076, "step": 520 }, { "epoch": 0.1633228840125392, "grad_norm": 3.4278814792633057, "learning_rate": 2.6050000000000004e-07, "loss": 8.9707, "step": 521 }, { "epoch": 0.16363636363636364, "grad_norm": 4.640440464019775, "learning_rate": 2.61e-07, "loss": 12.3595, "step": 522 }, { "epoch": 0.16394984326018808, "grad_norm": 4.346306800842285, "learning_rate": 2.615e-07, "loss": 11.8391, "step": 523 }, { "epoch": 0.16426332288401255, "grad_norm": 3.444598436355591, "learning_rate": 2.6200000000000004e-07, "loss": 8.5561, "step": 524 }, { "epoch": 0.164576802507837, "grad_norm": 3.9495432376861572, "learning_rate": 2.6250000000000003e-07, "loss": 11.1607, "step": 525 }, { "epoch": 0.16489028213166143, "grad_norm": 3.858132839202881, "learning_rate": 2.63e-07, "loss": 8.4288, "step": 526 }, { "epoch": 0.1652037617554859, "grad_norm": 4.880698204040527, "learning_rate": 2.635e-07, "loss": 17.5724, "step": 527 }, { "epoch": 0.16551724137931034, "grad_norm": 5.0397162437438965, "learning_rate": 2.6400000000000003e-07, "loss": 10.3173, "step": 528 }, { "epoch": 0.16583072100313478, "grad_norm": 3.1536805629730225, "learning_rate": 2.645e-07, "loss": 7.8392, "step": 529 }, { "epoch": 0.16614420062695925, "grad_norm": 2.8486595153808594, "learning_rate": 2.65e-07, "loss": 7.3892, "step": 530 }, { "epoch": 0.1664576802507837, "grad_norm": 3.6035428047180176, "learning_rate": 2.6550000000000004e-07, "loss": 8.3855, "step": 531 }, { "epoch": 0.16677115987460814, "grad_norm": 3.0228993892669678, "learning_rate": 2.66e-07, "loss": 7.0978, "step": 532 }, { "epoch": 0.1670846394984326, "grad_norm": 3.3057663440704346, "learning_rate": 2.665e-07, "loss": 8.4415, "step": 533 }, { "epoch": 0.16739811912225705, "grad_norm": 3.630830764770508, "learning_rate": 2.6700000000000005e-07, "loss": 7.5833, "step": 534 }, { "epoch": 0.1677115987460815, "grad_norm": 4.190485954284668, "learning_rate": 2.6750000000000003e-07, "loss": 6.7527, "step": 535 }, { "epoch": 0.16802507836990596, "grad_norm": 4.3880486488342285, "learning_rate": 2.68e-07, "loss": 10.2591, "step": 536 }, { "epoch": 0.1683385579937304, "grad_norm": 3.767117738723755, "learning_rate": 2.685e-07, "loss": 6.5155, "step": 537 }, { "epoch": 0.16865203761755485, "grad_norm": 4.589762210845947, "learning_rate": 2.6900000000000004e-07, "loss": 11.3011, "step": 538 }, { "epoch": 0.16896551724137931, "grad_norm": 2.933931350708008, "learning_rate": 2.695e-07, "loss": 7.7405, "step": 539 }, { "epoch": 0.16927899686520376, "grad_norm": 4.077798843383789, "learning_rate": 2.7e-07, "loss": 7.7766, "step": 540 }, { "epoch": 0.16959247648902823, "grad_norm": 4.123392581939697, "learning_rate": 2.7050000000000005e-07, "loss": 8.437, "step": 541 }, { "epoch": 0.16990595611285267, "grad_norm": 3.4657938480377197, "learning_rate": 2.7100000000000003e-07, "loss": 8.0518, "step": 542 }, { "epoch": 0.1702194357366771, "grad_norm": 3.4541122913360596, "learning_rate": 2.715e-07, "loss": 8.7644, "step": 543 }, { "epoch": 0.17053291536050158, "grad_norm": 4.604305267333984, "learning_rate": 2.72e-07, "loss": 10.5831, "step": 544 }, { "epoch": 0.17084639498432602, "grad_norm": 3.2504711151123047, "learning_rate": 2.7250000000000004e-07, "loss": 8.0843, "step": 545 }, { "epoch": 0.17115987460815046, "grad_norm": 3.421630382537842, "learning_rate": 2.73e-07, "loss": 7.3776, "step": 546 }, { "epoch": 0.17147335423197493, "grad_norm": 3.5182442665100098, "learning_rate": 2.735e-07, "loss": 7.4899, "step": 547 }, { "epoch": 0.17178683385579938, "grad_norm": 3.437476396560669, "learning_rate": 2.7400000000000004e-07, "loss": 8.6969, "step": 548 }, { "epoch": 0.17210031347962382, "grad_norm": 3.5565860271453857, "learning_rate": 2.7450000000000003e-07, "loss": 7.6131, "step": 549 }, { "epoch": 0.1724137931034483, "grad_norm": 4.052206516265869, "learning_rate": 2.75e-07, "loss": 10.5389, "step": 550 }, { "epoch": 0.17272727272727273, "grad_norm": 4.916564464569092, "learning_rate": 2.7550000000000005e-07, "loss": 13.2927, "step": 551 }, { "epoch": 0.17304075235109717, "grad_norm": 2.8071060180664062, "learning_rate": 2.7600000000000004e-07, "loss": 7.181, "step": 552 }, { "epoch": 0.17335423197492164, "grad_norm": 3.762843370437622, "learning_rate": 2.765e-07, "loss": 9.2195, "step": 553 }, { "epoch": 0.17366771159874608, "grad_norm": 4.453097820281982, "learning_rate": 2.77e-07, "loss": 11.1995, "step": 554 }, { "epoch": 0.17398119122257052, "grad_norm": 3.8368940353393555, "learning_rate": 2.7750000000000004e-07, "loss": 10.7022, "step": 555 }, { "epoch": 0.174294670846395, "grad_norm": 3.128065586090088, "learning_rate": 2.7800000000000003e-07, "loss": 7.396, "step": 556 }, { "epoch": 0.17460815047021944, "grad_norm": 3.999136447906494, "learning_rate": 2.785e-07, "loss": 9.3729, "step": 557 }, { "epoch": 0.17492163009404388, "grad_norm": 3.374950408935547, "learning_rate": 2.7900000000000005e-07, "loss": 6.9749, "step": 558 }, { "epoch": 0.17523510971786835, "grad_norm": 3.5031139850616455, "learning_rate": 2.7950000000000003e-07, "loss": 8.5784, "step": 559 }, { "epoch": 0.1755485893416928, "grad_norm": 3.912729263305664, "learning_rate": 2.8e-07, "loss": 8.9176, "step": 560 }, { "epoch": 0.17586206896551723, "grad_norm": 4.987295627593994, "learning_rate": 2.805e-07, "loss": 10.5289, "step": 561 }, { "epoch": 0.1761755485893417, "grad_norm": 5.127312183380127, "learning_rate": 2.8100000000000004e-07, "loss": 12.005, "step": 562 }, { "epoch": 0.17648902821316614, "grad_norm": 2.887765884399414, "learning_rate": 2.815e-07, "loss": 8.8117, "step": 563 }, { "epoch": 0.17680250783699059, "grad_norm": 3.11470890045166, "learning_rate": 2.82e-07, "loss": 9.1879, "step": 564 }, { "epoch": 0.17711598746081506, "grad_norm": 3.491626739501953, "learning_rate": 2.8250000000000005e-07, "loss": 8.4029, "step": 565 }, { "epoch": 0.1774294670846395, "grad_norm": 3.3320202827453613, "learning_rate": 2.8300000000000003e-07, "loss": 7.258, "step": 566 }, { "epoch": 0.17774294670846394, "grad_norm": 3.7208235263824463, "learning_rate": 2.835e-07, "loss": 9.4842, "step": 567 }, { "epoch": 0.1780564263322884, "grad_norm": 3.597604274749756, "learning_rate": 2.8400000000000005e-07, "loss": 8.5819, "step": 568 }, { "epoch": 0.17836990595611285, "grad_norm": 3.1107091903686523, "learning_rate": 2.8450000000000004e-07, "loss": 7.8507, "step": 569 }, { "epoch": 0.1786833855799373, "grad_norm": 2.587881565093994, "learning_rate": 2.85e-07, "loss": 7.2284, "step": 570 }, { "epoch": 0.17899686520376176, "grad_norm": 2.9738922119140625, "learning_rate": 2.855e-07, "loss": 8.5117, "step": 571 }, { "epoch": 0.1793103448275862, "grad_norm": 3.6646111011505127, "learning_rate": 2.8600000000000005e-07, "loss": 8.4424, "step": 572 }, { "epoch": 0.17962382445141065, "grad_norm": 3.749894380569458, "learning_rate": 2.8650000000000003e-07, "loss": 8.9965, "step": 573 }, { "epoch": 0.17993730407523512, "grad_norm": 2.9281930923461914, "learning_rate": 2.87e-07, "loss": 6.6291, "step": 574 }, { "epoch": 0.18025078369905956, "grad_norm": 5.141971111297607, "learning_rate": 2.8750000000000005e-07, "loss": 11.6456, "step": 575 }, { "epoch": 0.180564263322884, "grad_norm": 3.058166980743408, "learning_rate": 2.8800000000000004e-07, "loss": 7.7736, "step": 576 }, { "epoch": 0.18087774294670847, "grad_norm": 3.9422664642333984, "learning_rate": 2.885e-07, "loss": 11.6602, "step": 577 }, { "epoch": 0.1811912225705329, "grad_norm": 4.253895282745361, "learning_rate": 2.89e-07, "loss": 9.2913, "step": 578 }, { "epoch": 0.18150470219435735, "grad_norm": 3.428746461868286, "learning_rate": 2.8950000000000004e-07, "loss": 7.2803, "step": 579 }, { "epoch": 0.18181818181818182, "grad_norm": 2.983686685562134, "learning_rate": 2.9000000000000003e-07, "loss": 7.1757, "step": 580 }, { "epoch": 0.18213166144200627, "grad_norm": 3.508849859237671, "learning_rate": 2.905e-07, "loss": 8.3138, "step": 581 }, { "epoch": 0.18244514106583073, "grad_norm": 3.7815003395080566, "learning_rate": 2.9100000000000005e-07, "loss": 6.8592, "step": 582 }, { "epoch": 0.18275862068965518, "grad_norm": 3.492091417312622, "learning_rate": 2.9150000000000004e-07, "loss": 9.769, "step": 583 }, { "epoch": 0.18307210031347962, "grad_norm": 3.9170730113983154, "learning_rate": 2.92e-07, "loss": 6.7468, "step": 584 }, { "epoch": 0.1833855799373041, "grad_norm": 3.778157949447632, "learning_rate": 2.9250000000000006e-07, "loss": 7.6754, "step": 585 }, { "epoch": 0.18369905956112853, "grad_norm": 3.811721086502075, "learning_rate": 2.9300000000000004e-07, "loss": 9.9387, "step": 586 }, { "epoch": 0.18401253918495297, "grad_norm": 4.163631916046143, "learning_rate": 2.9350000000000003e-07, "loss": 8.2871, "step": 587 }, { "epoch": 0.18432601880877744, "grad_norm": 3.4269614219665527, "learning_rate": 2.94e-07, "loss": 7.114, "step": 588 }, { "epoch": 0.18463949843260188, "grad_norm": 3.990867853164673, "learning_rate": 2.9450000000000005e-07, "loss": 9.8128, "step": 589 }, { "epoch": 0.18495297805642633, "grad_norm": 3.4142584800720215, "learning_rate": 2.9500000000000003e-07, "loss": 7.7801, "step": 590 }, { "epoch": 0.1852664576802508, "grad_norm": 3.636627674102783, "learning_rate": 2.955e-07, "loss": 10.9509, "step": 591 }, { "epoch": 0.18557993730407524, "grad_norm": 4.59920072555542, "learning_rate": 2.9600000000000006e-07, "loss": 10.6065, "step": 592 }, { "epoch": 0.18589341692789968, "grad_norm": 3.2188518047332764, "learning_rate": 2.9650000000000004e-07, "loss": 7.1557, "step": 593 }, { "epoch": 0.18620689655172415, "grad_norm": 3.7860870361328125, "learning_rate": 2.97e-07, "loss": 7.7989, "step": 594 }, { "epoch": 0.1865203761755486, "grad_norm": 4.25967264175415, "learning_rate": 2.975e-07, "loss": 12.088, "step": 595 }, { "epoch": 0.18683385579937303, "grad_norm": 2.7574448585510254, "learning_rate": 2.9800000000000005e-07, "loss": 5.9382, "step": 596 }, { "epoch": 0.1871473354231975, "grad_norm": 3.17261004447937, "learning_rate": 2.9850000000000003e-07, "loss": 7.0805, "step": 597 }, { "epoch": 0.18746081504702194, "grad_norm": 4.3467793464660645, "learning_rate": 2.99e-07, "loss": 11.9778, "step": 598 }, { "epoch": 0.1877742946708464, "grad_norm": 4.424931526184082, "learning_rate": 2.9950000000000005e-07, "loss": 8.564, "step": 599 }, { "epoch": 0.18808777429467086, "grad_norm": 3.811779737472534, "learning_rate": 3.0000000000000004e-07, "loss": 7.9092, "step": 600 }, { "epoch": 0.1884012539184953, "grad_norm": 4.103071689605713, "learning_rate": 3.005e-07, "loss": 9.164, "step": 601 }, { "epoch": 0.18871473354231974, "grad_norm": 3.3931884765625, "learning_rate": 3.01e-07, "loss": 7.6855, "step": 602 }, { "epoch": 0.1890282131661442, "grad_norm": 3.917874574661255, "learning_rate": 3.0150000000000005e-07, "loss": 9.6698, "step": 603 }, { "epoch": 0.18934169278996865, "grad_norm": 4.343447685241699, "learning_rate": 3.0200000000000003e-07, "loss": 8.4255, "step": 604 }, { "epoch": 0.1896551724137931, "grad_norm": 4.0894927978515625, "learning_rate": 3.025e-07, "loss": 11.4714, "step": 605 }, { "epoch": 0.18996865203761756, "grad_norm": 4.6894073486328125, "learning_rate": 3.0300000000000005e-07, "loss": 10.4595, "step": 606 }, { "epoch": 0.190282131661442, "grad_norm": 2.9865052700042725, "learning_rate": 3.035e-07, "loss": 6.7574, "step": 607 }, { "epoch": 0.19059561128526645, "grad_norm": 3.2500112056732178, "learning_rate": 3.04e-07, "loss": 7.7867, "step": 608 }, { "epoch": 0.19090909090909092, "grad_norm": 3.837984323501587, "learning_rate": 3.0450000000000006e-07, "loss": 8.9823, "step": 609 }, { "epoch": 0.19122257053291536, "grad_norm": 5.894450664520264, "learning_rate": 3.0500000000000004e-07, "loss": 15.5915, "step": 610 }, { "epoch": 0.1915360501567398, "grad_norm": 3.60793137550354, "learning_rate": 3.0550000000000003e-07, "loss": 7.3988, "step": 611 }, { "epoch": 0.19184952978056427, "grad_norm": 3.7032573223114014, "learning_rate": 3.06e-07, "loss": 7.1858, "step": 612 }, { "epoch": 0.1921630094043887, "grad_norm": 3.515413522720337, "learning_rate": 3.0650000000000005e-07, "loss": 8.9439, "step": 613 }, { "epoch": 0.19247648902821315, "grad_norm": 4.349193096160889, "learning_rate": 3.0700000000000004e-07, "loss": 10.2601, "step": 614 }, { "epoch": 0.19278996865203762, "grad_norm": 3.940768241882324, "learning_rate": 3.075e-07, "loss": 10.1987, "step": 615 }, { "epoch": 0.19310344827586207, "grad_norm": 2.720353841781616, "learning_rate": 3.0800000000000006e-07, "loss": 6.7377, "step": 616 }, { "epoch": 0.1934169278996865, "grad_norm": 3.8752973079681396, "learning_rate": 3.085e-07, "loss": 8.5383, "step": 617 }, { "epoch": 0.19373040752351098, "grad_norm": 3.49699330329895, "learning_rate": 3.0900000000000003e-07, "loss": 8.1884, "step": 618 }, { "epoch": 0.19404388714733542, "grad_norm": 3.857430934906006, "learning_rate": 3.095e-07, "loss": 8.9309, "step": 619 }, { "epoch": 0.19435736677115986, "grad_norm": 3.196040153503418, "learning_rate": 3.1000000000000005e-07, "loss": 7.5568, "step": 620 }, { "epoch": 0.19467084639498433, "grad_norm": 3.7832367420196533, "learning_rate": 3.1050000000000003e-07, "loss": 9.6145, "step": 621 }, { "epoch": 0.19498432601880877, "grad_norm": 3.3154568672180176, "learning_rate": 3.11e-07, "loss": 7.3691, "step": 622 }, { "epoch": 0.19529780564263322, "grad_norm": 5.462655067443848, "learning_rate": 3.1150000000000006e-07, "loss": 12.7679, "step": 623 }, { "epoch": 0.19561128526645769, "grad_norm": 4.170969009399414, "learning_rate": 3.12e-07, "loss": 11.6538, "step": 624 }, { "epoch": 0.19592476489028213, "grad_norm": 5.751269817352295, "learning_rate": 3.125e-07, "loss": 13.9954, "step": 625 }, { "epoch": 0.1962382445141066, "grad_norm": 4.067015171051025, "learning_rate": 3.1300000000000006e-07, "loss": 6.5656, "step": 626 }, { "epoch": 0.19655172413793104, "grad_norm": 3.084688186645508, "learning_rate": 3.1350000000000005e-07, "loss": 8.2132, "step": 627 }, { "epoch": 0.19686520376175548, "grad_norm": 2.9886057376861572, "learning_rate": 3.14e-07, "loss": 6.7756, "step": 628 }, { "epoch": 0.19717868338557995, "grad_norm": 3.593235969543457, "learning_rate": 3.145e-07, "loss": 8.3709, "step": 629 }, { "epoch": 0.1974921630094044, "grad_norm": 3.363783597946167, "learning_rate": 3.15e-07, "loss": 10.0533, "step": 630 }, { "epoch": 0.19780564263322883, "grad_norm": 4.221756935119629, "learning_rate": 3.1550000000000004e-07, "loss": 11.3121, "step": 631 }, { "epoch": 0.1981191222570533, "grad_norm": 4.736835956573486, "learning_rate": 3.160000000000001e-07, "loss": 12.493, "step": 632 }, { "epoch": 0.19843260188087775, "grad_norm": 3.1203339099884033, "learning_rate": 3.165e-07, "loss": 7.1578, "step": 633 }, { "epoch": 0.1987460815047022, "grad_norm": 3.5580482482910156, "learning_rate": 3.17e-07, "loss": 7.742, "step": 634 }, { "epoch": 0.19905956112852666, "grad_norm": 3.4418790340423584, "learning_rate": 3.1750000000000003e-07, "loss": 9.445, "step": 635 }, { "epoch": 0.1993730407523511, "grad_norm": 3.2233057022094727, "learning_rate": 3.1800000000000007e-07, "loss": 8.102, "step": 636 }, { "epoch": 0.19968652037617554, "grad_norm": 3.335604190826416, "learning_rate": 3.1850000000000005e-07, "loss": 6.6202, "step": 637 }, { "epoch": 0.2, "grad_norm": 2.8534433841705322, "learning_rate": 3.19e-07, "loss": 7.2855, "step": 638 }, { "epoch": 0.20031347962382445, "grad_norm": 3.358614683151245, "learning_rate": 3.195e-07, "loss": 9.9256, "step": 639 }, { "epoch": 0.2006269592476489, "grad_norm": 3.0653884410858154, "learning_rate": 3.2e-07, "loss": 6.773, "step": 640 }, { "epoch": 0.20094043887147336, "grad_norm": 3.5670981407165527, "learning_rate": 3.2050000000000004e-07, "loss": 10.0116, "step": 641 }, { "epoch": 0.2012539184952978, "grad_norm": 3.6908740997314453, "learning_rate": 3.21e-07, "loss": 9.244, "step": 642 }, { "epoch": 0.20156739811912225, "grad_norm": 4.45837926864624, "learning_rate": 3.215e-07, "loss": 7.975, "step": 643 }, { "epoch": 0.20188087774294672, "grad_norm": 3.217262029647827, "learning_rate": 3.22e-07, "loss": 7.2521, "step": 644 }, { "epoch": 0.20219435736677116, "grad_norm": 3.4965760707855225, "learning_rate": 3.2250000000000004e-07, "loss": 7.2706, "step": 645 }, { "epoch": 0.2025078369905956, "grad_norm": 3.6670267581939697, "learning_rate": 3.2300000000000007e-07, "loss": 7.8659, "step": 646 }, { "epoch": 0.20282131661442007, "grad_norm": 3.4865596294403076, "learning_rate": 3.235e-07, "loss": 8.019, "step": 647 }, { "epoch": 0.2031347962382445, "grad_norm": 3.0378615856170654, "learning_rate": 3.24e-07, "loss": 8.2733, "step": 648 }, { "epoch": 0.20344827586206896, "grad_norm": 4.101233005523682, "learning_rate": 3.2450000000000003e-07, "loss": 8.759, "step": 649 }, { "epoch": 0.20376175548589343, "grad_norm": 3.027646064758301, "learning_rate": 3.25e-07, "loss": 7.4447, "step": 650 }, { "epoch": 0.20407523510971787, "grad_norm": 4.034641742706299, "learning_rate": 3.2550000000000005e-07, "loss": 10.331, "step": 651 }, { "epoch": 0.2043887147335423, "grad_norm": 3.4955785274505615, "learning_rate": 3.26e-07, "loss": 8.214, "step": 652 }, { "epoch": 0.20470219435736678, "grad_norm": 3.91501784324646, "learning_rate": 3.265e-07, "loss": 8.5537, "step": 653 }, { "epoch": 0.20501567398119122, "grad_norm": 3.4765992164611816, "learning_rate": 3.27e-07, "loss": 8.4838, "step": 654 }, { "epoch": 0.20532915360501566, "grad_norm": 2.9636855125427246, "learning_rate": 3.2750000000000004e-07, "loss": 7.3184, "step": 655 }, { "epoch": 0.20564263322884013, "grad_norm": 3.7056450843811035, "learning_rate": 3.280000000000001e-07, "loss": 8.578, "step": 656 }, { "epoch": 0.20595611285266457, "grad_norm": 5.734414100646973, "learning_rate": 3.285e-07, "loss": 19.8354, "step": 657 }, { "epoch": 0.20626959247648902, "grad_norm": 4.2713303565979, "learning_rate": 3.29e-07, "loss": 12.1838, "step": 658 }, { "epoch": 0.2065830721003135, "grad_norm": 3.7173306941986084, "learning_rate": 3.2950000000000003e-07, "loss": 8.0242, "step": 659 }, { "epoch": 0.20689655172413793, "grad_norm": 3.0381829738616943, "learning_rate": 3.3e-07, "loss": 7.6178, "step": 660 }, { "epoch": 0.20721003134796237, "grad_norm": 3.788667678833008, "learning_rate": 3.3050000000000005e-07, "loss": 9.3861, "step": 661 }, { "epoch": 0.20752351097178684, "grad_norm": 4.688927173614502, "learning_rate": 3.31e-07, "loss": 9.3514, "step": 662 }, { "epoch": 0.20783699059561128, "grad_norm": 4.714510440826416, "learning_rate": 3.315e-07, "loss": 11.6471, "step": 663 }, { "epoch": 0.20815047021943572, "grad_norm": 3.319627046585083, "learning_rate": 3.32e-07, "loss": 6.9164, "step": 664 }, { "epoch": 0.2084639498432602, "grad_norm": 4.469924449920654, "learning_rate": 3.3250000000000005e-07, "loss": 10.5332, "step": 665 }, { "epoch": 0.20877742946708464, "grad_norm": 3.59578537940979, "learning_rate": 3.330000000000001e-07, "loss": 7.8451, "step": 666 }, { "epoch": 0.20909090909090908, "grad_norm": 3.3554725646972656, "learning_rate": 3.335e-07, "loss": 8.2171, "step": 667 }, { "epoch": 0.20940438871473355, "grad_norm": 3.555884599685669, "learning_rate": 3.34e-07, "loss": 9.0251, "step": 668 }, { "epoch": 0.209717868338558, "grad_norm": 3.4159555435180664, "learning_rate": 3.3450000000000004e-07, "loss": 8.1913, "step": 669 }, { "epoch": 0.21003134796238246, "grad_norm": 3.6287314891815186, "learning_rate": 3.35e-07, "loss": 7.8455, "step": 670 }, { "epoch": 0.2103448275862069, "grad_norm": 5.647161960601807, "learning_rate": 3.3550000000000006e-07, "loss": 14.64, "step": 671 }, { "epoch": 0.21065830721003134, "grad_norm": 3.4719080924987793, "learning_rate": 3.36e-07, "loss": 8.0084, "step": 672 }, { "epoch": 0.2109717868338558, "grad_norm": 3.0406198501586914, "learning_rate": 3.3650000000000003e-07, "loss": 8.0089, "step": 673 }, { "epoch": 0.21128526645768025, "grad_norm": 3.4766266345977783, "learning_rate": 3.37e-07, "loss": 8.2851, "step": 674 }, { "epoch": 0.2115987460815047, "grad_norm": 4.269450664520264, "learning_rate": 3.3750000000000005e-07, "loss": 8.4858, "step": 675 }, { "epoch": 0.21191222570532917, "grad_norm": 4.416221618652344, "learning_rate": 3.38e-07, "loss": 8.1825, "step": 676 }, { "epoch": 0.2122257053291536, "grad_norm": 3.905118465423584, "learning_rate": 3.385e-07, "loss": 8.2382, "step": 677 }, { "epoch": 0.21253918495297805, "grad_norm": 3.801635980606079, "learning_rate": 3.39e-07, "loss": 8.5147, "step": 678 }, { "epoch": 0.21285266457680252, "grad_norm": 3.1189794540405273, "learning_rate": 3.3950000000000004e-07, "loss": 8.2367, "step": 679 }, { "epoch": 0.21316614420062696, "grad_norm": 3.4691970348358154, "learning_rate": 3.4000000000000003e-07, "loss": 8.0238, "step": 680 }, { "epoch": 0.2134796238244514, "grad_norm": 2.9724643230438232, "learning_rate": 3.405e-07, "loss": 7.9417, "step": 681 }, { "epoch": 0.21379310344827587, "grad_norm": 3.1940643787384033, "learning_rate": 3.41e-07, "loss": 8.5166, "step": 682 }, { "epoch": 0.21410658307210031, "grad_norm": 3.5346059799194336, "learning_rate": 3.4150000000000003e-07, "loss": 7.2445, "step": 683 }, { "epoch": 0.21442006269592476, "grad_norm": 3.558433771133423, "learning_rate": 3.42e-07, "loss": 6.4688, "step": 684 }, { "epoch": 0.21473354231974923, "grad_norm": 3.95771861076355, "learning_rate": 3.4250000000000006e-07, "loss": 9.9655, "step": 685 }, { "epoch": 0.21504702194357367, "grad_norm": 3.4017562866210938, "learning_rate": 3.43e-07, "loss": 8.1479, "step": 686 }, { "epoch": 0.2153605015673981, "grad_norm": 3.312255859375, "learning_rate": 3.435e-07, "loss": 7.3946, "step": 687 }, { "epoch": 0.21567398119122258, "grad_norm": 3.0924856662750244, "learning_rate": 3.44e-07, "loss": 7.1558, "step": 688 }, { "epoch": 0.21598746081504702, "grad_norm": 4.0401482582092285, "learning_rate": 3.4450000000000005e-07, "loss": 8.9632, "step": 689 }, { "epoch": 0.21630094043887146, "grad_norm": 3.2518508434295654, "learning_rate": 3.4500000000000003e-07, "loss": 7.0506, "step": 690 }, { "epoch": 0.21661442006269593, "grad_norm": 3.050063133239746, "learning_rate": 3.455e-07, "loss": 7.7591, "step": 691 }, { "epoch": 0.21692789968652038, "grad_norm": 4.7669997215271, "learning_rate": 3.46e-07, "loss": 9.7782, "step": 692 }, { "epoch": 0.21724137931034482, "grad_norm": 3.25658917427063, "learning_rate": 3.4650000000000004e-07, "loss": 7.9233, "step": 693 }, { "epoch": 0.2175548589341693, "grad_norm": 2.796995162963867, "learning_rate": 3.47e-07, "loss": 6.1858, "step": 694 }, { "epoch": 0.21786833855799373, "grad_norm": 3.696262836456299, "learning_rate": 3.4750000000000006e-07, "loss": 8.2826, "step": 695 }, { "epoch": 0.21818181818181817, "grad_norm": 4.308631896972656, "learning_rate": 3.48e-07, "loss": 12.5166, "step": 696 }, { "epoch": 0.21849529780564264, "grad_norm": 2.7071919441223145, "learning_rate": 3.4850000000000003e-07, "loss": 6.3807, "step": 697 }, { "epoch": 0.21880877742946708, "grad_norm": 4.019406795501709, "learning_rate": 3.49e-07, "loss": 11.1741, "step": 698 }, { "epoch": 0.21912225705329152, "grad_norm": 4.2864580154418945, "learning_rate": 3.4950000000000005e-07, "loss": 9.779, "step": 699 }, { "epoch": 0.219435736677116, "grad_norm": 3.9045979976654053, "learning_rate": 3.5000000000000004e-07, "loss": 6.9983, "step": 700 }, { "epoch": 0.21974921630094044, "grad_norm": 3.1046054363250732, "learning_rate": 3.505e-07, "loss": 7.6675, "step": 701 }, { "epoch": 0.22006269592476488, "grad_norm": 3.561466693878174, "learning_rate": 3.51e-07, "loss": 5.8858, "step": 702 }, { "epoch": 0.22037617554858935, "grad_norm": 3.2833735942840576, "learning_rate": 3.5150000000000004e-07, "loss": 6.8125, "step": 703 }, { "epoch": 0.2206896551724138, "grad_norm": 5.263795375823975, "learning_rate": 3.5200000000000003e-07, "loss": 11.6207, "step": 704 }, { "epoch": 0.22100313479623823, "grad_norm": 3.4245877265930176, "learning_rate": 3.525e-07, "loss": 8.0098, "step": 705 }, { "epoch": 0.2213166144200627, "grad_norm": 3.0912983417510986, "learning_rate": 3.53e-07, "loss": 7.7124, "step": 706 }, { "epoch": 0.22163009404388714, "grad_norm": 2.501897096633911, "learning_rate": 3.5350000000000004e-07, "loss": 6.3126, "step": 707 }, { "epoch": 0.22194357366771159, "grad_norm": 2.732363224029541, "learning_rate": 3.54e-07, "loss": 8.4596, "step": 708 }, { "epoch": 0.22225705329153606, "grad_norm": 2.866844892501831, "learning_rate": 3.5450000000000006e-07, "loss": 7.4127, "step": 709 }, { "epoch": 0.2225705329153605, "grad_norm": 3.562100410461426, "learning_rate": 3.55e-07, "loss": 10.6676, "step": 710 }, { "epoch": 0.22288401253918494, "grad_norm": 2.7351932525634766, "learning_rate": 3.5550000000000003e-07, "loss": 6.7216, "step": 711 }, { "epoch": 0.2231974921630094, "grad_norm": 2.929007053375244, "learning_rate": 3.56e-07, "loss": 7.5504, "step": 712 }, { "epoch": 0.22351097178683385, "grad_norm": 2.9847233295440674, "learning_rate": 3.5650000000000005e-07, "loss": 8.5679, "step": 713 }, { "epoch": 0.22382445141065832, "grad_norm": 3.290311336517334, "learning_rate": 3.5700000000000003e-07, "loss": 8.5614, "step": 714 }, { "epoch": 0.22413793103448276, "grad_norm": 3.176608085632324, "learning_rate": 3.575e-07, "loss": 7.1585, "step": 715 }, { "epoch": 0.2244514106583072, "grad_norm": 5.026858806610107, "learning_rate": 3.58e-07, "loss": 15.6788, "step": 716 }, { "epoch": 0.22476489028213167, "grad_norm": 2.588078260421753, "learning_rate": 3.5850000000000004e-07, "loss": 6.1721, "step": 717 }, { "epoch": 0.22507836990595612, "grad_norm": 3.75956654548645, "learning_rate": 3.59e-07, "loss": 9.8582, "step": 718 }, { "epoch": 0.22539184952978056, "grad_norm": 3.7098045349121094, "learning_rate": 3.5950000000000006e-07, "loss": 7.7692, "step": 719 }, { "epoch": 0.22570532915360503, "grad_norm": 3.2847020626068115, "learning_rate": 3.6e-07, "loss": 7.2862, "step": 720 }, { "epoch": 0.22601880877742947, "grad_norm": 4.180566310882568, "learning_rate": 3.6050000000000003e-07, "loss": 8.2584, "step": 721 }, { "epoch": 0.2263322884012539, "grad_norm": 3.886276960372925, "learning_rate": 3.61e-07, "loss": 8.167, "step": 722 }, { "epoch": 0.22664576802507838, "grad_norm": 3.5251870155334473, "learning_rate": 3.6150000000000005e-07, "loss": 8.091, "step": 723 }, { "epoch": 0.22695924764890282, "grad_norm": 3.8147459030151367, "learning_rate": 3.6200000000000004e-07, "loss": 8.1781, "step": 724 }, { "epoch": 0.22727272727272727, "grad_norm": 4.299187660217285, "learning_rate": 3.625e-07, "loss": 10.4068, "step": 725 }, { "epoch": 0.22758620689655173, "grad_norm": 4.2207255363464355, "learning_rate": 3.63e-07, "loss": 12.8631, "step": 726 }, { "epoch": 0.22789968652037618, "grad_norm": 4.004030704498291, "learning_rate": 3.6350000000000005e-07, "loss": 8.9691, "step": 727 }, { "epoch": 0.22821316614420062, "grad_norm": 4.197298526763916, "learning_rate": 3.6400000000000003e-07, "loss": 8.0809, "step": 728 }, { "epoch": 0.2285266457680251, "grad_norm": 4.978548049926758, "learning_rate": 3.6450000000000007e-07, "loss": 11.308, "step": 729 }, { "epoch": 0.22884012539184953, "grad_norm": 3.4799234867095947, "learning_rate": 3.65e-07, "loss": 8.5875, "step": 730 }, { "epoch": 0.22915360501567397, "grad_norm": 3.4249789714813232, "learning_rate": 3.6550000000000004e-07, "loss": 9.2064, "step": 731 }, { "epoch": 0.22946708463949844, "grad_norm": 2.496309995651245, "learning_rate": 3.66e-07, "loss": 6.6208, "step": 732 }, { "epoch": 0.22978056426332288, "grad_norm": 4.72249698638916, "learning_rate": 3.6650000000000006e-07, "loss": 7.5853, "step": 733 }, { "epoch": 0.23009404388714733, "grad_norm": 3.36556339263916, "learning_rate": 3.6700000000000004e-07, "loss": 6.5957, "step": 734 }, { "epoch": 0.2304075235109718, "grad_norm": 5.186474800109863, "learning_rate": 3.6750000000000003e-07, "loss": 19.8754, "step": 735 }, { "epoch": 0.23072100313479624, "grad_norm": 3.4509902000427246, "learning_rate": 3.68e-07, "loss": 8.3037, "step": 736 }, { "epoch": 0.23103448275862068, "grad_norm": 3.9138832092285156, "learning_rate": 3.6850000000000005e-07, "loss": 10.7141, "step": 737 }, { "epoch": 0.23134796238244515, "grad_norm": 4.275444030761719, "learning_rate": 3.6900000000000004e-07, "loss": 12.6663, "step": 738 }, { "epoch": 0.2316614420062696, "grad_norm": 2.733919858932495, "learning_rate": 3.695e-07, "loss": 7.8312, "step": 739 }, { "epoch": 0.23197492163009403, "grad_norm": 3.960728645324707, "learning_rate": 3.7e-07, "loss": 7.1807, "step": 740 }, { "epoch": 0.2322884012539185, "grad_norm": 3.9499926567077637, "learning_rate": 3.7050000000000004e-07, "loss": 8.1952, "step": 741 }, { "epoch": 0.23260188087774294, "grad_norm": 3.199276924133301, "learning_rate": 3.7100000000000003e-07, "loss": 8.8693, "step": 742 }, { "epoch": 0.2329153605015674, "grad_norm": 3.465644121170044, "learning_rate": 3.7150000000000006e-07, "loss": 8.3512, "step": 743 }, { "epoch": 0.23322884012539186, "grad_norm": 3.957305431365967, "learning_rate": 3.72e-07, "loss": 12.9835, "step": 744 }, { "epoch": 0.2335423197492163, "grad_norm": 4.241058349609375, "learning_rate": 3.7250000000000003e-07, "loss": 11.0323, "step": 745 }, { "epoch": 0.23385579937304074, "grad_norm": 3.187009572982788, "learning_rate": 3.73e-07, "loss": 8.9532, "step": 746 }, { "epoch": 0.2341692789968652, "grad_norm": 3.3454670906066895, "learning_rate": 3.7350000000000006e-07, "loss": 8.9809, "step": 747 }, { "epoch": 0.23448275862068965, "grad_norm": 4.55251407623291, "learning_rate": 3.7400000000000004e-07, "loss": 8.4751, "step": 748 }, { "epoch": 0.2347962382445141, "grad_norm": 3.5339887142181396, "learning_rate": 3.745e-07, "loss": 10.128, "step": 749 }, { "epoch": 0.23510971786833856, "grad_norm": 3.775538206100464, "learning_rate": 3.75e-07, "loss": 8.0116, "step": 750 }, { "epoch": 0.235423197492163, "grad_norm": 3.8628170490264893, "learning_rate": 3.7550000000000005e-07, "loss": 8.7356, "step": 751 }, { "epoch": 0.23573667711598745, "grad_norm": 3.2559432983398438, "learning_rate": 3.7600000000000003e-07, "loss": 7.3864, "step": 752 }, { "epoch": 0.23605015673981192, "grad_norm": 3.832037925720215, "learning_rate": 3.7650000000000007e-07, "loss": 10.6071, "step": 753 }, { "epoch": 0.23636363636363636, "grad_norm": 3.030069589614868, "learning_rate": 3.77e-07, "loss": 8.176, "step": 754 }, { "epoch": 0.23667711598746083, "grad_norm": 3.3893072605133057, "learning_rate": 3.7750000000000004e-07, "loss": 7.696, "step": 755 }, { "epoch": 0.23699059561128527, "grad_norm": 3.5850815773010254, "learning_rate": 3.78e-07, "loss": 8.6539, "step": 756 }, { "epoch": 0.2373040752351097, "grad_norm": 3.6070199012756348, "learning_rate": 3.7850000000000006e-07, "loss": 9.2962, "step": 757 }, { "epoch": 0.23761755485893418, "grad_norm": 3.547919273376465, "learning_rate": 3.7900000000000005e-07, "loss": 10.0205, "step": 758 }, { "epoch": 0.23793103448275862, "grad_norm": 3.6676182746887207, "learning_rate": 3.7950000000000003e-07, "loss": 7.0754, "step": 759 }, { "epoch": 0.23824451410658307, "grad_norm": 4.715800762176514, "learning_rate": 3.8e-07, "loss": 13.3314, "step": 760 }, { "epoch": 0.23855799373040754, "grad_norm": 3.9477193355560303, "learning_rate": 3.8050000000000005e-07, "loss": 8.7171, "step": 761 }, { "epoch": 0.23887147335423198, "grad_norm": 4.012876987457275, "learning_rate": 3.8100000000000004e-07, "loss": 8.9844, "step": 762 }, { "epoch": 0.23918495297805642, "grad_norm": 3.585956335067749, "learning_rate": 3.815000000000001e-07, "loss": 8.8991, "step": 763 }, { "epoch": 0.2394984326018809, "grad_norm": 3.68974232673645, "learning_rate": 3.82e-07, "loss": 8.506, "step": 764 }, { "epoch": 0.23981191222570533, "grad_norm": 3.4120147228240967, "learning_rate": 3.8250000000000004e-07, "loss": 7.4377, "step": 765 }, { "epoch": 0.24012539184952977, "grad_norm": 2.9373185634613037, "learning_rate": 3.8300000000000003e-07, "loss": 6.9149, "step": 766 }, { "epoch": 0.24043887147335424, "grad_norm": 6.263192176818848, "learning_rate": 3.8350000000000007e-07, "loss": 15.8049, "step": 767 }, { "epoch": 0.24075235109717869, "grad_norm": 3.48296856880188, "learning_rate": 3.84e-07, "loss": 7.886, "step": 768 }, { "epoch": 0.24106583072100313, "grad_norm": 3.5808095932006836, "learning_rate": 3.8450000000000004e-07, "loss": 8.8138, "step": 769 }, { "epoch": 0.2413793103448276, "grad_norm": 4.42182731628418, "learning_rate": 3.85e-07, "loss": 11.1607, "step": 770 }, { "epoch": 0.24169278996865204, "grad_norm": 3.0944645404815674, "learning_rate": 3.8550000000000006e-07, "loss": 8.216, "step": 771 }, { "epoch": 0.24200626959247648, "grad_norm": 4.562041282653809, "learning_rate": 3.8600000000000004e-07, "loss": 10.2914, "step": 772 }, { "epoch": 0.24231974921630095, "grad_norm": 3.654736042022705, "learning_rate": 3.865e-07, "loss": 7.9714, "step": 773 }, { "epoch": 0.2426332288401254, "grad_norm": 3.496178150177002, "learning_rate": 3.87e-07, "loss": 8.5865, "step": 774 }, { "epoch": 0.24294670846394983, "grad_norm": 3.801335096359253, "learning_rate": 3.8750000000000005e-07, "loss": 8.9361, "step": 775 }, { "epoch": 0.2432601880877743, "grad_norm": 4.17310094833374, "learning_rate": 3.8800000000000003e-07, "loss": 7.906, "step": 776 }, { "epoch": 0.24357366771159875, "grad_norm": 3.4103686809539795, "learning_rate": 3.8850000000000007e-07, "loss": 8.7571, "step": 777 }, { "epoch": 0.2438871473354232, "grad_norm": 3.6416525840759277, "learning_rate": 3.89e-07, "loss": 7.8949, "step": 778 }, { "epoch": 0.24420062695924766, "grad_norm": 4.429583549499512, "learning_rate": 3.8950000000000004e-07, "loss": 7.3726, "step": 779 }, { "epoch": 0.2445141065830721, "grad_norm": 3.5960888862609863, "learning_rate": 3.9e-07, "loss": 8.3664, "step": 780 }, { "epoch": 0.24482758620689654, "grad_norm": 6.068627834320068, "learning_rate": 3.9050000000000006e-07, "loss": 14.7055, "step": 781 }, { "epoch": 0.245141065830721, "grad_norm": 4.344872951507568, "learning_rate": 3.9100000000000005e-07, "loss": 10.0272, "step": 782 }, { "epoch": 0.24545454545454545, "grad_norm": 3.6392626762390137, "learning_rate": 3.915e-07, "loss": 6.46, "step": 783 }, { "epoch": 0.2457680250783699, "grad_norm": 3.415555238723755, "learning_rate": 3.92e-07, "loss": 8.4829, "step": 784 }, { "epoch": 0.24608150470219436, "grad_norm": 4.941524982452393, "learning_rate": 3.9250000000000005e-07, "loss": 11.3501, "step": 785 }, { "epoch": 0.2463949843260188, "grad_norm": 3.191941738128662, "learning_rate": 3.9300000000000004e-07, "loss": 8.494, "step": 786 }, { "epoch": 0.24670846394984325, "grad_norm": 3.4646642208099365, "learning_rate": 3.935000000000001e-07, "loss": 7.9202, "step": 787 }, { "epoch": 0.24702194357366772, "grad_norm": 3.47352933883667, "learning_rate": 3.94e-07, "loss": 10.1235, "step": 788 }, { "epoch": 0.24733542319749216, "grad_norm": 4.042746543884277, "learning_rate": 3.9450000000000005e-07, "loss": 7.8208, "step": 789 }, { "epoch": 0.2476489028213166, "grad_norm": 3.284029006958008, "learning_rate": 3.9500000000000003e-07, "loss": 7.8496, "step": 790 }, { "epoch": 0.24796238244514107, "grad_norm": 4.137977600097656, "learning_rate": 3.9550000000000007e-07, "loss": 11.9702, "step": 791 }, { "epoch": 0.2482758620689655, "grad_norm": 3.6068592071533203, "learning_rate": 3.9600000000000005e-07, "loss": 6.9573, "step": 792 }, { "epoch": 0.24858934169278996, "grad_norm": 3.264119863510132, "learning_rate": 3.965e-07, "loss": 7.5779, "step": 793 }, { "epoch": 0.24890282131661443, "grad_norm": 3.505993366241455, "learning_rate": 3.97e-07, "loss": 7.7447, "step": 794 }, { "epoch": 0.24921630094043887, "grad_norm": 3.9648635387420654, "learning_rate": 3.9750000000000006e-07, "loss": 11.2637, "step": 795 }, { "epoch": 0.2495297805642633, "grad_norm": 3.1940486431121826, "learning_rate": 3.9800000000000004e-07, "loss": 8.9283, "step": 796 }, { "epoch": 0.24984326018808778, "grad_norm": 3.5564422607421875, "learning_rate": 3.985e-07, "loss": 7.2672, "step": 797 }, { "epoch": 0.2501567398119122, "grad_norm": 3.626217842102051, "learning_rate": 3.99e-07, "loss": 8.51, "step": 798 }, { "epoch": 0.2501567398119122, "eval_loss": 28.84930992126465, "eval_runtime": 20.5251, "eval_samples_per_second": 130.913, "eval_steps_per_second": 8.185, "step": 798 }, { "epoch": 0.25047021943573666, "grad_norm": 3.097006320953369, "learning_rate": 3.9950000000000005e-07, "loss": 7.4537, "step": 799 }, { "epoch": 0.2507836990595611, "grad_norm": 3.5833940505981445, "learning_rate": 4.0000000000000003e-07, "loss": 10.0463, "step": 800 }, { "epoch": 0.2510971786833856, "grad_norm": 4.196984767913818, "learning_rate": 4.0050000000000007e-07, "loss": 7.771, "step": 801 }, { "epoch": 0.25141065830721004, "grad_norm": 4.240922927856445, "learning_rate": 4.01e-07, "loss": 10.3631, "step": 802 }, { "epoch": 0.2517241379310345, "grad_norm": 3.425544023513794, "learning_rate": 4.015e-07, "loss": 6.2936, "step": 803 }, { "epoch": 0.25203761755485893, "grad_norm": 3.8338463306427, "learning_rate": 4.02e-07, "loss": 11.4034, "step": 804 }, { "epoch": 0.25235109717868337, "grad_norm": 3.4138238430023193, "learning_rate": 4.0250000000000006e-07, "loss": 6.9629, "step": 805 }, { "epoch": 0.2526645768025078, "grad_norm": 3.1681156158447266, "learning_rate": 4.0300000000000005e-07, "loss": 7.616, "step": 806 }, { "epoch": 0.2529780564263323, "grad_norm": 13.125308990478516, "learning_rate": 4.035e-07, "loss": 7.871, "step": 807 }, { "epoch": 0.25329153605015675, "grad_norm": 5.784317493438721, "learning_rate": 4.04e-07, "loss": 20.8958, "step": 808 }, { "epoch": 0.2536050156739812, "grad_norm": 3.021087646484375, "learning_rate": 4.0450000000000006e-07, "loss": 7.3132, "step": 809 }, { "epoch": 0.25391849529780564, "grad_norm": 2.8129184246063232, "learning_rate": 4.0500000000000004e-07, "loss": 7.4504, "step": 810 }, { "epoch": 0.2542319749216301, "grad_norm": 3.036388874053955, "learning_rate": 4.055000000000001e-07, "loss": 6.5679, "step": 811 }, { "epoch": 0.2545454545454545, "grad_norm": 3.112868547439575, "learning_rate": 4.06e-07, "loss": 8.2039, "step": 812 }, { "epoch": 0.254858934169279, "grad_norm": 3.948190212249756, "learning_rate": 4.065e-07, "loss": 9.2969, "step": 813 }, { "epoch": 0.25517241379310346, "grad_norm": 3.823413133621216, "learning_rate": 4.0700000000000003e-07, "loss": 8.2135, "step": 814 }, { "epoch": 0.2554858934169279, "grad_norm": 3.8010671138763428, "learning_rate": 4.0750000000000007e-07, "loss": 8.7767, "step": 815 }, { "epoch": 0.25579937304075234, "grad_norm": 2.9464848041534424, "learning_rate": 4.0800000000000005e-07, "loss": 6.3003, "step": 816 }, { "epoch": 0.2561128526645768, "grad_norm": 3.2114903926849365, "learning_rate": 4.085e-07, "loss": 8.0724, "step": 817 }, { "epoch": 0.2564263322884012, "grad_norm": 3.254835844039917, "learning_rate": 4.09e-07, "loss": 7.6632, "step": 818 }, { "epoch": 0.2567398119122257, "grad_norm": 4.548694133758545, "learning_rate": 4.0950000000000006e-07, "loss": 12.1637, "step": 819 }, { "epoch": 0.25705329153605017, "grad_norm": 4.109635829925537, "learning_rate": 4.1000000000000004e-07, "loss": 9.4519, "step": 820 }, { "epoch": 0.2573667711598746, "grad_norm": 4.340848922729492, "learning_rate": 4.105000000000001e-07, "loss": 8.5242, "step": 821 }, { "epoch": 0.25768025078369905, "grad_norm": 3.260594367980957, "learning_rate": 4.11e-07, "loss": 7.83, "step": 822 }, { "epoch": 0.2579937304075235, "grad_norm": 5.5108771324157715, "learning_rate": 4.115e-07, "loss": 10.5884, "step": 823 }, { "epoch": 0.258307210031348, "grad_norm": 3.490274429321289, "learning_rate": 4.1200000000000004e-07, "loss": 8.7728, "step": 824 }, { "epoch": 0.25862068965517243, "grad_norm": 6.578031539916992, "learning_rate": 4.125000000000001e-07, "loss": 10.9723, "step": 825 }, { "epoch": 0.2589341692789969, "grad_norm": 2.774808406829834, "learning_rate": 4.1300000000000006e-07, "loss": 6.9211, "step": 826 }, { "epoch": 0.2592476489028213, "grad_norm": 5.915481090545654, "learning_rate": 4.135e-07, "loss": 9.1981, "step": 827 }, { "epoch": 0.25956112852664576, "grad_norm": 2.8038599491119385, "learning_rate": 4.1400000000000003e-07, "loss": 6.1127, "step": 828 }, { "epoch": 0.2598746081504702, "grad_norm": 3.501262903213501, "learning_rate": 4.1450000000000007e-07, "loss": 6.5746, "step": 829 }, { "epoch": 0.2601880877742947, "grad_norm": 3.5788586139678955, "learning_rate": 4.1500000000000005e-07, "loss": 8.6381, "step": 830 }, { "epoch": 0.26050156739811914, "grad_norm": 5.020801544189453, "learning_rate": 4.155e-07, "loss": 16.7279, "step": 831 }, { "epoch": 0.2608150470219436, "grad_norm": 3.9327170848846436, "learning_rate": 4.16e-07, "loss": 12.7287, "step": 832 }, { "epoch": 0.261128526645768, "grad_norm": 3.2840352058410645, "learning_rate": 4.165e-07, "loss": 6.839, "step": 833 }, { "epoch": 0.26144200626959246, "grad_norm": 3.750833511352539, "learning_rate": 4.1700000000000004e-07, "loss": 11.1242, "step": 834 }, { "epoch": 0.2617554858934169, "grad_norm": 3.46420955657959, "learning_rate": 4.175000000000001e-07, "loss": 9.7763, "step": 835 }, { "epoch": 0.2620689655172414, "grad_norm": 3.476485252380371, "learning_rate": 4.18e-07, "loss": 7.1691, "step": 836 }, { "epoch": 0.26238244514106585, "grad_norm": 4.025050163269043, "learning_rate": 4.185e-07, "loss": 9.9421, "step": 837 }, { "epoch": 0.2626959247648903, "grad_norm": 3.0116124153137207, "learning_rate": 4.1900000000000003e-07, "loss": 6.3215, "step": 838 }, { "epoch": 0.26300940438871473, "grad_norm": 4.668792247772217, "learning_rate": 4.1950000000000007e-07, "loss": 11.0969, "step": 839 }, { "epoch": 0.26332288401253917, "grad_norm": 2.963440418243408, "learning_rate": 4.2000000000000006e-07, "loss": 7.2545, "step": 840 }, { "epoch": 0.2636363636363636, "grad_norm": 4.21551513671875, "learning_rate": 4.205e-07, "loss": 8.2629, "step": 841 }, { "epoch": 0.2639498432601881, "grad_norm": 3.958061456680298, "learning_rate": 4.21e-07, "loss": 9.7021, "step": 842 }, { "epoch": 0.26426332288401255, "grad_norm": 2.9558305740356445, "learning_rate": 4.215e-07, "loss": 7.3453, "step": 843 }, { "epoch": 0.264576802507837, "grad_norm": 4.2135233879089355, "learning_rate": 4.2200000000000005e-07, "loss": 8.955, "step": 844 }, { "epoch": 0.26489028213166144, "grad_norm": 3.562573194503784, "learning_rate": 4.225000000000001e-07, "loss": 7.2757, "step": 845 }, { "epoch": 0.2652037617554859, "grad_norm": 3.2057554721832275, "learning_rate": 4.23e-07, "loss": 8.2123, "step": 846 }, { "epoch": 0.2655172413793103, "grad_norm": 3.901542901992798, "learning_rate": 4.235e-07, "loss": 9.1358, "step": 847 }, { "epoch": 0.2658307210031348, "grad_norm": 3.1347365379333496, "learning_rate": 4.2400000000000004e-07, "loss": 6.5317, "step": 848 }, { "epoch": 0.26614420062695926, "grad_norm": 5.042209625244141, "learning_rate": 4.245000000000001e-07, "loss": 14.4307, "step": 849 }, { "epoch": 0.2664576802507837, "grad_norm": 3.5455801486968994, "learning_rate": 4.2500000000000006e-07, "loss": 8.8807, "step": 850 }, { "epoch": 0.26677115987460814, "grad_norm": 3.4751904010772705, "learning_rate": 4.255e-07, "loss": 8.8506, "step": 851 }, { "epoch": 0.2670846394984326, "grad_norm": 3.6751580238342285, "learning_rate": 4.2600000000000003e-07, "loss": 10.4928, "step": 852 }, { "epoch": 0.267398119122257, "grad_norm": 4.02175760269165, "learning_rate": 4.265e-07, "loss": 6.9068, "step": 853 }, { "epoch": 0.2677115987460815, "grad_norm": 7.32761287689209, "learning_rate": 4.2700000000000005e-07, "loss": 17.0699, "step": 854 }, { "epoch": 0.26802507836990597, "grad_norm": 3.607837200164795, "learning_rate": 4.275000000000001e-07, "loss": 8.021, "step": 855 }, { "epoch": 0.2683385579937304, "grad_norm": 2.8955557346343994, "learning_rate": 4.28e-07, "loss": 8.8195, "step": 856 }, { "epoch": 0.26865203761755485, "grad_norm": 3.2017621994018555, "learning_rate": 4.285e-07, "loss": 7.1978, "step": 857 }, { "epoch": 0.2689655172413793, "grad_norm": 3.7214231491088867, "learning_rate": 4.2900000000000004e-07, "loss": 7.2386, "step": 858 }, { "epoch": 0.26927899686520373, "grad_norm": 3.3874855041503906, "learning_rate": 4.295000000000001e-07, "loss": 7.6419, "step": 859 }, { "epoch": 0.26959247648902823, "grad_norm": 4.008450508117676, "learning_rate": 4.3e-07, "loss": 8.0012, "step": 860 }, { "epoch": 0.2699059561128527, "grad_norm": 3.3975911140441895, "learning_rate": 4.305e-07, "loss": 7.3739, "step": 861 }, { "epoch": 0.2702194357366771, "grad_norm": 3.656843423843384, "learning_rate": 4.3100000000000003e-07, "loss": 8.8982, "step": 862 }, { "epoch": 0.27053291536050156, "grad_norm": 3.2853686809539795, "learning_rate": 4.315e-07, "loss": 7.6732, "step": 863 }, { "epoch": 0.270846394984326, "grad_norm": 4.232872009277344, "learning_rate": 4.3200000000000006e-07, "loss": 10.4912, "step": 864 }, { "epoch": 0.2711598746081505, "grad_norm": 3.6398162841796875, "learning_rate": 4.325e-07, "loss": 9.274, "step": 865 }, { "epoch": 0.27147335423197494, "grad_norm": 2.7679269313812256, "learning_rate": 4.33e-07, "loss": 7.0839, "step": 866 }, { "epoch": 0.2717868338557994, "grad_norm": 3.5400402545928955, "learning_rate": 4.335e-07, "loss": 8.8406, "step": 867 }, { "epoch": 0.2721003134796238, "grad_norm": 3.0687062740325928, "learning_rate": 4.3400000000000005e-07, "loss": 7.36, "step": 868 }, { "epoch": 0.27241379310344827, "grad_norm": 3.573641061782837, "learning_rate": 4.345000000000001e-07, "loss": 8.0488, "step": 869 }, { "epoch": 0.2727272727272727, "grad_norm": 4.390239238739014, "learning_rate": 4.35e-07, "loss": 12.9096, "step": 870 }, { "epoch": 0.2730407523510972, "grad_norm": 3.314115524291992, "learning_rate": 4.355e-07, "loss": 7.1066, "step": 871 }, { "epoch": 0.27335423197492165, "grad_norm": 5.46345853805542, "learning_rate": 4.3600000000000004e-07, "loss": 11.487, "step": 872 }, { "epoch": 0.2736677115987461, "grad_norm": 5.023271083831787, "learning_rate": 4.365e-07, "loss": 10.0573, "step": 873 }, { "epoch": 0.27398119122257053, "grad_norm": 2.962111234664917, "learning_rate": 4.3700000000000006e-07, "loss": 8.829, "step": 874 }, { "epoch": 0.274294670846395, "grad_norm": 3.1559054851531982, "learning_rate": 4.375e-07, "loss": 7.1875, "step": 875 }, { "epoch": 0.2746081504702194, "grad_norm": 3.6950178146362305, "learning_rate": 4.3800000000000003e-07, "loss": 8.4211, "step": 876 }, { "epoch": 0.2749216300940439, "grad_norm": 4.661724090576172, "learning_rate": 4.385e-07, "loss": 11.424, "step": 877 }, { "epoch": 0.27523510971786835, "grad_norm": 3.7680742740631104, "learning_rate": 4.3900000000000005e-07, "loss": 9.1259, "step": 878 }, { "epoch": 0.2755485893416928, "grad_norm": 3.409186601638794, "learning_rate": 4.395000000000001e-07, "loss": 6.5314, "step": 879 }, { "epoch": 0.27586206896551724, "grad_norm": 3.2190394401550293, "learning_rate": 4.4e-07, "loss": 6.4504, "step": 880 }, { "epoch": 0.2761755485893417, "grad_norm": 4.012719631195068, "learning_rate": 4.405e-07, "loss": 9.4085, "step": 881 }, { "epoch": 0.2764890282131661, "grad_norm": 4.595925807952881, "learning_rate": 4.4100000000000004e-07, "loss": 13.1328, "step": 882 }, { "epoch": 0.2768025078369906, "grad_norm": 4.065830230712891, "learning_rate": 4.4150000000000003e-07, "loss": 8.3007, "step": 883 }, { "epoch": 0.27711598746081506, "grad_norm": 4.229946613311768, "learning_rate": 4.4200000000000007e-07, "loss": 11.1843, "step": 884 }, { "epoch": 0.2774294670846395, "grad_norm": 3.8973495960235596, "learning_rate": 4.425e-07, "loss": 7.978, "step": 885 }, { "epoch": 0.27774294670846394, "grad_norm": 2.952505111694336, "learning_rate": 4.4300000000000004e-07, "loss": 8.5853, "step": 886 }, { "epoch": 0.2780564263322884, "grad_norm": 2.839816093444824, "learning_rate": 4.435e-07, "loss": 6.6168, "step": 887 }, { "epoch": 0.27836990595611283, "grad_norm": 4.309621334075928, "learning_rate": 4.4400000000000006e-07, "loss": 10.9452, "step": 888 }, { "epoch": 0.2786833855799373, "grad_norm": 5.065432071685791, "learning_rate": 4.445000000000001e-07, "loss": 9.2203, "step": 889 }, { "epoch": 0.27899686520376177, "grad_norm": 4.246471881866455, "learning_rate": 4.4500000000000003e-07, "loss": 7.9399, "step": 890 }, { "epoch": 0.2793103448275862, "grad_norm": 5.130437850952148, "learning_rate": 4.455e-07, "loss": 11.7158, "step": 891 }, { "epoch": 0.27962382445141065, "grad_norm": 3.858509063720703, "learning_rate": 4.4600000000000005e-07, "loss": 9.7248, "step": 892 }, { "epoch": 0.2799373040752351, "grad_norm": 3.6608774662017822, "learning_rate": 4.4650000000000003e-07, "loss": 12.7871, "step": 893 }, { "epoch": 0.28025078369905954, "grad_norm": 4.863644123077393, "learning_rate": 4.47e-07, "loss": 9.9666, "step": 894 }, { "epoch": 0.28056426332288403, "grad_norm": 3.754490613937378, "learning_rate": 4.475e-07, "loss": 7.9063, "step": 895 }, { "epoch": 0.2808777429467085, "grad_norm": 3.4950146675109863, "learning_rate": 4.4800000000000004e-07, "loss": 8.867, "step": 896 }, { "epoch": 0.2811912225705329, "grad_norm": 3.564337968826294, "learning_rate": 4.485e-07, "loss": 7.4725, "step": 897 }, { "epoch": 0.28150470219435736, "grad_norm": 2.9661481380462646, "learning_rate": 4.4900000000000006e-07, "loss": 6.9609, "step": 898 }, { "epoch": 0.2818181818181818, "grad_norm": 3.3152127265930176, "learning_rate": 4.495e-07, "loss": 7.2334, "step": 899 }, { "epoch": 0.28213166144200624, "grad_norm": 3.4572389125823975, "learning_rate": 4.5000000000000003e-07, "loss": 7.8584, "step": 900 }, { "epoch": 0.28244514106583074, "grad_norm": 3.333373546600342, "learning_rate": 4.505e-07, "loss": 6.6077, "step": 901 }, { "epoch": 0.2827586206896552, "grad_norm": 4.405636310577393, "learning_rate": 4.5100000000000005e-07, "loss": 11.1708, "step": 902 }, { "epoch": 0.2830721003134796, "grad_norm": 3.0989999771118164, "learning_rate": 4.5150000000000004e-07, "loss": 7.3427, "step": 903 }, { "epoch": 0.28338557993730407, "grad_norm": 3.1780450344085693, "learning_rate": 4.52e-07, "loss": 7.8806, "step": 904 }, { "epoch": 0.2836990595611285, "grad_norm": 3.923069953918457, "learning_rate": 4.525e-07, "loss": 9.7537, "step": 905 }, { "epoch": 0.284012539184953, "grad_norm": 4.220728874206543, "learning_rate": 4.5300000000000005e-07, "loss": 9.528, "step": 906 }, { "epoch": 0.28432601880877745, "grad_norm": 3.3002758026123047, "learning_rate": 4.5350000000000003e-07, "loss": 6.0225, "step": 907 }, { "epoch": 0.2846394984326019, "grad_norm": 4.327773571014404, "learning_rate": 4.5400000000000007e-07, "loss": 8.0762, "step": 908 }, { "epoch": 0.28495297805642633, "grad_norm": 3.128321409225464, "learning_rate": 4.545e-07, "loss": 8.3285, "step": 909 }, { "epoch": 0.2852664576802508, "grad_norm": 3.2514123916625977, "learning_rate": 4.5500000000000004e-07, "loss": 8.2311, "step": 910 }, { "epoch": 0.2855799373040752, "grad_norm": 5.544764518737793, "learning_rate": 4.555e-07, "loss": 11.3571, "step": 911 }, { "epoch": 0.2858934169278997, "grad_norm": 5.904582977294922, "learning_rate": 4.5600000000000006e-07, "loss": 18.3895, "step": 912 }, { "epoch": 0.28620689655172415, "grad_norm": 3.38208270072937, "learning_rate": 4.5650000000000004e-07, "loss": 9.4236, "step": 913 }, { "epoch": 0.2865203761755486, "grad_norm": 3.433018445968628, "learning_rate": 4.5700000000000003e-07, "loss": 11.2537, "step": 914 }, { "epoch": 0.28683385579937304, "grad_norm": 4.306755542755127, "learning_rate": 4.575e-07, "loss": 11.701, "step": 915 }, { "epoch": 0.2871473354231975, "grad_norm": 3.296084403991699, "learning_rate": 4.5800000000000005e-07, "loss": 6.5184, "step": 916 }, { "epoch": 0.2874608150470219, "grad_norm": 4.437682628631592, "learning_rate": 4.5850000000000004e-07, "loss": 8.7946, "step": 917 }, { "epoch": 0.2877742946708464, "grad_norm": 4.028353214263916, "learning_rate": 4.5900000000000007e-07, "loss": 9.0222, "step": 918 }, { "epoch": 0.28808777429467086, "grad_norm": 2.981740951538086, "learning_rate": 4.595e-07, "loss": 6.3495, "step": 919 }, { "epoch": 0.2884012539184953, "grad_norm": 3.5462889671325684, "learning_rate": 4.6000000000000004e-07, "loss": 9.6387, "step": 920 }, { "epoch": 0.28871473354231975, "grad_norm": 3.0416839122772217, "learning_rate": 4.6050000000000003e-07, "loss": 5.743, "step": 921 }, { "epoch": 0.2890282131661442, "grad_norm": 2.538724660873413, "learning_rate": 4.6100000000000006e-07, "loss": 6.5413, "step": 922 }, { "epoch": 0.28934169278996863, "grad_norm": 3.508538007736206, "learning_rate": 4.615e-07, "loss": 9.6875, "step": 923 }, { "epoch": 0.2896551724137931, "grad_norm": 3.4180221557617188, "learning_rate": 4.6200000000000003e-07, "loss": 7.9855, "step": 924 }, { "epoch": 0.28996865203761757, "grad_norm": 3.5554192066192627, "learning_rate": 4.625e-07, "loss": 7.1513, "step": 925 }, { "epoch": 0.290282131661442, "grad_norm": 3.860269784927368, "learning_rate": 4.6300000000000006e-07, "loss": 9.0043, "step": 926 }, { "epoch": 0.29059561128526645, "grad_norm": 3.593224287033081, "learning_rate": 4.6350000000000004e-07, "loss": 8.9827, "step": 927 }, { "epoch": 0.2909090909090909, "grad_norm": 4.743045330047607, "learning_rate": 4.64e-07, "loss": 12.2761, "step": 928 }, { "epoch": 0.29122257053291534, "grad_norm": 3.506476640701294, "learning_rate": 4.645e-07, "loss": 9.6425, "step": 929 }, { "epoch": 0.29153605015673983, "grad_norm": 3.420703887939453, "learning_rate": 4.6500000000000005e-07, "loss": 7.238, "step": 930 }, { "epoch": 0.2918495297805643, "grad_norm": 8.075421333312988, "learning_rate": 4.6550000000000003e-07, "loss": 19.2958, "step": 931 }, { "epoch": 0.2921630094043887, "grad_norm": 4.11147928237915, "learning_rate": 4.6600000000000007e-07, "loss": 7.9719, "step": 932 }, { "epoch": 0.29247648902821316, "grad_norm": 4.024860382080078, "learning_rate": 4.665e-07, "loss": 8.5926, "step": 933 }, { "epoch": 0.2927899686520376, "grad_norm": 3.3219480514526367, "learning_rate": 4.6700000000000004e-07, "loss": 8.4054, "step": 934 }, { "epoch": 0.29310344827586204, "grad_norm": 3.3816304206848145, "learning_rate": 4.675e-07, "loss": 6.9335, "step": 935 }, { "epoch": 0.29341692789968654, "grad_norm": 3.370476484298706, "learning_rate": 4.6800000000000006e-07, "loss": 8.5852, "step": 936 }, { "epoch": 0.293730407523511, "grad_norm": 3.3092455863952637, "learning_rate": 4.6850000000000005e-07, "loss": 8.164, "step": 937 }, { "epoch": 0.2940438871473354, "grad_norm": 3.3833985328674316, "learning_rate": 4.6900000000000003e-07, "loss": 7.7537, "step": 938 }, { "epoch": 0.29435736677115987, "grad_norm": 4.879644870758057, "learning_rate": 4.695e-07, "loss": 9.2243, "step": 939 }, { "epoch": 0.2946708463949843, "grad_norm": 3.931847333908081, "learning_rate": 4.7000000000000005e-07, "loss": 9.3823, "step": 940 }, { "epoch": 0.29498432601880875, "grad_norm": 3.1454243659973145, "learning_rate": 4.7050000000000004e-07, "loss": 6.3691, "step": 941 }, { "epoch": 0.29529780564263325, "grad_norm": 3.438483953475952, "learning_rate": 4.710000000000001e-07, "loss": 7.0824, "step": 942 }, { "epoch": 0.2956112852664577, "grad_norm": 3.97963809967041, "learning_rate": 4.715e-07, "loss": 9.5606, "step": 943 }, { "epoch": 0.29592476489028213, "grad_norm": 5.063857078552246, "learning_rate": 4.7200000000000004e-07, "loss": 12.8541, "step": 944 }, { "epoch": 0.2962382445141066, "grad_norm": 4.647433280944824, "learning_rate": 4.7250000000000003e-07, "loss": 11.1892, "step": 945 }, { "epoch": 0.296551724137931, "grad_norm": 4.020174026489258, "learning_rate": 4.7300000000000007e-07, "loss": 9.2035, "step": 946 }, { "epoch": 0.29686520376175546, "grad_norm": 3.719456434249878, "learning_rate": 4.7350000000000005e-07, "loss": 7.5178, "step": 947 }, { "epoch": 0.29717868338557996, "grad_norm": 3.1977577209472656, "learning_rate": 4.7400000000000004e-07, "loss": 6.5933, "step": 948 }, { "epoch": 0.2974921630094044, "grad_norm": 2.9651682376861572, "learning_rate": 4.745e-07, "loss": 6.2385, "step": 949 }, { "epoch": 0.29780564263322884, "grad_norm": 4.075620174407959, "learning_rate": 4.7500000000000006e-07, "loss": 13.3818, "step": 950 }, { "epoch": 0.2981191222570533, "grad_norm": 4.676611423492432, "learning_rate": 4.7550000000000004e-07, "loss": 13.5246, "step": 951 }, { "epoch": 0.2984326018808777, "grad_norm": 4.3568291664123535, "learning_rate": 4.760000000000001e-07, "loss": 10.122, "step": 952 }, { "epoch": 0.2987460815047022, "grad_norm": 3.4638936519622803, "learning_rate": 4.765e-07, "loss": 9.9878, "step": 953 }, { "epoch": 0.29905956112852666, "grad_norm": 3.4983904361724854, "learning_rate": 4.77e-07, "loss": 7.968, "step": 954 }, { "epoch": 0.2993730407523511, "grad_norm": 2.9241857528686523, "learning_rate": 4.775000000000001e-07, "loss": 6.1837, "step": 955 }, { "epoch": 0.29968652037617555, "grad_norm": 5.027895927429199, "learning_rate": 4.78e-07, "loss": 10.7868, "step": 956 }, { "epoch": 0.3, "grad_norm": 3.0963079929351807, "learning_rate": 4.785000000000001e-07, "loss": 9.2105, "step": 957 }, { "epoch": 0.30031347962382443, "grad_norm": 3.367241144180298, "learning_rate": 4.79e-07, "loss": 8.6549, "step": 958 }, { "epoch": 0.30062695924764893, "grad_norm": 3.1908528804779053, "learning_rate": 4.795e-07, "loss": 9.3335, "step": 959 }, { "epoch": 0.30094043887147337, "grad_norm": 3.797089099884033, "learning_rate": 4.800000000000001e-07, "loss": 10.5798, "step": 960 }, { "epoch": 0.3012539184952978, "grad_norm": 3.816810131072998, "learning_rate": 4.805000000000001e-07, "loss": 8.424, "step": 961 }, { "epoch": 0.30156739811912225, "grad_norm": 2.98754620552063, "learning_rate": 4.81e-07, "loss": 6.592, "step": 962 }, { "epoch": 0.3018808777429467, "grad_norm": 7.078648090362549, "learning_rate": 4.815000000000001e-07, "loss": 20.1471, "step": 963 }, { "epoch": 0.30219435736677114, "grad_norm": 3.7833642959594727, "learning_rate": 4.82e-07, "loss": 8.8326, "step": 964 }, { "epoch": 0.30250783699059564, "grad_norm": 2.790283203125, "learning_rate": 4.825e-07, "loss": 6.6206, "step": 965 }, { "epoch": 0.3028213166144201, "grad_norm": 4.206464767456055, "learning_rate": 4.830000000000001e-07, "loss": 10.1851, "step": 966 }, { "epoch": 0.3031347962382445, "grad_norm": 4.659297943115234, "learning_rate": 4.835e-07, "loss": 11.0335, "step": 967 }, { "epoch": 0.30344827586206896, "grad_norm": 4.047898769378662, "learning_rate": 4.84e-07, "loss": 10.0668, "step": 968 }, { "epoch": 0.3037617554858934, "grad_norm": 4.826545715332031, "learning_rate": 4.845000000000001e-07, "loss": 9.4059, "step": 969 }, { "epoch": 0.30407523510971785, "grad_norm": 3.3105831146240234, "learning_rate": 4.85e-07, "loss": 9.7796, "step": 970 }, { "epoch": 0.30438871473354234, "grad_norm": 3.3956053256988525, "learning_rate": 4.855e-07, "loss": 8.4835, "step": 971 }, { "epoch": 0.3047021943573668, "grad_norm": 3.9917337894439697, "learning_rate": 4.86e-07, "loss": 12.5467, "step": 972 }, { "epoch": 0.3050156739811912, "grad_norm": 2.869931221008301, "learning_rate": 4.865e-07, "loss": 7.641, "step": 973 }, { "epoch": 0.30532915360501567, "grad_norm": 4.201666355133057, "learning_rate": 4.870000000000001e-07, "loss": 8.2918, "step": 974 }, { "epoch": 0.3056426332288401, "grad_norm": 3.3518385887145996, "learning_rate": 4.875000000000001e-07, "loss": 10.0857, "step": 975 }, { "epoch": 0.30595611285266455, "grad_norm": 3.7927677631378174, "learning_rate": 4.88e-07, "loss": 8.8694, "step": 976 }, { "epoch": 0.30626959247648905, "grad_norm": 2.8095505237579346, "learning_rate": 4.885000000000001e-07, "loss": 6.6404, "step": 977 }, { "epoch": 0.3065830721003135, "grad_norm": 3.649327516555786, "learning_rate": 4.89e-07, "loss": 8.3927, "step": 978 }, { "epoch": 0.30689655172413793, "grad_norm": 3.178901195526123, "learning_rate": 4.895e-07, "loss": 7.3435, "step": 979 }, { "epoch": 0.3072100313479624, "grad_norm": 3.161583185195923, "learning_rate": 4.900000000000001e-07, "loss": 8.2137, "step": 980 }, { "epoch": 0.3075235109717868, "grad_norm": 3.098173141479492, "learning_rate": 4.905000000000001e-07, "loss": 7.279, "step": 981 }, { "epoch": 0.30783699059561126, "grad_norm": 2.761258363723755, "learning_rate": 4.91e-07, "loss": 6.147, "step": 982 }, { "epoch": 0.30815047021943576, "grad_norm": 2.6040921211242676, "learning_rate": 4.915000000000001e-07, "loss": 6.242, "step": 983 }, { "epoch": 0.3084639498432602, "grad_norm": 3.172119617462158, "learning_rate": 4.92e-07, "loss": 6.6438, "step": 984 }, { "epoch": 0.30877742946708464, "grad_norm": 3.180616617202759, "learning_rate": 4.925e-07, "loss": 7.9243, "step": 985 }, { "epoch": 0.3090909090909091, "grad_norm": 5.080252170562744, "learning_rate": 4.93e-07, "loss": 13.2643, "step": 986 }, { "epoch": 0.3094043887147335, "grad_norm": 3.6815905570983887, "learning_rate": 4.935e-07, "loss": 10.5387, "step": 987 }, { "epoch": 0.30971786833855797, "grad_norm": 5.263372898101807, "learning_rate": 4.940000000000001e-07, "loss": 16.638, "step": 988 }, { "epoch": 0.31003134796238246, "grad_norm": 2.8743481636047363, "learning_rate": 4.945000000000001e-07, "loss": 9.2944, "step": 989 }, { "epoch": 0.3103448275862069, "grad_norm": 5.4028096199035645, "learning_rate": 4.95e-07, "loss": 13.0038, "step": 990 }, { "epoch": 0.31065830721003135, "grad_norm": 3.379136323928833, "learning_rate": 4.955e-07, "loss": 7.5712, "step": 991 }, { "epoch": 0.3109717868338558, "grad_norm": 3.4020957946777344, "learning_rate": 4.96e-07, "loss": 8.871, "step": 992 }, { "epoch": 0.31128526645768023, "grad_norm": 4.882453918457031, "learning_rate": 4.965e-07, "loss": 12.0715, "step": 993 }, { "epoch": 0.31159874608150473, "grad_norm": 5.079671859741211, "learning_rate": 4.970000000000001e-07, "loss": 14.1268, "step": 994 }, { "epoch": 0.31191222570532917, "grad_norm": 4.953401565551758, "learning_rate": 4.975000000000001e-07, "loss": 9.5667, "step": 995 }, { "epoch": 0.3122257053291536, "grad_norm": 3.9187986850738525, "learning_rate": 4.98e-07, "loss": 8.5076, "step": 996 }, { "epoch": 0.31253918495297806, "grad_norm": 3.059863805770874, "learning_rate": 4.985000000000001e-07, "loss": 6.8895, "step": 997 }, { "epoch": 0.3128526645768025, "grad_norm": 3.27205228805542, "learning_rate": 4.99e-07, "loss": 7.4944, "step": 998 }, { "epoch": 0.31316614420062694, "grad_norm": 4.175319194793701, "learning_rate": 4.995e-07, "loss": 10.0083, "step": 999 }, { "epoch": 0.31347962382445144, "grad_norm": 4.775247097015381, "learning_rate": 5.000000000000001e-07, "loss": 10.5517, "step": 1000 }, { "epoch": 0.3137931034482759, "grad_norm": 4.07278299331665, "learning_rate": 5.005e-07, "loss": 10.0079, "step": 1001 }, { "epoch": 0.3141065830721003, "grad_norm": 3.5604565143585205, "learning_rate": 5.01e-07, "loss": 9.4462, "step": 1002 }, { "epoch": 0.31442006269592476, "grad_norm": 4.0364990234375, "learning_rate": 5.015000000000001e-07, "loss": 10.3931, "step": 1003 }, { "epoch": 0.3147335423197492, "grad_norm": 4.032431125640869, "learning_rate": 5.02e-07, "loss": 8.9061, "step": 1004 }, { "epoch": 0.31504702194357365, "grad_norm": 3.2184982299804688, "learning_rate": 5.025000000000001e-07, "loss": 7.8915, "step": 1005 }, { "epoch": 0.31536050156739814, "grad_norm": 3.669436454772949, "learning_rate": 5.03e-07, "loss": 8.4085, "step": 1006 }, { "epoch": 0.3156739811912226, "grad_norm": 3.7609059810638428, "learning_rate": 5.035e-07, "loss": 10.8993, "step": 1007 }, { "epoch": 0.315987460815047, "grad_norm": 3.2515149116516113, "learning_rate": 5.040000000000001e-07, "loss": 10.0338, "step": 1008 }, { "epoch": 0.31630094043887147, "grad_norm": 4.648274898529053, "learning_rate": 5.045000000000001e-07, "loss": 11.952, "step": 1009 }, { "epoch": 0.3166144200626959, "grad_norm": 3.8987960815429688, "learning_rate": 5.05e-07, "loss": 8.06, "step": 1010 }, { "epoch": 0.31692789968652035, "grad_norm": 3.405416965484619, "learning_rate": 5.055e-07, "loss": 6.5823, "step": 1011 }, { "epoch": 0.31724137931034485, "grad_norm": 4.502613067626953, "learning_rate": 5.06e-07, "loss": 13.0741, "step": 1012 }, { "epoch": 0.3175548589341693, "grad_norm": 3.1743547916412354, "learning_rate": 5.065e-07, "loss": 6.6114, "step": 1013 }, { "epoch": 0.31786833855799373, "grad_norm": 3.0800437927246094, "learning_rate": 5.070000000000001e-07, "loss": 7.7446, "step": 1014 }, { "epoch": 0.3181818181818182, "grad_norm": 4.586092472076416, "learning_rate": 5.075000000000001e-07, "loss": 12.149, "step": 1015 }, { "epoch": 0.3184952978056426, "grad_norm": 4.496054649353027, "learning_rate": 5.08e-07, "loss": 9.6337, "step": 1016 }, { "epoch": 0.31880877742946706, "grad_norm": 2.7630536556243896, "learning_rate": 5.085000000000001e-07, "loss": 6.5648, "step": 1017 }, { "epoch": 0.31912225705329156, "grad_norm": 4.153172016143799, "learning_rate": 5.09e-07, "loss": 8.3805, "step": 1018 }, { "epoch": 0.319435736677116, "grad_norm": 3.8220574855804443, "learning_rate": 5.095000000000001e-07, "loss": 8.8453, "step": 1019 }, { "epoch": 0.31974921630094044, "grad_norm": 3.7293951511383057, "learning_rate": 5.1e-07, "loss": 9.2166, "step": 1020 }, { "epoch": 0.3200626959247649, "grad_norm": 3.5755627155303955, "learning_rate": 5.105e-07, "loss": 8.9008, "step": 1021 }, { "epoch": 0.3203761755485893, "grad_norm": 3.9862473011016846, "learning_rate": 5.110000000000001e-07, "loss": 11.2926, "step": 1022 }, { "epoch": 0.32068965517241377, "grad_norm": 3.1247386932373047, "learning_rate": 5.115000000000001e-07, "loss": 7.0044, "step": 1023 }, { "epoch": 0.32100313479623827, "grad_norm": 3.4025156497955322, "learning_rate": 5.12e-07, "loss": 8.6599, "step": 1024 }, { "epoch": 0.3213166144200627, "grad_norm": 3.8938844203948975, "learning_rate": 5.125e-07, "loss": 7.4781, "step": 1025 }, { "epoch": 0.32163009404388715, "grad_norm": 3.694965124130249, "learning_rate": 5.13e-07, "loss": 8.2075, "step": 1026 }, { "epoch": 0.3219435736677116, "grad_norm": 3.9770612716674805, "learning_rate": 5.135e-07, "loss": 6.7056, "step": 1027 }, { "epoch": 0.32225705329153603, "grad_norm": 3.3346900939941406, "learning_rate": 5.140000000000001e-07, "loss": 6.8695, "step": 1028 }, { "epoch": 0.3225705329153605, "grad_norm": 4.177748680114746, "learning_rate": 5.145000000000001e-07, "loss": 11.471, "step": 1029 }, { "epoch": 0.322884012539185, "grad_norm": 2.610774517059326, "learning_rate": 5.15e-07, "loss": 7.1106, "step": 1030 }, { "epoch": 0.3231974921630094, "grad_norm": 4.45445442199707, "learning_rate": 5.155e-07, "loss": 9.852, "step": 1031 }, { "epoch": 0.32351097178683386, "grad_norm": 3.7908337116241455, "learning_rate": 5.16e-07, "loss": 7.8356, "step": 1032 }, { "epoch": 0.3238244514106583, "grad_norm": 4.410386562347412, "learning_rate": 5.165e-07, "loss": 9.0971, "step": 1033 }, { "epoch": 0.32413793103448274, "grad_norm": 3.542137384414673, "learning_rate": 5.170000000000001e-07, "loss": 11.263, "step": 1034 }, { "epoch": 0.32445141065830724, "grad_norm": 4.1727681159973145, "learning_rate": 5.175e-07, "loss": 13.1233, "step": 1035 }, { "epoch": 0.3247648902821317, "grad_norm": 3.6058053970336914, "learning_rate": 5.180000000000001e-07, "loss": 9.9525, "step": 1036 }, { "epoch": 0.3250783699059561, "grad_norm": 4.465059757232666, "learning_rate": 5.185000000000001e-07, "loss": 9.9752, "step": 1037 }, { "epoch": 0.32539184952978056, "grad_norm": 4.481200695037842, "learning_rate": 5.19e-07, "loss": 9.631, "step": 1038 }, { "epoch": 0.325705329153605, "grad_norm": 4.167693138122559, "learning_rate": 5.195000000000001e-07, "loss": 10.6723, "step": 1039 }, { "epoch": 0.32601880877742945, "grad_norm": 4.743021011352539, "learning_rate": 5.2e-07, "loss": 10.3528, "step": 1040 }, { "epoch": 0.32633228840125394, "grad_norm": 3.419243812561035, "learning_rate": 5.205e-07, "loss": 6.2905, "step": 1041 }, { "epoch": 0.3266457680250784, "grad_norm": 3.0460777282714844, "learning_rate": 5.210000000000001e-07, "loss": 6.7207, "step": 1042 }, { "epoch": 0.32695924764890283, "grad_norm": 3.9700310230255127, "learning_rate": 5.215000000000001e-07, "loss": 8.2999, "step": 1043 }, { "epoch": 0.32727272727272727, "grad_norm": 3.008124351501465, "learning_rate": 5.22e-07, "loss": 7.1742, "step": 1044 }, { "epoch": 0.3275862068965517, "grad_norm": 3.2137064933776855, "learning_rate": 5.225e-07, "loss": 6.8929, "step": 1045 }, { "epoch": 0.32789968652037615, "grad_norm": 4.896831035614014, "learning_rate": 5.23e-07, "loss": 11.7279, "step": 1046 }, { "epoch": 0.32821316614420065, "grad_norm": 3.5546038150787354, "learning_rate": 5.235e-07, "loss": 7.8732, "step": 1047 }, { "epoch": 0.3285266457680251, "grad_norm": 3.5171022415161133, "learning_rate": 5.240000000000001e-07, "loss": 8.6413, "step": 1048 }, { "epoch": 0.32884012539184954, "grad_norm": 3.2662153244018555, "learning_rate": 5.245e-07, "loss": 8.4025, "step": 1049 }, { "epoch": 0.329153605015674, "grad_norm": 4.3712005615234375, "learning_rate": 5.250000000000001e-07, "loss": 8.6959, "step": 1050 }, { "epoch": 0.3294670846394984, "grad_norm": 3.437635660171509, "learning_rate": 5.255e-07, "loss": 9.4805, "step": 1051 }, { "epoch": 0.32978056426332286, "grad_norm": 4.1790876388549805, "learning_rate": 5.26e-07, "loss": 8.4143, "step": 1052 }, { "epoch": 0.33009404388714736, "grad_norm": 3.861347198486328, "learning_rate": 5.265000000000001e-07, "loss": 7.995, "step": 1053 }, { "epoch": 0.3304075235109718, "grad_norm": 4.775915622711182, "learning_rate": 5.27e-07, "loss": 13.2288, "step": 1054 }, { "epoch": 0.33072100313479624, "grad_norm": 3.08349347114563, "learning_rate": 5.275e-07, "loss": 8.4152, "step": 1055 }, { "epoch": 0.3310344827586207, "grad_norm": 4.555473804473877, "learning_rate": 5.280000000000001e-07, "loss": 9.4847, "step": 1056 }, { "epoch": 0.3313479623824451, "grad_norm": 4.14658260345459, "learning_rate": 5.285000000000001e-07, "loss": 9.6713, "step": 1057 }, { "epoch": 0.33166144200626957, "grad_norm": 3.729527711868286, "learning_rate": 5.29e-07, "loss": 9.5404, "step": 1058 }, { "epoch": 0.33197492163009407, "grad_norm": 4.110138893127441, "learning_rate": 5.295e-07, "loss": 7.4485, "step": 1059 }, { "epoch": 0.3322884012539185, "grad_norm": 3.4189460277557373, "learning_rate": 5.3e-07, "loss": 9.1945, "step": 1060 }, { "epoch": 0.33260188087774295, "grad_norm": 4.435842990875244, "learning_rate": 5.305e-07, "loss": 9.3644, "step": 1061 }, { "epoch": 0.3329153605015674, "grad_norm": 3.913818836212158, "learning_rate": 5.310000000000001e-07, "loss": 10.439, "step": 1062 }, { "epoch": 0.33322884012539183, "grad_norm": 5.079927921295166, "learning_rate": 5.315000000000001e-07, "loss": 11.0371, "step": 1063 }, { "epoch": 0.3335423197492163, "grad_norm": 3.028818368911743, "learning_rate": 5.32e-07, "loss": 6.7478, "step": 1064 }, { "epoch": 0.3338557993730408, "grad_norm": 2.9255435466766357, "learning_rate": 5.325e-07, "loss": 6.9333, "step": 1065 }, { "epoch": 0.3341692789968652, "grad_norm": 3.4844298362731934, "learning_rate": 5.33e-07, "loss": 8.3449, "step": 1066 }, { "epoch": 0.33448275862068966, "grad_norm": 4.424320697784424, "learning_rate": 5.335000000000001e-07, "loss": 11.034, "step": 1067 }, { "epoch": 0.3347962382445141, "grad_norm": 4.413206577301025, "learning_rate": 5.340000000000001e-07, "loss": 9.8029, "step": 1068 }, { "epoch": 0.33510971786833854, "grad_norm": 3.4747884273529053, "learning_rate": 5.345e-07, "loss": 8.9154, "step": 1069 }, { "epoch": 0.335423197492163, "grad_norm": 3.6665916442871094, "learning_rate": 5.350000000000001e-07, "loss": 7.1011, "step": 1070 }, { "epoch": 0.3357366771159875, "grad_norm": 2.8998830318450928, "learning_rate": 5.355e-07, "loss": 7.7268, "step": 1071 }, { "epoch": 0.3360501567398119, "grad_norm": 3.9192614555358887, "learning_rate": 5.36e-07, "loss": 10.9592, "step": 1072 }, { "epoch": 0.33636363636363636, "grad_norm": 2.7648117542266846, "learning_rate": 5.365000000000001e-07, "loss": 7.8875, "step": 1073 }, { "epoch": 0.3366771159874608, "grad_norm": 4.6224751472473145, "learning_rate": 5.37e-07, "loss": 10.7205, "step": 1074 }, { "epoch": 0.33699059561128525, "grad_norm": 3.7434327602386475, "learning_rate": 5.375e-07, "loss": 9.4757, "step": 1075 }, { "epoch": 0.3373040752351097, "grad_norm": 2.9543545246124268, "learning_rate": 5.380000000000001e-07, "loss": 6.5248, "step": 1076 }, { "epoch": 0.3376175548589342, "grad_norm": 4.7394185066223145, "learning_rate": 5.385000000000001e-07, "loss": 16.7804, "step": 1077 }, { "epoch": 0.33793103448275863, "grad_norm": 3.574186325073242, "learning_rate": 5.39e-07, "loss": 7.1065, "step": 1078 }, { "epoch": 0.33824451410658307, "grad_norm": 2.6284937858581543, "learning_rate": 5.395e-07, "loss": 6.8801, "step": 1079 }, { "epoch": 0.3385579937304075, "grad_norm": 3.36057710647583, "learning_rate": 5.4e-07, "loss": 10.2646, "step": 1080 }, { "epoch": 0.33887147335423196, "grad_norm": 4.6945624351501465, "learning_rate": 5.405000000000001e-07, "loss": 12.4643, "step": 1081 }, { "epoch": 0.33918495297805645, "grad_norm": 3.1167168617248535, "learning_rate": 5.410000000000001e-07, "loss": 6.7228, "step": 1082 }, { "epoch": 0.3394984326018809, "grad_norm": 3.513232469558716, "learning_rate": 5.415e-07, "loss": 9.1163, "step": 1083 }, { "epoch": 0.33981191222570534, "grad_norm": 5.232317924499512, "learning_rate": 5.420000000000001e-07, "loss": 11.4712, "step": 1084 }, { "epoch": 0.3401253918495298, "grad_norm": 3.7332653999328613, "learning_rate": 5.425e-07, "loss": 7.6378, "step": 1085 }, { "epoch": 0.3404388714733542, "grad_norm": 4.457612037658691, "learning_rate": 5.43e-07, "loss": 11.9674, "step": 1086 }, { "epoch": 0.34075235109717866, "grad_norm": 3.196615695953369, "learning_rate": 5.435000000000001e-07, "loss": 8.66, "step": 1087 }, { "epoch": 0.34106583072100316, "grad_norm": 4.207455635070801, "learning_rate": 5.44e-07, "loss": 8.5547, "step": 1088 }, { "epoch": 0.3413793103448276, "grad_norm": 3.803297519683838, "learning_rate": 5.445e-07, "loss": 8.1143, "step": 1089 }, { "epoch": 0.34169278996865204, "grad_norm": 3.9146108627319336, "learning_rate": 5.450000000000001e-07, "loss": 8.9048, "step": 1090 }, { "epoch": 0.3420062695924765, "grad_norm": 3.0338478088378906, "learning_rate": 5.455e-07, "loss": 7.1311, "step": 1091 }, { "epoch": 0.34231974921630093, "grad_norm": 3.301464557647705, "learning_rate": 5.46e-07, "loss": 7.2672, "step": 1092 }, { "epoch": 0.34263322884012537, "grad_norm": 3.3662900924682617, "learning_rate": 5.465e-07, "loss": 9.0924, "step": 1093 }, { "epoch": 0.34294670846394987, "grad_norm": 3.946822166442871, "learning_rate": 5.47e-07, "loss": 10.3786, "step": 1094 }, { "epoch": 0.3432601880877743, "grad_norm": 3.5544939041137695, "learning_rate": 5.475e-07, "loss": 7.2466, "step": 1095 }, { "epoch": 0.34357366771159875, "grad_norm": 3.7304489612579346, "learning_rate": 5.480000000000001e-07, "loss": 9.4879, "step": 1096 }, { "epoch": 0.3438871473354232, "grad_norm": 4.137592315673828, "learning_rate": 5.485e-07, "loss": 8.2506, "step": 1097 }, { "epoch": 0.34420062695924764, "grad_norm": 3.8458902835845947, "learning_rate": 5.490000000000001e-07, "loss": 10.3391, "step": 1098 }, { "epoch": 0.3445141065830721, "grad_norm": 3.944636106491089, "learning_rate": 5.495e-07, "loss": 10.2381, "step": 1099 }, { "epoch": 0.3448275862068966, "grad_norm": 2.935556411743164, "learning_rate": 5.5e-07, "loss": 6.7504, "step": 1100 }, { "epoch": 0.345141065830721, "grad_norm": 2.839891195297241, "learning_rate": 5.505000000000001e-07, "loss": 6.2424, "step": 1101 }, { "epoch": 0.34545454545454546, "grad_norm": 3.0521023273468018, "learning_rate": 5.510000000000001e-07, "loss": 8.1019, "step": 1102 }, { "epoch": 0.3457680250783699, "grad_norm": 2.7549774646759033, "learning_rate": 5.515e-07, "loss": 6.9908, "step": 1103 }, { "epoch": 0.34608150470219434, "grad_norm": 2.460770845413208, "learning_rate": 5.520000000000001e-07, "loss": 6.4588, "step": 1104 }, { "epoch": 0.3463949843260188, "grad_norm": 3.0610549449920654, "learning_rate": 5.525e-07, "loss": 7.1453, "step": 1105 }, { "epoch": 0.3467084639498433, "grad_norm": 3.7821202278137207, "learning_rate": 5.53e-07, "loss": 8.4413, "step": 1106 }, { "epoch": 0.3470219435736677, "grad_norm": 3.272524356842041, "learning_rate": 5.535000000000001e-07, "loss": 7.5479, "step": 1107 }, { "epoch": 0.34733542319749217, "grad_norm": 4.5854082107543945, "learning_rate": 5.54e-07, "loss": 12.7431, "step": 1108 }, { "epoch": 0.3476489028213166, "grad_norm": 3.382383346557617, "learning_rate": 5.545e-07, "loss": 8.0938, "step": 1109 }, { "epoch": 0.34796238244514105, "grad_norm": 4.590867042541504, "learning_rate": 5.550000000000001e-07, "loss": 9.5862, "step": 1110 }, { "epoch": 0.3482758620689655, "grad_norm": 3.3720743656158447, "learning_rate": 5.555e-07, "loss": 7.8215, "step": 1111 }, { "epoch": 0.34858934169279, "grad_norm": 3.0238683223724365, "learning_rate": 5.560000000000001e-07, "loss": 8.9804, "step": 1112 }, { "epoch": 0.34890282131661443, "grad_norm": 3.9639229774475098, "learning_rate": 5.565e-07, "loss": 9.2357, "step": 1113 }, { "epoch": 0.3492163009404389, "grad_norm": 2.6964738368988037, "learning_rate": 5.57e-07, "loss": 7.0609, "step": 1114 }, { "epoch": 0.3495297805642633, "grad_norm": 4.804142475128174, "learning_rate": 5.575000000000001e-07, "loss": 13.9845, "step": 1115 }, { "epoch": 0.34984326018808776, "grad_norm": 4.389916896820068, "learning_rate": 5.580000000000001e-07, "loss": 7.7403, "step": 1116 }, { "epoch": 0.3501567398119122, "grad_norm": 4.200212478637695, "learning_rate": 5.585e-07, "loss": 9.193, "step": 1117 }, { "epoch": 0.3504702194357367, "grad_norm": 4.732844829559326, "learning_rate": 5.590000000000001e-07, "loss": 13.0463, "step": 1118 }, { "epoch": 0.35078369905956114, "grad_norm": 2.878286123275757, "learning_rate": 5.595e-07, "loss": 6.3315, "step": 1119 }, { "epoch": 0.3510971786833856, "grad_norm": 4.683932781219482, "learning_rate": 5.6e-07, "loss": 10.1464, "step": 1120 }, { "epoch": 0.35141065830721, "grad_norm": 3.83774995803833, "learning_rate": 5.605000000000001e-07, "loss": 8.5081, "step": 1121 }, { "epoch": 0.35172413793103446, "grad_norm": 3.030034303665161, "learning_rate": 5.61e-07, "loss": 6.6782, "step": 1122 }, { "epoch": 0.35203761755485896, "grad_norm": 3.7640645503997803, "learning_rate": 5.615e-07, "loss": 7.6323, "step": 1123 }, { "epoch": 0.3523510971786834, "grad_norm": 3.636522054672241, "learning_rate": 5.620000000000001e-07, "loss": 8.9132, "step": 1124 }, { "epoch": 0.35266457680250785, "grad_norm": 5.256652355194092, "learning_rate": 5.625e-07, "loss": 13.0476, "step": 1125 }, { "epoch": 0.3529780564263323, "grad_norm": 3.7114644050598145, "learning_rate": 5.63e-07, "loss": 7.7855, "step": 1126 }, { "epoch": 0.35329153605015673, "grad_norm": 4.469037055969238, "learning_rate": 5.635e-07, "loss": 10.8488, "step": 1127 }, { "epoch": 0.35360501567398117, "grad_norm": 4.428511619567871, "learning_rate": 5.64e-07, "loss": 10.8417, "step": 1128 }, { "epoch": 0.35391849529780567, "grad_norm": 3.272402048110962, "learning_rate": 5.645000000000001e-07, "loss": 7.0908, "step": 1129 }, { "epoch": 0.3542319749216301, "grad_norm": 4.514960765838623, "learning_rate": 5.650000000000001e-07, "loss": 11.1969, "step": 1130 }, { "epoch": 0.35454545454545455, "grad_norm": 3.0566494464874268, "learning_rate": 5.655e-07, "loss": 6.9586, "step": 1131 }, { "epoch": 0.354858934169279, "grad_norm": 4.796217441558838, "learning_rate": 5.660000000000001e-07, "loss": 12.4509, "step": 1132 }, { "epoch": 0.35517241379310344, "grad_norm": 3.8200011253356934, "learning_rate": 5.665e-07, "loss": 7.6372, "step": 1133 }, { "epoch": 0.3554858934169279, "grad_norm": 4.99943733215332, "learning_rate": 5.67e-07, "loss": 16.3218, "step": 1134 }, { "epoch": 0.3557993730407524, "grad_norm": 4.074048042297363, "learning_rate": 5.675000000000001e-07, "loss": 11.7632, "step": 1135 }, { "epoch": 0.3561128526645768, "grad_norm": 3.519336700439453, "learning_rate": 5.680000000000001e-07, "loss": 12.5035, "step": 1136 }, { "epoch": 0.35642633228840126, "grad_norm": 3.5221669673919678, "learning_rate": 5.685e-07, "loss": 8.2566, "step": 1137 }, { "epoch": 0.3567398119122257, "grad_norm": 3.0210797786712646, "learning_rate": 5.690000000000001e-07, "loss": 7.6131, "step": 1138 }, { "epoch": 0.35705329153605014, "grad_norm": 3.848649740219116, "learning_rate": 5.695e-07, "loss": 9.2893, "step": 1139 }, { "epoch": 0.3573667711598746, "grad_norm": 3.4947142601013184, "learning_rate": 5.7e-07, "loss": 10.0174, "step": 1140 }, { "epoch": 0.3576802507836991, "grad_norm": 4.418102741241455, "learning_rate": 5.705e-07, "loss": 10.2442, "step": 1141 }, { "epoch": 0.3579937304075235, "grad_norm": 3.0057661533355713, "learning_rate": 5.71e-07, "loss": 6.8693, "step": 1142 }, { "epoch": 0.35830721003134797, "grad_norm": 5.070820331573486, "learning_rate": 5.715000000000001e-07, "loss": 12.646, "step": 1143 }, { "epoch": 0.3586206896551724, "grad_norm": 3.685112714767456, "learning_rate": 5.720000000000001e-07, "loss": 8.9385, "step": 1144 }, { "epoch": 0.35893416927899685, "grad_norm": 3.591017007827759, "learning_rate": 5.725e-07, "loss": 7.2102, "step": 1145 }, { "epoch": 0.3592476489028213, "grad_norm": 3.66312575340271, "learning_rate": 5.730000000000001e-07, "loss": 8.0567, "step": 1146 }, { "epoch": 0.3595611285266458, "grad_norm": 3.2532527446746826, "learning_rate": 5.735e-07, "loss": 7.5575, "step": 1147 }, { "epoch": 0.35987460815047023, "grad_norm": 2.452608108520508, "learning_rate": 5.74e-07, "loss": 6.1812, "step": 1148 }, { "epoch": 0.3601880877742947, "grad_norm": 4.427051067352295, "learning_rate": 5.745000000000001e-07, "loss": 9.7562, "step": 1149 }, { "epoch": 0.3605015673981191, "grad_norm": 4.048511981964111, "learning_rate": 5.750000000000001e-07, "loss": 7.2065, "step": 1150 }, { "epoch": 0.36081504702194356, "grad_norm": 4.313746929168701, "learning_rate": 5.755e-07, "loss": 12.0608, "step": 1151 }, { "epoch": 0.361128526645768, "grad_norm": 3.703238010406494, "learning_rate": 5.760000000000001e-07, "loss": 8.8626, "step": 1152 }, { "epoch": 0.3614420062695925, "grad_norm": 3.321089744567871, "learning_rate": 5.765e-07, "loss": 7.7751, "step": 1153 }, { "epoch": 0.36175548589341694, "grad_norm": 3.596169948577881, "learning_rate": 5.77e-07, "loss": 7.2633, "step": 1154 }, { "epoch": 0.3620689655172414, "grad_norm": 3.9182851314544678, "learning_rate": 5.775000000000001e-07, "loss": 8.9107, "step": 1155 }, { "epoch": 0.3623824451410658, "grad_norm": 2.809251070022583, "learning_rate": 5.78e-07, "loss": 7.5999, "step": 1156 }, { "epoch": 0.36269592476489027, "grad_norm": 3.9834253787994385, "learning_rate": 5.785e-07, "loss": 10.8149, "step": 1157 }, { "epoch": 0.3630094043887147, "grad_norm": 3.9162814617156982, "learning_rate": 5.790000000000001e-07, "loss": 9.0022, "step": 1158 }, { "epoch": 0.3633228840125392, "grad_norm": 3.7181437015533447, "learning_rate": 5.795e-07, "loss": 12.1042, "step": 1159 }, { "epoch": 0.36363636363636365, "grad_norm": 3.6745083332061768, "learning_rate": 5.800000000000001e-07, "loss": 10.9387, "step": 1160 }, { "epoch": 0.3639498432601881, "grad_norm": 3.6251866817474365, "learning_rate": 5.805e-07, "loss": 9.3794, "step": 1161 }, { "epoch": 0.36426332288401253, "grad_norm": 3.9675116539001465, "learning_rate": 5.81e-07, "loss": 7.4843, "step": 1162 }, { "epoch": 0.364576802507837, "grad_norm": 4.0218915939331055, "learning_rate": 5.815000000000001e-07, "loss": 11.3793, "step": 1163 }, { "epoch": 0.36489028213166147, "grad_norm": 4.011270999908447, "learning_rate": 5.820000000000001e-07, "loss": 8.5104, "step": 1164 }, { "epoch": 0.3652037617554859, "grad_norm": 3.468784809112549, "learning_rate": 5.825e-07, "loss": 7.195, "step": 1165 }, { "epoch": 0.36551724137931035, "grad_norm": 5.364189624786377, "learning_rate": 5.830000000000001e-07, "loss": 19.2203, "step": 1166 }, { "epoch": 0.3658307210031348, "grad_norm": 4.526871681213379, "learning_rate": 5.835e-07, "loss": 11.444, "step": 1167 }, { "epoch": 0.36614420062695924, "grad_norm": 3.9016737937927246, "learning_rate": 5.84e-07, "loss": 11.455, "step": 1168 }, { "epoch": 0.3664576802507837, "grad_norm": 3.691763401031494, "learning_rate": 5.845000000000001e-07, "loss": 8.6588, "step": 1169 }, { "epoch": 0.3667711598746082, "grad_norm": 3.842839241027832, "learning_rate": 5.850000000000001e-07, "loss": 8.317, "step": 1170 }, { "epoch": 0.3670846394984326, "grad_norm": 3.8796255588531494, "learning_rate": 5.855e-07, "loss": 9.903, "step": 1171 }, { "epoch": 0.36739811912225706, "grad_norm": 4.093709468841553, "learning_rate": 5.860000000000001e-07, "loss": 9.6562, "step": 1172 }, { "epoch": 0.3677115987460815, "grad_norm": 4.154341220855713, "learning_rate": 5.865e-07, "loss": 9.7709, "step": 1173 }, { "epoch": 0.36802507836990594, "grad_norm": 5.882456302642822, "learning_rate": 5.870000000000001e-07, "loss": 16.238, "step": 1174 }, { "epoch": 0.3683385579937304, "grad_norm": 4.124493598937988, "learning_rate": 5.875e-07, "loss": 9.4565, "step": 1175 }, { "epoch": 0.3686520376175549, "grad_norm": 3.4049675464630127, "learning_rate": 5.88e-07, "loss": 6.7959, "step": 1176 }, { "epoch": 0.3689655172413793, "grad_norm": 3.196589708328247, "learning_rate": 5.885000000000001e-07, "loss": 6.6816, "step": 1177 }, { "epoch": 0.36927899686520377, "grad_norm": 4.004443645477295, "learning_rate": 5.890000000000001e-07, "loss": 9.0686, "step": 1178 }, { "epoch": 0.3695924764890282, "grad_norm": 3.7830114364624023, "learning_rate": 5.895e-07, "loss": 8.4215, "step": 1179 }, { "epoch": 0.36990595611285265, "grad_norm": 16.382652282714844, "learning_rate": 5.900000000000001e-07, "loss": 8.5894, "step": 1180 }, { "epoch": 0.3702194357366771, "grad_norm": 4.150466442108154, "learning_rate": 5.905e-07, "loss": 9.9268, "step": 1181 }, { "epoch": 0.3705329153605016, "grad_norm": 4.039095401763916, "learning_rate": 5.91e-07, "loss": 7.6783, "step": 1182 }, { "epoch": 0.37084639498432603, "grad_norm": 4.113837242126465, "learning_rate": 5.915000000000001e-07, "loss": 10.9473, "step": 1183 }, { "epoch": 0.3711598746081505, "grad_norm": 4.3714447021484375, "learning_rate": 5.920000000000001e-07, "loss": 8.6596, "step": 1184 }, { "epoch": 0.3714733542319749, "grad_norm": 4.696350574493408, "learning_rate": 5.925e-07, "loss": 11.0915, "step": 1185 }, { "epoch": 0.37178683385579936, "grad_norm": 3.4268805980682373, "learning_rate": 5.930000000000001e-07, "loss": 8.7898, "step": 1186 }, { "epoch": 0.3721003134796238, "grad_norm": 3.7892448902130127, "learning_rate": 5.935e-07, "loss": 7.9897, "step": 1187 }, { "epoch": 0.3724137931034483, "grad_norm": 3.5174129009246826, "learning_rate": 5.94e-07, "loss": 8.9607, "step": 1188 }, { "epoch": 0.37272727272727274, "grad_norm": 3.5492613315582275, "learning_rate": 5.945000000000001e-07, "loss": 8.2248, "step": 1189 }, { "epoch": 0.3730407523510972, "grad_norm": 3.0557780265808105, "learning_rate": 5.95e-07, "loss": 7.4525, "step": 1190 }, { "epoch": 0.3733542319749216, "grad_norm": 3.610521078109741, "learning_rate": 5.955000000000001e-07, "loss": 8.6173, "step": 1191 }, { "epoch": 0.37366771159874607, "grad_norm": 5.024316310882568, "learning_rate": 5.960000000000001e-07, "loss": 9.0102, "step": 1192 }, { "epoch": 0.3739811912225705, "grad_norm": 3.3854024410247803, "learning_rate": 5.965e-07, "loss": 9.3159, "step": 1193 }, { "epoch": 0.374294670846395, "grad_norm": 3.692943572998047, "learning_rate": 5.970000000000001e-07, "loss": 6.6683, "step": 1194 }, { "epoch": 0.37460815047021945, "grad_norm": 4.656247138977051, "learning_rate": 5.975e-07, "loss": 9.6465, "step": 1195 }, { "epoch": 0.3749216300940439, "grad_norm": 2.7945048809051514, "learning_rate": 5.98e-07, "loss": 7.6828, "step": 1196 }, { "epoch": 0.37523510971786833, "grad_norm": 3.6382293701171875, "learning_rate": 5.985000000000001e-07, "loss": 8.2837, "step": 1197 }, { "epoch": 0.3755485893416928, "grad_norm": 4.916844844818115, "learning_rate": 5.990000000000001e-07, "loss": 10.6478, "step": 1198 }, { "epoch": 0.3758620689655172, "grad_norm": 2.6404612064361572, "learning_rate": 5.995e-07, "loss": 5.8349, "step": 1199 }, { "epoch": 0.3761755485893417, "grad_norm": 3.963083505630493, "learning_rate": 6.000000000000001e-07, "loss": 7.5912, "step": 1200 }, { "epoch": 0.37648902821316615, "grad_norm": 3.9304614067077637, "learning_rate": 6.005e-07, "loss": 9.4547, "step": 1201 }, { "epoch": 0.3768025078369906, "grad_norm": 3.4377453327178955, "learning_rate": 6.01e-07, "loss": 9.7902, "step": 1202 }, { "epoch": 0.37711598746081504, "grad_norm": 3.773444414138794, "learning_rate": 6.015000000000001e-07, "loss": 9.5372, "step": 1203 }, { "epoch": 0.3774294670846395, "grad_norm": 3.075242280960083, "learning_rate": 6.02e-07, "loss": 8.3831, "step": 1204 }, { "epoch": 0.3777429467084639, "grad_norm": 3.2014951705932617, "learning_rate": 6.025000000000001e-07, "loss": 7.9814, "step": 1205 }, { "epoch": 0.3780564263322884, "grad_norm": 3.752372980117798, "learning_rate": 6.030000000000001e-07, "loss": 8.109, "step": 1206 }, { "epoch": 0.37836990595611286, "grad_norm": 4.027518272399902, "learning_rate": 6.035e-07, "loss": 7.6091, "step": 1207 }, { "epoch": 0.3786833855799373, "grad_norm": 3.438729763031006, "learning_rate": 6.040000000000001e-07, "loss": 6.8037, "step": 1208 }, { "epoch": 0.37899686520376175, "grad_norm": 4.082209587097168, "learning_rate": 6.045e-07, "loss": 9.9197, "step": 1209 }, { "epoch": 0.3793103448275862, "grad_norm": 5.536364555358887, "learning_rate": 6.05e-07, "loss": 13.404, "step": 1210 }, { "epoch": 0.3796238244514107, "grad_norm": 3.8817527294158936, "learning_rate": 6.055000000000001e-07, "loss": 11.2098, "step": 1211 }, { "epoch": 0.3799373040752351, "grad_norm": 3.6680006980895996, "learning_rate": 6.060000000000001e-07, "loss": 8.0359, "step": 1212 }, { "epoch": 0.38025078369905957, "grad_norm": 4.336513519287109, "learning_rate": 6.065e-07, "loss": 10.526, "step": 1213 }, { "epoch": 0.380564263322884, "grad_norm": 3.7376954555511475, "learning_rate": 6.07e-07, "loss": 9.7812, "step": 1214 }, { "epoch": 0.38087774294670845, "grad_norm": 4.394288539886475, "learning_rate": 6.075e-07, "loss": 10.5658, "step": 1215 }, { "epoch": 0.3811912225705329, "grad_norm": 3.9178755283355713, "learning_rate": 6.08e-07, "loss": 9.1735, "step": 1216 }, { "epoch": 0.3815047021943574, "grad_norm": 3.385538339614868, "learning_rate": 6.085000000000001e-07, "loss": 8.2474, "step": 1217 }, { "epoch": 0.38181818181818183, "grad_norm": 3.3162765502929688, "learning_rate": 6.090000000000001e-07, "loss": 7.6934, "step": 1218 }, { "epoch": 0.3821316614420063, "grad_norm": 4.181975364685059, "learning_rate": 6.095e-07, "loss": 9.205, "step": 1219 }, { "epoch": 0.3824451410658307, "grad_norm": 3.707218647003174, "learning_rate": 6.100000000000001e-07, "loss": 7.6344, "step": 1220 }, { "epoch": 0.38275862068965516, "grad_norm": 3.0025179386138916, "learning_rate": 6.105e-07, "loss": 8.5384, "step": 1221 }, { "epoch": 0.3830721003134796, "grad_norm": 3.2136359214782715, "learning_rate": 6.110000000000001e-07, "loss": 7.3391, "step": 1222 }, { "epoch": 0.3833855799373041, "grad_norm": 3.87955904006958, "learning_rate": 6.115000000000001e-07, "loss": 9.051, "step": 1223 }, { "epoch": 0.38369905956112854, "grad_norm": 2.9862728118896484, "learning_rate": 6.12e-07, "loss": 6.7873, "step": 1224 }, { "epoch": 0.384012539184953, "grad_norm": 4.336016654968262, "learning_rate": 6.125000000000001e-07, "loss": 8.1016, "step": 1225 }, { "epoch": 0.3843260188087774, "grad_norm": 3.270927906036377, "learning_rate": 6.130000000000001e-07, "loss": 9.3506, "step": 1226 }, { "epoch": 0.38463949843260187, "grad_norm": 4.06575870513916, "learning_rate": 6.135e-07, "loss": 8.5931, "step": 1227 }, { "epoch": 0.3849529780564263, "grad_norm": 3.6191561222076416, "learning_rate": 6.140000000000001e-07, "loss": 8.4115, "step": 1228 }, { "epoch": 0.3852664576802508, "grad_norm": 2.8012547492980957, "learning_rate": 6.145e-07, "loss": 7.0199, "step": 1229 }, { "epoch": 0.38557993730407525, "grad_norm": 4.054346084594727, "learning_rate": 6.15e-07, "loss": 9.9699, "step": 1230 }, { "epoch": 0.3858934169278997, "grad_norm": 4.556372165679932, "learning_rate": 6.155000000000001e-07, "loss": 14.4437, "step": 1231 }, { "epoch": 0.38620689655172413, "grad_norm": 4.824500560760498, "learning_rate": 6.160000000000001e-07, "loss": 10.5535, "step": 1232 }, { "epoch": 0.3865203761755486, "grad_norm": 4.935982704162598, "learning_rate": 6.165e-07, "loss": 8.3582, "step": 1233 }, { "epoch": 0.386833855799373, "grad_norm": 8.828557014465332, "learning_rate": 6.17e-07, "loss": 19.3321, "step": 1234 }, { "epoch": 0.3871473354231975, "grad_norm": 2.967073678970337, "learning_rate": 6.175e-07, "loss": 6.7905, "step": 1235 }, { "epoch": 0.38746081504702196, "grad_norm": 4.4314799308776855, "learning_rate": 6.180000000000001e-07, "loss": 10.2915, "step": 1236 }, { "epoch": 0.3877742946708464, "grad_norm": 3.622016191482544, "learning_rate": 6.185000000000001e-07, "loss": 9.6284, "step": 1237 }, { "epoch": 0.38808777429467084, "grad_norm": 3.8630599975585938, "learning_rate": 6.19e-07, "loss": 7.4754, "step": 1238 }, { "epoch": 0.3884012539184953, "grad_norm": 3.7421834468841553, "learning_rate": 6.195000000000001e-07, "loss": 8.5903, "step": 1239 }, { "epoch": 0.3887147335423197, "grad_norm": 3.313210964202881, "learning_rate": 6.200000000000001e-07, "loss": 9.3886, "step": 1240 }, { "epoch": 0.3890282131661442, "grad_norm": 3.029578924179077, "learning_rate": 6.205e-07, "loss": 7.3684, "step": 1241 }, { "epoch": 0.38934169278996866, "grad_norm": 4.0817670822143555, "learning_rate": 6.210000000000001e-07, "loss": 8.6238, "step": 1242 }, { "epoch": 0.3896551724137931, "grad_norm": 3.1163382530212402, "learning_rate": 6.215e-07, "loss": 7.8326, "step": 1243 }, { "epoch": 0.38996865203761755, "grad_norm": 3.418872833251953, "learning_rate": 6.22e-07, "loss": 9.6686, "step": 1244 }, { "epoch": 0.390282131661442, "grad_norm": 3.265856981277466, "learning_rate": 6.225000000000001e-07, "loss": 6.8915, "step": 1245 }, { "epoch": 0.39059561128526643, "grad_norm": 2.5802524089813232, "learning_rate": 6.230000000000001e-07, "loss": 5.6296, "step": 1246 }, { "epoch": 0.39090909090909093, "grad_norm": 3.2665584087371826, "learning_rate": 6.235e-07, "loss": 7.1075, "step": 1247 }, { "epoch": 0.39122257053291537, "grad_norm": 4.159616470336914, "learning_rate": 6.24e-07, "loss": 11.453, "step": 1248 }, { "epoch": 0.3915360501567398, "grad_norm": 4.0924482345581055, "learning_rate": 6.245e-07, "loss": 9.2337, "step": 1249 }, { "epoch": 0.39184952978056425, "grad_norm": 3.432142972946167, "learning_rate": 6.25e-07, "loss": 8.3582, "step": 1250 }, { "epoch": 0.3921630094043887, "grad_norm": 3.7760117053985596, "learning_rate": 6.255e-07, "loss": 10.0285, "step": 1251 }, { "epoch": 0.3924764890282132, "grad_norm": 3.153862237930298, "learning_rate": 6.260000000000001e-07, "loss": 7.4688, "step": 1252 }, { "epoch": 0.39278996865203764, "grad_norm": 3.4726884365081787, "learning_rate": 6.265000000000001e-07, "loss": 9.7888, "step": 1253 }, { "epoch": 0.3931034482758621, "grad_norm": 2.944378614425659, "learning_rate": 6.270000000000001e-07, "loss": 7.6955, "step": 1254 }, { "epoch": 0.3934169278996865, "grad_norm": 3.5531530380249023, "learning_rate": 6.275e-07, "loss": 6.9592, "step": 1255 }, { "epoch": 0.39373040752351096, "grad_norm": 4.495529651641846, "learning_rate": 6.28e-07, "loss": 13.8164, "step": 1256 }, { "epoch": 0.3940438871473354, "grad_norm": 3.5309019088745117, "learning_rate": 6.285000000000001e-07, "loss": 8.0689, "step": 1257 }, { "epoch": 0.3943573667711599, "grad_norm": 5.7128682136535645, "learning_rate": 6.29e-07, "loss": 14.1, "step": 1258 }, { "epoch": 0.39467084639498434, "grad_norm": 2.9314987659454346, "learning_rate": 6.295000000000001e-07, "loss": 6.5813, "step": 1259 }, { "epoch": 0.3949843260188088, "grad_norm": 4.196994781494141, "learning_rate": 6.3e-07, "loss": 11.5749, "step": 1260 }, { "epoch": 0.3952978056426332, "grad_norm": 3.2020435333251953, "learning_rate": 6.305e-07, "loss": 7.9098, "step": 1261 }, { "epoch": 0.39561128526645767, "grad_norm": 4.530463218688965, "learning_rate": 6.310000000000001e-07, "loss": 13.4221, "step": 1262 }, { "epoch": 0.3959247648902821, "grad_norm": 3.823436975479126, "learning_rate": 6.315e-07, "loss": 10.8347, "step": 1263 }, { "epoch": 0.3962382445141066, "grad_norm": 4.340320110321045, "learning_rate": 6.320000000000002e-07, "loss": 10.6129, "step": 1264 }, { "epoch": 0.39655172413793105, "grad_norm": 4.320498466491699, "learning_rate": 6.325000000000001e-07, "loss": 9.6761, "step": 1265 }, { "epoch": 0.3968652037617555, "grad_norm": 3.8908495903015137, "learning_rate": 6.33e-07, "loss": 9.3212, "step": 1266 }, { "epoch": 0.39717868338557993, "grad_norm": 4.79443359375, "learning_rate": 6.335000000000001e-07, "loss": 8.4957, "step": 1267 }, { "epoch": 0.3974921630094044, "grad_norm": 3.5879569053649902, "learning_rate": 6.34e-07, "loss": 7.9517, "step": 1268 }, { "epoch": 0.3978056426332288, "grad_norm": 3.9859533309936523, "learning_rate": 6.345000000000001e-07, "loss": 10.1546, "step": 1269 }, { "epoch": 0.3981191222570533, "grad_norm": 4.457554817199707, "learning_rate": 6.350000000000001e-07, "loss": 11.0071, "step": 1270 }, { "epoch": 0.39843260188087776, "grad_norm": 3.4207284450531006, "learning_rate": 6.355e-07, "loss": 7.5192, "step": 1271 }, { "epoch": 0.3987460815047022, "grad_norm": 2.9599740505218506, "learning_rate": 6.360000000000001e-07, "loss": 8.1417, "step": 1272 }, { "epoch": 0.39905956112852664, "grad_norm": 5.587130069732666, "learning_rate": 6.365000000000001e-07, "loss": 13.4308, "step": 1273 }, { "epoch": 0.3993730407523511, "grad_norm": 4.257179260253906, "learning_rate": 6.370000000000001e-07, "loss": 9.8754, "step": 1274 }, { "epoch": 0.3996865203761755, "grad_norm": 3.8205981254577637, "learning_rate": 6.375e-07, "loss": 7.3584, "step": 1275 }, { "epoch": 0.4, "grad_norm": 4.358394145965576, "learning_rate": 6.38e-07, "loss": 8.0118, "step": 1276 }, { "epoch": 0.40031347962382446, "grad_norm": 3.71220326423645, "learning_rate": 6.385000000000001e-07, "loss": 10.123, "step": 1277 }, { "epoch": 0.4006269592476489, "grad_norm": 3.896798610687256, "learning_rate": 6.39e-07, "loss": 7.8256, "step": 1278 }, { "epoch": 0.40094043887147335, "grad_norm": 3.1705808639526367, "learning_rate": 6.395000000000001e-07, "loss": 7.985, "step": 1279 }, { "epoch": 0.4012539184952978, "grad_norm": 5.2914838790893555, "learning_rate": 6.4e-07, "loss": 10.3964, "step": 1280 }, { "epoch": 0.40156739811912223, "grad_norm": 3.0464425086975098, "learning_rate": 6.405e-07, "loss": 6.6271, "step": 1281 }, { "epoch": 0.40188087774294673, "grad_norm": 4.177611827850342, "learning_rate": 6.410000000000001e-07, "loss": 12.3688, "step": 1282 }, { "epoch": 0.40219435736677117, "grad_norm": 3.6782820224761963, "learning_rate": 6.415e-07, "loss": 7.3459, "step": 1283 }, { "epoch": 0.4025078369905956, "grad_norm": 3.16139817237854, "learning_rate": 6.42e-07, "loss": 7.4459, "step": 1284 }, { "epoch": 0.40282131661442006, "grad_norm": 3.282700538635254, "learning_rate": 6.425000000000001e-07, "loss": 6.6329, "step": 1285 }, { "epoch": 0.4031347962382445, "grad_norm": 3.4450936317443848, "learning_rate": 6.43e-07, "loss": 9.4665, "step": 1286 }, { "epoch": 0.40344827586206894, "grad_norm": 2.994030475616455, "learning_rate": 6.435000000000001e-07, "loss": 6.7493, "step": 1287 }, { "epoch": 0.40376175548589344, "grad_norm": 4.3425822257995605, "learning_rate": 6.44e-07, "loss": 9.9379, "step": 1288 }, { "epoch": 0.4040752351097179, "grad_norm": 4.892250061035156, "learning_rate": 6.445e-07, "loss": 11.3928, "step": 1289 }, { "epoch": 0.4043887147335423, "grad_norm": 3.246832847595215, "learning_rate": 6.450000000000001e-07, "loss": 7.554, "step": 1290 }, { "epoch": 0.40470219435736676, "grad_norm": 4.307924270629883, "learning_rate": 6.455e-07, "loss": 11.633, "step": 1291 }, { "epoch": 0.4050156739811912, "grad_norm": 3.2229695320129395, "learning_rate": 6.460000000000001e-07, "loss": 6.5894, "step": 1292 }, { "epoch": 0.40532915360501565, "grad_norm": 4.491090774536133, "learning_rate": 6.465000000000001e-07, "loss": 12.8065, "step": 1293 }, { "epoch": 0.40564263322884014, "grad_norm": 3.604726552963257, "learning_rate": 6.47e-07, "loss": 7.5955, "step": 1294 }, { "epoch": 0.4059561128526646, "grad_norm": 2.8530771732330322, "learning_rate": 6.475e-07, "loss": 6.6698, "step": 1295 }, { "epoch": 0.406269592476489, "grad_norm": 3.601217746734619, "learning_rate": 6.48e-07, "loss": 7.8026, "step": 1296 }, { "epoch": 0.40658307210031347, "grad_norm": 3.475271701812744, "learning_rate": 6.485000000000001e-07, "loss": 7.1792, "step": 1297 }, { "epoch": 0.4068965517241379, "grad_norm": 3.568582057952881, "learning_rate": 6.490000000000001e-07, "loss": 8.6444, "step": 1298 }, { "epoch": 0.4072100313479624, "grad_norm": 3.6762585639953613, "learning_rate": 6.495e-07, "loss": 7.4771, "step": 1299 }, { "epoch": 0.40752351097178685, "grad_norm": 4.190446376800537, "learning_rate": 6.5e-07, "loss": 12.6181, "step": 1300 }, { "epoch": 0.4078369905956113, "grad_norm": 3.1675896644592285, "learning_rate": 6.505000000000001e-07, "loss": 6.2409, "step": 1301 }, { "epoch": 0.40815047021943573, "grad_norm": 4.8222856521606445, "learning_rate": 6.510000000000001e-07, "loss": 11.7033, "step": 1302 }, { "epoch": 0.4084639498432602, "grad_norm": 4.360724925994873, "learning_rate": 6.515e-07, "loss": 13.757, "step": 1303 }, { "epoch": 0.4087774294670846, "grad_norm": 3.622544288635254, "learning_rate": 6.52e-07, "loss": 9.4418, "step": 1304 }, { "epoch": 0.4090909090909091, "grad_norm": 3.5663022994995117, "learning_rate": 6.525000000000001e-07, "loss": 7.6926, "step": 1305 }, { "epoch": 0.40940438871473356, "grad_norm": 4.033201694488525, "learning_rate": 6.53e-07, "loss": 8.5123, "step": 1306 }, { "epoch": 0.409717868338558, "grad_norm": 3.480999708175659, "learning_rate": 6.535000000000001e-07, "loss": 7.2778, "step": 1307 }, { "epoch": 0.41003134796238244, "grad_norm": 3.1945245265960693, "learning_rate": 6.54e-07, "loss": 7.371, "step": 1308 }, { "epoch": 0.4103448275862069, "grad_norm": 3.4284608364105225, "learning_rate": 6.545e-07, "loss": 7.7181, "step": 1309 }, { "epoch": 0.4106583072100313, "grad_norm": 6.255939483642578, "learning_rate": 6.550000000000001e-07, "loss": 13.9958, "step": 1310 }, { "epoch": 0.4109717868338558, "grad_norm": 3.8257150650024414, "learning_rate": 6.555e-07, "loss": 9.3615, "step": 1311 }, { "epoch": 0.41128526645768027, "grad_norm": 5.354909420013428, "learning_rate": 6.560000000000002e-07, "loss": 10.8434, "step": 1312 }, { "epoch": 0.4115987460815047, "grad_norm": 4.636758804321289, "learning_rate": 6.565000000000001e-07, "loss": 10.2746, "step": 1313 }, { "epoch": 0.41191222570532915, "grad_norm": 4.180166721343994, "learning_rate": 6.57e-07, "loss": 10.871, "step": 1314 }, { "epoch": 0.4122257053291536, "grad_norm": 4.635682106018066, "learning_rate": 6.575000000000001e-07, "loss": 9.849, "step": 1315 }, { "epoch": 0.41253918495297803, "grad_norm": 3.4844233989715576, "learning_rate": 6.58e-07, "loss": 10.2331, "step": 1316 }, { "epoch": 0.41285266457680253, "grad_norm": 2.8132760524749756, "learning_rate": 6.585000000000001e-07, "loss": 6.6116, "step": 1317 }, { "epoch": 0.413166144200627, "grad_norm": 3.6411001682281494, "learning_rate": 6.590000000000001e-07, "loss": 9.5894, "step": 1318 }, { "epoch": 0.4134796238244514, "grad_norm": 3.4417355060577393, "learning_rate": 6.595e-07, "loss": 8.5706, "step": 1319 }, { "epoch": 0.41379310344827586, "grad_norm": 4.843432903289795, "learning_rate": 6.6e-07, "loss": 8.9815, "step": 1320 }, { "epoch": 0.4141065830721003, "grad_norm": 4.353475093841553, "learning_rate": 6.605000000000001e-07, "loss": 7.6671, "step": 1321 }, { "epoch": 0.41442006269592474, "grad_norm": 3.1126198768615723, "learning_rate": 6.610000000000001e-07, "loss": 7.3491, "step": 1322 }, { "epoch": 0.41473354231974924, "grad_norm": 2.952798366546631, "learning_rate": 6.615e-07, "loss": 6.6309, "step": 1323 }, { "epoch": 0.4150470219435737, "grad_norm": 3.8614158630371094, "learning_rate": 6.62e-07, "loss": 7.1688, "step": 1324 }, { "epoch": 0.4153605015673981, "grad_norm": 3.264158248901367, "learning_rate": 6.625000000000001e-07, "loss": 7.4095, "step": 1325 }, { "epoch": 0.41567398119122256, "grad_norm": 4.060593605041504, "learning_rate": 6.63e-07, "loss": 11.2259, "step": 1326 }, { "epoch": 0.415987460815047, "grad_norm": 3.954174757003784, "learning_rate": 6.635000000000001e-07, "loss": 9.4012, "step": 1327 }, { "epoch": 0.41630094043887145, "grad_norm": 3.335627555847168, "learning_rate": 6.64e-07, "loss": 7.6803, "step": 1328 }, { "epoch": 0.41661442006269594, "grad_norm": 3.3831787109375, "learning_rate": 6.645000000000001e-07, "loss": 8.848, "step": 1329 }, { "epoch": 0.4169278996865204, "grad_norm": 4.064548492431641, "learning_rate": 6.650000000000001e-07, "loss": 7.7122, "step": 1330 }, { "epoch": 0.41724137931034483, "grad_norm": 3.576258659362793, "learning_rate": 6.655e-07, "loss": 9.0345, "step": 1331 }, { "epoch": 0.41755485893416927, "grad_norm": 4.037924766540527, "learning_rate": 6.660000000000002e-07, "loss": 10.0141, "step": 1332 }, { "epoch": 0.4178683385579937, "grad_norm": 4.519690990447998, "learning_rate": 6.665000000000001e-07, "loss": 7.9251, "step": 1333 }, { "epoch": 0.41818181818181815, "grad_norm": 3.73274827003479, "learning_rate": 6.67e-07, "loss": 9.2519, "step": 1334 }, { "epoch": 0.41849529780564265, "grad_norm": 3.267875909805298, "learning_rate": 6.675000000000001e-07, "loss": 6.2191, "step": 1335 }, { "epoch": 0.4188087774294671, "grad_norm": 3.341024160385132, "learning_rate": 6.68e-07, "loss": 7.4184, "step": 1336 }, { "epoch": 0.41912225705329154, "grad_norm": 2.8991432189941406, "learning_rate": 6.685000000000001e-07, "loss": 7.4803, "step": 1337 }, { "epoch": 0.419435736677116, "grad_norm": 4.8145928382873535, "learning_rate": 6.690000000000001e-07, "loss": 10.2224, "step": 1338 }, { "epoch": 0.4197492163009404, "grad_norm": 3.761445999145508, "learning_rate": 6.695e-07, "loss": 8.4562, "step": 1339 }, { "epoch": 0.4200626959247649, "grad_norm": 3.892420768737793, "learning_rate": 6.7e-07, "loss": 9.1571, "step": 1340 }, { "epoch": 0.42037617554858936, "grad_norm": 5.118911266326904, "learning_rate": 6.705000000000001e-07, "loss": 12.1094, "step": 1341 }, { "epoch": 0.4206896551724138, "grad_norm": 3.2898852825164795, "learning_rate": 6.710000000000001e-07, "loss": 7.4492, "step": 1342 }, { "epoch": 0.42100313479623824, "grad_norm": 2.7074368000030518, "learning_rate": 6.715e-07, "loss": 6.0427, "step": 1343 }, { "epoch": 0.4213166144200627, "grad_norm": 4.675681114196777, "learning_rate": 6.72e-07, "loss": 13.983, "step": 1344 }, { "epoch": 0.4216300940438871, "grad_norm": 3.761418104171753, "learning_rate": 6.725000000000001e-07, "loss": 7.1159, "step": 1345 }, { "epoch": 0.4219435736677116, "grad_norm": 3.7398788928985596, "learning_rate": 6.730000000000001e-07, "loss": 8.9778, "step": 1346 }, { "epoch": 0.42225705329153607, "grad_norm": 4.614981651306152, "learning_rate": 6.735e-07, "loss": 11.8006, "step": 1347 }, { "epoch": 0.4225705329153605, "grad_norm": 4.014955043792725, "learning_rate": 6.74e-07, "loss": 7.9856, "step": 1348 }, { "epoch": 0.42288401253918495, "grad_norm": 4.2952470779418945, "learning_rate": 6.745000000000001e-07, "loss": 8.6141, "step": 1349 }, { "epoch": 0.4231974921630094, "grad_norm": 4.254025459289551, "learning_rate": 6.750000000000001e-07, "loss": 13.2271, "step": 1350 }, { "epoch": 0.42351097178683383, "grad_norm": 3.196575403213501, "learning_rate": 6.755e-07, "loss": 6.6306, "step": 1351 }, { "epoch": 0.42382445141065833, "grad_norm": 4.765541076660156, "learning_rate": 6.76e-07, "loss": 12.8829, "step": 1352 }, { "epoch": 0.4241379310344828, "grad_norm": 4.769258499145508, "learning_rate": 6.765000000000001e-07, "loss": 10.9679, "step": 1353 }, { "epoch": 0.4244514106583072, "grad_norm": 3.533536195755005, "learning_rate": 6.77e-07, "loss": 6.7036, "step": 1354 }, { "epoch": 0.42476489028213166, "grad_norm": 4.482390403747559, "learning_rate": 6.775000000000001e-07, "loss": 9.1064, "step": 1355 }, { "epoch": 0.4250783699059561, "grad_norm": 3.877317428588867, "learning_rate": 6.78e-07, "loss": 9.4161, "step": 1356 }, { "epoch": 0.42539184952978054, "grad_norm": 4.770163059234619, "learning_rate": 6.784999999999999e-07, "loss": 13.1058, "step": 1357 }, { "epoch": 0.42570532915360504, "grad_norm": 3.1010990142822266, "learning_rate": 6.790000000000001e-07, "loss": 7.6001, "step": 1358 }, { "epoch": 0.4260188087774295, "grad_norm": 4.240891456604004, "learning_rate": 6.795e-07, "loss": 10.6859, "step": 1359 }, { "epoch": 0.4263322884012539, "grad_norm": 3.357872486114502, "learning_rate": 6.800000000000001e-07, "loss": 9.1908, "step": 1360 }, { "epoch": 0.42664576802507836, "grad_norm": 4.015371322631836, "learning_rate": 6.805000000000001e-07, "loss": 7.1454, "step": 1361 }, { "epoch": 0.4269592476489028, "grad_norm": 3.248347520828247, "learning_rate": 6.81e-07, "loss": 7.6464, "step": 1362 }, { "epoch": 0.42727272727272725, "grad_norm": 3.975553512573242, "learning_rate": 6.815000000000001e-07, "loss": 9.1993, "step": 1363 }, { "epoch": 0.42758620689655175, "grad_norm": 5.927193641662598, "learning_rate": 6.82e-07, "loss": 9.7591, "step": 1364 }, { "epoch": 0.4278996865203762, "grad_norm": 3.059760808944702, "learning_rate": 6.825000000000001e-07, "loss": 6.5011, "step": 1365 }, { "epoch": 0.42821316614420063, "grad_norm": 3.896256923675537, "learning_rate": 6.830000000000001e-07, "loss": 7.2387, "step": 1366 }, { "epoch": 0.42852664576802507, "grad_norm": 3.3756370544433594, "learning_rate": 6.835e-07, "loss": 7.5067, "step": 1367 }, { "epoch": 0.4288401253918495, "grad_norm": 4.085315704345703, "learning_rate": 6.84e-07, "loss": 7.0529, "step": 1368 }, { "epoch": 0.42915360501567396, "grad_norm": 2.654603958129883, "learning_rate": 6.845000000000001e-07, "loss": 7.7151, "step": 1369 }, { "epoch": 0.42946708463949845, "grad_norm": 4.527529716491699, "learning_rate": 6.850000000000001e-07, "loss": 10.333, "step": 1370 }, { "epoch": 0.4297805642633229, "grad_norm": 3.930635690689087, "learning_rate": 6.855e-07, "loss": 9.8118, "step": 1371 }, { "epoch": 0.43009404388714734, "grad_norm": 4.306950092315674, "learning_rate": 6.86e-07, "loss": 9.8894, "step": 1372 }, { "epoch": 0.4304075235109718, "grad_norm": 3.4206173419952393, "learning_rate": 6.865000000000001e-07, "loss": 8.8873, "step": 1373 }, { "epoch": 0.4307210031347962, "grad_norm": 4.430827617645264, "learning_rate": 6.87e-07, "loss": 9.3459, "step": 1374 }, { "epoch": 0.43103448275862066, "grad_norm": 3.1537222862243652, "learning_rate": 6.875000000000001e-07, "loss": 7.5838, "step": 1375 }, { "epoch": 0.43134796238244516, "grad_norm": 4.20504093170166, "learning_rate": 6.88e-07, "loss": 12.7391, "step": 1376 }, { "epoch": 0.4316614420062696, "grad_norm": 3.653794050216675, "learning_rate": 6.885e-07, "loss": 8.7142, "step": 1377 }, { "epoch": 0.43197492163009404, "grad_norm": 4.225579261779785, "learning_rate": 6.890000000000001e-07, "loss": 10.9947, "step": 1378 }, { "epoch": 0.4322884012539185, "grad_norm": 3.8407626152038574, "learning_rate": 6.895e-07, "loss": 9.6915, "step": 1379 }, { "epoch": 0.43260188087774293, "grad_norm": 3.124218463897705, "learning_rate": 6.900000000000001e-07, "loss": 6.9618, "step": 1380 }, { "epoch": 0.4329153605015674, "grad_norm": 3.4793970584869385, "learning_rate": 6.905000000000001e-07, "loss": 10.846, "step": 1381 }, { "epoch": 0.43322884012539187, "grad_norm": 3.0507242679595947, "learning_rate": 6.91e-07, "loss": 8.2012, "step": 1382 }, { "epoch": 0.4335423197492163, "grad_norm": 3.7710156440734863, "learning_rate": 6.915000000000001e-07, "loss": 9.654, "step": 1383 }, { "epoch": 0.43385579937304075, "grad_norm": 3.3001532554626465, "learning_rate": 6.92e-07, "loss": 6.7221, "step": 1384 }, { "epoch": 0.4341692789968652, "grad_norm": 4.092366695404053, "learning_rate": 6.925000000000001e-07, "loss": 10.6304, "step": 1385 }, { "epoch": 0.43448275862068964, "grad_norm": 3.2531421184539795, "learning_rate": 6.930000000000001e-07, "loss": 7.9767, "step": 1386 }, { "epoch": 0.43479623824451413, "grad_norm": 3.08742094039917, "learning_rate": 6.935e-07, "loss": 7.9412, "step": 1387 }, { "epoch": 0.4351097178683386, "grad_norm": 4.16629695892334, "learning_rate": 6.94e-07, "loss": 7.9362, "step": 1388 }, { "epoch": 0.435423197492163, "grad_norm": 3.6127755641937256, "learning_rate": 6.945000000000001e-07, "loss": 8.0281, "step": 1389 }, { "epoch": 0.43573667711598746, "grad_norm": 3.736285924911499, "learning_rate": 6.950000000000001e-07, "loss": 9.2018, "step": 1390 }, { "epoch": 0.4360501567398119, "grad_norm": 5.327370643615723, "learning_rate": 6.955000000000001e-07, "loss": 10.193, "step": 1391 }, { "epoch": 0.43636363636363634, "grad_norm": 3.426323413848877, "learning_rate": 6.96e-07, "loss": 8.9392, "step": 1392 }, { "epoch": 0.43667711598746084, "grad_norm": 4.5006937980651855, "learning_rate": 6.965000000000001e-07, "loss": 11.3513, "step": 1393 }, { "epoch": 0.4369905956112853, "grad_norm": 3.4478158950805664, "learning_rate": 6.970000000000001e-07, "loss": 8.5876, "step": 1394 }, { "epoch": 0.4373040752351097, "grad_norm": 3.8687758445739746, "learning_rate": 6.975000000000001e-07, "loss": 8.731, "step": 1395 }, { "epoch": 0.43761755485893417, "grad_norm": 3.383852958679199, "learning_rate": 6.98e-07, "loss": 7.7361, "step": 1396 }, { "epoch": 0.4379310344827586, "grad_norm": 3.9315686225891113, "learning_rate": 6.985e-07, "loss": 9.827, "step": 1397 }, { "epoch": 0.43824451410658305, "grad_norm": 3.924696683883667, "learning_rate": 6.990000000000001e-07, "loss": 7.6312, "step": 1398 }, { "epoch": 0.43855799373040755, "grad_norm": 3.3615076541900635, "learning_rate": 6.995e-07, "loss": 7.7659, "step": 1399 }, { "epoch": 0.438871473354232, "grad_norm": 3.5409042835235596, "learning_rate": 7.000000000000001e-07, "loss": 6.6861, "step": 1400 }, { "epoch": 0.43918495297805643, "grad_norm": 3.4962658882141113, "learning_rate": 7.005000000000001e-07, "loss": 8.9106, "step": 1401 }, { "epoch": 0.4394984326018809, "grad_norm": 4.335007667541504, "learning_rate": 7.01e-07, "loss": 7.0246, "step": 1402 }, { "epoch": 0.4398119122257053, "grad_norm": 3.592299222946167, "learning_rate": 7.015000000000001e-07, "loss": 9.6732, "step": 1403 }, { "epoch": 0.44012539184952976, "grad_norm": 5.375201225280762, "learning_rate": 7.02e-07, "loss": 8.865, "step": 1404 }, { "epoch": 0.44043887147335425, "grad_norm": 5.488976955413818, "learning_rate": 7.025000000000002e-07, "loss": 14.8031, "step": 1405 }, { "epoch": 0.4407523510971787, "grad_norm": 3.3288614749908447, "learning_rate": 7.030000000000001e-07, "loss": 8.8429, "step": 1406 }, { "epoch": 0.44106583072100314, "grad_norm": 3.75602388381958, "learning_rate": 7.035e-07, "loss": 7.5947, "step": 1407 }, { "epoch": 0.4413793103448276, "grad_norm": 2.7701900005340576, "learning_rate": 7.040000000000001e-07, "loss": 7.4473, "step": 1408 }, { "epoch": 0.441692789968652, "grad_norm": 4.249903678894043, "learning_rate": 7.045000000000001e-07, "loss": 8.4978, "step": 1409 }, { "epoch": 0.44200626959247646, "grad_norm": 3.3370988368988037, "learning_rate": 7.05e-07, "loss": 6.707, "step": 1410 }, { "epoch": 0.44231974921630096, "grad_norm": 5.536358833312988, "learning_rate": 7.055000000000001e-07, "loss": 11.3998, "step": 1411 }, { "epoch": 0.4426332288401254, "grad_norm": 4.11740779876709, "learning_rate": 7.06e-07, "loss": 8.3269, "step": 1412 }, { "epoch": 0.44294670846394985, "grad_norm": 4.711095809936523, "learning_rate": 7.065000000000001e-07, "loss": 12.2608, "step": 1413 }, { "epoch": 0.4432601880877743, "grad_norm": 3.0188283920288086, "learning_rate": 7.070000000000001e-07, "loss": 7.8679, "step": 1414 }, { "epoch": 0.44357366771159873, "grad_norm": 3.6578001976013184, "learning_rate": 7.075e-07, "loss": 7.9547, "step": 1415 }, { "epoch": 0.44388714733542317, "grad_norm": 3.710144519805908, "learning_rate": 7.08e-07, "loss": 9.1333, "step": 1416 }, { "epoch": 0.44420062695924767, "grad_norm": 3.580639600753784, "learning_rate": 7.085e-07, "loss": 8.6172, "step": 1417 }, { "epoch": 0.4445141065830721, "grad_norm": 3.4201769828796387, "learning_rate": 7.090000000000001e-07, "loss": 7.38, "step": 1418 }, { "epoch": 0.44482758620689655, "grad_norm": 3.8317136764526367, "learning_rate": 7.095e-07, "loss": 7.1807, "step": 1419 }, { "epoch": 0.445141065830721, "grad_norm": 6.48163366317749, "learning_rate": 7.1e-07, "loss": 15.1615, "step": 1420 }, { "epoch": 0.44545454545454544, "grad_norm": 3.3979268074035645, "learning_rate": 7.105000000000001e-07, "loss": 6.6834, "step": 1421 }, { "epoch": 0.4457680250783699, "grad_norm": 3.5724081993103027, "learning_rate": 7.110000000000001e-07, "loss": 8.6715, "step": 1422 }, { "epoch": 0.4460815047021944, "grad_norm": 3.738381862640381, "learning_rate": 7.115000000000001e-07, "loss": 7.0441, "step": 1423 }, { "epoch": 0.4463949843260188, "grad_norm": 4.122878551483154, "learning_rate": 7.12e-07, "loss": 7.2443, "step": 1424 }, { "epoch": 0.44670846394984326, "grad_norm": 3.6147820949554443, "learning_rate": 7.125e-07, "loss": 7.8883, "step": 1425 }, { "epoch": 0.4470219435736677, "grad_norm": 3.86806058883667, "learning_rate": 7.130000000000001e-07, "loss": 8.0316, "step": 1426 }, { "epoch": 0.44733542319749214, "grad_norm": 4.159463882446289, "learning_rate": 7.135e-07, "loss": 7.2277, "step": 1427 }, { "epoch": 0.44764890282131664, "grad_norm": 2.5668580532073975, "learning_rate": 7.140000000000001e-07, "loss": 6.9613, "step": 1428 }, { "epoch": 0.4479623824451411, "grad_norm": 4.1336469650268555, "learning_rate": 7.145000000000001e-07, "loss": 9.4702, "step": 1429 }, { "epoch": 0.4482758620689655, "grad_norm": 6.645508289337158, "learning_rate": 7.15e-07, "loss": 15.8057, "step": 1430 }, { "epoch": 0.44858934169278997, "grad_norm": 4.979160785675049, "learning_rate": 7.155000000000001e-07, "loss": 15.3406, "step": 1431 }, { "epoch": 0.4489028213166144, "grad_norm": 3.6304969787597656, "learning_rate": 7.16e-07, "loss": 8.6424, "step": 1432 }, { "epoch": 0.44921630094043885, "grad_norm": 4.989882469177246, "learning_rate": 7.165000000000001e-07, "loss": 11.6095, "step": 1433 }, { "epoch": 0.44952978056426335, "grad_norm": 3.8582489490509033, "learning_rate": 7.170000000000001e-07, "loss": 8.5874, "step": 1434 }, { "epoch": 0.4498432601880878, "grad_norm": 4.933996200561523, "learning_rate": 7.175e-07, "loss": 11.0086, "step": 1435 }, { "epoch": 0.45015673981191223, "grad_norm": 4.105413913726807, "learning_rate": 7.18e-07, "loss": 10.1414, "step": 1436 }, { "epoch": 0.4504702194357367, "grad_norm": 4.856784820556641, "learning_rate": 7.185e-07, "loss": 10.7219, "step": 1437 }, { "epoch": 0.4507836990595611, "grad_norm": 3.4442543983459473, "learning_rate": 7.190000000000001e-07, "loss": 6.6924, "step": 1438 }, { "epoch": 0.45109717868338556, "grad_norm": 3.018460988998413, "learning_rate": 7.195000000000001e-07, "loss": 8.6876, "step": 1439 }, { "epoch": 0.45141065830721006, "grad_norm": 3.129737377166748, "learning_rate": 7.2e-07, "loss": 7.0111, "step": 1440 }, { "epoch": 0.4517241379310345, "grad_norm": 4.270967483520508, "learning_rate": 7.205000000000001e-07, "loss": 9.3915, "step": 1441 }, { "epoch": 0.45203761755485894, "grad_norm": 3.9212779998779297, "learning_rate": 7.210000000000001e-07, "loss": 10.3279, "step": 1442 }, { "epoch": 0.4523510971786834, "grad_norm": 3.4876608848571777, "learning_rate": 7.215000000000001e-07, "loss": 6.606, "step": 1443 }, { "epoch": 0.4526645768025078, "grad_norm": 3.4680469036102295, "learning_rate": 7.22e-07, "loss": 7.3311, "step": 1444 }, { "epoch": 0.45297805642633227, "grad_norm": 3.6926703453063965, "learning_rate": 7.225e-07, "loss": 7.5257, "step": 1445 }, { "epoch": 0.45329153605015676, "grad_norm": 6.416301727294922, "learning_rate": 7.230000000000001e-07, "loss": 14.6459, "step": 1446 }, { "epoch": 0.4536050156739812, "grad_norm": 4.126363277435303, "learning_rate": 7.235e-07, "loss": 8.369, "step": 1447 }, { "epoch": 0.45391849529780565, "grad_norm": 3.9787609577178955, "learning_rate": 7.240000000000001e-07, "loss": 8.115, "step": 1448 }, { "epoch": 0.4542319749216301, "grad_norm": 3.5446219444274902, "learning_rate": 7.245000000000001e-07, "loss": 7.3203, "step": 1449 }, { "epoch": 0.45454545454545453, "grad_norm": 3.7280256748199463, "learning_rate": 7.25e-07, "loss": 9.8608, "step": 1450 }, { "epoch": 0.454858934169279, "grad_norm": 3.1473164558410645, "learning_rate": 7.255000000000001e-07, "loss": 7.5801, "step": 1451 }, { "epoch": 0.45517241379310347, "grad_norm": 3.243011951446533, "learning_rate": 7.26e-07, "loss": 6.5251, "step": 1452 }, { "epoch": 0.4554858934169279, "grad_norm": 3.912731170654297, "learning_rate": 7.265000000000002e-07, "loss": 9.8108, "step": 1453 }, { "epoch": 0.45579937304075235, "grad_norm": 3.3148555755615234, "learning_rate": 7.270000000000001e-07, "loss": 9.7591, "step": 1454 }, { "epoch": 0.4561128526645768, "grad_norm": 3.5249342918395996, "learning_rate": 7.275e-07, "loss": 7.1523, "step": 1455 }, { "epoch": 0.45642633228840124, "grad_norm": 3.985668420791626, "learning_rate": 7.280000000000001e-07, "loss": 7.9672, "step": 1456 }, { "epoch": 0.4567398119122257, "grad_norm": 3.4252846240997314, "learning_rate": 7.285e-07, "loss": 8.2703, "step": 1457 }, { "epoch": 0.4570532915360502, "grad_norm": 5.43139123916626, "learning_rate": 7.290000000000001e-07, "loss": 9.2351, "step": 1458 }, { "epoch": 0.4573667711598746, "grad_norm": 5.333076000213623, "learning_rate": 7.295000000000001e-07, "loss": 13.8034, "step": 1459 }, { "epoch": 0.45768025078369906, "grad_norm": 3.3597097396850586, "learning_rate": 7.3e-07, "loss": 11.3477, "step": 1460 }, { "epoch": 0.4579937304075235, "grad_norm": 2.962592601776123, "learning_rate": 7.305000000000001e-07, "loss": 8.5858, "step": 1461 }, { "epoch": 0.45830721003134794, "grad_norm": 2.7471611499786377, "learning_rate": 7.310000000000001e-07, "loss": 6.0348, "step": 1462 }, { "epoch": 0.4586206896551724, "grad_norm": 4.322995662689209, "learning_rate": 7.315000000000001e-07, "loss": 9.9356, "step": 1463 }, { "epoch": 0.4589341692789969, "grad_norm": 3.186292886734009, "learning_rate": 7.32e-07, "loss": 8.0499, "step": 1464 }, { "epoch": 0.4592476489028213, "grad_norm": 2.747281789779663, "learning_rate": 7.325e-07, "loss": 6.545, "step": 1465 }, { "epoch": 0.45956112852664577, "grad_norm": 4.15464448928833, "learning_rate": 7.330000000000001e-07, "loss": 8.2943, "step": 1466 }, { "epoch": 0.4598746081504702, "grad_norm": 3.4695398807525635, "learning_rate": 7.335e-07, "loss": 7.7409, "step": 1467 }, { "epoch": 0.46018808777429465, "grad_norm": 3.8150601387023926, "learning_rate": 7.340000000000001e-07, "loss": 9.3436, "step": 1468 }, { "epoch": 0.46050156739811915, "grad_norm": 3.744868278503418, "learning_rate": 7.345000000000001e-07, "loss": 9.7803, "step": 1469 }, { "epoch": 0.4608150470219436, "grad_norm": 3.8790037631988525, "learning_rate": 7.350000000000001e-07, "loss": 8.0119, "step": 1470 }, { "epoch": 0.46112852664576803, "grad_norm": 3.8017265796661377, "learning_rate": 7.355000000000001e-07, "loss": 7.3153, "step": 1471 }, { "epoch": 0.4614420062695925, "grad_norm": 4.845677375793457, "learning_rate": 7.36e-07, "loss": 12.615, "step": 1472 }, { "epoch": 0.4617554858934169, "grad_norm": 4.192208766937256, "learning_rate": 7.365e-07, "loss": 7.3787, "step": 1473 }, { "epoch": 0.46206896551724136, "grad_norm": 4.240219593048096, "learning_rate": 7.370000000000001e-07, "loss": 8.5848, "step": 1474 }, { "epoch": 0.46238244514106586, "grad_norm": 4.029890537261963, "learning_rate": 7.375e-07, "loss": 9.3911, "step": 1475 }, { "epoch": 0.4626959247648903, "grad_norm": 3.331756591796875, "learning_rate": 7.380000000000001e-07, "loss": 6.3178, "step": 1476 }, { "epoch": 0.46300940438871474, "grad_norm": 3.7580671310424805, "learning_rate": 7.385e-07, "loss": 9.2023, "step": 1477 }, { "epoch": 0.4633228840125392, "grad_norm": 4.716964244842529, "learning_rate": 7.39e-07, "loss": 11.6383, "step": 1478 }, { "epoch": 0.4636363636363636, "grad_norm": 3.466538906097412, "learning_rate": 7.395000000000001e-07, "loss": 8.556, "step": 1479 }, { "epoch": 0.46394984326018807, "grad_norm": 3.618819236755371, "learning_rate": 7.4e-07, "loss": 8.6228, "step": 1480 }, { "epoch": 0.46426332288401256, "grad_norm": 3.9370975494384766, "learning_rate": 7.405000000000002e-07, "loss": 8.9446, "step": 1481 }, { "epoch": 0.464576802507837, "grad_norm": 3.3497865200042725, "learning_rate": 7.410000000000001e-07, "loss": 8.4896, "step": 1482 }, { "epoch": 0.46489028213166145, "grad_norm": 3.9014110565185547, "learning_rate": 7.415e-07, "loss": 9.6106, "step": 1483 }, { "epoch": 0.4652037617554859, "grad_norm": 3.2826168537139893, "learning_rate": 7.420000000000001e-07, "loss": 7.3699, "step": 1484 }, { "epoch": 0.46551724137931033, "grad_norm": 4.081614971160889, "learning_rate": 7.425e-07, "loss": 8.7689, "step": 1485 }, { "epoch": 0.4658307210031348, "grad_norm": 3.3145084381103516, "learning_rate": 7.430000000000001e-07, "loss": 7.7831, "step": 1486 }, { "epoch": 0.46614420062695927, "grad_norm": 5.257850170135498, "learning_rate": 7.435000000000001e-07, "loss": 14.4335, "step": 1487 }, { "epoch": 0.4664576802507837, "grad_norm": 3.631011486053467, "learning_rate": 7.44e-07, "loss": 8.0316, "step": 1488 }, { "epoch": 0.46677115987460815, "grad_norm": 3.9170570373535156, "learning_rate": 7.445000000000001e-07, "loss": 7.9331, "step": 1489 }, { "epoch": 0.4670846394984326, "grad_norm": 3.710536003112793, "learning_rate": 7.450000000000001e-07, "loss": 10.0156, "step": 1490 }, { "epoch": 0.46739811912225704, "grad_norm": 3.666468858718872, "learning_rate": 7.455000000000001e-07, "loss": 7.2938, "step": 1491 }, { "epoch": 0.4677115987460815, "grad_norm": 4.270216941833496, "learning_rate": 7.46e-07, "loss": 10.1365, "step": 1492 }, { "epoch": 0.468025078369906, "grad_norm": 4.941109657287598, "learning_rate": 7.465e-07, "loss": 10.8911, "step": 1493 }, { "epoch": 0.4683385579937304, "grad_norm": 3.449632167816162, "learning_rate": 7.470000000000001e-07, "loss": 8.6447, "step": 1494 }, { "epoch": 0.46865203761755486, "grad_norm": 2.9346587657928467, "learning_rate": 7.475e-07, "loss": 6.7992, "step": 1495 }, { "epoch": 0.4689655172413793, "grad_norm": 2.9172539710998535, "learning_rate": 7.480000000000001e-07, "loss": 7.0059, "step": 1496 }, { "epoch": 0.46927899686520375, "grad_norm": 3.648991346359253, "learning_rate": 7.485e-07, "loss": 8.9688, "step": 1497 }, { "epoch": 0.4695924764890282, "grad_norm": 3.0724036693573, "learning_rate": 7.49e-07, "loss": 7.4737, "step": 1498 }, { "epoch": 0.4699059561128527, "grad_norm": 3.7592673301696777, "learning_rate": 7.495000000000001e-07, "loss": 9.0526, "step": 1499 }, { "epoch": 0.4702194357366771, "grad_norm": 3.5673487186431885, "learning_rate": 7.5e-07, "loss": 6.967, "step": 1500 }, { "epoch": 0.47053291536050157, "grad_norm": 4.531203269958496, "learning_rate": 7.505000000000002e-07, "loss": 8.6275, "step": 1501 }, { "epoch": 0.470846394984326, "grad_norm": 4.227149486541748, "learning_rate": 7.510000000000001e-07, "loss": 10.3291, "step": 1502 }, { "epoch": 0.47115987460815045, "grad_norm": 3.532550096511841, "learning_rate": 7.515e-07, "loss": 9.2704, "step": 1503 }, { "epoch": 0.4714733542319749, "grad_norm": 3.5547232627868652, "learning_rate": 7.520000000000001e-07, "loss": 8.9358, "step": 1504 }, { "epoch": 0.4717868338557994, "grad_norm": 2.87225604057312, "learning_rate": 7.525e-07, "loss": 7.2077, "step": 1505 }, { "epoch": 0.47210031347962383, "grad_norm": 3.5084965229034424, "learning_rate": 7.530000000000001e-07, "loss": 7.6485, "step": 1506 }, { "epoch": 0.4724137931034483, "grad_norm": 4.405672550201416, "learning_rate": 7.535000000000001e-07, "loss": 9.7914, "step": 1507 }, { "epoch": 0.4727272727272727, "grad_norm": 3.6262218952178955, "learning_rate": 7.54e-07, "loss": 5.8048, "step": 1508 }, { "epoch": 0.47304075235109716, "grad_norm": 4.285321235656738, "learning_rate": 7.545000000000001e-07, "loss": 8.9892, "step": 1509 }, { "epoch": 0.47335423197492166, "grad_norm": 3.221437692642212, "learning_rate": 7.550000000000001e-07, "loss": 8.5896, "step": 1510 }, { "epoch": 0.4736677115987461, "grad_norm": 3.349771022796631, "learning_rate": 7.555000000000001e-07, "loss": 7.2775, "step": 1511 }, { "epoch": 0.47398119122257054, "grad_norm": 4.374917984008789, "learning_rate": 7.56e-07, "loss": 10.1504, "step": 1512 }, { "epoch": 0.474294670846395, "grad_norm": 4.312491416931152, "learning_rate": 7.565e-07, "loss": 9.6675, "step": 1513 }, { "epoch": 0.4746081504702194, "grad_norm": 4.160287380218506, "learning_rate": 7.570000000000001e-07, "loss": 7.7446, "step": 1514 }, { "epoch": 0.47492163009404387, "grad_norm": 4.5663909912109375, "learning_rate": 7.575000000000001e-07, "loss": 10.4192, "step": 1515 }, { "epoch": 0.47523510971786836, "grad_norm": 3.4202966690063477, "learning_rate": 7.580000000000001e-07, "loss": 7.1886, "step": 1516 }, { "epoch": 0.4755485893416928, "grad_norm": 4.065593719482422, "learning_rate": 7.585e-07, "loss": 8.5182, "step": 1517 }, { "epoch": 0.47586206896551725, "grad_norm": 4.027133464813232, "learning_rate": 7.590000000000001e-07, "loss": 8.4011, "step": 1518 }, { "epoch": 0.4761755485893417, "grad_norm": 3.671494722366333, "learning_rate": 7.595000000000001e-07, "loss": 7.6953, "step": 1519 }, { "epoch": 0.47648902821316613, "grad_norm": 4.113628387451172, "learning_rate": 7.6e-07, "loss": 9.0673, "step": 1520 }, { "epoch": 0.4768025078369906, "grad_norm": 4.1456756591796875, "learning_rate": 7.605000000000002e-07, "loss": 8.4153, "step": 1521 }, { "epoch": 0.47711598746081507, "grad_norm": 4.083249568939209, "learning_rate": 7.610000000000001e-07, "loss": 9.7206, "step": 1522 }, { "epoch": 0.4774294670846395, "grad_norm": 4.154261589050293, "learning_rate": 7.615e-07, "loss": 8.6472, "step": 1523 }, { "epoch": 0.47774294670846396, "grad_norm": 4.028384208679199, "learning_rate": 7.620000000000001e-07, "loss": 7.9002, "step": 1524 }, { "epoch": 0.4780564263322884, "grad_norm": 4.162931442260742, "learning_rate": 7.625e-07, "loss": 9.9488, "step": 1525 }, { "epoch": 0.47836990595611284, "grad_norm": 3.495849132537842, "learning_rate": 7.630000000000001e-07, "loss": 9.5747, "step": 1526 }, { "epoch": 0.4786833855799373, "grad_norm": 3.8134756088256836, "learning_rate": 7.635000000000001e-07, "loss": 8.6538, "step": 1527 }, { "epoch": 0.4789968652037618, "grad_norm": 3.9398064613342285, "learning_rate": 7.64e-07, "loss": 8.3008, "step": 1528 }, { "epoch": 0.4793103448275862, "grad_norm": 3.440537691116333, "learning_rate": 7.645000000000002e-07, "loss": 6.1765, "step": 1529 }, { "epoch": 0.47962382445141066, "grad_norm": 4.416642665863037, "learning_rate": 7.650000000000001e-07, "loss": 9.6123, "step": 1530 }, { "epoch": 0.4799373040752351, "grad_norm": 3.9015026092529297, "learning_rate": 7.655000000000001e-07, "loss": 8.8604, "step": 1531 }, { "epoch": 0.48025078369905955, "grad_norm": 3.634904623031616, "learning_rate": 7.660000000000001e-07, "loss": 8.6111, "step": 1532 }, { "epoch": 0.480564263322884, "grad_norm": 3.1789402961730957, "learning_rate": 7.665e-07, "loss": 8.7108, "step": 1533 }, { "epoch": 0.4808777429467085, "grad_norm": 3.556788206100464, "learning_rate": 7.670000000000001e-07, "loss": 7.9619, "step": 1534 }, { "epoch": 0.48119122257053293, "grad_norm": 3.4481894969940186, "learning_rate": 7.675000000000001e-07, "loss": 8.4755, "step": 1535 }, { "epoch": 0.48150470219435737, "grad_norm": 3.0415053367614746, "learning_rate": 7.68e-07, "loss": 6.9378, "step": 1536 }, { "epoch": 0.4818181818181818, "grad_norm": 3.7473058700561523, "learning_rate": 7.685e-07, "loss": 10.5121, "step": 1537 }, { "epoch": 0.48213166144200625, "grad_norm": 4.390431880950928, "learning_rate": 7.690000000000001e-07, "loss": 8.8402, "step": 1538 }, { "epoch": 0.4824451410658307, "grad_norm": 3.8549649715423584, "learning_rate": 7.695000000000001e-07, "loss": 8.73, "step": 1539 }, { "epoch": 0.4827586206896552, "grad_norm": 5.701721668243408, "learning_rate": 7.7e-07, "loss": 9.7011, "step": 1540 }, { "epoch": 0.48307210031347964, "grad_norm": 3.738424777984619, "learning_rate": 7.705e-07, "loss": 8.1286, "step": 1541 }, { "epoch": 0.4833855799373041, "grad_norm": 3.1514458656311035, "learning_rate": 7.710000000000001e-07, "loss": 7.7808, "step": 1542 }, { "epoch": 0.4836990595611285, "grad_norm": 2.9332237243652344, "learning_rate": 7.715e-07, "loss": 7.3622, "step": 1543 }, { "epoch": 0.48401253918495296, "grad_norm": 3.4852969646453857, "learning_rate": 7.720000000000001e-07, "loss": 7.6669, "step": 1544 }, { "epoch": 0.4843260188087774, "grad_norm": 7.102890968322754, "learning_rate": 7.725e-07, "loss": 7.0981, "step": 1545 }, { "epoch": 0.4846394984326019, "grad_norm": 3.0128979682922363, "learning_rate": 7.73e-07, "loss": 6.5679, "step": 1546 }, { "epoch": 0.48495297805642634, "grad_norm": 3.7915170192718506, "learning_rate": 7.735000000000001e-07, "loss": 7.2064, "step": 1547 }, { "epoch": 0.4852664576802508, "grad_norm": 2.735111713409424, "learning_rate": 7.74e-07, "loss": 6.365, "step": 1548 }, { "epoch": 0.4855799373040752, "grad_norm": 3.264190435409546, "learning_rate": 7.745000000000002e-07, "loss": 8.0042, "step": 1549 }, { "epoch": 0.48589341692789967, "grad_norm": 4.327232837677002, "learning_rate": 7.750000000000001e-07, "loss": 9.7193, "step": 1550 }, { "epoch": 0.4862068965517241, "grad_norm": 3.012519359588623, "learning_rate": 7.755e-07, "loss": 6.3156, "step": 1551 }, { "epoch": 0.4865203761755486, "grad_norm": 4.424871921539307, "learning_rate": 7.760000000000001e-07, "loss": 8.8488, "step": 1552 }, { "epoch": 0.48683385579937305, "grad_norm": 3.7431857585906982, "learning_rate": 7.765e-07, "loss": 8.0868, "step": 1553 }, { "epoch": 0.4871473354231975, "grad_norm": 3.6024186611175537, "learning_rate": 7.770000000000001e-07, "loss": 10.3315, "step": 1554 }, { "epoch": 0.48746081504702193, "grad_norm": 3.3676459789276123, "learning_rate": 7.775000000000001e-07, "loss": 6.5772, "step": 1555 }, { "epoch": 0.4877742946708464, "grad_norm": 4.312803745269775, "learning_rate": 7.78e-07, "loss": 8.3249, "step": 1556 }, { "epoch": 0.4880877742946709, "grad_norm": 4.562367916107178, "learning_rate": 7.785e-07, "loss": 7.975, "step": 1557 }, { "epoch": 0.4884012539184953, "grad_norm": 3.697195053100586, "learning_rate": 7.790000000000001e-07, "loss": 7.3428, "step": 1558 }, { "epoch": 0.48871473354231976, "grad_norm": 4.236298561096191, "learning_rate": 7.795000000000001e-07, "loss": 10.2044, "step": 1559 }, { "epoch": 0.4890282131661442, "grad_norm": 3.0572011470794678, "learning_rate": 7.8e-07, "loss": 6.726, "step": 1560 }, { "epoch": 0.48934169278996864, "grad_norm": 3.513315439224243, "learning_rate": 7.805e-07, "loss": 7.963, "step": 1561 }, { "epoch": 0.4896551724137931, "grad_norm": 4.3329548835754395, "learning_rate": 7.810000000000001e-07, "loss": 7.9515, "step": 1562 }, { "epoch": 0.4899686520376176, "grad_norm": 4.270005226135254, "learning_rate": 7.815000000000001e-07, "loss": 10.5315, "step": 1563 }, { "epoch": 0.490282131661442, "grad_norm": 4.627399921417236, "learning_rate": 7.820000000000001e-07, "loss": 7.8852, "step": 1564 }, { "epoch": 0.49059561128526646, "grad_norm": 3.8905036449432373, "learning_rate": 7.825e-07, "loss": 8.9447, "step": 1565 }, { "epoch": 0.4909090909090909, "grad_norm": 4.1380181312561035, "learning_rate": 7.83e-07, "loss": 9.1815, "step": 1566 }, { "epoch": 0.49122257053291535, "grad_norm": 3.6653265953063965, "learning_rate": 7.835000000000001e-07, "loss": 8.7808, "step": 1567 }, { "epoch": 0.4915360501567398, "grad_norm": 3.8347866535186768, "learning_rate": 7.84e-07, "loss": 8.8488, "step": 1568 }, { "epoch": 0.4918495297805643, "grad_norm": 3.9519078731536865, "learning_rate": 7.845000000000001e-07, "loss": 8.6361, "step": 1569 }, { "epoch": 0.49216300940438873, "grad_norm": 4.5167036056518555, "learning_rate": 7.850000000000001e-07, "loss": 12.1671, "step": 1570 }, { "epoch": 0.49247648902821317, "grad_norm": 4.719188690185547, "learning_rate": 7.855e-07, "loss": 8.6138, "step": 1571 }, { "epoch": 0.4927899686520376, "grad_norm": 4.4409284591674805, "learning_rate": 7.860000000000001e-07, "loss": 9.6423, "step": 1572 }, { "epoch": 0.49310344827586206, "grad_norm": 2.9862163066864014, "learning_rate": 7.865e-07, "loss": 6.1838, "step": 1573 }, { "epoch": 0.4934169278996865, "grad_norm": 4.9992146492004395, "learning_rate": 7.870000000000002e-07, "loss": 11.4322, "step": 1574 }, { "epoch": 0.493730407523511, "grad_norm": 4.074015140533447, "learning_rate": 7.875000000000001e-07, "loss": 9.6542, "step": 1575 }, { "epoch": 0.49404388714733544, "grad_norm": 3.1027674674987793, "learning_rate": 7.88e-07, "loss": 6.8902, "step": 1576 }, { "epoch": 0.4943573667711599, "grad_norm": 4.539718151092529, "learning_rate": 7.885e-07, "loss": 11.3049, "step": 1577 }, { "epoch": 0.4946708463949843, "grad_norm": 4.0925140380859375, "learning_rate": 7.890000000000001e-07, "loss": 8.0069, "step": 1578 }, { "epoch": 0.49498432601880876, "grad_norm": 4.172740936279297, "learning_rate": 7.895000000000001e-07, "loss": 7.8851, "step": 1579 }, { "epoch": 0.4952978056426332, "grad_norm": 3.7345142364501953, "learning_rate": 7.900000000000001e-07, "loss": 9.2792, "step": 1580 }, { "epoch": 0.4956112852664577, "grad_norm": 3.628817081451416, "learning_rate": 7.905e-07, "loss": 6.9987, "step": 1581 }, { "epoch": 0.49592476489028214, "grad_norm": 4.242038249969482, "learning_rate": 7.910000000000001e-07, "loss": 9.6288, "step": 1582 }, { "epoch": 0.4962382445141066, "grad_norm": 5.484152793884277, "learning_rate": 7.915000000000001e-07, "loss": 15.2138, "step": 1583 }, { "epoch": 0.496551724137931, "grad_norm": 4.404213905334473, "learning_rate": 7.920000000000001e-07, "loss": 8.7437, "step": 1584 }, { "epoch": 0.49686520376175547, "grad_norm": 3.855804920196533, "learning_rate": 7.925e-07, "loss": 8.6907, "step": 1585 }, { "epoch": 0.4971786833855799, "grad_norm": 6.219444274902344, "learning_rate": 7.93e-07, "loss": 10.0004, "step": 1586 }, { "epoch": 0.4974921630094044, "grad_norm": 5.004199981689453, "learning_rate": 7.935000000000001e-07, "loss": 12.7213, "step": 1587 }, { "epoch": 0.49780564263322885, "grad_norm": 4.671052932739258, "learning_rate": 7.94e-07, "loss": 11.9351, "step": 1588 }, { "epoch": 0.4981191222570533, "grad_norm": 4.2331438064575195, "learning_rate": 7.945000000000001e-07, "loss": 9.8539, "step": 1589 }, { "epoch": 0.49843260188087773, "grad_norm": 3.8226304054260254, "learning_rate": 7.950000000000001e-07, "loss": 8.1552, "step": 1590 }, { "epoch": 0.4987460815047022, "grad_norm": 4.827603340148926, "learning_rate": 7.955e-07, "loss": 11.9278, "step": 1591 }, { "epoch": 0.4990595611285266, "grad_norm": 4.008607864379883, "learning_rate": 7.960000000000001e-07, "loss": 9.1185, "step": 1592 }, { "epoch": 0.4993730407523511, "grad_norm": 3.332667112350464, "learning_rate": 7.965e-07, "loss": 7.4391, "step": 1593 }, { "epoch": 0.49968652037617556, "grad_norm": 4.7476677894592285, "learning_rate": 7.97e-07, "loss": 7.911, "step": 1594 }, { "epoch": 0.5, "grad_norm": 3.6339681148529053, "learning_rate": 7.975000000000001e-07, "loss": 8.406, "step": 1595 }, { "epoch": 0.5003134796238244, "grad_norm": 3.5920093059539795, "learning_rate": 7.98e-07, "loss": 7.1252, "step": 1596 }, { "epoch": 0.5003134796238244, "eval_loss": 28.69378662109375, "eval_runtime": 20.7573, "eval_samples_per_second": 129.449, "eval_steps_per_second": 8.094, "step": 1596 }, { "epoch": 0.5006269592476489, "grad_norm": 3.7933077812194824, "learning_rate": 7.985000000000001e-07, "loss": 7.7692, "step": 1597 }, { "epoch": 0.5009404388714733, "grad_norm": 3.132864475250244, "learning_rate": 7.990000000000001e-07, "loss": 7.6955, "step": 1598 }, { "epoch": 0.5012539184952978, "grad_norm": 3.2161154747009277, "learning_rate": 7.995e-07, "loss": 7.817, "step": 1599 }, { "epoch": 0.5015673981191222, "grad_norm": 3.270275115966797, "learning_rate": 8.000000000000001e-07, "loss": 8.0105, "step": 1600 }, { "epoch": 0.5018808777429468, "grad_norm": 3.4156415462493896, "learning_rate": 8.005e-07, "loss": 7.2844, "step": 1601 }, { "epoch": 0.5021943573667712, "grad_norm": 3.0608019828796387, "learning_rate": 8.010000000000001e-07, "loss": 7.7339, "step": 1602 }, { "epoch": 0.5025078369905956, "grad_norm": 3.434061050415039, "learning_rate": 8.015000000000001e-07, "loss": 8.1699, "step": 1603 }, { "epoch": 0.5028213166144201, "grad_norm": 3.004152536392212, "learning_rate": 8.02e-07, "loss": 6.8241, "step": 1604 }, { "epoch": 0.5031347962382445, "grad_norm": 3.3219809532165527, "learning_rate": 8.025e-07, "loss": 7.6745, "step": 1605 }, { "epoch": 0.503448275862069, "grad_norm": 4.926275730133057, "learning_rate": 8.03e-07, "loss": 11.5191, "step": 1606 }, { "epoch": 0.5037617554858934, "grad_norm": 3.7799763679504395, "learning_rate": 8.035000000000001e-07, "loss": 8.9354, "step": 1607 }, { "epoch": 0.5040752351097179, "grad_norm": 3.3203251361846924, "learning_rate": 8.04e-07, "loss": 7.2962, "step": 1608 }, { "epoch": 0.5043887147335423, "grad_norm": 4.6373419761657715, "learning_rate": 8.045e-07, "loss": 11.3006, "step": 1609 }, { "epoch": 0.5047021943573667, "grad_norm": 4.2450103759765625, "learning_rate": 8.050000000000001e-07, "loss": 7.3383, "step": 1610 }, { "epoch": 0.5050156739811912, "grad_norm": 3.985623359680176, "learning_rate": 8.055000000000001e-07, "loss": 8.2701, "step": 1611 }, { "epoch": 0.5053291536050156, "grad_norm": 3.601292848587036, "learning_rate": 8.060000000000001e-07, "loss": 8.3857, "step": 1612 }, { "epoch": 0.5056426332288402, "grad_norm": 5.563617706298828, "learning_rate": 8.065e-07, "loss": 14.2597, "step": 1613 }, { "epoch": 0.5059561128526646, "grad_norm": 4.677281856536865, "learning_rate": 8.07e-07, "loss": 9.532, "step": 1614 }, { "epoch": 0.5062695924764891, "grad_norm": 3.749307632446289, "learning_rate": 8.075000000000001e-07, "loss": 6.9486, "step": 1615 }, { "epoch": 0.5065830721003135, "grad_norm": 4.544729232788086, "learning_rate": 8.08e-07, "loss": 8.8941, "step": 1616 }, { "epoch": 0.506896551724138, "grad_norm": 3.591256856918335, "learning_rate": 8.085000000000001e-07, "loss": 8.471, "step": 1617 }, { "epoch": 0.5072100313479624, "grad_norm": 5.5906548500061035, "learning_rate": 8.090000000000001e-07, "loss": 10.5195, "step": 1618 }, { "epoch": 0.5075235109717868, "grad_norm": 4.689754009246826, "learning_rate": 8.095e-07, "loss": 8.9176, "step": 1619 }, { "epoch": 0.5078369905956113, "grad_norm": 4.00483512878418, "learning_rate": 8.100000000000001e-07, "loss": 8.361, "step": 1620 }, { "epoch": 0.5081504702194357, "grad_norm": 3.282796621322632, "learning_rate": 8.105e-07, "loss": 7.4275, "step": 1621 }, { "epoch": 0.5084639498432602, "grad_norm": 4.235138416290283, "learning_rate": 8.110000000000002e-07, "loss": 9.8603, "step": 1622 }, { "epoch": 0.5087774294670846, "grad_norm": 4.118296146392822, "learning_rate": 8.115000000000001e-07, "loss": 8.8349, "step": 1623 }, { "epoch": 0.509090909090909, "grad_norm": 3.592233180999756, "learning_rate": 8.12e-07, "loss": 9.0909, "step": 1624 }, { "epoch": 0.5094043887147336, "grad_norm": 3.8250372409820557, "learning_rate": 8.125000000000001e-07, "loss": 6.3841, "step": 1625 }, { "epoch": 0.509717868338558, "grad_norm": 3.962688446044922, "learning_rate": 8.13e-07, "loss": 7.9938, "step": 1626 }, { "epoch": 0.5100313479623825, "grad_norm": 4.621908664703369, "learning_rate": 8.135000000000001e-07, "loss": 9.8957, "step": 1627 }, { "epoch": 0.5103448275862069, "grad_norm": 3.104255199432373, "learning_rate": 8.140000000000001e-07, "loss": 6.7121, "step": 1628 }, { "epoch": 0.5106583072100314, "grad_norm": 4.97641134262085, "learning_rate": 8.145e-07, "loss": 6.1975, "step": 1629 }, { "epoch": 0.5109717868338558, "grad_norm": 4.67917013168335, "learning_rate": 8.150000000000001e-07, "loss": 13.774, "step": 1630 }, { "epoch": 0.5112852664576802, "grad_norm": 3.4128177165985107, "learning_rate": 8.155000000000001e-07, "loss": 8.6059, "step": 1631 }, { "epoch": 0.5115987460815047, "grad_norm": 4.750978469848633, "learning_rate": 8.160000000000001e-07, "loss": 11.319, "step": 1632 }, { "epoch": 0.5119122257053291, "grad_norm": 3.3610668182373047, "learning_rate": 8.165e-07, "loss": 6.3177, "step": 1633 }, { "epoch": 0.5122257053291536, "grad_norm": 4.7112956047058105, "learning_rate": 8.17e-07, "loss": 10.9001, "step": 1634 }, { "epoch": 0.512539184952978, "grad_norm": 4.46808385848999, "learning_rate": 8.175000000000001e-07, "loss": 12.3748, "step": 1635 }, { "epoch": 0.5128526645768025, "grad_norm": 3.6378445625305176, "learning_rate": 8.18e-07, "loss": 9.1623, "step": 1636 }, { "epoch": 0.513166144200627, "grad_norm": 4.917815685272217, "learning_rate": 8.185000000000001e-07, "loss": 11.5192, "step": 1637 }, { "epoch": 0.5134796238244514, "grad_norm": 3.1884193420410156, "learning_rate": 8.190000000000001e-07, "loss": 6.7961, "step": 1638 }, { "epoch": 0.5137931034482759, "grad_norm": 2.6258697509765625, "learning_rate": 8.195e-07, "loss": 6.3646, "step": 1639 }, { "epoch": 0.5141065830721003, "grad_norm": 4.169460296630859, "learning_rate": 8.200000000000001e-07, "loss": 9.467, "step": 1640 }, { "epoch": 0.5144200626959248, "grad_norm": 5.105269432067871, "learning_rate": 8.205e-07, "loss": 13.4161, "step": 1641 }, { "epoch": 0.5147335423197492, "grad_norm": 3.549649238586426, "learning_rate": 8.210000000000002e-07, "loss": 7.6424, "step": 1642 }, { "epoch": 0.5150470219435737, "grad_norm": 5.451201915740967, "learning_rate": 8.215000000000001e-07, "loss": 10.404, "step": 1643 }, { "epoch": 0.5153605015673981, "grad_norm": 4.793301582336426, "learning_rate": 8.22e-07, "loss": 10.6138, "step": 1644 }, { "epoch": 0.5156739811912225, "grad_norm": 4.120952129364014, "learning_rate": 8.225000000000001e-07, "loss": 9.8365, "step": 1645 }, { "epoch": 0.515987460815047, "grad_norm": 5.643743991851807, "learning_rate": 8.23e-07, "loss": 13.1447, "step": 1646 }, { "epoch": 0.5163009404388714, "grad_norm": 3.094728469848633, "learning_rate": 8.235000000000001e-07, "loss": 6.3482, "step": 1647 }, { "epoch": 0.516614420062696, "grad_norm": 5.247632026672363, "learning_rate": 8.240000000000001e-07, "loss": 12.6728, "step": 1648 }, { "epoch": 0.5169278996865204, "grad_norm": 4.007214069366455, "learning_rate": 8.245e-07, "loss": 9.3752, "step": 1649 }, { "epoch": 0.5172413793103449, "grad_norm": 3.778515577316284, "learning_rate": 8.250000000000001e-07, "loss": 8.7075, "step": 1650 }, { "epoch": 0.5175548589341693, "grad_norm": 2.9957754611968994, "learning_rate": 8.255000000000001e-07, "loss": 6.3369, "step": 1651 }, { "epoch": 0.5178683385579937, "grad_norm": 3.4570438861846924, "learning_rate": 8.260000000000001e-07, "loss": 7.8577, "step": 1652 }, { "epoch": 0.5181818181818182, "grad_norm": 3.390144109725952, "learning_rate": 8.265e-07, "loss": 8.0684, "step": 1653 }, { "epoch": 0.5184952978056426, "grad_norm": 3.808957815170288, "learning_rate": 8.27e-07, "loss": 8.0672, "step": 1654 }, { "epoch": 0.5188087774294671, "grad_norm": 3.636155843734741, "learning_rate": 8.275000000000001e-07, "loss": 9.3789, "step": 1655 }, { "epoch": 0.5191222570532915, "grad_norm": 4.033257007598877, "learning_rate": 8.280000000000001e-07, "loss": 8.0115, "step": 1656 }, { "epoch": 0.519435736677116, "grad_norm": 4.070474624633789, "learning_rate": 8.285e-07, "loss": 8.4374, "step": 1657 }, { "epoch": 0.5197492163009404, "grad_norm": 3.276840925216675, "learning_rate": 8.290000000000001e-07, "loss": 6.5463, "step": 1658 }, { "epoch": 0.5200626959247648, "grad_norm": 3.784796714782715, "learning_rate": 8.295000000000001e-07, "loss": 9.6518, "step": 1659 }, { "epoch": 0.5203761755485894, "grad_norm": 3.181030035018921, "learning_rate": 8.300000000000001e-07, "loss": 7.1291, "step": 1660 }, { "epoch": 0.5206896551724138, "grad_norm": 6.969316005706787, "learning_rate": 8.305e-07, "loss": 13.7243, "step": 1661 }, { "epoch": 0.5210031347962383, "grad_norm": 3.7494585514068604, "learning_rate": 8.31e-07, "loss": 7.9837, "step": 1662 }, { "epoch": 0.5213166144200627, "grad_norm": 4.4347991943359375, "learning_rate": 8.315000000000001e-07, "loss": 9.8557, "step": 1663 }, { "epoch": 0.5216300940438872, "grad_norm": 3.5605807304382324, "learning_rate": 8.32e-07, "loss": 8.4987, "step": 1664 }, { "epoch": 0.5219435736677116, "grad_norm": 3.945876359939575, "learning_rate": 8.325000000000001e-07, "loss": 9.7023, "step": 1665 }, { "epoch": 0.522257053291536, "grad_norm": 3.840888500213623, "learning_rate": 8.33e-07, "loss": 10.5396, "step": 1666 }, { "epoch": 0.5225705329153605, "grad_norm": 5.052422523498535, "learning_rate": 8.335e-07, "loss": 10.9094, "step": 1667 }, { "epoch": 0.5228840125391849, "grad_norm": 3.726701259613037, "learning_rate": 8.340000000000001e-07, "loss": 8.6335, "step": 1668 }, { "epoch": 0.5231974921630094, "grad_norm": 3.632563591003418, "learning_rate": 8.345e-07, "loss": 7.3737, "step": 1669 }, { "epoch": 0.5235109717868338, "grad_norm": 5.3117499351501465, "learning_rate": 8.350000000000002e-07, "loss": 9.7193, "step": 1670 }, { "epoch": 0.5238244514106583, "grad_norm": 4.223608493804932, "learning_rate": 8.355000000000001e-07, "loss": 8.8684, "step": 1671 }, { "epoch": 0.5241379310344828, "grad_norm": 4.2938151359558105, "learning_rate": 8.36e-07, "loss": 7.9856, "step": 1672 }, { "epoch": 0.5244514106583072, "grad_norm": 3.227950096130371, "learning_rate": 8.365000000000001e-07, "loss": 8.0511, "step": 1673 }, { "epoch": 0.5247648902821317, "grad_norm": 3.6287143230438232, "learning_rate": 8.37e-07, "loss": 6.8761, "step": 1674 }, { "epoch": 0.5250783699059561, "grad_norm": 3.5546576976776123, "learning_rate": 8.375000000000001e-07, "loss": 8.3474, "step": 1675 }, { "epoch": 0.5253918495297806, "grad_norm": 4.312641143798828, "learning_rate": 8.380000000000001e-07, "loss": 8.7992, "step": 1676 }, { "epoch": 0.525705329153605, "grad_norm": 4.029053688049316, "learning_rate": 8.385e-07, "loss": 8.269, "step": 1677 }, { "epoch": 0.5260188087774295, "grad_norm": 4.90485143661499, "learning_rate": 8.390000000000001e-07, "loss": 10.3717, "step": 1678 }, { "epoch": 0.5263322884012539, "grad_norm": 3.8778631687164307, "learning_rate": 8.395000000000001e-07, "loss": 11.3539, "step": 1679 }, { "epoch": 0.5266457680250783, "grad_norm": 4.429601669311523, "learning_rate": 8.400000000000001e-07, "loss": 10.0634, "step": 1680 }, { "epoch": 0.5269592476489028, "grad_norm": 5.119757175445557, "learning_rate": 8.405e-07, "loss": 10.9194, "step": 1681 }, { "epoch": 0.5272727272727272, "grad_norm": 8.258536338806152, "learning_rate": 8.41e-07, "loss": 22.845, "step": 1682 }, { "epoch": 0.5275862068965518, "grad_norm": 4.698957920074463, "learning_rate": 8.415000000000001e-07, "loss": 8.8281, "step": 1683 }, { "epoch": 0.5278996865203762, "grad_norm": 3.764852285385132, "learning_rate": 8.42e-07, "loss": 7.234, "step": 1684 }, { "epoch": 0.5282131661442007, "grad_norm": 6.866782188415527, "learning_rate": 8.425000000000001e-07, "loss": 9.7779, "step": 1685 }, { "epoch": 0.5285266457680251, "grad_norm": 4.264431476593018, "learning_rate": 8.43e-07, "loss": 10.7163, "step": 1686 }, { "epoch": 0.5288401253918495, "grad_norm": 5.4701995849609375, "learning_rate": 8.435000000000001e-07, "loss": 11.002, "step": 1687 }, { "epoch": 0.529153605015674, "grad_norm": 4.967133522033691, "learning_rate": 8.440000000000001e-07, "loss": 9.6692, "step": 1688 }, { "epoch": 0.5294670846394984, "grad_norm": 3.6275758743286133, "learning_rate": 8.445e-07, "loss": 6.5804, "step": 1689 }, { "epoch": 0.5297805642633229, "grad_norm": 4.6618170738220215, "learning_rate": 8.450000000000002e-07, "loss": 8.9379, "step": 1690 }, { "epoch": 0.5300940438871473, "grad_norm": 3.9030559062957764, "learning_rate": 8.455000000000001e-07, "loss": 7.8233, "step": 1691 }, { "epoch": 0.5304075235109718, "grad_norm": 3.4137344360351562, "learning_rate": 8.46e-07, "loss": 7.6667, "step": 1692 }, { "epoch": 0.5307210031347962, "grad_norm": 3.2711963653564453, "learning_rate": 8.465000000000001e-07, "loss": 7.6183, "step": 1693 }, { "epoch": 0.5310344827586206, "grad_norm": 5.124555587768555, "learning_rate": 8.47e-07, "loss": 16.8092, "step": 1694 }, { "epoch": 0.5313479623824452, "grad_norm": 3.800910711288452, "learning_rate": 8.475000000000001e-07, "loss": 8.8713, "step": 1695 }, { "epoch": 0.5316614420062696, "grad_norm": 4.457700252532959, "learning_rate": 8.480000000000001e-07, "loss": 8.7624, "step": 1696 }, { "epoch": 0.5319749216300941, "grad_norm": 3.4324142932891846, "learning_rate": 8.485e-07, "loss": 7.641, "step": 1697 }, { "epoch": 0.5322884012539185, "grad_norm": 3.523561477661133, "learning_rate": 8.490000000000002e-07, "loss": 7.0024, "step": 1698 }, { "epoch": 0.532601880877743, "grad_norm": 3.4999146461486816, "learning_rate": 8.495000000000001e-07, "loss": 8.2928, "step": 1699 }, { "epoch": 0.5329153605015674, "grad_norm": 4.3573222160339355, "learning_rate": 8.500000000000001e-07, "loss": 8.0377, "step": 1700 }, { "epoch": 0.5332288401253918, "grad_norm": 3.5811939239501953, "learning_rate": 8.505e-07, "loss": 7.2941, "step": 1701 }, { "epoch": 0.5335423197492163, "grad_norm": 3.876803398132324, "learning_rate": 8.51e-07, "loss": 9.4216, "step": 1702 }, { "epoch": 0.5338557993730407, "grad_norm": 2.872718572616577, "learning_rate": 8.515000000000001e-07, "loss": 6.6356, "step": 1703 }, { "epoch": 0.5341692789968652, "grad_norm": 4.38058614730835, "learning_rate": 8.520000000000001e-07, "loss": 9.6664, "step": 1704 }, { "epoch": 0.5344827586206896, "grad_norm": 4.397758960723877, "learning_rate": 8.525000000000001e-07, "loss": 10.8439, "step": 1705 }, { "epoch": 0.534796238244514, "grad_norm": 3.576199769973755, "learning_rate": 8.53e-07, "loss": 7.8423, "step": 1706 }, { "epoch": 0.5351097178683386, "grad_norm": 3.0761783123016357, "learning_rate": 8.535000000000001e-07, "loss": 7.3114, "step": 1707 }, { "epoch": 0.535423197492163, "grad_norm": 4.053654670715332, "learning_rate": 8.540000000000001e-07, "loss": 8.3691, "step": 1708 }, { "epoch": 0.5357366771159875, "grad_norm": 4.436123847961426, "learning_rate": 8.545e-07, "loss": 12.1063, "step": 1709 }, { "epoch": 0.5360501567398119, "grad_norm": 5.484961986541748, "learning_rate": 8.550000000000002e-07, "loss": 10.3859, "step": 1710 }, { "epoch": 0.5363636363636364, "grad_norm": 4.26839542388916, "learning_rate": 8.555000000000001e-07, "loss": 8.8806, "step": 1711 }, { "epoch": 0.5366771159874608, "grad_norm": 5.425708770751953, "learning_rate": 8.56e-07, "loss": 11.4141, "step": 1712 }, { "epoch": 0.5369905956112853, "grad_norm": 4.29015588760376, "learning_rate": 8.565000000000001e-07, "loss": 9.4032, "step": 1713 }, { "epoch": 0.5373040752351097, "grad_norm": 3.599637508392334, "learning_rate": 8.57e-07, "loss": 8.3692, "step": 1714 }, { "epoch": 0.5376175548589341, "grad_norm": 4.9885358810424805, "learning_rate": 8.575000000000002e-07, "loss": 9.0797, "step": 1715 }, { "epoch": 0.5379310344827586, "grad_norm": 3.783560037612915, "learning_rate": 8.580000000000001e-07, "loss": 7.1796, "step": 1716 }, { "epoch": 0.538244514106583, "grad_norm": 8.088088035583496, "learning_rate": 8.585e-07, "loss": 19.7025, "step": 1717 }, { "epoch": 0.5385579937304075, "grad_norm": 2.911607503890991, "learning_rate": 8.590000000000002e-07, "loss": 6.2524, "step": 1718 }, { "epoch": 0.538871473354232, "grad_norm": 3.7300970554351807, "learning_rate": 8.595000000000001e-07, "loss": 8.4442, "step": 1719 }, { "epoch": 0.5391849529780565, "grad_norm": 3.8973944187164307, "learning_rate": 8.6e-07, "loss": 9.406, "step": 1720 }, { "epoch": 0.5394984326018809, "grad_norm": 4.2937397956848145, "learning_rate": 8.605000000000001e-07, "loss": 6.9903, "step": 1721 }, { "epoch": 0.5398119122257053, "grad_norm": 4.311274528503418, "learning_rate": 8.61e-07, "loss": 7.9766, "step": 1722 }, { "epoch": 0.5401253918495298, "grad_norm": 3.7893688678741455, "learning_rate": 8.615000000000001e-07, "loss": 8.0692, "step": 1723 }, { "epoch": 0.5404388714733542, "grad_norm": 3.4368412494659424, "learning_rate": 8.620000000000001e-07, "loss": 7.6627, "step": 1724 }, { "epoch": 0.5407523510971787, "grad_norm": 4.653461933135986, "learning_rate": 8.625e-07, "loss": 8.2418, "step": 1725 }, { "epoch": 0.5410658307210031, "grad_norm": 3.554748058319092, "learning_rate": 8.63e-07, "loss": 7.7693, "step": 1726 }, { "epoch": 0.5413793103448276, "grad_norm": 3.596076726913452, "learning_rate": 8.635000000000001e-07, "loss": 11.7324, "step": 1727 }, { "epoch": 0.541692789968652, "grad_norm": 5.29031229019165, "learning_rate": 8.640000000000001e-07, "loss": 10.1403, "step": 1728 }, { "epoch": 0.5420062695924764, "grad_norm": 4.6957807540893555, "learning_rate": 8.645e-07, "loss": 10.0583, "step": 1729 }, { "epoch": 0.542319749216301, "grad_norm": 3.0998942852020264, "learning_rate": 8.65e-07, "loss": 7.5078, "step": 1730 }, { "epoch": 0.5426332288401254, "grad_norm": 6.108078479766846, "learning_rate": 8.655000000000001e-07, "loss": 15.5995, "step": 1731 }, { "epoch": 0.5429467084639499, "grad_norm": 4.639706611633301, "learning_rate": 8.66e-07, "loss": 8.0792, "step": 1732 }, { "epoch": 0.5432601880877743, "grad_norm": 3.6627628803253174, "learning_rate": 8.665000000000001e-07, "loss": 8.4311, "step": 1733 }, { "epoch": 0.5435736677115988, "grad_norm": 3.9771621227264404, "learning_rate": 8.67e-07, "loss": 8.8372, "step": 1734 }, { "epoch": 0.5438871473354232, "grad_norm": 3.9450392723083496, "learning_rate": 8.675000000000001e-07, "loss": 10.2992, "step": 1735 }, { "epoch": 0.5442006269592476, "grad_norm": 3.019969940185547, "learning_rate": 8.680000000000001e-07, "loss": 7.7473, "step": 1736 }, { "epoch": 0.5445141065830721, "grad_norm": 5.00993013381958, "learning_rate": 8.685e-07, "loss": 10.7672, "step": 1737 }, { "epoch": 0.5448275862068965, "grad_norm": 4.663172245025635, "learning_rate": 8.690000000000002e-07, "loss": 8.9194, "step": 1738 }, { "epoch": 0.545141065830721, "grad_norm": 5.089305400848389, "learning_rate": 8.695000000000001e-07, "loss": 10.1471, "step": 1739 }, { "epoch": 0.5454545454545454, "grad_norm": 8.746878623962402, "learning_rate": 8.7e-07, "loss": 11.3008, "step": 1740 }, { "epoch": 0.5457680250783699, "grad_norm": 3.6730334758758545, "learning_rate": 8.705000000000001e-07, "loss": 9.8861, "step": 1741 }, { "epoch": 0.5460815047021944, "grad_norm": 3.33270263671875, "learning_rate": 8.71e-07, "loss": 8.7378, "step": 1742 }, { "epoch": 0.5463949843260189, "grad_norm": 3.6083526611328125, "learning_rate": 8.715000000000001e-07, "loss": 7.0979, "step": 1743 }, { "epoch": 0.5467084639498433, "grad_norm": 4.226643085479736, "learning_rate": 8.720000000000001e-07, "loss": 10.3873, "step": 1744 }, { "epoch": 0.5470219435736677, "grad_norm": 5.096066951751709, "learning_rate": 8.725e-07, "loss": 10.6816, "step": 1745 }, { "epoch": 0.5473354231974922, "grad_norm": 5.229349613189697, "learning_rate": 8.73e-07, "loss": 11.0695, "step": 1746 }, { "epoch": 0.5476489028213166, "grad_norm": 3.8301353454589844, "learning_rate": 8.735000000000001e-07, "loss": 8.6425, "step": 1747 }, { "epoch": 0.5479623824451411, "grad_norm": 3.517941951751709, "learning_rate": 8.740000000000001e-07, "loss": 6.2092, "step": 1748 }, { "epoch": 0.5482758620689655, "grad_norm": 5.241222858428955, "learning_rate": 8.745000000000001e-07, "loss": 11.7764, "step": 1749 }, { "epoch": 0.54858934169279, "grad_norm": 5.1557793617248535, "learning_rate": 8.75e-07, "loss": 13.8192, "step": 1750 }, { "epoch": 0.5489028213166144, "grad_norm": 4.522706508636475, "learning_rate": 8.755000000000001e-07, "loss": 9.2798, "step": 1751 }, { "epoch": 0.5492163009404388, "grad_norm": 5.382571697235107, "learning_rate": 8.760000000000001e-07, "loss": 11.1607, "step": 1752 }, { "epoch": 0.5495297805642633, "grad_norm": 3.164684295654297, "learning_rate": 8.765000000000001e-07, "loss": 6.6911, "step": 1753 }, { "epoch": 0.5498432601880878, "grad_norm": 3.7033884525299072, "learning_rate": 8.77e-07, "loss": 7.5505, "step": 1754 }, { "epoch": 0.5501567398119123, "grad_norm": 4.4705424308776855, "learning_rate": 8.775000000000001e-07, "loss": 10.6266, "step": 1755 }, { "epoch": 0.5504702194357367, "grad_norm": 3.576754331588745, "learning_rate": 8.780000000000001e-07, "loss": 7.6148, "step": 1756 }, { "epoch": 0.5507836990595611, "grad_norm": 6.662974834442139, "learning_rate": 8.785e-07, "loss": 15.6583, "step": 1757 }, { "epoch": 0.5510971786833856, "grad_norm": 4.8018364906311035, "learning_rate": 8.790000000000002e-07, "loss": 7.7867, "step": 1758 }, { "epoch": 0.55141065830721, "grad_norm": 3.5706562995910645, "learning_rate": 8.795000000000001e-07, "loss": 7.1267, "step": 1759 }, { "epoch": 0.5517241379310345, "grad_norm": 3.8161635398864746, "learning_rate": 8.8e-07, "loss": 8.0858, "step": 1760 }, { "epoch": 0.5520376175548589, "grad_norm": 4.2101898193359375, "learning_rate": 8.805000000000001e-07, "loss": 8.4719, "step": 1761 }, { "epoch": 0.5523510971786834, "grad_norm": 3.5980849266052246, "learning_rate": 8.81e-07, "loss": 8.5086, "step": 1762 }, { "epoch": 0.5526645768025078, "grad_norm": 4.349874973297119, "learning_rate": 8.815000000000002e-07, "loss": 9.6007, "step": 1763 }, { "epoch": 0.5529780564263322, "grad_norm": 4.534544467926025, "learning_rate": 8.820000000000001e-07, "loss": 9.1542, "step": 1764 }, { "epoch": 0.5532915360501567, "grad_norm": 5.753028392791748, "learning_rate": 8.825e-07, "loss": 13.3121, "step": 1765 }, { "epoch": 0.5536050156739812, "grad_norm": 4.646644592285156, "learning_rate": 8.830000000000001e-07, "loss": 10.1693, "step": 1766 }, { "epoch": 0.5539184952978057, "grad_norm": 4.152806758880615, "learning_rate": 8.835000000000001e-07, "loss": 9.3472, "step": 1767 }, { "epoch": 0.5542319749216301, "grad_norm": 4.156120300292969, "learning_rate": 8.840000000000001e-07, "loss": 8.9546, "step": 1768 }, { "epoch": 0.5545454545454546, "grad_norm": 4.695617198944092, "learning_rate": 8.845000000000001e-07, "loss": 9.1984, "step": 1769 }, { "epoch": 0.554858934169279, "grad_norm": 3.514662742614746, "learning_rate": 8.85e-07, "loss": 6.9892, "step": 1770 }, { "epoch": 0.5551724137931034, "grad_norm": 6.180057048797607, "learning_rate": 8.855000000000001e-07, "loss": 15.8813, "step": 1771 }, { "epoch": 0.5554858934169279, "grad_norm": 3.279160261154175, "learning_rate": 8.860000000000001e-07, "loss": 6.86, "step": 1772 }, { "epoch": 0.5557993730407523, "grad_norm": 4.369133472442627, "learning_rate": 8.865000000000001e-07, "loss": 9.133, "step": 1773 }, { "epoch": 0.5561128526645768, "grad_norm": 3.11153507232666, "learning_rate": 8.87e-07, "loss": 7.517, "step": 1774 }, { "epoch": 0.5564263322884012, "grad_norm": 4.086194038391113, "learning_rate": 8.875000000000001e-07, "loss": 8.7146, "step": 1775 }, { "epoch": 0.5567398119122257, "grad_norm": 4.669196605682373, "learning_rate": 8.880000000000001e-07, "loss": 9.4316, "step": 1776 }, { "epoch": 0.5570532915360502, "grad_norm": 4.278200626373291, "learning_rate": 8.885e-07, "loss": 9.8683, "step": 1777 }, { "epoch": 0.5573667711598747, "grad_norm": 6.216943740844727, "learning_rate": 8.890000000000002e-07, "loss": 12.5523, "step": 1778 }, { "epoch": 0.5576802507836991, "grad_norm": 4.141219139099121, "learning_rate": 8.895000000000001e-07, "loss": 9.5758, "step": 1779 }, { "epoch": 0.5579937304075235, "grad_norm": 5.539915561676025, "learning_rate": 8.900000000000001e-07, "loss": 13.5515, "step": 1780 }, { "epoch": 0.558307210031348, "grad_norm": 4.460906505584717, "learning_rate": 8.905000000000001e-07, "loss": 9.7846, "step": 1781 }, { "epoch": 0.5586206896551724, "grad_norm": 3.381852388381958, "learning_rate": 8.91e-07, "loss": 7.3451, "step": 1782 }, { "epoch": 0.5589341692789969, "grad_norm": 4.935987949371338, "learning_rate": 8.915e-07, "loss": 9.0696, "step": 1783 }, { "epoch": 0.5592476489028213, "grad_norm": 4.209127426147461, "learning_rate": 8.920000000000001e-07, "loss": 8.1537, "step": 1784 }, { "epoch": 0.5595611285266457, "grad_norm": 3.22871994972229, "learning_rate": 8.925e-07, "loss": 7.8978, "step": 1785 }, { "epoch": 0.5598746081504702, "grad_norm": 4.8778510093688965, "learning_rate": 8.930000000000001e-07, "loss": 9.9476, "step": 1786 }, { "epoch": 0.5601880877742946, "grad_norm": 4.3461384773254395, "learning_rate": 8.935000000000001e-07, "loss": 8.4605, "step": 1787 }, { "epoch": 0.5605015673981191, "grad_norm": 4.09140157699585, "learning_rate": 8.94e-07, "loss": 8.6371, "step": 1788 }, { "epoch": 0.5608150470219436, "grad_norm": 3.154021739959717, "learning_rate": 8.945000000000001e-07, "loss": 6.6153, "step": 1789 }, { "epoch": 0.5611285266457681, "grad_norm": 5.04860782623291, "learning_rate": 8.95e-07, "loss": 10.1237, "step": 1790 }, { "epoch": 0.5614420062695925, "grad_norm": 3.98972487449646, "learning_rate": 8.955000000000002e-07, "loss": 7.8599, "step": 1791 }, { "epoch": 0.561755485893417, "grad_norm": 3.527597665786743, "learning_rate": 8.960000000000001e-07, "loss": 8.2753, "step": 1792 }, { "epoch": 0.5620689655172414, "grad_norm": 6.0412139892578125, "learning_rate": 8.965e-07, "loss": 13.1168, "step": 1793 }, { "epoch": 0.5623824451410658, "grad_norm": 4.94082498550415, "learning_rate": 8.97e-07, "loss": 10.4559, "step": 1794 }, { "epoch": 0.5626959247648903, "grad_norm": 4.965950012207031, "learning_rate": 8.975000000000001e-07, "loss": 10.237, "step": 1795 }, { "epoch": 0.5630094043887147, "grad_norm": 4.55895471572876, "learning_rate": 8.980000000000001e-07, "loss": 10.7508, "step": 1796 }, { "epoch": 0.5633228840125392, "grad_norm": 4.183038234710693, "learning_rate": 8.985000000000001e-07, "loss": 7.4745, "step": 1797 }, { "epoch": 0.5636363636363636, "grad_norm": 3.6337876319885254, "learning_rate": 8.99e-07, "loss": 6.8318, "step": 1798 }, { "epoch": 0.563949843260188, "grad_norm": 5.689560890197754, "learning_rate": 8.995000000000001e-07, "loss": 12.2413, "step": 1799 }, { "epoch": 0.5642633228840125, "grad_norm": 3.8097832202911377, "learning_rate": 9.000000000000001e-07, "loss": 7.0029, "step": 1800 }, { "epoch": 0.564576802507837, "grad_norm": 4.039278030395508, "learning_rate": 9.005000000000001e-07, "loss": 9.2906, "step": 1801 }, { "epoch": 0.5648902821316615, "grad_norm": 3.7648391723632812, "learning_rate": 9.01e-07, "loss": 7.9313, "step": 1802 }, { "epoch": 0.5652037617554859, "grad_norm": 4.256327152252197, "learning_rate": 9.015e-07, "loss": 6.9056, "step": 1803 }, { "epoch": 0.5655172413793104, "grad_norm": 4.904327869415283, "learning_rate": 9.020000000000001e-07, "loss": 9.3173, "step": 1804 }, { "epoch": 0.5658307210031348, "grad_norm": 3.761366367340088, "learning_rate": 9.025e-07, "loss": 10.9158, "step": 1805 }, { "epoch": 0.5661442006269592, "grad_norm": 4.512961387634277, "learning_rate": 9.030000000000001e-07, "loss": 9.3172, "step": 1806 }, { "epoch": 0.5664576802507837, "grad_norm": 4.454610824584961, "learning_rate": 9.035000000000001e-07, "loss": 9.806, "step": 1807 }, { "epoch": 0.5667711598746081, "grad_norm": 3.878605842590332, "learning_rate": 9.04e-07, "loss": 7.8319, "step": 1808 }, { "epoch": 0.5670846394984326, "grad_norm": 3.3651275634765625, "learning_rate": 9.045000000000001e-07, "loss": 7.7877, "step": 1809 }, { "epoch": 0.567398119122257, "grad_norm": 4.225704669952393, "learning_rate": 9.05e-07, "loss": 8.8007, "step": 1810 }, { "epoch": 0.5677115987460815, "grad_norm": 4.0525617599487305, "learning_rate": 9.055000000000002e-07, "loss": 8.9639, "step": 1811 }, { "epoch": 0.568025078369906, "grad_norm": 3.8242380619049072, "learning_rate": 9.060000000000001e-07, "loss": 8.4361, "step": 1812 }, { "epoch": 0.5683385579937305, "grad_norm": 6.60628080368042, "learning_rate": 9.065e-07, "loss": 9.6387, "step": 1813 }, { "epoch": 0.5686520376175549, "grad_norm": 4.6100616455078125, "learning_rate": 9.070000000000001e-07, "loss": 8.7753, "step": 1814 }, { "epoch": 0.5689655172413793, "grad_norm": 3.9851086139678955, "learning_rate": 9.075000000000001e-07, "loss": 7.7566, "step": 1815 }, { "epoch": 0.5692789968652038, "grad_norm": 5.08043098449707, "learning_rate": 9.080000000000001e-07, "loss": 12.0445, "step": 1816 }, { "epoch": 0.5695924764890282, "grad_norm": 8.151008605957031, "learning_rate": 9.085000000000001e-07, "loss": 12.3116, "step": 1817 }, { "epoch": 0.5699059561128527, "grad_norm": 5.475290775299072, "learning_rate": 9.09e-07, "loss": 12.5451, "step": 1818 }, { "epoch": 0.5702194357366771, "grad_norm": 3.203784227371216, "learning_rate": 9.095000000000001e-07, "loss": 6.5738, "step": 1819 }, { "epoch": 0.5705329153605015, "grad_norm": 3.322894334793091, "learning_rate": 9.100000000000001e-07, "loss": 7.036, "step": 1820 }, { "epoch": 0.570846394984326, "grad_norm": 5.2033772468566895, "learning_rate": 9.105000000000001e-07, "loss": 11.3802, "step": 1821 }, { "epoch": 0.5711598746081504, "grad_norm": 3.5129075050354004, "learning_rate": 9.11e-07, "loss": 5.7609, "step": 1822 }, { "epoch": 0.5714733542319749, "grad_norm": 3.822200059890747, "learning_rate": 9.115e-07, "loss": 8.4694, "step": 1823 }, { "epoch": 0.5717868338557994, "grad_norm": 5.213927268981934, "learning_rate": 9.120000000000001e-07, "loss": 10.8033, "step": 1824 }, { "epoch": 0.5721003134796239, "grad_norm": 3.970292806625366, "learning_rate": 9.125e-07, "loss": 8.1297, "step": 1825 }, { "epoch": 0.5724137931034483, "grad_norm": 3.255862236022949, "learning_rate": 9.130000000000001e-07, "loss": 6.6128, "step": 1826 }, { "epoch": 0.5727272727272728, "grad_norm": 5.0875139236450195, "learning_rate": 9.135000000000001e-07, "loss": 10.7206, "step": 1827 }, { "epoch": 0.5730407523510972, "grad_norm": 4.447558403015137, "learning_rate": 9.140000000000001e-07, "loss": 8.3421, "step": 1828 }, { "epoch": 0.5733542319749216, "grad_norm": 4.044737815856934, "learning_rate": 9.145000000000001e-07, "loss": 7.2425, "step": 1829 }, { "epoch": 0.5736677115987461, "grad_norm": 2.927945852279663, "learning_rate": 9.15e-07, "loss": 6.6691, "step": 1830 }, { "epoch": 0.5739811912225705, "grad_norm": 3.3857815265655518, "learning_rate": 9.155000000000002e-07, "loss": 6.6235, "step": 1831 }, { "epoch": 0.574294670846395, "grad_norm": 4.590656757354736, "learning_rate": 9.160000000000001e-07, "loss": 11.8911, "step": 1832 }, { "epoch": 0.5746081504702194, "grad_norm": 4.849060535430908, "learning_rate": 9.165e-07, "loss": 9.1651, "step": 1833 }, { "epoch": 0.5749216300940438, "grad_norm": 3.661118984222412, "learning_rate": 9.170000000000001e-07, "loss": 8.3145, "step": 1834 }, { "epoch": 0.5752351097178683, "grad_norm": 4.550508975982666, "learning_rate": 9.175000000000001e-07, "loss": 10.9553, "step": 1835 }, { "epoch": 0.5755485893416928, "grad_norm": 4.856189250946045, "learning_rate": 9.180000000000001e-07, "loss": 10.9995, "step": 1836 }, { "epoch": 0.5758620689655173, "grad_norm": 3.962833881378174, "learning_rate": 9.185000000000001e-07, "loss": 7.6791, "step": 1837 }, { "epoch": 0.5761755485893417, "grad_norm": 3.818290948867798, "learning_rate": 9.19e-07, "loss": 9.4411, "step": 1838 }, { "epoch": 0.5764890282131662, "grad_norm": 4.265783309936523, "learning_rate": 9.195000000000002e-07, "loss": 8.6175, "step": 1839 }, { "epoch": 0.5768025078369906, "grad_norm": 4.399896621704102, "learning_rate": 9.200000000000001e-07, "loss": 8.0593, "step": 1840 }, { "epoch": 0.577115987460815, "grad_norm": 3.487797975540161, "learning_rate": 9.205000000000001e-07, "loss": 7.8994, "step": 1841 }, { "epoch": 0.5774294670846395, "grad_norm": 4.428309440612793, "learning_rate": 9.210000000000001e-07, "loss": 7.2103, "step": 1842 }, { "epoch": 0.5777429467084639, "grad_norm": 4.569345951080322, "learning_rate": 9.215e-07, "loss": 9.6218, "step": 1843 }, { "epoch": 0.5780564263322884, "grad_norm": 3.990476369857788, "learning_rate": 9.220000000000001e-07, "loss": 9.4368, "step": 1844 }, { "epoch": 0.5783699059561128, "grad_norm": 3.2364110946655273, "learning_rate": 9.225000000000001e-07, "loss": 6.377, "step": 1845 }, { "epoch": 0.5786833855799373, "grad_norm": 4.474930763244629, "learning_rate": 9.23e-07, "loss": 8.6538, "step": 1846 }, { "epoch": 0.5789968652037617, "grad_norm": 4.316197872161865, "learning_rate": 9.235000000000001e-07, "loss": 9.8697, "step": 1847 }, { "epoch": 0.5793103448275863, "grad_norm": 3.6721043586730957, "learning_rate": 9.240000000000001e-07, "loss": 7.6369, "step": 1848 }, { "epoch": 0.5796238244514107, "grad_norm": 4.066787242889404, "learning_rate": 9.245000000000001e-07, "loss": 8.1153, "step": 1849 }, { "epoch": 0.5799373040752351, "grad_norm": 4.483874320983887, "learning_rate": 9.25e-07, "loss": 7.9151, "step": 1850 }, { "epoch": 0.5802507836990596, "grad_norm": 6.05754280090332, "learning_rate": 9.255e-07, "loss": 11.5638, "step": 1851 }, { "epoch": 0.580564263322884, "grad_norm": 4.692145824432373, "learning_rate": 9.260000000000001e-07, "loss": 9.0891, "step": 1852 }, { "epoch": 0.5808777429467085, "grad_norm": 4.905162334442139, "learning_rate": 9.265e-07, "loss": 11.009, "step": 1853 }, { "epoch": 0.5811912225705329, "grad_norm": 3.9991772174835205, "learning_rate": 9.270000000000001e-07, "loss": 7.1689, "step": 1854 }, { "epoch": 0.5815047021943573, "grad_norm": 3.545417308807373, "learning_rate": 9.275000000000001e-07, "loss": 8.8178, "step": 1855 }, { "epoch": 0.5818181818181818, "grad_norm": 3.4547829627990723, "learning_rate": 9.28e-07, "loss": 6.4124, "step": 1856 }, { "epoch": 0.5821316614420062, "grad_norm": 5.388890266418457, "learning_rate": 9.285000000000001e-07, "loss": 9.8751, "step": 1857 }, { "epoch": 0.5824451410658307, "grad_norm": 3.773027181625366, "learning_rate": 9.29e-07, "loss": 9.1356, "step": 1858 }, { "epoch": 0.5827586206896552, "grad_norm": 3.524904727935791, "learning_rate": 9.295000000000002e-07, "loss": 7.9023, "step": 1859 }, { "epoch": 0.5830721003134797, "grad_norm": 3.681853771209717, "learning_rate": 9.300000000000001e-07, "loss": 7.6041, "step": 1860 }, { "epoch": 0.5833855799373041, "grad_norm": 4.124235153198242, "learning_rate": 9.305e-07, "loss": 10.3233, "step": 1861 }, { "epoch": 0.5836990595611286, "grad_norm": 4.318082809448242, "learning_rate": 9.310000000000001e-07, "loss": 10.4379, "step": 1862 }, { "epoch": 0.584012539184953, "grad_norm": 3.5927541255950928, "learning_rate": 9.315e-07, "loss": 7.0752, "step": 1863 }, { "epoch": 0.5843260188087774, "grad_norm": 3.7160086631774902, "learning_rate": 9.320000000000001e-07, "loss": 8.0991, "step": 1864 }, { "epoch": 0.5846394984326019, "grad_norm": 4.717742919921875, "learning_rate": 9.325000000000001e-07, "loss": 8.6703, "step": 1865 }, { "epoch": 0.5849529780564263, "grad_norm": 4.089090347290039, "learning_rate": 9.33e-07, "loss": 8.6572, "step": 1866 }, { "epoch": 0.5852664576802508, "grad_norm": 5.522655010223389, "learning_rate": 9.335000000000001e-07, "loss": 9.8966, "step": 1867 }, { "epoch": 0.5855799373040752, "grad_norm": 3.4123260974884033, "learning_rate": 9.340000000000001e-07, "loss": 7.0431, "step": 1868 }, { "epoch": 0.5858934169278996, "grad_norm": 4.160542011260986, "learning_rate": 9.345000000000001e-07, "loss": 8.1701, "step": 1869 }, { "epoch": 0.5862068965517241, "grad_norm": 3.6532208919525146, "learning_rate": 9.35e-07, "loss": 10.5839, "step": 1870 }, { "epoch": 0.5865203761755486, "grad_norm": 4.788308143615723, "learning_rate": 9.355e-07, "loss": 9.6103, "step": 1871 }, { "epoch": 0.5868338557993731, "grad_norm": 3.656114339828491, "learning_rate": 9.360000000000001e-07, "loss": 9.0917, "step": 1872 }, { "epoch": 0.5871473354231975, "grad_norm": 3.7740187644958496, "learning_rate": 9.365000000000001e-07, "loss": 10.3359, "step": 1873 }, { "epoch": 0.587460815047022, "grad_norm": 3.203486680984497, "learning_rate": 9.370000000000001e-07, "loss": 6.0882, "step": 1874 }, { "epoch": 0.5877742946708464, "grad_norm": 4.1461920738220215, "learning_rate": 9.375000000000001e-07, "loss": 10.5749, "step": 1875 }, { "epoch": 0.5880877742946709, "grad_norm": 4.353470802307129, "learning_rate": 9.380000000000001e-07, "loss": 8.5529, "step": 1876 }, { "epoch": 0.5884012539184953, "grad_norm": 4.317278861999512, "learning_rate": 9.385000000000001e-07, "loss": 9.1286, "step": 1877 }, { "epoch": 0.5887147335423197, "grad_norm": 4.176400661468506, "learning_rate": 9.39e-07, "loss": 9.2643, "step": 1878 }, { "epoch": 0.5890282131661442, "grad_norm": 3.544487953186035, "learning_rate": 9.395000000000002e-07, "loss": 7.9035, "step": 1879 }, { "epoch": 0.5893416927899686, "grad_norm": 4.3520731925964355, "learning_rate": 9.400000000000001e-07, "loss": 8.9988, "step": 1880 }, { "epoch": 0.5896551724137931, "grad_norm": 4.328791618347168, "learning_rate": 9.405e-07, "loss": 8.2385, "step": 1881 }, { "epoch": 0.5899686520376175, "grad_norm": 4.532482147216797, "learning_rate": 9.410000000000001e-07, "loss": 10.0327, "step": 1882 }, { "epoch": 0.590282131661442, "grad_norm": 4.309502124786377, "learning_rate": 9.415e-07, "loss": 9.5955, "step": 1883 }, { "epoch": 0.5905956112852665, "grad_norm": 9.216413497924805, "learning_rate": 9.420000000000002e-07, "loss": 17.2207, "step": 1884 }, { "epoch": 0.5909090909090909, "grad_norm": 4.302111625671387, "learning_rate": 9.425000000000001e-07, "loss": 8.134, "step": 1885 }, { "epoch": 0.5912225705329154, "grad_norm": 3.478489637374878, "learning_rate": 9.43e-07, "loss": 7.1565, "step": 1886 }, { "epoch": 0.5915360501567398, "grad_norm": 5.942442893981934, "learning_rate": 9.435000000000002e-07, "loss": 16.1045, "step": 1887 }, { "epoch": 0.5918495297805643, "grad_norm": 4.080247402191162, "learning_rate": 9.440000000000001e-07, "loss": 10.6923, "step": 1888 }, { "epoch": 0.5921630094043887, "grad_norm": 4.32432222366333, "learning_rate": 9.445000000000001e-07, "loss": 8.9229, "step": 1889 }, { "epoch": 0.5924764890282131, "grad_norm": 3.346331834793091, "learning_rate": 9.450000000000001e-07, "loss": 6.9437, "step": 1890 }, { "epoch": 0.5927899686520376, "grad_norm": 8.710538864135742, "learning_rate": 9.455e-07, "loss": 18.3767, "step": 1891 }, { "epoch": 0.593103448275862, "grad_norm": 6.308311462402344, "learning_rate": 9.460000000000001e-07, "loss": 13.8071, "step": 1892 }, { "epoch": 0.5934169278996865, "grad_norm": 4.4475321769714355, "learning_rate": 9.465000000000001e-07, "loss": 8.8487, "step": 1893 }, { "epoch": 0.5937304075235109, "grad_norm": 5.897817611694336, "learning_rate": 9.470000000000001e-07, "loss": 11.9756, "step": 1894 }, { "epoch": 0.5940438871473355, "grad_norm": 5.22391414642334, "learning_rate": 9.475e-07, "loss": 7.3174, "step": 1895 }, { "epoch": 0.5943573667711599, "grad_norm": 4.082650184631348, "learning_rate": 9.480000000000001e-07, "loss": 7.8726, "step": 1896 }, { "epoch": 0.5946708463949844, "grad_norm": 4.346643447875977, "learning_rate": 9.485000000000001e-07, "loss": 8.1732, "step": 1897 }, { "epoch": 0.5949843260188088, "grad_norm": 6.866816997528076, "learning_rate": 9.49e-07, "loss": 12.5577, "step": 1898 }, { "epoch": 0.5952978056426332, "grad_norm": 3.5348422527313232, "learning_rate": 9.495000000000002e-07, "loss": 6.8785, "step": 1899 }, { "epoch": 0.5956112852664577, "grad_norm": 3.0132737159729004, "learning_rate": 9.500000000000001e-07, "loss": 5.57, "step": 1900 }, { "epoch": 0.5959247648902821, "grad_norm": 3.286238193511963, "learning_rate": 9.505e-07, "loss": 8.4462, "step": 1901 }, { "epoch": 0.5962382445141066, "grad_norm": 5.34050989151001, "learning_rate": 9.510000000000001e-07, "loss": 8.19, "step": 1902 }, { "epoch": 0.596551724137931, "grad_norm": 3.889528751373291, "learning_rate": 9.515e-07, "loss": 7.5043, "step": 1903 }, { "epoch": 0.5968652037617554, "grad_norm": 3.9055964946746826, "learning_rate": 9.520000000000002e-07, "loss": 8.6483, "step": 1904 }, { "epoch": 0.5971786833855799, "grad_norm": 3.298340320587158, "learning_rate": 9.525000000000001e-07, "loss": 6.4088, "step": 1905 }, { "epoch": 0.5974921630094044, "grad_norm": 3.5493357181549072, "learning_rate": 9.53e-07, "loss": 7.4659, "step": 1906 }, { "epoch": 0.5978056426332289, "grad_norm": 4.918447494506836, "learning_rate": 9.535000000000002e-07, "loss": 9.6589, "step": 1907 }, { "epoch": 0.5981191222570533, "grad_norm": 3.490692615509033, "learning_rate": 9.54e-07, "loss": 7.0492, "step": 1908 }, { "epoch": 0.5984326018808778, "grad_norm": 4.562289237976074, "learning_rate": 9.545e-07, "loss": 9.4843, "step": 1909 }, { "epoch": 0.5987460815047022, "grad_norm": 5.618401050567627, "learning_rate": 9.550000000000002e-07, "loss": 15.0284, "step": 1910 }, { "epoch": 0.5990595611285267, "grad_norm": 5.000195503234863, "learning_rate": 9.555e-07, "loss": 10.3012, "step": 1911 }, { "epoch": 0.5993730407523511, "grad_norm": 4.634150505065918, "learning_rate": 9.56e-07, "loss": 8.0312, "step": 1912 }, { "epoch": 0.5996865203761755, "grad_norm": 4.144935131072998, "learning_rate": 9.565e-07, "loss": 7.6425, "step": 1913 }, { "epoch": 0.6, "grad_norm": 6.016616344451904, "learning_rate": 9.570000000000001e-07, "loss": 10.0162, "step": 1914 }, { "epoch": 0.6003134796238244, "grad_norm": 4.132763862609863, "learning_rate": 9.575000000000001e-07, "loss": 7.813, "step": 1915 }, { "epoch": 0.6006269592476489, "grad_norm": 5.254676818847656, "learning_rate": 9.58e-07, "loss": 9.9075, "step": 1916 }, { "epoch": 0.6009404388714733, "grad_norm": 3.9892544746398926, "learning_rate": 9.585000000000002e-07, "loss": 8.0493, "step": 1917 }, { "epoch": 0.6012539184952979, "grad_norm": 4.14557409286499, "learning_rate": 9.59e-07, "loss": 8.4252, "step": 1918 }, { "epoch": 0.6015673981191223, "grad_norm": 5.171091079711914, "learning_rate": 9.595e-07, "loss": 9.5018, "step": 1919 }, { "epoch": 0.6018808777429467, "grad_norm": 5.105014801025391, "learning_rate": 9.600000000000001e-07, "loss": 8.7117, "step": 1920 }, { "epoch": 0.6021943573667712, "grad_norm": 4.491677284240723, "learning_rate": 9.605e-07, "loss": 9.022, "step": 1921 }, { "epoch": 0.6025078369905956, "grad_norm": 4.195937156677246, "learning_rate": 9.610000000000002e-07, "loss": 8.2956, "step": 1922 }, { "epoch": 0.6028213166144201, "grad_norm": 2.9639382362365723, "learning_rate": 9.615e-07, "loss": 6.2187, "step": 1923 }, { "epoch": 0.6031347962382445, "grad_norm": 5.277063369750977, "learning_rate": 9.62e-07, "loss": 9.245, "step": 1924 }, { "epoch": 0.603448275862069, "grad_norm": 3.676309585571289, "learning_rate": 9.625e-07, "loss": 7.6557, "step": 1925 }, { "epoch": 0.6037617554858934, "grad_norm": 5.051148891448975, "learning_rate": 9.630000000000001e-07, "loss": 11.8118, "step": 1926 }, { "epoch": 0.6040752351097178, "grad_norm": 4.08391809463501, "learning_rate": 9.635000000000002e-07, "loss": 8.8092, "step": 1927 }, { "epoch": 0.6043887147335423, "grad_norm": 4.592138290405273, "learning_rate": 9.64e-07, "loss": 8.8694, "step": 1928 }, { "epoch": 0.6047021943573667, "grad_norm": 6.512213230133057, "learning_rate": 9.645e-07, "loss": 10.43, "step": 1929 }, { "epoch": 0.6050156739811913, "grad_norm": 5.282738208770752, "learning_rate": 9.65e-07, "loss": 11.6431, "step": 1930 }, { "epoch": 0.6053291536050157, "grad_norm": 3.4257993698120117, "learning_rate": 9.655000000000001e-07, "loss": 7.8864, "step": 1931 }, { "epoch": 0.6056426332288402, "grad_norm": 3.866560459136963, "learning_rate": 9.660000000000002e-07, "loss": 7.6513, "step": 1932 }, { "epoch": 0.6059561128526646, "grad_norm": 4.288784503936768, "learning_rate": 9.665e-07, "loss": 8.4351, "step": 1933 }, { "epoch": 0.606269592476489, "grad_norm": 5.792541980743408, "learning_rate": 9.67e-07, "loss": 9.6173, "step": 1934 }, { "epoch": 0.6065830721003135, "grad_norm": 2.695124626159668, "learning_rate": 9.675e-07, "loss": 6.1397, "step": 1935 }, { "epoch": 0.6068965517241379, "grad_norm": 4.266729354858398, "learning_rate": 9.68e-07, "loss": 7.2699, "step": 1936 }, { "epoch": 0.6072100313479624, "grad_norm": 3.4654881954193115, "learning_rate": 9.685000000000001e-07, "loss": 6.0948, "step": 1937 }, { "epoch": 0.6075235109717868, "grad_norm": 4.1121954917907715, "learning_rate": 9.690000000000002e-07, "loss": 8.0899, "step": 1938 }, { "epoch": 0.6078369905956112, "grad_norm": 4.078955173492432, "learning_rate": 9.695e-07, "loss": 7.1175, "step": 1939 }, { "epoch": 0.6081504702194357, "grad_norm": 7.904345512390137, "learning_rate": 9.7e-07, "loss": 14.9369, "step": 1940 }, { "epoch": 0.6084639498432602, "grad_norm": 3.5593857765197754, "learning_rate": 9.705e-07, "loss": 6.9494, "step": 1941 }, { "epoch": 0.6087774294670847, "grad_norm": 3.992682933807373, "learning_rate": 9.71e-07, "loss": 8.6618, "step": 1942 }, { "epoch": 0.6090909090909091, "grad_norm": 3.1125354766845703, "learning_rate": 9.715000000000001e-07, "loss": 5.9992, "step": 1943 }, { "epoch": 0.6094043887147336, "grad_norm": 3.825096845626831, "learning_rate": 9.72e-07, "loss": 7.6425, "step": 1944 }, { "epoch": 0.609717868338558, "grad_norm": 3.6271681785583496, "learning_rate": 9.725e-07, "loss": 7.7853, "step": 1945 }, { "epoch": 0.6100313479623825, "grad_norm": 6.750822067260742, "learning_rate": 9.73e-07, "loss": 16.4377, "step": 1946 }, { "epoch": 0.6103448275862069, "grad_norm": 3.6327602863311768, "learning_rate": 9.735e-07, "loss": 8.6631, "step": 1947 }, { "epoch": 0.6106583072100313, "grad_norm": 3.878876209259033, "learning_rate": 9.740000000000001e-07, "loss": 6.7651, "step": 1948 }, { "epoch": 0.6109717868338558, "grad_norm": 3.9936046600341797, "learning_rate": 9.745e-07, "loss": 7.4598, "step": 1949 }, { "epoch": 0.6112852664576802, "grad_norm": 3.6358838081359863, "learning_rate": 9.750000000000002e-07, "loss": 6.9454, "step": 1950 }, { "epoch": 0.6115987460815047, "grad_norm": 4.248730182647705, "learning_rate": 9.755e-07, "loss": 8.5024, "step": 1951 }, { "epoch": 0.6119122257053291, "grad_norm": 3.337886095046997, "learning_rate": 9.76e-07, "loss": 6.4199, "step": 1952 }, { "epoch": 0.6122257053291537, "grad_norm": 3.8960623741149902, "learning_rate": 9.765e-07, "loss": 7.2547, "step": 1953 }, { "epoch": 0.6125391849529781, "grad_norm": 3.7697343826293945, "learning_rate": 9.770000000000001e-07, "loss": 8.3044, "step": 1954 }, { "epoch": 0.6128526645768025, "grad_norm": 4.274561882019043, "learning_rate": 9.775000000000002e-07, "loss": 9.8826, "step": 1955 }, { "epoch": 0.613166144200627, "grad_norm": 4.1365461349487305, "learning_rate": 9.78e-07, "loss": 6.5002, "step": 1956 }, { "epoch": 0.6134796238244514, "grad_norm": 4.854091167449951, "learning_rate": 9.785000000000002e-07, "loss": 9.7826, "step": 1957 }, { "epoch": 0.6137931034482759, "grad_norm": 4.134148597717285, "learning_rate": 9.79e-07, "loss": 7.5184, "step": 1958 }, { "epoch": 0.6141065830721003, "grad_norm": 3.301070213317871, "learning_rate": 9.795000000000001e-07, "loss": 7.5854, "step": 1959 }, { "epoch": 0.6144200626959248, "grad_norm": 5.0885820388793945, "learning_rate": 9.800000000000001e-07, "loss": 12.4143, "step": 1960 }, { "epoch": 0.6147335423197492, "grad_norm": 3.778027296066284, "learning_rate": 9.805e-07, "loss": 6.6442, "step": 1961 }, { "epoch": 0.6150470219435736, "grad_norm": 4.023343086242676, "learning_rate": 9.810000000000002e-07, "loss": 9.6106, "step": 1962 }, { "epoch": 0.6153605015673981, "grad_norm": 4.735710144042969, "learning_rate": 9.815e-07, "loss": 8.8034, "step": 1963 }, { "epoch": 0.6156739811912225, "grad_norm": 4.353292942047119, "learning_rate": 9.82e-07, "loss": 9.1107, "step": 1964 }, { "epoch": 0.6159874608150471, "grad_norm": 4.989565849304199, "learning_rate": 9.825000000000001e-07, "loss": 9.7556, "step": 1965 }, { "epoch": 0.6163009404388715, "grad_norm": 3.5673139095306396, "learning_rate": 9.830000000000002e-07, "loss": 6.4914, "step": 1966 }, { "epoch": 0.616614420062696, "grad_norm": 4.158758640289307, "learning_rate": 9.835000000000002e-07, "loss": 7.9705, "step": 1967 }, { "epoch": 0.6169278996865204, "grad_norm": 7.679725170135498, "learning_rate": 9.84e-07, "loss": 9.2285, "step": 1968 }, { "epoch": 0.6172413793103448, "grad_norm": 5.259065628051758, "learning_rate": 9.845e-07, "loss": 12.6733, "step": 1969 }, { "epoch": 0.6175548589341693, "grad_norm": 4.931044578552246, "learning_rate": 9.85e-07, "loss": 9.4881, "step": 1970 }, { "epoch": 0.6178683385579937, "grad_norm": 4.3828125, "learning_rate": 9.855000000000001e-07, "loss": 8.3451, "step": 1971 }, { "epoch": 0.6181818181818182, "grad_norm": 5.446940898895264, "learning_rate": 9.86e-07, "loss": 8.3187, "step": 1972 }, { "epoch": 0.6184952978056426, "grad_norm": 5.320723533630371, "learning_rate": 9.865e-07, "loss": 9.4323, "step": 1973 }, { "epoch": 0.618808777429467, "grad_norm": 4.248854637145996, "learning_rate": 9.87e-07, "loss": 7.6606, "step": 1974 }, { "epoch": 0.6191222570532915, "grad_norm": 4.689657211303711, "learning_rate": 9.875e-07, "loss": 6.7403, "step": 1975 }, { "epoch": 0.6194357366771159, "grad_norm": 4.264988422393799, "learning_rate": 9.880000000000001e-07, "loss": 9.1217, "step": 1976 }, { "epoch": 0.6197492163009405, "grad_norm": 5.480489253997803, "learning_rate": 9.885e-07, "loss": 8.779, "step": 1977 }, { "epoch": 0.6200626959247649, "grad_norm": 4.787638187408447, "learning_rate": 9.890000000000002e-07, "loss": 8.9536, "step": 1978 }, { "epoch": 0.6203761755485894, "grad_norm": 4.072272300720215, "learning_rate": 9.895e-07, "loss": 8.668, "step": 1979 }, { "epoch": 0.6206896551724138, "grad_norm": 3.5888800621032715, "learning_rate": 9.9e-07, "loss": 7.6795, "step": 1980 }, { "epoch": 0.6210031347962383, "grad_norm": 5.570566177368164, "learning_rate": 9.905e-07, "loss": 12.1096, "step": 1981 }, { "epoch": 0.6213166144200627, "grad_norm": 4.423348426818848, "learning_rate": 9.91e-07, "loss": 9.1405, "step": 1982 }, { "epoch": 0.6216300940438871, "grad_norm": 4.6284356117248535, "learning_rate": 9.915000000000002e-07, "loss": 7.2074, "step": 1983 }, { "epoch": 0.6219435736677116, "grad_norm": 4.048377990722656, "learning_rate": 9.92e-07, "loss": 7.6424, "step": 1984 }, { "epoch": 0.622257053291536, "grad_norm": 3.3591959476470947, "learning_rate": 9.925e-07, "loss": 7.673, "step": 1985 }, { "epoch": 0.6225705329153605, "grad_norm": 3.9147262573242188, "learning_rate": 9.93e-07, "loss": 7.3318, "step": 1986 }, { "epoch": 0.6228840125391849, "grad_norm": 3.4501211643218994, "learning_rate": 9.935e-07, "loss": 7.3048, "step": 1987 }, { "epoch": 0.6231974921630095, "grad_norm": 3.9171688556671143, "learning_rate": 9.940000000000001e-07, "loss": 7.2895, "step": 1988 }, { "epoch": 0.6235109717868339, "grad_norm": 4.911311626434326, "learning_rate": 9.945e-07, "loss": 7.0245, "step": 1989 }, { "epoch": 0.6238244514106583, "grad_norm": 3.89208722114563, "learning_rate": 9.950000000000002e-07, "loss": 9.1831, "step": 1990 }, { "epoch": 0.6241379310344828, "grad_norm": 5.180747032165527, "learning_rate": 9.955e-07, "loss": 9.1415, "step": 1991 }, { "epoch": 0.6244514106583072, "grad_norm": 4.259772777557373, "learning_rate": 9.96e-07, "loss": 9.0887, "step": 1992 }, { "epoch": 0.6247648902821317, "grad_norm": 3.5109479427337646, "learning_rate": 9.965000000000001e-07, "loss": 7.4708, "step": 1993 }, { "epoch": 0.6250783699059561, "grad_norm": 3.7264599800109863, "learning_rate": 9.970000000000002e-07, "loss": 8.0124, "step": 1994 }, { "epoch": 0.6253918495297806, "grad_norm": 6.5949249267578125, "learning_rate": 9.975000000000002e-07, "loss": 12.249, "step": 1995 }, { "epoch": 0.625705329153605, "grad_norm": 5.1157450675964355, "learning_rate": 9.98e-07, "loss": 10.3741, "step": 1996 }, { "epoch": 0.6260188087774294, "grad_norm": 4.158992767333984, "learning_rate": 9.985e-07, "loss": 7.6302, "step": 1997 }, { "epoch": 0.6263322884012539, "grad_norm": 5.8107428550720215, "learning_rate": 9.99e-07, "loss": 14.3595, "step": 1998 }, { "epoch": 0.6266457680250783, "grad_norm": 5.154208660125732, "learning_rate": 9.995000000000001e-07, "loss": 8.8518, "step": 1999 }, { "epoch": 0.6269592476489029, "grad_norm": 4.65903902053833, "learning_rate": 1.0000000000000002e-06, "loss": 7.9164, "step": 2000 }, { "epoch": 0.6272727272727273, "grad_norm": 4.164371013641357, "learning_rate": 1.0005e-06, "loss": 8.2315, "step": 2001 }, { "epoch": 0.6275862068965518, "grad_norm": 4.161435604095459, "learning_rate": 1.001e-06, "loss": 9.4845, "step": 2002 }, { "epoch": 0.6278996865203762, "grad_norm": 6.261343002319336, "learning_rate": 1.0015e-06, "loss": 14.0749, "step": 2003 }, { "epoch": 0.6282131661442006, "grad_norm": 6.03980827331543, "learning_rate": 1.002e-06, "loss": 12.252, "step": 2004 }, { "epoch": 0.6285266457680251, "grad_norm": 4.8788862228393555, "learning_rate": 1.0025000000000001e-06, "loss": 8.6274, "step": 2005 }, { "epoch": 0.6288401253918495, "grad_norm": 4.340915679931641, "learning_rate": 1.0030000000000002e-06, "loss": 8.0098, "step": 2006 }, { "epoch": 0.629153605015674, "grad_norm": 4.664811611175537, "learning_rate": 1.0035e-06, "loss": 8.8534, "step": 2007 }, { "epoch": 0.6294670846394984, "grad_norm": 3.558058977127075, "learning_rate": 1.004e-06, "loss": 7.5891, "step": 2008 }, { "epoch": 0.6297805642633229, "grad_norm": 5.759692668914795, "learning_rate": 1.0045e-06, "loss": 10.9539, "step": 2009 }, { "epoch": 0.6300940438871473, "grad_norm": 3.947463035583496, "learning_rate": 1.0050000000000001e-06, "loss": 6.9227, "step": 2010 }, { "epoch": 0.6304075235109717, "grad_norm": 4.328242778778076, "learning_rate": 1.0055000000000002e-06, "loss": 8.2226, "step": 2011 }, { "epoch": 0.6307210031347963, "grad_norm": 5.028584957122803, "learning_rate": 1.006e-06, "loss": 11.2504, "step": 2012 }, { "epoch": 0.6310344827586207, "grad_norm": 5.828540325164795, "learning_rate": 1.0065e-06, "loss": 12.0494, "step": 2013 }, { "epoch": 0.6313479623824452, "grad_norm": 5.526819705963135, "learning_rate": 1.007e-06, "loss": 13.3581, "step": 2014 }, { "epoch": 0.6316614420062696, "grad_norm": 4.654907703399658, "learning_rate": 1.0075e-06, "loss": 9.1679, "step": 2015 }, { "epoch": 0.631974921630094, "grad_norm": 4.357663154602051, "learning_rate": 1.0080000000000001e-06, "loss": 8.1994, "step": 2016 }, { "epoch": 0.6322884012539185, "grad_norm": 5.075762748718262, "learning_rate": 1.0085e-06, "loss": 10.5631, "step": 2017 }, { "epoch": 0.6326018808777429, "grad_norm": 3.278012275695801, "learning_rate": 1.0090000000000002e-06, "loss": 7.5705, "step": 2018 }, { "epoch": 0.6329153605015674, "grad_norm": 3.894144058227539, "learning_rate": 1.0095e-06, "loss": 7.3649, "step": 2019 }, { "epoch": 0.6332288401253918, "grad_norm": 3.737834930419922, "learning_rate": 1.01e-06, "loss": 8.4081, "step": 2020 }, { "epoch": 0.6335423197492163, "grad_norm": 3.789785623550415, "learning_rate": 1.0105000000000001e-06, "loss": 6.3057, "step": 2021 }, { "epoch": 0.6338557993730407, "grad_norm": 3.5118370056152344, "learning_rate": 1.011e-06, "loss": 5.7457, "step": 2022 }, { "epoch": 0.6341692789968651, "grad_norm": 3.056246519088745, "learning_rate": 1.0115000000000002e-06, "loss": 5.266, "step": 2023 }, { "epoch": 0.6344827586206897, "grad_norm": 4.797995567321777, "learning_rate": 1.012e-06, "loss": 7.2787, "step": 2024 }, { "epoch": 0.6347962382445141, "grad_norm": 3.8254222869873047, "learning_rate": 1.0125e-06, "loss": 7.5975, "step": 2025 }, { "epoch": 0.6351097178683386, "grad_norm": 3.7210936546325684, "learning_rate": 1.013e-06, "loss": 7.1766, "step": 2026 }, { "epoch": 0.635423197492163, "grad_norm": 4.653789520263672, "learning_rate": 1.0135000000000001e-06, "loss": 8.7087, "step": 2027 }, { "epoch": 0.6357366771159875, "grad_norm": 5.211480617523193, "learning_rate": 1.0140000000000002e-06, "loss": 11.3491, "step": 2028 }, { "epoch": 0.6360501567398119, "grad_norm": 4.071037292480469, "learning_rate": 1.0145e-06, "loss": 9.2332, "step": 2029 }, { "epoch": 0.6363636363636364, "grad_norm": 3.9774158000946045, "learning_rate": 1.0150000000000002e-06, "loss": 7.939, "step": 2030 }, { "epoch": 0.6366771159874608, "grad_norm": 4.108770370483398, "learning_rate": 1.0155e-06, "loss": 8.5832, "step": 2031 }, { "epoch": 0.6369905956112852, "grad_norm": 3.7771849632263184, "learning_rate": 1.016e-06, "loss": 7.9219, "step": 2032 }, { "epoch": 0.6373040752351097, "grad_norm": 4.883484363555908, "learning_rate": 1.0165000000000001e-06, "loss": 7.4839, "step": 2033 }, { "epoch": 0.6376175548589341, "grad_norm": 5.540393829345703, "learning_rate": 1.0170000000000002e-06, "loss": 11.3517, "step": 2034 }, { "epoch": 0.6379310344827587, "grad_norm": 3.476276397705078, "learning_rate": 1.0175e-06, "loss": 5.4489, "step": 2035 }, { "epoch": 0.6382445141065831, "grad_norm": 4.134171485900879, "learning_rate": 1.018e-06, "loss": 9.0043, "step": 2036 }, { "epoch": 0.6385579937304076, "grad_norm": 3.8782660961151123, "learning_rate": 1.0185e-06, "loss": 7.7151, "step": 2037 }, { "epoch": 0.638871473354232, "grad_norm": 5.3437347412109375, "learning_rate": 1.0190000000000001e-06, "loss": 8.5708, "step": 2038 }, { "epoch": 0.6391849529780564, "grad_norm": 4.901749610900879, "learning_rate": 1.0195000000000001e-06, "loss": 9.5016, "step": 2039 }, { "epoch": 0.6394984326018809, "grad_norm": 9.612685203552246, "learning_rate": 1.02e-06, "loss": 25.8075, "step": 2040 }, { "epoch": 0.6398119122257053, "grad_norm": 4.421562671661377, "learning_rate": 1.0205e-06, "loss": 8.9058, "step": 2041 }, { "epoch": 0.6401253918495298, "grad_norm": 4.578634738922119, "learning_rate": 1.021e-06, "loss": 8.4515, "step": 2042 }, { "epoch": 0.6404388714733542, "grad_norm": 5.676279067993164, "learning_rate": 1.0215e-06, "loss": 12.9783, "step": 2043 }, { "epoch": 0.6407523510971787, "grad_norm": 4.57780647277832, "learning_rate": 1.0220000000000001e-06, "loss": 7.5248, "step": 2044 }, { "epoch": 0.6410658307210031, "grad_norm": 4.466324806213379, "learning_rate": 1.0225e-06, "loss": 9.4472, "step": 2045 }, { "epoch": 0.6413793103448275, "grad_norm": 4.786015510559082, "learning_rate": 1.0230000000000002e-06, "loss": 7.8918, "step": 2046 }, { "epoch": 0.6416927899686521, "grad_norm": 4.978856086730957, "learning_rate": 1.0235e-06, "loss": 8.9204, "step": 2047 }, { "epoch": 0.6420062695924765, "grad_norm": 3.4786014556884766, "learning_rate": 1.024e-06, "loss": 7.3538, "step": 2048 }, { "epoch": 0.642319749216301, "grad_norm": 6.084647178649902, "learning_rate": 1.0245e-06, "loss": 8.8253, "step": 2049 }, { "epoch": 0.6426332288401254, "grad_norm": 5.3138556480407715, "learning_rate": 1.025e-06, "loss": 11.3613, "step": 2050 }, { "epoch": 0.6429467084639499, "grad_norm": 3.573272943496704, "learning_rate": 1.0255000000000002e-06, "loss": 7.1191, "step": 2051 }, { "epoch": 0.6432601880877743, "grad_norm": 4.325623035430908, "learning_rate": 1.026e-06, "loss": 6.4057, "step": 2052 }, { "epoch": 0.6435736677115987, "grad_norm": 6.643167972564697, "learning_rate": 1.0265e-06, "loss": 16.8502, "step": 2053 }, { "epoch": 0.6438871473354232, "grad_norm": 4.033902645111084, "learning_rate": 1.027e-06, "loss": 7.7638, "step": 2054 }, { "epoch": 0.6442006269592476, "grad_norm": 3.4600517749786377, "learning_rate": 1.0275000000000001e-06, "loss": 6.4728, "step": 2055 }, { "epoch": 0.6445141065830721, "grad_norm": 5.4869818687438965, "learning_rate": 1.0280000000000002e-06, "loss": 11.0305, "step": 2056 }, { "epoch": 0.6448275862068965, "grad_norm": 2.9913864135742188, "learning_rate": 1.0285e-06, "loss": 6.2851, "step": 2057 }, { "epoch": 0.645141065830721, "grad_norm": 4.9674296379089355, "learning_rate": 1.0290000000000002e-06, "loss": 10.5091, "step": 2058 }, { "epoch": 0.6454545454545455, "grad_norm": 4.117265224456787, "learning_rate": 1.0295e-06, "loss": 7.9589, "step": 2059 }, { "epoch": 0.64576802507837, "grad_norm": 4.783974647521973, "learning_rate": 1.03e-06, "loss": 10.0756, "step": 2060 }, { "epoch": 0.6460815047021944, "grad_norm": 3.9700472354888916, "learning_rate": 1.0305000000000001e-06, "loss": 8.5293, "step": 2061 }, { "epoch": 0.6463949843260188, "grad_norm": 3.448155641555786, "learning_rate": 1.031e-06, "loss": 6.8082, "step": 2062 }, { "epoch": 0.6467084639498433, "grad_norm": 5.751345157623291, "learning_rate": 1.0315000000000002e-06, "loss": 10.7494, "step": 2063 }, { "epoch": 0.6470219435736677, "grad_norm": 4.36277961730957, "learning_rate": 1.032e-06, "loss": 7.5388, "step": 2064 }, { "epoch": 0.6473354231974922, "grad_norm": 4.9406633377075195, "learning_rate": 1.0325e-06, "loss": 8.7483, "step": 2065 }, { "epoch": 0.6476489028213166, "grad_norm": 3.610996961593628, "learning_rate": 1.033e-06, "loss": 5.993, "step": 2066 }, { "epoch": 0.647962382445141, "grad_norm": 8.09903621673584, "learning_rate": 1.0335000000000001e-06, "loss": 18.3466, "step": 2067 }, { "epoch": 0.6482758620689655, "grad_norm": 4.765588760375977, "learning_rate": 1.0340000000000002e-06, "loss": 9.1752, "step": 2068 }, { "epoch": 0.6485893416927899, "grad_norm": 5.2749199867248535, "learning_rate": 1.0345e-06, "loss": 11.2824, "step": 2069 }, { "epoch": 0.6489028213166145, "grad_norm": 5.744815826416016, "learning_rate": 1.035e-06, "loss": 12.4292, "step": 2070 }, { "epoch": 0.6492163009404389, "grad_norm": 4.47613525390625, "learning_rate": 1.0355e-06, "loss": 7.8614, "step": 2071 }, { "epoch": 0.6495297805642634, "grad_norm": 6.598586082458496, "learning_rate": 1.0360000000000001e-06, "loss": 11.9815, "step": 2072 }, { "epoch": 0.6498432601880878, "grad_norm": 5.2339606285095215, "learning_rate": 1.0365000000000002e-06, "loss": 11.3129, "step": 2073 }, { "epoch": 0.6501567398119122, "grad_norm": 4.304049015045166, "learning_rate": 1.0370000000000002e-06, "loss": 7.8178, "step": 2074 }, { "epoch": 0.6504702194357367, "grad_norm": 5.00603723526001, "learning_rate": 1.0375e-06, "loss": 10.3162, "step": 2075 }, { "epoch": 0.6507836990595611, "grad_norm": 3.639469623565674, "learning_rate": 1.038e-06, "loss": 6.6105, "step": 2076 }, { "epoch": 0.6510971786833856, "grad_norm": 4.230920314788818, "learning_rate": 1.0385e-06, "loss": 7.4984, "step": 2077 }, { "epoch": 0.65141065830721, "grad_norm": 4.557835578918457, "learning_rate": 1.0390000000000001e-06, "loss": 8.9535, "step": 2078 }, { "epoch": 0.6517241379310345, "grad_norm": 4.792398452758789, "learning_rate": 1.0395000000000002e-06, "loss": 8.1143, "step": 2079 }, { "epoch": 0.6520376175548589, "grad_norm": 4.397486209869385, "learning_rate": 1.04e-06, "loss": 9.0494, "step": 2080 }, { "epoch": 0.6523510971786833, "grad_norm": 6.484209060668945, "learning_rate": 1.0405e-06, "loss": 11.6424, "step": 2081 }, { "epoch": 0.6526645768025079, "grad_norm": 5.75565767288208, "learning_rate": 1.041e-06, "loss": 10.6868, "step": 2082 }, { "epoch": 0.6529780564263323, "grad_norm": 4.253330707550049, "learning_rate": 1.0415000000000001e-06, "loss": 8.2279, "step": 2083 }, { "epoch": 0.6532915360501568, "grad_norm": 5.125704288482666, "learning_rate": 1.0420000000000001e-06, "loss": 8.9277, "step": 2084 }, { "epoch": 0.6536050156739812, "grad_norm": 4.079887390136719, "learning_rate": 1.0425e-06, "loss": 7.9315, "step": 2085 }, { "epoch": 0.6539184952978057, "grad_norm": 4.4778361320495605, "learning_rate": 1.0430000000000002e-06, "loss": 8.606, "step": 2086 }, { "epoch": 0.6542319749216301, "grad_norm": 4.556799411773682, "learning_rate": 1.0435e-06, "loss": 6.0047, "step": 2087 }, { "epoch": 0.6545454545454545, "grad_norm": 6.23499870300293, "learning_rate": 1.044e-06, "loss": 11.1014, "step": 2088 }, { "epoch": 0.654858934169279, "grad_norm": 5.879521369934082, "learning_rate": 1.0445000000000001e-06, "loss": 11.7058, "step": 2089 }, { "epoch": 0.6551724137931034, "grad_norm": 6.175170421600342, "learning_rate": 1.045e-06, "loss": 8.7771, "step": 2090 }, { "epoch": 0.6554858934169279, "grad_norm": 4.931567192077637, "learning_rate": 1.0455000000000002e-06, "loss": 10.7524, "step": 2091 }, { "epoch": 0.6557993730407523, "grad_norm": 5.440428733825684, "learning_rate": 1.046e-06, "loss": 7.6057, "step": 2092 }, { "epoch": 0.6561128526645768, "grad_norm": 4.918017387390137, "learning_rate": 1.0465e-06, "loss": 10.069, "step": 2093 }, { "epoch": 0.6564263322884013, "grad_norm": 4.340334892272949, "learning_rate": 1.047e-06, "loss": 8.5315, "step": 2094 }, { "epoch": 0.6567398119122257, "grad_norm": 4.921428680419922, "learning_rate": 1.0475000000000001e-06, "loss": 9.7294, "step": 2095 }, { "epoch": 0.6570532915360502, "grad_norm": 6.337207794189453, "learning_rate": 1.0480000000000002e-06, "loss": 9.9782, "step": 2096 }, { "epoch": 0.6573667711598746, "grad_norm": 4.316909313201904, "learning_rate": 1.0485e-06, "loss": 6.4938, "step": 2097 }, { "epoch": 0.6576802507836991, "grad_norm": 4.156019687652588, "learning_rate": 1.049e-06, "loss": 7.8225, "step": 2098 }, { "epoch": 0.6579937304075235, "grad_norm": 3.3352489471435547, "learning_rate": 1.0495e-06, "loss": 6.4576, "step": 2099 }, { "epoch": 0.658307210031348, "grad_norm": 4.093966007232666, "learning_rate": 1.0500000000000001e-06, "loss": 7.3799, "step": 2100 }, { "epoch": 0.6586206896551724, "grad_norm": 3.7547051906585693, "learning_rate": 1.0505000000000001e-06, "loss": 8.6161, "step": 2101 }, { "epoch": 0.6589341692789968, "grad_norm": 4.8714447021484375, "learning_rate": 1.051e-06, "loss": 8.6568, "step": 2102 }, { "epoch": 0.6592476489028213, "grad_norm": 4.202094078063965, "learning_rate": 1.0515e-06, "loss": 7.711, "step": 2103 }, { "epoch": 0.6595611285266457, "grad_norm": 5.062414169311523, "learning_rate": 1.052e-06, "loss": 8.4271, "step": 2104 }, { "epoch": 0.6598746081504702, "grad_norm": 5.461309909820557, "learning_rate": 1.0525e-06, "loss": 8.2217, "step": 2105 }, { "epoch": 0.6601880877742947, "grad_norm": 4.929152965545654, "learning_rate": 1.0530000000000001e-06, "loss": 7.7316, "step": 2106 }, { "epoch": 0.6605015673981192, "grad_norm": 5.310903072357178, "learning_rate": 1.0535000000000002e-06, "loss": 9.4138, "step": 2107 }, { "epoch": 0.6608150470219436, "grad_norm": 5.781913757324219, "learning_rate": 1.054e-06, "loss": 8.4, "step": 2108 }, { "epoch": 0.661128526645768, "grad_norm": 4.272099494934082, "learning_rate": 1.0545e-06, "loss": 7.6243, "step": 2109 }, { "epoch": 0.6614420062695925, "grad_norm": 3.9875686168670654, "learning_rate": 1.055e-06, "loss": 7.8158, "step": 2110 }, { "epoch": 0.6617554858934169, "grad_norm": 7.379361629486084, "learning_rate": 1.0555e-06, "loss": 13.4704, "step": 2111 }, { "epoch": 0.6620689655172414, "grad_norm": 8.117541313171387, "learning_rate": 1.0560000000000001e-06, "loss": 16.404, "step": 2112 }, { "epoch": 0.6623824451410658, "grad_norm": 3.9060027599334717, "learning_rate": 1.0565e-06, "loss": 5.9239, "step": 2113 }, { "epoch": 0.6626959247648903, "grad_norm": 4.950360298156738, "learning_rate": 1.0570000000000002e-06, "loss": 7.5477, "step": 2114 }, { "epoch": 0.6630094043887147, "grad_norm": 6.759565353393555, "learning_rate": 1.0575e-06, "loss": 13.3141, "step": 2115 }, { "epoch": 0.6633228840125391, "grad_norm": 4.705567836761475, "learning_rate": 1.058e-06, "loss": 10.2262, "step": 2116 }, { "epoch": 0.6636363636363637, "grad_norm": 6.787057876586914, "learning_rate": 1.0585000000000001e-06, "loss": 10.6113, "step": 2117 }, { "epoch": 0.6639498432601881, "grad_norm": 4.983616828918457, "learning_rate": 1.059e-06, "loss": 10.8316, "step": 2118 }, { "epoch": 0.6642633228840126, "grad_norm": 3.8625426292419434, "learning_rate": 1.0595000000000002e-06, "loss": 7.7674, "step": 2119 }, { "epoch": 0.664576802507837, "grad_norm": 4.234352111816406, "learning_rate": 1.06e-06, "loss": 7.263, "step": 2120 }, { "epoch": 0.6648902821316615, "grad_norm": 4.677432537078857, "learning_rate": 1.0605e-06, "loss": 7.1481, "step": 2121 }, { "epoch": 0.6652037617554859, "grad_norm": 4.3010029792785645, "learning_rate": 1.061e-06, "loss": 6.9077, "step": 2122 }, { "epoch": 0.6655172413793103, "grad_norm": 3.675344228744507, "learning_rate": 1.0615000000000001e-06, "loss": 6.9315, "step": 2123 }, { "epoch": 0.6658307210031348, "grad_norm": 5.514710903167725, "learning_rate": 1.0620000000000002e-06, "loss": 11.0089, "step": 2124 }, { "epoch": 0.6661442006269592, "grad_norm": 4.249388694763184, "learning_rate": 1.0625e-06, "loss": 8.537, "step": 2125 }, { "epoch": 0.6664576802507837, "grad_norm": 4.913691997528076, "learning_rate": 1.0630000000000002e-06, "loss": 8.3072, "step": 2126 }, { "epoch": 0.6667711598746081, "grad_norm": 7.471773624420166, "learning_rate": 1.0635e-06, "loss": 9.3104, "step": 2127 }, { "epoch": 0.6670846394984326, "grad_norm": 5.314687728881836, "learning_rate": 1.064e-06, "loss": 9.9356, "step": 2128 }, { "epoch": 0.6673981191222571, "grad_norm": 5.965895175933838, "learning_rate": 1.0645000000000001e-06, "loss": 12.0117, "step": 2129 }, { "epoch": 0.6677115987460815, "grad_norm": 4.372379779815674, "learning_rate": 1.065e-06, "loss": 7.607, "step": 2130 }, { "epoch": 0.668025078369906, "grad_norm": 4.424569606781006, "learning_rate": 1.0655000000000002e-06, "loss": 10.169, "step": 2131 }, { "epoch": 0.6683385579937304, "grad_norm": 3.1224844455718994, "learning_rate": 1.066e-06, "loss": 6.0792, "step": 2132 }, { "epoch": 0.6686520376175549, "grad_norm": 5.132580280303955, "learning_rate": 1.0665e-06, "loss": 9.2276, "step": 2133 }, { "epoch": 0.6689655172413793, "grad_norm": 4.87095832824707, "learning_rate": 1.0670000000000001e-06, "loss": 7.84, "step": 2134 }, { "epoch": 0.6692789968652038, "grad_norm": 5.659262657165527, "learning_rate": 1.0675000000000002e-06, "loss": 8.4573, "step": 2135 }, { "epoch": 0.6695924764890282, "grad_norm": 4.755374908447266, "learning_rate": 1.0680000000000002e-06, "loss": 10.6835, "step": 2136 }, { "epoch": 0.6699059561128526, "grad_norm": 5.089763641357422, "learning_rate": 1.0685e-06, "loss": 7.9605, "step": 2137 }, { "epoch": 0.6702194357366771, "grad_norm": 8.315079689025879, "learning_rate": 1.069e-06, "loss": 11.3606, "step": 2138 }, { "epoch": 0.6705329153605015, "grad_norm": 5.036296367645264, "learning_rate": 1.0695e-06, "loss": 9.4513, "step": 2139 }, { "epoch": 0.670846394984326, "grad_norm": 4.769467353820801, "learning_rate": 1.0700000000000001e-06, "loss": 7.5086, "step": 2140 }, { "epoch": 0.6711598746081505, "grad_norm": 5.184993743896484, "learning_rate": 1.0705000000000002e-06, "loss": 10.644, "step": 2141 }, { "epoch": 0.671473354231975, "grad_norm": 3.7264373302459717, "learning_rate": 1.071e-06, "loss": 6.8944, "step": 2142 }, { "epoch": 0.6717868338557994, "grad_norm": 4.51856803894043, "learning_rate": 1.0715e-06, "loss": 7.1033, "step": 2143 }, { "epoch": 0.6721003134796238, "grad_norm": 4.620069980621338, "learning_rate": 1.072e-06, "loss": 9.0823, "step": 2144 }, { "epoch": 0.6724137931034483, "grad_norm": 4.350764274597168, "learning_rate": 1.0725000000000001e-06, "loss": 7.9142, "step": 2145 }, { "epoch": 0.6727272727272727, "grad_norm": 5.366765022277832, "learning_rate": 1.0730000000000001e-06, "loss": 7.2714, "step": 2146 }, { "epoch": 0.6730407523510972, "grad_norm": 4.46659517288208, "learning_rate": 1.0735000000000002e-06, "loss": 8.6368, "step": 2147 }, { "epoch": 0.6733542319749216, "grad_norm": 4.17587947845459, "learning_rate": 1.074e-06, "loss": 6.8412, "step": 2148 }, { "epoch": 0.673667711598746, "grad_norm": 4.389501571655273, "learning_rate": 1.0745e-06, "loss": 5.929, "step": 2149 }, { "epoch": 0.6739811912225705, "grad_norm": 3.766752004623413, "learning_rate": 1.075e-06, "loss": 7.428, "step": 2150 }, { "epoch": 0.6742946708463949, "grad_norm": 4.460997581481934, "learning_rate": 1.0755000000000001e-06, "loss": 8.5694, "step": 2151 }, { "epoch": 0.6746081504702194, "grad_norm": 3.7466864585876465, "learning_rate": 1.0760000000000002e-06, "loss": 6.5698, "step": 2152 }, { "epoch": 0.6749216300940439, "grad_norm": 5.410111904144287, "learning_rate": 1.0765e-06, "loss": 13.9963, "step": 2153 }, { "epoch": 0.6752351097178684, "grad_norm": 6.007702827453613, "learning_rate": 1.0770000000000002e-06, "loss": 11.1922, "step": 2154 }, { "epoch": 0.6755485893416928, "grad_norm": 3.324204206466675, "learning_rate": 1.0775e-06, "loss": 7.0282, "step": 2155 }, { "epoch": 0.6758620689655173, "grad_norm": 6.3619914054870605, "learning_rate": 1.078e-06, "loss": 8.3813, "step": 2156 }, { "epoch": 0.6761755485893417, "grad_norm": 5.453836441040039, "learning_rate": 1.0785000000000001e-06, "loss": 8.7658, "step": 2157 }, { "epoch": 0.6764890282131661, "grad_norm": 4.859471321105957, "learning_rate": 1.079e-06, "loss": 7.7485, "step": 2158 }, { "epoch": 0.6768025078369906, "grad_norm": 5.14132022857666, "learning_rate": 1.0795000000000002e-06, "loss": 10.0406, "step": 2159 }, { "epoch": 0.677115987460815, "grad_norm": 4.963798522949219, "learning_rate": 1.08e-06, "loss": 9.1328, "step": 2160 }, { "epoch": 0.6774294670846395, "grad_norm": 5.451442718505859, "learning_rate": 1.0805e-06, "loss": 7.1831, "step": 2161 }, { "epoch": 0.6777429467084639, "grad_norm": 4.889834403991699, "learning_rate": 1.0810000000000001e-06, "loss": 8.8367, "step": 2162 }, { "epoch": 0.6780564263322884, "grad_norm": 3.9947681427001953, "learning_rate": 1.0815000000000001e-06, "loss": 6.2821, "step": 2163 }, { "epoch": 0.6783699059561129, "grad_norm": 6.472747325897217, "learning_rate": 1.0820000000000002e-06, "loss": 12.1777, "step": 2164 }, { "epoch": 0.6786833855799373, "grad_norm": 3.9725639820098877, "learning_rate": 1.0825e-06, "loss": 7.2472, "step": 2165 }, { "epoch": 0.6789968652037618, "grad_norm": 4.466004371643066, "learning_rate": 1.083e-06, "loss": 7.1208, "step": 2166 }, { "epoch": 0.6793103448275862, "grad_norm": 8.549521446228027, "learning_rate": 1.0835e-06, "loss": 14.6298, "step": 2167 }, { "epoch": 0.6796238244514107, "grad_norm": 3.372896671295166, "learning_rate": 1.0840000000000001e-06, "loss": 6.0872, "step": 2168 }, { "epoch": 0.6799373040752351, "grad_norm": 4.142289161682129, "learning_rate": 1.0845000000000002e-06, "loss": 6.0638, "step": 2169 }, { "epoch": 0.6802507836990596, "grad_norm": 4.548405170440674, "learning_rate": 1.085e-06, "loss": 9.5781, "step": 2170 }, { "epoch": 0.680564263322884, "grad_norm": 4.637272357940674, "learning_rate": 1.0855e-06, "loss": 6.9768, "step": 2171 }, { "epoch": 0.6808777429467084, "grad_norm": 4.340859889984131, "learning_rate": 1.086e-06, "loss": 6.6182, "step": 2172 }, { "epoch": 0.6811912225705329, "grad_norm": 4.831057548522949, "learning_rate": 1.0865e-06, "loss": 8.176, "step": 2173 }, { "epoch": 0.6815047021943573, "grad_norm": 4.174441337585449, "learning_rate": 1.0870000000000001e-06, "loss": 9.2991, "step": 2174 }, { "epoch": 0.6818181818181818, "grad_norm": 4.800502300262451, "learning_rate": 1.0875000000000002e-06, "loss": 8.0076, "step": 2175 }, { "epoch": 0.6821316614420063, "grad_norm": 5.333560466766357, "learning_rate": 1.088e-06, "loss": 8.8977, "step": 2176 }, { "epoch": 0.6824451410658308, "grad_norm": 4.281090259552002, "learning_rate": 1.0885e-06, "loss": 6.8888, "step": 2177 }, { "epoch": 0.6827586206896552, "grad_norm": 7.420187950134277, "learning_rate": 1.089e-06, "loss": 12.2294, "step": 2178 }, { "epoch": 0.6830721003134796, "grad_norm": 5.143314838409424, "learning_rate": 1.0895000000000001e-06, "loss": 8.4805, "step": 2179 }, { "epoch": 0.6833855799373041, "grad_norm": 5.837392807006836, "learning_rate": 1.0900000000000002e-06, "loss": 9.5465, "step": 2180 }, { "epoch": 0.6836990595611285, "grad_norm": 4.490211009979248, "learning_rate": 1.0905e-06, "loss": 7.1872, "step": 2181 }, { "epoch": 0.684012539184953, "grad_norm": 5.512997150421143, "learning_rate": 1.091e-06, "loss": 10.6141, "step": 2182 }, { "epoch": 0.6843260188087774, "grad_norm": 5.654632568359375, "learning_rate": 1.0915e-06, "loss": 10.529, "step": 2183 }, { "epoch": 0.6846394984326019, "grad_norm": 3.9920573234558105, "learning_rate": 1.092e-06, "loss": 7.4162, "step": 2184 }, { "epoch": 0.6849529780564263, "grad_norm": 9.621570587158203, "learning_rate": 1.0925000000000001e-06, "loss": 17.1348, "step": 2185 }, { "epoch": 0.6852664576802507, "grad_norm": 4.3478007316589355, "learning_rate": 1.093e-06, "loss": 7.5085, "step": 2186 }, { "epoch": 0.6855799373040752, "grad_norm": 4.477773666381836, "learning_rate": 1.0935000000000002e-06, "loss": 8.6712, "step": 2187 }, { "epoch": 0.6858934169278997, "grad_norm": 4.364738941192627, "learning_rate": 1.094e-06, "loss": 8.0755, "step": 2188 }, { "epoch": 0.6862068965517242, "grad_norm": 5.307512283325195, "learning_rate": 1.0945e-06, "loss": 9.8482, "step": 2189 }, { "epoch": 0.6865203761755486, "grad_norm": 5.022215366363525, "learning_rate": 1.095e-06, "loss": 9.1433, "step": 2190 }, { "epoch": 0.6868338557993731, "grad_norm": 5.021544933319092, "learning_rate": 1.0955e-06, "loss": 8.4613, "step": 2191 }, { "epoch": 0.6871473354231975, "grad_norm": 3.7242307662963867, "learning_rate": 1.0960000000000002e-06, "loss": 6.4761, "step": 2192 }, { "epoch": 0.687460815047022, "grad_norm": 4.364880561828613, "learning_rate": 1.0965e-06, "loss": 8.0514, "step": 2193 }, { "epoch": 0.6877742946708464, "grad_norm": 6.270079612731934, "learning_rate": 1.097e-06, "loss": 7.7713, "step": 2194 }, { "epoch": 0.6880877742946708, "grad_norm": 4.493875026702881, "learning_rate": 1.0975e-06, "loss": 7.1808, "step": 2195 }, { "epoch": 0.6884012539184953, "grad_norm": 5.605881214141846, "learning_rate": 1.0980000000000001e-06, "loss": 8.8931, "step": 2196 }, { "epoch": 0.6887147335423197, "grad_norm": 5.781085014343262, "learning_rate": 1.0985000000000002e-06, "loss": 11.2424, "step": 2197 }, { "epoch": 0.6890282131661442, "grad_norm": 6.997844219207764, "learning_rate": 1.099e-06, "loss": 14.7351, "step": 2198 }, { "epoch": 0.6893416927899687, "grad_norm": 5.427860260009766, "learning_rate": 1.0995000000000002e-06, "loss": 8.0223, "step": 2199 }, { "epoch": 0.6896551724137931, "grad_norm": 5.856205940246582, "learning_rate": 1.1e-06, "loss": 10.776, "step": 2200 }, { "epoch": 0.6899686520376176, "grad_norm": 5.698307037353516, "learning_rate": 1.1005e-06, "loss": 10.4188, "step": 2201 }, { "epoch": 0.690282131661442, "grad_norm": 3.5812063217163086, "learning_rate": 1.1010000000000001e-06, "loss": 5.7494, "step": 2202 }, { "epoch": 0.6905956112852665, "grad_norm": 6.817094802856445, "learning_rate": 1.1015000000000002e-06, "loss": 15.1615, "step": 2203 }, { "epoch": 0.6909090909090909, "grad_norm": 5.15540075302124, "learning_rate": 1.1020000000000002e-06, "loss": 7.7424, "step": 2204 }, { "epoch": 0.6912225705329154, "grad_norm": 5.062628746032715, "learning_rate": 1.1025e-06, "loss": 6.6624, "step": 2205 }, { "epoch": 0.6915360501567398, "grad_norm": 4.202182769775391, "learning_rate": 1.103e-06, "loss": 7.4259, "step": 2206 }, { "epoch": 0.6918495297805642, "grad_norm": 4.714681148529053, "learning_rate": 1.1035000000000001e-06, "loss": 9.4869, "step": 2207 }, { "epoch": 0.6921630094043887, "grad_norm": 5.397708415985107, "learning_rate": 1.1040000000000001e-06, "loss": 8.4847, "step": 2208 }, { "epoch": 0.6924764890282131, "grad_norm": 6.186098098754883, "learning_rate": 1.1045000000000002e-06, "loss": 9.6202, "step": 2209 }, { "epoch": 0.6927899686520376, "grad_norm": 5.049215793609619, "learning_rate": 1.105e-06, "loss": 8.4358, "step": 2210 }, { "epoch": 0.6931034482758621, "grad_norm": 4.837051868438721, "learning_rate": 1.1055e-06, "loss": 7.3944, "step": 2211 }, { "epoch": 0.6934169278996866, "grad_norm": 5.781193256378174, "learning_rate": 1.106e-06, "loss": 10.1386, "step": 2212 }, { "epoch": 0.693730407523511, "grad_norm": 4.360887050628662, "learning_rate": 1.1065000000000001e-06, "loss": 7.0723, "step": 2213 }, { "epoch": 0.6940438871473354, "grad_norm": 4.668460845947266, "learning_rate": 1.1070000000000002e-06, "loss": 8.563, "step": 2214 }, { "epoch": 0.6943573667711599, "grad_norm": 5.356121063232422, "learning_rate": 1.1075000000000002e-06, "loss": 9.5143, "step": 2215 }, { "epoch": 0.6946708463949843, "grad_norm": 4.979715824127197, "learning_rate": 1.108e-06, "loss": 8.1418, "step": 2216 }, { "epoch": 0.6949843260188088, "grad_norm": 4.595643520355225, "learning_rate": 1.1085e-06, "loss": 9.48, "step": 2217 }, { "epoch": 0.6952978056426332, "grad_norm": 5.082688331604004, "learning_rate": 1.109e-06, "loss": 7.4409, "step": 2218 }, { "epoch": 0.6956112852664577, "grad_norm": 5.961426258087158, "learning_rate": 1.1095e-06, "loss": 9.6929, "step": 2219 }, { "epoch": 0.6959247648902821, "grad_norm": 4.507264137268066, "learning_rate": 1.1100000000000002e-06, "loss": 9.0472, "step": 2220 }, { "epoch": 0.6962382445141065, "grad_norm": 6.450530052185059, "learning_rate": 1.1105e-06, "loss": 11.311, "step": 2221 }, { "epoch": 0.696551724137931, "grad_norm": 6.884239673614502, "learning_rate": 1.111e-06, "loss": 9.9046, "step": 2222 }, { "epoch": 0.6968652037617555, "grad_norm": 4.918006420135498, "learning_rate": 1.1115e-06, "loss": 7.7746, "step": 2223 }, { "epoch": 0.69717868338558, "grad_norm": 3.9736998081207275, "learning_rate": 1.1120000000000001e-06, "loss": 8.2501, "step": 2224 }, { "epoch": 0.6974921630094044, "grad_norm": 5.021994590759277, "learning_rate": 1.1125000000000001e-06, "loss": 9.8031, "step": 2225 }, { "epoch": 0.6978056426332289, "grad_norm": 4.272434711456299, "learning_rate": 1.113e-06, "loss": 6.6057, "step": 2226 }, { "epoch": 0.6981191222570533, "grad_norm": 4.612613677978516, "learning_rate": 1.1135000000000002e-06, "loss": 8.5593, "step": 2227 }, { "epoch": 0.6984326018808777, "grad_norm": 6.331014633178711, "learning_rate": 1.114e-06, "loss": 8.7987, "step": 2228 }, { "epoch": 0.6987460815047022, "grad_norm": 6.129283905029297, "learning_rate": 1.1145e-06, "loss": 8.4672, "step": 2229 }, { "epoch": 0.6990595611285266, "grad_norm": 4.61047887802124, "learning_rate": 1.1150000000000001e-06, "loss": 9.5344, "step": 2230 }, { "epoch": 0.6993730407523511, "grad_norm": 4.625089168548584, "learning_rate": 1.1155e-06, "loss": 8.4914, "step": 2231 }, { "epoch": 0.6996865203761755, "grad_norm": 7.176419258117676, "learning_rate": 1.1160000000000002e-06, "loss": 10.456, "step": 2232 }, { "epoch": 0.7, "grad_norm": 4.757896900177002, "learning_rate": 1.1165e-06, "loss": 8.0939, "step": 2233 }, { "epoch": 0.7003134796238244, "grad_norm": 4.703888893127441, "learning_rate": 1.117e-06, "loss": 8.9324, "step": 2234 }, { "epoch": 0.700626959247649, "grad_norm": 5.082801342010498, "learning_rate": 1.1175e-06, "loss": 7.9512, "step": 2235 }, { "epoch": 0.7009404388714734, "grad_norm": 7.074932098388672, "learning_rate": 1.1180000000000001e-06, "loss": 10.5051, "step": 2236 }, { "epoch": 0.7012539184952978, "grad_norm": 5.920054912567139, "learning_rate": 1.1185000000000002e-06, "loss": 10.0657, "step": 2237 }, { "epoch": 0.7015673981191223, "grad_norm": 4.017765045166016, "learning_rate": 1.119e-06, "loss": 7.4091, "step": 2238 }, { "epoch": 0.7018808777429467, "grad_norm": 4.405153274536133, "learning_rate": 1.1195e-06, "loss": 8.9839, "step": 2239 }, { "epoch": 0.7021943573667712, "grad_norm": 5.551645755767822, "learning_rate": 1.12e-06, "loss": 8.2674, "step": 2240 }, { "epoch": 0.7025078369905956, "grad_norm": 5.920106887817383, "learning_rate": 1.1205000000000001e-06, "loss": 10.3074, "step": 2241 }, { "epoch": 0.70282131661442, "grad_norm": 6.09196138381958, "learning_rate": 1.1210000000000002e-06, "loss": 9.587, "step": 2242 }, { "epoch": 0.7031347962382445, "grad_norm": 6.698037624359131, "learning_rate": 1.1215000000000002e-06, "loss": 10.1409, "step": 2243 }, { "epoch": 0.7034482758620689, "grad_norm": 6.3669962882995605, "learning_rate": 1.122e-06, "loss": 10.6705, "step": 2244 }, { "epoch": 0.7037617554858934, "grad_norm": 4.820910930633545, "learning_rate": 1.1225e-06, "loss": 10.7634, "step": 2245 }, { "epoch": 0.7040752351097179, "grad_norm": 6.152073860168457, "learning_rate": 1.123e-06, "loss": 9.9762, "step": 2246 }, { "epoch": 0.7043887147335424, "grad_norm": 6.534127235412598, "learning_rate": 1.1235000000000001e-06, "loss": 9.1076, "step": 2247 }, { "epoch": 0.7047021943573668, "grad_norm": 4.584049224853516, "learning_rate": 1.1240000000000002e-06, "loss": 6.3946, "step": 2248 }, { "epoch": 0.7050156739811912, "grad_norm": 6.2237653732299805, "learning_rate": 1.1245e-06, "loss": 10.3125, "step": 2249 }, { "epoch": 0.7053291536050157, "grad_norm": 7.915231227874756, "learning_rate": 1.125e-06, "loss": 11.7329, "step": 2250 }, { "epoch": 0.7056426332288401, "grad_norm": 5.413792610168457, "learning_rate": 1.1255e-06, "loss": 7.3437, "step": 2251 }, { "epoch": 0.7059561128526646, "grad_norm": 6.203031063079834, "learning_rate": 1.126e-06, "loss": 8.8131, "step": 2252 }, { "epoch": 0.706269592476489, "grad_norm": 5.079891681671143, "learning_rate": 1.1265000000000001e-06, "loss": 8.0062, "step": 2253 }, { "epoch": 0.7065830721003135, "grad_norm": 5.417928218841553, "learning_rate": 1.127e-06, "loss": 7.8545, "step": 2254 }, { "epoch": 0.7068965517241379, "grad_norm": 8.853837966918945, "learning_rate": 1.1275000000000002e-06, "loss": 12.7805, "step": 2255 }, { "epoch": 0.7072100313479623, "grad_norm": 5.80386209487915, "learning_rate": 1.128e-06, "loss": 10.7414, "step": 2256 }, { "epoch": 0.7075235109717868, "grad_norm": 5.670801639556885, "learning_rate": 1.1285e-06, "loss": 8.2194, "step": 2257 }, { "epoch": 0.7078369905956113, "grad_norm": 3.81965708732605, "learning_rate": 1.1290000000000001e-06, "loss": 6.5429, "step": 2258 }, { "epoch": 0.7081504702194358, "grad_norm": 5.524906635284424, "learning_rate": 1.1295e-06, "loss": 10.1136, "step": 2259 }, { "epoch": 0.7084639498432602, "grad_norm": 7.865563869476318, "learning_rate": 1.1300000000000002e-06, "loss": 11.7016, "step": 2260 }, { "epoch": 0.7087774294670847, "grad_norm": 4.466341495513916, "learning_rate": 1.1305e-06, "loss": 6.4666, "step": 2261 }, { "epoch": 0.7090909090909091, "grad_norm": 4.703956127166748, "learning_rate": 1.131e-06, "loss": 7.191, "step": 2262 }, { "epoch": 0.7094043887147335, "grad_norm": 4.839312553405762, "learning_rate": 1.1315e-06, "loss": 8.4636, "step": 2263 }, { "epoch": 0.709717868338558, "grad_norm": 6.889242172241211, "learning_rate": 1.1320000000000001e-06, "loss": 9.0166, "step": 2264 }, { "epoch": 0.7100313479623824, "grad_norm": 5.956830978393555, "learning_rate": 1.1325000000000002e-06, "loss": 7.6537, "step": 2265 }, { "epoch": 0.7103448275862069, "grad_norm": 3.948118209838867, "learning_rate": 1.133e-06, "loss": 6.8683, "step": 2266 }, { "epoch": 0.7106583072100313, "grad_norm": 5.981534004211426, "learning_rate": 1.1335000000000002e-06, "loss": 9.1839, "step": 2267 }, { "epoch": 0.7109717868338558, "grad_norm": 6.075943470001221, "learning_rate": 1.134e-06, "loss": 8.9451, "step": 2268 }, { "epoch": 0.7112852664576802, "grad_norm": 7.1982879638671875, "learning_rate": 1.1345000000000001e-06, "loss": 10.2837, "step": 2269 }, { "epoch": 0.7115987460815048, "grad_norm": 4.65557861328125, "learning_rate": 1.1350000000000001e-06, "loss": 7.0759, "step": 2270 }, { "epoch": 0.7119122257053292, "grad_norm": 7.2853007316589355, "learning_rate": 1.1355e-06, "loss": 10.6811, "step": 2271 }, { "epoch": 0.7122257053291536, "grad_norm": 4.625248432159424, "learning_rate": 1.1360000000000002e-06, "loss": 6.6552, "step": 2272 }, { "epoch": 0.7125391849529781, "grad_norm": 5.515805244445801, "learning_rate": 1.1365e-06, "loss": 11.0095, "step": 2273 }, { "epoch": 0.7128526645768025, "grad_norm": 8.214932441711426, "learning_rate": 1.137e-06, "loss": 9.5016, "step": 2274 }, { "epoch": 0.713166144200627, "grad_norm": 4.511641979217529, "learning_rate": 1.1375000000000001e-06, "loss": 9.456, "step": 2275 }, { "epoch": 0.7134796238244514, "grad_norm": 4.926652908325195, "learning_rate": 1.1380000000000002e-06, "loss": 6.4554, "step": 2276 }, { "epoch": 0.7137931034482758, "grad_norm": 5.701582908630371, "learning_rate": 1.1385000000000002e-06, "loss": 7.8086, "step": 2277 }, { "epoch": 0.7141065830721003, "grad_norm": 5.361356735229492, "learning_rate": 1.139e-06, "loss": 5.5772, "step": 2278 }, { "epoch": 0.7144200626959247, "grad_norm": 5.268955707550049, "learning_rate": 1.1395e-06, "loss": 6.9328, "step": 2279 }, { "epoch": 0.7147335423197492, "grad_norm": 6.054418087005615, "learning_rate": 1.14e-06, "loss": 11.8742, "step": 2280 }, { "epoch": 0.7150470219435736, "grad_norm": 6.118603229522705, "learning_rate": 1.1405000000000001e-06, "loss": 8.7145, "step": 2281 }, { "epoch": 0.7153605015673982, "grad_norm": 5.574398517608643, "learning_rate": 1.141e-06, "loss": 8.354, "step": 2282 }, { "epoch": 0.7156739811912226, "grad_norm": 8.102057456970215, "learning_rate": 1.1415000000000002e-06, "loss": 9.6707, "step": 2283 }, { "epoch": 0.715987460815047, "grad_norm": 6.044636249542236, "learning_rate": 1.142e-06, "loss": 8.4545, "step": 2284 }, { "epoch": 0.7163009404388715, "grad_norm": 5.610233783721924, "learning_rate": 1.1425e-06, "loss": 6.494, "step": 2285 }, { "epoch": 0.7166144200626959, "grad_norm": 7.252960681915283, "learning_rate": 1.1430000000000001e-06, "loss": 9.9181, "step": 2286 }, { "epoch": 0.7169278996865204, "grad_norm": 7.041637897491455, "learning_rate": 1.1435e-06, "loss": 9.8605, "step": 2287 }, { "epoch": 0.7172413793103448, "grad_norm": 7.529675006866455, "learning_rate": 1.1440000000000002e-06, "loss": 9.2948, "step": 2288 }, { "epoch": 0.7175548589341693, "grad_norm": 6.867233753204346, "learning_rate": 1.1445e-06, "loss": 10.663, "step": 2289 }, { "epoch": 0.7178683385579937, "grad_norm": 4.269197463989258, "learning_rate": 1.145e-06, "loss": 6.5775, "step": 2290 }, { "epoch": 0.7181818181818181, "grad_norm": 5.0431060791015625, "learning_rate": 1.1455e-06, "loss": 8.0019, "step": 2291 }, { "epoch": 0.7184952978056426, "grad_norm": 5.916482925415039, "learning_rate": 1.1460000000000001e-06, "loss": 9.4506, "step": 2292 }, { "epoch": 0.7188087774294671, "grad_norm": 8.643074035644531, "learning_rate": 1.1465000000000002e-06, "loss": 11.3396, "step": 2293 }, { "epoch": 0.7191222570532916, "grad_norm": 4.569228649139404, "learning_rate": 1.147e-06, "loss": 6.5566, "step": 2294 }, { "epoch": 0.719435736677116, "grad_norm": 5.326075077056885, "learning_rate": 1.1475000000000002e-06, "loss": 8.4556, "step": 2295 }, { "epoch": 0.7197492163009405, "grad_norm": 5.192141532897949, "learning_rate": 1.148e-06, "loss": 8.5234, "step": 2296 }, { "epoch": 0.7200626959247649, "grad_norm": 4.826154708862305, "learning_rate": 1.1485e-06, "loss": 6.925, "step": 2297 }, { "epoch": 0.7203761755485893, "grad_norm": 6.822600841522217, "learning_rate": 1.1490000000000001e-06, "loss": 12.9681, "step": 2298 }, { "epoch": 0.7206896551724138, "grad_norm": 7.6909003257751465, "learning_rate": 1.1495e-06, "loss": 10.0613, "step": 2299 }, { "epoch": 0.7210031347962382, "grad_norm": 6.721874713897705, "learning_rate": 1.1500000000000002e-06, "loss": 11.6405, "step": 2300 }, { "epoch": 0.7213166144200627, "grad_norm": 5.8059163093566895, "learning_rate": 1.1505e-06, "loss": 8.9701, "step": 2301 }, { "epoch": 0.7216300940438871, "grad_norm": 8.074931144714355, "learning_rate": 1.151e-06, "loss": 9.3916, "step": 2302 }, { "epoch": 0.7219435736677116, "grad_norm": 7.6194987297058105, "learning_rate": 1.1515000000000001e-06, "loss": 11.8245, "step": 2303 }, { "epoch": 0.722257053291536, "grad_norm": 4.169922351837158, "learning_rate": 1.1520000000000002e-06, "loss": 6.3245, "step": 2304 }, { "epoch": 0.7225705329153606, "grad_norm": 7.672272682189941, "learning_rate": 1.1525000000000002e-06, "loss": 10.7136, "step": 2305 }, { "epoch": 0.722884012539185, "grad_norm": 6.1154656410217285, "learning_rate": 1.153e-06, "loss": 10.13, "step": 2306 }, { "epoch": 0.7231974921630094, "grad_norm": 5.131744861602783, "learning_rate": 1.1535e-06, "loss": 6.4488, "step": 2307 }, { "epoch": 0.7235109717868339, "grad_norm": 4.942484378814697, "learning_rate": 1.154e-06, "loss": 7.9078, "step": 2308 }, { "epoch": 0.7238244514106583, "grad_norm": 6.5893449783325195, "learning_rate": 1.1545000000000001e-06, "loss": 10.2193, "step": 2309 }, { "epoch": 0.7241379310344828, "grad_norm": 5.4219136238098145, "learning_rate": 1.1550000000000002e-06, "loss": 7.0708, "step": 2310 }, { "epoch": 0.7244514106583072, "grad_norm": 8.25130558013916, "learning_rate": 1.1555e-06, "loss": 10.7496, "step": 2311 }, { "epoch": 0.7247648902821316, "grad_norm": 4.836966514587402, "learning_rate": 1.156e-06, "loss": 6.6638, "step": 2312 }, { "epoch": 0.7250783699059561, "grad_norm": 7.104246139526367, "learning_rate": 1.1565e-06, "loss": 8.5781, "step": 2313 }, { "epoch": 0.7253918495297805, "grad_norm": 5.00009822845459, "learning_rate": 1.157e-06, "loss": 6.4871, "step": 2314 }, { "epoch": 0.725705329153605, "grad_norm": 4.757653713226318, "learning_rate": 1.1575000000000001e-06, "loss": 7.6189, "step": 2315 }, { "epoch": 0.7260188087774294, "grad_norm": 6.125999450683594, "learning_rate": 1.1580000000000002e-06, "loss": 7.1317, "step": 2316 }, { "epoch": 0.726332288401254, "grad_norm": 6.397932052612305, "learning_rate": 1.1585e-06, "loss": 7.738, "step": 2317 }, { "epoch": 0.7266457680250784, "grad_norm": 5.191574573516846, "learning_rate": 1.159e-06, "loss": 7.8031, "step": 2318 }, { "epoch": 0.7269592476489029, "grad_norm": 6.020807266235352, "learning_rate": 1.1595e-06, "loss": 9.9713, "step": 2319 }, { "epoch": 0.7272727272727273, "grad_norm": 4.571470260620117, "learning_rate": 1.1600000000000001e-06, "loss": 5.8561, "step": 2320 }, { "epoch": 0.7275862068965517, "grad_norm": 6.151414394378662, "learning_rate": 1.1605000000000002e-06, "loss": 8.3401, "step": 2321 }, { "epoch": 0.7278996865203762, "grad_norm": 5.342029094696045, "learning_rate": 1.161e-06, "loss": 7.4286, "step": 2322 }, { "epoch": 0.7282131661442006, "grad_norm": 8.054902076721191, "learning_rate": 1.1615000000000002e-06, "loss": 11.2857, "step": 2323 }, { "epoch": 0.7285266457680251, "grad_norm": 8.631732940673828, "learning_rate": 1.162e-06, "loss": 11.7461, "step": 2324 }, { "epoch": 0.7288401253918495, "grad_norm": 5.86944580078125, "learning_rate": 1.1625e-06, "loss": 8.2415, "step": 2325 }, { "epoch": 0.729153605015674, "grad_norm": 6.775248050689697, "learning_rate": 1.1630000000000001e-06, "loss": 9.1465, "step": 2326 }, { "epoch": 0.7294670846394984, "grad_norm": 5.014388561248779, "learning_rate": 1.1635e-06, "loss": 9.4963, "step": 2327 }, { "epoch": 0.7297805642633229, "grad_norm": 6.137758731842041, "learning_rate": 1.1640000000000002e-06, "loss": 7.47, "step": 2328 }, { "epoch": 0.7300940438871474, "grad_norm": 7.857049942016602, "learning_rate": 1.1645e-06, "loss": 12.0102, "step": 2329 }, { "epoch": 0.7304075235109718, "grad_norm": 5.5146002769470215, "learning_rate": 1.165e-06, "loss": 7.7632, "step": 2330 }, { "epoch": 0.7307210031347963, "grad_norm": 5.327179431915283, "learning_rate": 1.1655000000000001e-06, "loss": 9.2846, "step": 2331 }, { "epoch": 0.7310344827586207, "grad_norm": 5.694344997406006, "learning_rate": 1.1660000000000001e-06, "loss": 8.7609, "step": 2332 }, { "epoch": 0.7313479623824451, "grad_norm": 6.549810886383057, "learning_rate": 1.1665000000000002e-06, "loss": 8.8996, "step": 2333 }, { "epoch": 0.7316614420062696, "grad_norm": 4.765985488891602, "learning_rate": 1.167e-06, "loss": 7.4456, "step": 2334 }, { "epoch": 0.731974921630094, "grad_norm": 4.749044418334961, "learning_rate": 1.1675000000000003e-06, "loss": 6.8727, "step": 2335 }, { "epoch": 0.7322884012539185, "grad_norm": 6.0622711181640625, "learning_rate": 1.168e-06, "loss": 8.9842, "step": 2336 }, { "epoch": 0.7326018808777429, "grad_norm": 5.587085723876953, "learning_rate": 1.1685000000000001e-06, "loss": 7.9121, "step": 2337 }, { "epoch": 0.7329153605015674, "grad_norm": 6.6463189125061035, "learning_rate": 1.1690000000000002e-06, "loss": 7.1128, "step": 2338 }, { "epoch": 0.7332288401253918, "grad_norm": 6.599662780761719, "learning_rate": 1.1695e-06, "loss": 8.0378, "step": 2339 }, { "epoch": 0.7335423197492164, "grad_norm": 5.061799049377441, "learning_rate": 1.1700000000000002e-06, "loss": 6.4478, "step": 2340 }, { "epoch": 0.7338557993730408, "grad_norm": 5.785806655883789, "learning_rate": 1.1705e-06, "loss": 10.5814, "step": 2341 }, { "epoch": 0.7341692789968652, "grad_norm": 5.782971382141113, "learning_rate": 1.171e-06, "loss": 6.4085, "step": 2342 }, { "epoch": 0.7344827586206897, "grad_norm": 5.601132392883301, "learning_rate": 1.1715000000000001e-06, "loss": 9.9022, "step": 2343 }, { "epoch": 0.7347962382445141, "grad_norm": 5.016490459442139, "learning_rate": 1.1720000000000002e-06, "loss": 7.146, "step": 2344 }, { "epoch": 0.7351097178683386, "grad_norm": 5.3097310066223145, "learning_rate": 1.1725e-06, "loss": 7.8484, "step": 2345 }, { "epoch": 0.735423197492163, "grad_norm": 5.761673927307129, "learning_rate": 1.173e-06, "loss": 6.9101, "step": 2346 }, { "epoch": 0.7357366771159874, "grad_norm": 6.155106544494629, "learning_rate": 1.1735e-06, "loss": 7.6222, "step": 2347 }, { "epoch": 0.7360501567398119, "grad_norm": 5.324407577514648, "learning_rate": 1.1740000000000001e-06, "loss": 7.18, "step": 2348 }, { "epoch": 0.7363636363636363, "grad_norm": 5.507112503051758, "learning_rate": 1.1745000000000001e-06, "loss": 7.4581, "step": 2349 }, { "epoch": 0.7366771159874608, "grad_norm": 6.135045528411865, "learning_rate": 1.175e-06, "loss": 8.3784, "step": 2350 }, { "epoch": 0.7369905956112852, "grad_norm": 5.607395172119141, "learning_rate": 1.1755e-06, "loss": 8.0636, "step": 2351 }, { "epoch": 0.7373040752351098, "grad_norm": 4.155460834503174, "learning_rate": 1.176e-06, "loss": 7.7036, "step": 2352 }, { "epoch": 0.7376175548589342, "grad_norm": 6.254929542541504, "learning_rate": 1.1765e-06, "loss": 7.4921, "step": 2353 }, { "epoch": 0.7379310344827587, "grad_norm": 7.738031387329102, "learning_rate": 1.1770000000000001e-06, "loss": 8.3682, "step": 2354 }, { "epoch": 0.7382445141065831, "grad_norm": 4.532285213470459, "learning_rate": 1.1775e-06, "loss": 6.7809, "step": 2355 }, { "epoch": 0.7385579937304075, "grad_norm": 5.4931640625, "learning_rate": 1.1780000000000002e-06, "loss": 8.2121, "step": 2356 }, { "epoch": 0.738871473354232, "grad_norm": 5.828548908233643, "learning_rate": 1.1785e-06, "loss": 7.4208, "step": 2357 }, { "epoch": 0.7391849529780564, "grad_norm": 8.793325424194336, "learning_rate": 1.179e-06, "loss": 10.5305, "step": 2358 }, { "epoch": 0.7394984326018809, "grad_norm": 7.540804386138916, "learning_rate": 1.1795e-06, "loss": 10.3455, "step": 2359 }, { "epoch": 0.7398119122257053, "grad_norm": 7.395792007446289, "learning_rate": 1.1800000000000001e-06, "loss": 8.6099, "step": 2360 }, { "epoch": 0.7401253918495297, "grad_norm": 7.034842491149902, "learning_rate": 1.1805000000000002e-06, "loss": 12.9946, "step": 2361 }, { "epoch": 0.7404388714733542, "grad_norm": 6.157832145690918, "learning_rate": 1.181e-06, "loss": 9.0294, "step": 2362 }, { "epoch": 0.7407523510971786, "grad_norm": 5.958568096160889, "learning_rate": 1.1815000000000002e-06, "loss": 9.4815, "step": 2363 }, { "epoch": 0.7410658307210032, "grad_norm": 5.166853427886963, "learning_rate": 1.182e-06, "loss": 7.8122, "step": 2364 }, { "epoch": 0.7413793103448276, "grad_norm": 6.027432441711426, "learning_rate": 1.1825000000000001e-06, "loss": 9.6608, "step": 2365 }, { "epoch": 0.7416927899686521, "grad_norm": 8.379853248596191, "learning_rate": 1.1830000000000002e-06, "loss": 10.2204, "step": 2366 }, { "epoch": 0.7420062695924765, "grad_norm": 6.882002353668213, "learning_rate": 1.1835e-06, "loss": 9.1155, "step": 2367 }, { "epoch": 0.742319749216301, "grad_norm": 5.61342191696167, "learning_rate": 1.1840000000000002e-06, "loss": 9.3164, "step": 2368 }, { "epoch": 0.7426332288401254, "grad_norm": 7.321317195892334, "learning_rate": 1.1845e-06, "loss": 7.8616, "step": 2369 }, { "epoch": 0.7429467084639498, "grad_norm": 5.494234561920166, "learning_rate": 1.185e-06, "loss": 6.7372, "step": 2370 }, { "epoch": 0.7432601880877743, "grad_norm": 6.378052234649658, "learning_rate": 1.1855000000000001e-06, "loss": 8.2405, "step": 2371 }, { "epoch": 0.7435736677115987, "grad_norm": 9.557308197021484, "learning_rate": 1.1860000000000002e-06, "loss": 12.0775, "step": 2372 }, { "epoch": 0.7438871473354232, "grad_norm": 7.571987152099609, "learning_rate": 1.1865000000000002e-06, "loss": 9.8201, "step": 2373 }, { "epoch": 0.7442006269592476, "grad_norm": 5.4964213371276855, "learning_rate": 1.187e-06, "loss": 6.8091, "step": 2374 }, { "epoch": 0.7445141065830722, "grad_norm": 5.78679895401001, "learning_rate": 1.1875e-06, "loss": 7.4889, "step": 2375 }, { "epoch": 0.7448275862068966, "grad_norm": 5.633636474609375, "learning_rate": 1.188e-06, "loss": 6.6415, "step": 2376 }, { "epoch": 0.745141065830721, "grad_norm": 6.5983428955078125, "learning_rate": 1.1885000000000001e-06, "loss": 7.7574, "step": 2377 }, { "epoch": 0.7454545454545455, "grad_norm": 9.485814094543457, "learning_rate": 1.1890000000000002e-06, "loss": 12.1369, "step": 2378 }, { "epoch": 0.7457680250783699, "grad_norm": 6.415633201599121, "learning_rate": 1.1895e-06, "loss": 7.9947, "step": 2379 }, { "epoch": 0.7460815047021944, "grad_norm": 6.807877063751221, "learning_rate": 1.19e-06, "loss": 8.5465, "step": 2380 }, { "epoch": 0.7463949843260188, "grad_norm": 6.27053165435791, "learning_rate": 1.1905e-06, "loss": 7.1279, "step": 2381 }, { "epoch": 0.7467084639498432, "grad_norm": 5.453633785247803, "learning_rate": 1.1910000000000001e-06, "loss": 7.1849, "step": 2382 }, { "epoch": 0.7470219435736677, "grad_norm": 6.075678825378418, "learning_rate": 1.1915000000000002e-06, "loss": 8.3007, "step": 2383 }, { "epoch": 0.7473354231974921, "grad_norm": 10.42825984954834, "learning_rate": 1.1920000000000002e-06, "loss": 10.672, "step": 2384 }, { "epoch": 0.7476489028213166, "grad_norm": 5.732131481170654, "learning_rate": 1.1925e-06, "loss": 6.1683, "step": 2385 }, { "epoch": 0.747962382445141, "grad_norm": 6.508482456207275, "learning_rate": 1.193e-06, "loss": 9.2361, "step": 2386 }, { "epoch": 0.7482758620689656, "grad_norm": 5.837035179138184, "learning_rate": 1.1935e-06, "loss": 7.1001, "step": 2387 }, { "epoch": 0.74858934169279, "grad_norm": 9.754945755004883, "learning_rate": 1.1940000000000001e-06, "loss": 10.2582, "step": 2388 }, { "epoch": 0.7489028213166145, "grad_norm": 5.980183124542236, "learning_rate": 1.1945000000000002e-06, "loss": 8.5382, "step": 2389 }, { "epoch": 0.7492163009404389, "grad_norm": 5.945464134216309, "learning_rate": 1.195e-06, "loss": 9.5149, "step": 2390 }, { "epoch": 0.7495297805642633, "grad_norm": 6.858036041259766, "learning_rate": 1.1955e-06, "loss": 9.1617, "step": 2391 }, { "epoch": 0.7498432601880878, "grad_norm": 7.403122901916504, "learning_rate": 1.196e-06, "loss": 9.4436, "step": 2392 }, { "epoch": 0.7501567398119122, "grad_norm": 5.871950626373291, "learning_rate": 1.1965000000000001e-06, "loss": 7.0343, "step": 2393 }, { "epoch": 0.7504702194357367, "grad_norm": 9.108241081237793, "learning_rate": 1.1970000000000001e-06, "loss": 11.0054, "step": 2394 }, { "epoch": 0.7504702194357367, "eval_loss": 27.862751007080078, "eval_runtime": 20.9206, "eval_samples_per_second": 128.438, "eval_steps_per_second": 8.03, "step": 2394 }, { "epoch": 0.7507836990595611, "grad_norm": 6.919027805328369, "learning_rate": 1.1975e-06, "loss": 8.3136, "step": 2395 }, { "epoch": 0.7510971786833855, "grad_norm": 5.883973598480225, "learning_rate": 1.1980000000000002e-06, "loss": 7.4282, "step": 2396 }, { "epoch": 0.75141065830721, "grad_norm": 6.176295280456543, "learning_rate": 1.1985e-06, "loss": 8.8492, "step": 2397 }, { "epoch": 0.7517241379310344, "grad_norm": 6.810075283050537, "learning_rate": 1.199e-06, "loss": 9.706, "step": 2398 }, { "epoch": 0.752037617554859, "grad_norm": 5.519807815551758, "learning_rate": 1.1995000000000001e-06, "loss": 8.259, "step": 2399 }, { "epoch": 0.7523510971786834, "grad_norm": 4.797094821929932, "learning_rate": 1.2000000000000002e-06, "loss": 7.2727, "step": 2400 }, { "epoch": 0.7526645768025079, "grad_norm": 6.261948108673096, "learning_rate": 1.2005000000000002e-06, "loss": 9.45, "step": 2401 }, { "epoch": 0.7529780564263323, "grad_norm": 5.316954135894775, "learning_rate": 1.201e-06, "loss": 6.7442, "step": 2402 }, { "epoch": 0.7532915360501568, "grad_norm": 6.721187591552734, "learning_rate": 1.2015000000000003e-06, "loss": 7.6002, "step": 2403 }, { "epoch": 0.7536050156739812, "grad_norm": 5.482544898986816, "learning_rate": 1.202e-06, "loss": 6.7844, "step": 2404 }, { "epoch": 0.7539184952978056, "grad_norm": 5.34820556640625, "learning_rate": 1.2025000000000001e-06, "loss": 7.4373, "step": 2405 }, { "epoch": 0.7542319749216301, "grad_norm": 7.171077728271484, "learning_rate": 1.2030000000000002e-06, "loss": 7.6944, "step": 2406 }, { "epoch": 0.7545454545454545, "grad_norm": 5.950673580169678, "learning_rate": 1.2035e-06, "loss": 8.4961, "step": 2407 }, { "epoch": 0.754858934169279, "grad_norm": 7.122339725494385, "learning_rate": 1.204e-06, "loss": 8.6441, "step": 2408 }, { "epoch": 0.7551724137931034, "grad_norm": 8.164888381958008, "learning_rate": 1.2045e-06, "loss": 10.9008, "step": 2409 }, { "epoch": 0.7554858934169278, "grad_norm": 8.712209701538086, "learning_rate": 1.2050000000000001e-06, "loss": 11.4492, "step": 2410 }, { "epoch": 0.7557993730407524, "grad_norm": 5.684695720672607, "learning_rate": 1.2055000000000001e-06, "loss": 7.7799, "step": 2411 }, { "epoch": 0.7561128526645768, "grad_norm": 5.714209079742432, "learning_rate": 1.2060000000000002e-06, "loss": 8.4314, "step": 2412 }, { "epoch": 0.7564263322884013, "grad_norm": 5.888979434967041, "learning_rate": 1.2065e-06, "loss": 6.7525, "step": 2413 }, { "epoch": 0.7567398119122257, "grad_norm": 7.8231425285339355, "learning_rate": 1.207e-06, "loss": 12.5721, "step": 2414 }, { "epoch": 0.7570532915360502, "grad_norm": 6.4952898025512695, "learning_rate": 1.2075e-06, "loss": 10.1876, "step": 2415 }, { "epoch": 0.7573667711598746, "grad_norm": 6.374924182891846, "learning_rate": 1.2080000000000001e-06, "loss": 7.2975, "step": 2416 }, { "epoch": 0.757680250783699, "grad_norm": 5.737400531768799, "learning_rate": 1.2085000000000002e-06, "loss": 8.371, "step": 2417 }, { "epoch": 0.7579937304075235, "grad_norm": 5.425909042358398, "learning_rate": 1.209e-06, "loss": 6.2901, "step": 2418 }, { "epoch": 0.7583072100313479, "grad_norm": 7.00061559677124, "learning_rate": 1.2095e-06, "loss": 6.5716, "step": 2419 }, { "epoch": 0.7586206896551724, "grad_norm": 9.082880020141602, "learning_rate": 1.21e-06, "loss": 9.7631, "step": 2420 }, { "epoch": 0.7589341692789968, "grad_norm": 5.837677478790283, "learning_rate": 1.2105e-06, "loss": 7.0652, "step": 2421 }, { "epoch": 0.7592476489028214, "grad_norm": 10.264467239379883, "learning_rate": 1.2110000000000001e-06, "loss": 13.0527, "step": 2422 }, { "epoch": 0.7595611285266458, "grad_norm": 6.860038757324219, "learning_rate": 1.2115e-06, "loss": 8.581, "step": 2423 }, { "epoch": 0.7598746081504703, "grad_norm": 8.064723014831543, "learning_rate": 1.2120000000000002e-06, "loss": 7.2021, "step": 2424 }, { "epoch": 0.7601880877742947, "grad_norm": 5.945399761199951, "learning_rate": 1.2125e-06, "loss": 8.6018, "step": 2425 }, { "epoch": 0.7605015673981191, "grad_norm": 8.428067207336426, "learning_rate": 1.213e-06, "loss": 9.2994, "step": 2426 }, { "epoch": 0.7608150470219436, "grad_norm": 6.145665168762207, "learning_rate": 1.2135000000000001e-06, "loss": 7.1988, "step": 2427 }, { "epoch": 0.761128526645768, "grad_norm": 7.332749843597412, "learning_rate": 1.214e-06, "loss": 8.2611, "step": 2428 }, { "epoch": 0.7614420062695925, "grad_norm": 13.179669380187988, "learning_rate": 1.2145000000000002e-06, "loss": 15.9689, "step": 2429 }, { "epoch": 0.7617554858934169, "grad_norm": 5.4208197593688965, "learning_rate": 1.215e-06, "loss": 6.9613, "step": 2430 }, { "epoch": 0.7620689655172413, "grad_norm": 5.367392063140869, "learning_rate": 1.2155e-06, "loss": 6.4245, "step": 2431 }, { "epoch": 0.7623824451410658, "grad_norm": 5.860224723815918, "learning_rate": 1.216e-06, "loss": 6.6126, "step": 2432 }, { "epoch": 0.7626959247648902, "grad_norm": 10.526330947875977, "learning_rate": 1.2165000000000001e-06, "loss": 13.7313, "step": 2433 }, { "epoch": 0.7630094043887148, "grad_norm": 5.663205623626709, "learning_rate": 1.2170000000000002e-06, "loss": 8.9684, "step": 2434 }, { "epoch": 0.7633228840125392, "grad_norm": 7.089934825897217, "learning_rate": 1.2175e-06, "loss": 7.9986, "step": 2435 }, { "epoch": 0.7636363636363637, "grad_norm": 5.609120845794678, "learning_rate": 1.2180000000000002e-06, "loss": 6.8416, "step": 2436 }, { "epoch": 0.7639498432601881, "grad_norm": 6.426872730255127, "learning_rate": 1.2185e-06, "loss": 7.9864, "step": 2437 }, { "epoch": 0.7642633228840126, "grad_norm": 8.020556449890137, "learning_rate": 1.219e-06, "loss": 8.7561, "step": 2438 }, { "epoch": 0.764576802507837, "grad_norm": 14.095486640930176, "learning_rate": 1.2195000000000001e-06, "loss": 14.5934, "step": 2439 }, { "epoch": 0.7648902821316614, "grad_norm": 6.615420341491699, "learning_rate": 1.2200000000000002e-06, "loss": 6.5653, "step": 2440 }, { "epoch": 0.7652037617554859, "grad_norm": 6.877127647399902, "learning_rate": 1.2205000000000002e-06, "loss": 9.0697, "step": 2441 }, { "epoch": 0.7655172413793103, "grad_norm": 7.938243389129639, "learning_rate": 1.221e-06, "loss": 7.9308, "step": 2442 }, { "epoch": 0.7658307210031348, "grad_norm": 5.383735656738281, "learning_rate": 1.2215e-06, "loss": 6.4287, "step": 2443 }, { "epoch": 0.7661442006269592, "grad_norm": 12.071537971496582, "learning_rate": 1.2220000000000001e-06, "loss": 9.1741, "step": 2444 }, { "epoch": 0.7664576802507836, "grad_norm": 8.097783088684082, "learning_rate": 1.2225000000000002e-06, "loss": 9.6469, "step": 2445 }, { "epoch": 0.7667711598746082, "grad_norm": 10.031937599182129, "learning_rate": 1.2230000000000002e-06, "loss": 11.3807, "step": 2446 }, { "epoch": 0.7670846394984326, "grad_norm": 6.421173572540283, "learning_rate": 1.2235e-06, "loss": 7.5599, "step": 2447 }, { "epoch": 0.7673981191222571, "grad_norm": 7.706267356872559, "learning_rate": 1.224e-06, "loss": 9.3446, "step": 2448 }, { "epoch": 0.7677115987460815, "grad_norm": 8.22382926940918, "learning_rate": 1.2245e-06, "loss": 10.4658, "step": 2449 }, { "epoch": 0.768025078369906, "grad_norm": 9.478100776672363, "learning_rate": 1.2250000000000001e-06, "loss": 9.6859, "step": 2450 }, { "epoch": 0.7683385579937304, "grad_norm": 10.563023567199707, "learning_rate": 1.2255000000000002e-06, "loss": 9.8905, "step": 2451 }, { "epoch": 0.7686520376175549, "grad_norm": 6.427543640136719, "learning_rate": 1.2260000000000002e-06, "loss": 7.7129, "step": 2452 }, { "epoch": 0.7689655172413793, "grad_norm": 7.049678325653076, "learning_rate": 1.2265e-06, "loss": 8.6615, "step": 2453 }, { "epoch": 0.7692789968652037, "grad_norm": 8.58466911315918, "learning_rate": 1.227e-06, "loss": 8.6398, "step": 2454 }, { "epoch": 0.7695924764890282, "grad_norm": 5.85331392288208, "learning_rate": 1.2275000000000001e-06, "loss": 7.8588, "step": 2455 }, { "epoch": 0.7699059561128526, "grad_norm": 7.215764999389648, "learning_rate": 1.2280000000000001e-06, "loss": 8.4039, "step": 2456 }, { "epoch": 0.7702194357366771, "grad_norm": 13.46800422668457, "learning_rate": 1.2285000000000002e-06, "loss": 11.4974, "step": 2457 }, { "epoch": 0.7705329153605016, "grad_norm": 7.22824764251709, "learning_rate": 1.229e-06, "loss": 8.1564, "step": 2458 }, { "epoch": 0.770846394984326, "grad_norm": 6.372166156768799, "learning_rate": 1.2295e-06, "loss": 6.8613, "step": 2459 }, { "epoch": 0.7711598746081505, "grad_norm": 8.008846282958984, "learning_rate": 1.23e-06, "loss": 8.2465, "step": 2460 }, { "epoch": 0.7714733542319749, "grad_norm": 10.947956085205078, "learning_rate": 1.2305000000000001e-06, "loss": 11.4562, "step": 2461 }, { "epoch": 0.7717868338557994, "grad_norm": 7.392462253570557, "learning_rate": 1.2310000000000002e-06, "loss": 9.5973, "step": 2462 }, { "epoch": 0.7721003134796238, "grad_norm": 6.9039082527160645, "learning_rate": 1.2315e-06, "loss": 7.2863, "step": 2463 }, { "epoch": 0.7724137931034483, "grad_norm": 6.517736911773682, "learning_rate": 1.2320000000000002e-06, "loss": 6.8215, "step": 2464 }, { "epoch": 0.7727272727272727, "grad_norm": 6.294345378875732, "learning_rate": 1.2325e-06, "loss": 6.9459, "step": 2465 }, { "epoch": 0.7730407523510971, "grad_norm": 7.055344581604004, "learning_rate": 1.233e-06, "loss": 7.9461, "step": 2466 }, { "epoch": 0.7733542319749216, "grad_norm": 8.949790954589844, "learning_rate": 1.2335000000000001e-06, "loss": 10.2358, "step": 2467 }, { "epoch": 0.773667711598746, "grad_norm": 12.344834327697754, "learning_rate": 1.234e-06, "loss": 11.9022, "step": 2468 }, { "epoch": 0.7739811912225706, "grad_norm": 13.830575942993164, "learning_rate": 1.2345000000000002e-06, "loss": 17.4536, "step": 2469 }, { "epoch": 0.774294670846395, "grad_norm": 10.294656753540039, "learning_rate": 1.235e-06, "loss": 7.7727, "step": 2470 }, { "epoch": 0.7746081504702195, "grad_norm": 6.9053874015808105, "learning_rate": 1.2355e-06, "loss": 8.134, "step": 2471 }, { "epoch": 0.7749216300940439, "grad_norm": 8.573596954345703, "learning_rate": 1.2360000000000001e-06, "loss": 9.5387, "step": 2472 }, { "epoch": 0.7752351097178684, "grad_norm": 8.290820121765137, "learning_rate": 1.2365000000000001e-06, "loss": 11.1582, "step": 2473 }, { "epoch": 0.7755485893416928, "grad_norm": 7.78951358795166, "learning_rate": 1.2370000000000002e-06, "loss": 9.6967, "step": 2474 }, { "epoch": 0.7758620689655172, "grad_norm": 9.113204956054688, "learning_rate": 1.2375e-06, "loss": 8.3096, "step": 2475 }, { "epoch": 0.7761755485893417, "grad_norm": 8.999139785766602, "learning_rate": 1.238e-06, "loss": 8.9482, "step": 2476 }, { "epoch": 0.7764890282131661, "grad_norm": 7.273940563201904, "learning_rate": 1.2385e-06, "loss": 7.3428, "step": 2477 }, { "epoch": 0.7768025078369906, "grad_norm": 9.114079475402832, "learning_rate": 1.2390000000000001e-06, "loss": 9.9958, "step": 2478 }, { "epoch": 0.777115987460815, "grad_norm": 9.299750328063965, "learning_rate": 1.2395000000000002e-06, "loss": 9.2364, "step": 2479 }, { "epoch": 0.7774294670846394, "grad_norm": 7.594107151031494, "learning_rate": 1.2400000000000002e-06, "loss": 8.9441, "step": 2480 }, { "epoch": 0.777742946708464, "grad_norm": 15.933536529541016, "learning_rate": 1.2405e-06, "loss": 15.7072, "step": 2481 }, { "epoch": 0.7780564263322884, "grad_norm": 6.593420028686523, "learning_rate": 1.241e-06, "loss": 7.2379, "step": 2482 }, { "epoch": 0.7783699059561129, "grad_norm": 10.504799842834473, "learning_rate": 1.2415e-06, "loss": 8.9226, "step": 2483 }, { "epoch": 0.7786833855799373, "grad_norm": 7.769631385803223, "learning_rate": 1.2420000000000001e-06, "loss": 7.9376, "step": 2484 }, { "epoch": 0.7789968652037618, "grad_norm": 9.674359321594238, "learning_rate": 1.2425000000000002e-06, "loss": 6.172, "step": 2485 }, { "epoch": 0.7793103448275862, "grad_norm": 10.36740493774414, "learning_rate": 1.243e-06, "loss": 10.3587, "step": 2486 }, { "epoch": 0.7796238244514107, "grad_norm": 10.151678085327148, "learning_rate": 1.2435e-06, "loss": 8.8859, "step": 2487 }, { "epoch": 0.7799373040752351, "grad_norm": 5.526301860809326, "learning_rate": 1.244e-06, "loss": 6.5573, "step": 2488 }, { "epoch": 0.7802507836990595, "grad_norm": 5.857382774353027, "learning_rate": 1.2445000000000001e-06, "loss": 6.264, "step": 2489 }, { "epoch": 0.780564263322884, "grad_norm": 7.313165187835693, "learning_rate": 1.2450000000000002e-06, "loss": 6.587, "step": 2490 }, { "epoch": 0.7808777429467084, "grad_norm": 10.243111610412598, "learning_rate": 1.2455e-06, "loss": 9.8566, "step": 2491 }, { "epoch": 0.7811912225705329, "grad_norm": 8.093012809753418, "learning_rate": 1.2460000000000002e-06, "loss": 7.5616, "step": 2492 }, { "epoch": 0.7815047021943574, "grad_norm": 9.839873313903809, "learning_rate": 1.2465e-06, "loss": 9.569, "step": 2493 }, { "epoch": 0.7818181818181819, "grad_norm": 9.529997825622559, "learning_rate": 1.247e-06, "loss": 8.8561, "step": 2494 }, { "epoch": 0.7821316614420063, "grad_norm": 7.144129276275635, "learning_rate": 1.2475000000000001e-06, "loss": 6.8548, "step": 2495 }, { "epoch": 0.7824451410658307, "grad_norm": 9.005447387695312, "learning_rate": 1.248e-06, "loss": 9.8978, "step": 2496 }, { "epoch": 0.7827586206896552, "grad_norm": 6.3512797355651855, "learning_rate": 1.2485000000000002e-06, "loss": 7.3197, "step": 2497 }, { "epoch": 0.7830721003134796, "grad_norm": 13.88167667388916, "learning_rate": 1.249e-06, "loss": 10.638, "step": 2498 }, { "epoch": 0.7833855799373041, "grad_norm": 8.67345905303955, "learning_rate": 1.2495e-06, "loss": 8.4422, "step": 2499 }, { "epoch": 0.7836990595611285, "grad_norm": 7.372941970825195, "learning_rate": 1.25e-06, "loss": 7.1321, "step": 2500 }, { "epoch": 0.784012539184953, "grad_norm": 7.833611488342285, "learning_rate": 1.2505000000000001e-06, "loss": 7.717, "step": 2501 }, { "epoch": 0.7843260188087774, "grad_norm": 7.675827980041504, "learning_rate": 1.251e-06, "loss": 6.8507, "step": 2502 }, { "epoch": 0.7846394984326018, "grad_norm": 7.610905170440674, "learning_rate": 1.2515000000000002e-06, "loss": 7.2601, "step": 2503 }, { "epoch": 0.7849529780564264, "grad_norm": 8.251374244689941, "learning_rate": 1.2520000000000003e-06, "loss": 7.8607, "step": 2504 }, { "epoch": 0.7852664576802508, "grad_norm": 8.651074409484863, "learning_rate": 1.2525e-06, "loss": 9.5066, "step": 2505 }, { "epoch": 0.7855799373040753, "grad_norm": 8.352408409118652, "learning_rate": 1.2530000000000001e-06, "loss": 6.5008, "step": 2506 }, { "epoch": 0.7858934169278997, "grad_norm": 8.274674415588379, "learning_rate": 1.2535e-06, "loss": 6.7467, "step": 2507 }, { "epoch": 0.7862068965517242, "grad_norm": 14.473176956176758, "learning_rate": 1.2540000000000002e-06, "loss": 14.1415, "step": 2508 }, { "epoch": 0.7865203761755486, "grad_norm": 7.974168300628662, "learning_rate": 1.2545000000000002e-06, "loss": 7.2298, "step": 2509 }, { "epoch": 0.786833855799373, "grad_norm": 7.596717357635498, "learning_rate": 1.255e-06, "loss": 7.7122, "step": 2510 }, { "epoch": 0.7871473354231975, "grad_norm": 11.445423126220703, "learning_rate": 1.2555e-06, "loss": 10.7785, "step": 2511 }, { "epoch": 0.7874608150470219, "grad_norm": 7.318839073181152, "learning_rate": 1.256e-06, "loss": 7.747, "step": 2512 }, { "epoch": 0.7877742946708464, "grad_norm": 8.10612964630127, "learning_rate": 1.2565000000000002e-06, "loss": 7.516, "step": 2513 }, { "epoch": 0.7880877742946708, "grad_norm": 6.949522972106934, "learning_rate": 1.2570000000000002e-06, "loss": 8.044, "step": 2514 }, { "epoch": 0.7884012539184952, "grad_norm": 10.29315185546875, "learning_rate": 1.2575e-06, "loss": 9.012, "step": 2515 }, { "epoch": 0.7887147335423198, "grad_norm": 8.57775592803955, "learning_rate": 1.258e-06, "loss": 6.2847, "step": 2516 }, { "epoch": 0.7890282131661442, "grad_norm": 11.213354110717773, "learning_rate": 1.2584999999999999e-06, "loss": 10.073, "step": 2517 }, { "epoch": 0.7893416927899687, "grad_norm": 12.789294242858887, "learning_rate": 1.2590000000000001e-06, "loss": 12.4063, "step": 2518 }, { "epoch": 0.7896551724137931, "grad_norm": 6.868427753448486, "learning_rate": 1.2595000000000002e-06, "loss": 5.4573, "step": 2519 }, { "epoch": 0.7899686520376176, "grad_norm": 6.975144863128662, "learning_rate": 1.26e-06, "loss": 7.0482, "step": 2520 }, { "epoch": 0.790282131661442, "grad_norm": 9.508519172668457, "learning_rate": 1.2605e-06, "loss": 9.9895, "step": 2521 }, { "epoch": 0.7905956112852665, "grad_norm": 12.231802940368652, "learning_rate": 1.261e-06, "loss": 10.782, "step": 2522 }, { "epoch": 0.7909090909090909, "grad_norm": 7.636473178863525, "learning_rate": 1.2615000000000001e-06, "loss": 7.4447, "step": 2523 }, { "epoch": 0.7912225705329153, "grad_norm": 14.089887619018555, "learning_rate": 1.2620000000000002e-06, "loss": 11.8825, "step": 2524 }, { "epoch": 0.7915360501567398, "grad_norm": 11.20523738861084, "learning_rate": 1.2625000000000002e-06, "loss": 10.0402, "step": 2525 }, { "epoch": 0.7918495297805642, "grad_norm": 10.182788848876953, "learning_rate": 1.263e-06, "loss": 7.4292, "step": 2526 }, { "epoch": 0.7921630094043887, "grad_norm": 8.2036771774292, "learning_rate": 1.2635e-06, "loss": 7.3361, "step": 2527 }, { "epoch": 0.7924764890282132, "grad_norm": 9.289824485778809, "learning_rate": 1.2640000000000003e-06, "loss": 7.6616, "step": 2528 }, { "epoch": 0.7927899686520377, "grad_norm": 10.154131889343262, "learning_rate": 1.2645000000000001e-06, "loss": 9.678, "step": 2529 }, { "epoch": 0.7931034482758621, "grad_norm": 8.044170379638672, "learning_rate": 1.2650000000000002e-06, "loss": 7.7272, "step": 2530 }, { "epoch": 0.7934169278996865, "grad_norm": 14.741924285888672, "learning_rate": 1.2655e-06, "loss": 11.6074, "step": 2531 }, { "epoch": 0.793730407523511, "grad_norm": 12.883913040161133, "learning_rate": 1.266e-06, "loss": 7.45, "step": 2532 }, { "epoch": 0.7940438871473354, "grad_norm": 14.291522026062012, "learning_rate": 1.2665000000000003e-06, "loss": 8.8233, "step": 2533 }, { "epoch": 0.7943573667711599, "grad_norm": 10.920492172241211, "learning_rate": 1.2670000000000001e-06, "loss": 8.2044, "step": 2534 }, { "epoch": 0.7946708463949843, "grad_norm": 9.552704811096191, "learning_rate": 1.2675000000000001e-06, "loss": 7.1354, "step": 2535 }, { "epoch": 0.7949843260188088, "grad_norm": 13.507481575012207, "learning_rate": 1.268e-06, "loss": 11.1275, "step": 2536 }, { "epoch": 0.7952978056426332, "grad_norm": 8.592397689819336, "learning_rate": 1.2685e-06, "loss": 6.2247, "step": 2537 }, { "epoch": 0.7956112852664576, "grad_norm": 8.421808242797852, "learning_rate": 1.2690000000000003e-06, "loss": 6.4027, "step": 2538 }, { "epoch": 0.7959247648902821, "grad_norm": 8.067540168762207, "learning_rate": 1.2695e-06, "loss": 8.3162, "step": 2539 }, { "epoch": 0.7962382445141066, "grad_norm": 8.634953498840332, "learning_rate": 1.2700000000000001e-06, "loss": 6.8064, "step": 2540 }, { "epoch": 0.7965517241379311, "grad_norm": 13.059370994567871, "learning_rate": 1.2705000000000002e-06, "loss": 10.4157, "step": 2541 }, { "epoch": 0.7968652037617555, "grad_norm": 7.6555304527282715, "learning_rate": 1.271e-06, "loss": 6.2527, "step": 2542 }, { "epoch": 0.79717868338558, "grad_norm": 9.67294692993164, "learning_rate": 1.2715000000000002e-06, "loss": 6.2133, "step": 2543 }, { "epoch": 0.7974921630094044, "grad_norm": 8.20207691192627, "learning_rate": 1.2720000000000003e-06, "loss": 7.4998, "step": 2544 }, { "epoch": 0.7978056426332288, "grad_norm": 9.94020938873291, "learning_rate": 1.2725e-06, "loss": 8.4282, "step": 2545 }, { "epoch": 0.7981191222570533, "grad_norm": 7.922036647796631, "learning_rate": 1.2730000000000001e-06, "loss": 6.9887, "step": 2546 }, { "epoch": 0.7984326018808777, "grad_norm": 11.113056182861328, "learning_rate": 1.2735e-06, "loss": 7.7985, "step": 2547 }, { "epoch": 0.7987460815047022, "grad_norm": 12.114452362060547, "learning_rate": 1.2740000000000002e-06, "loss": 8.92, "step": 2548 }, { "epoch": 0.7990595611285266, "grad_norm": 12.305609703063965, "learning_rate": 1.2745000000000002e-06, "loss": 10.3891, "step": 2549 }, { "epoch": 0.799373040752351, "grad_norm": 7.7950215339660645, "learning_rate": 1.275e-06, "loss": 6.9325, "step": 2550 }, { "epoch": 0.7996865203761756, "grad_norm": 12.936772346496582, "learning_rate": 1.2755000000000001e-06, "loss": 9.9608, "step": 2551 }, { "epoch": 0.8, "grad_norm": 9.42872428894043, "learning_rate": 1.276e-06, "loss": 7.2778, "step": 2552 }, { "epoch": 0.8003134796238245, "grad_norm": 12.10912036895752, "learning_rate": 1.2765000000000002e-06, "loss": 8.8379, "step": 2553 }, { "epoch": 0.8006269592476489, "grad_norm": 10.744391441345215, "learning_rate": 1.2770000000000002e-06, "loss": 8.0975, "step": 2554 }, { "epoch": 0.8009404388714734, "grad_norm": 9.00346565246582, "learning_rate": 1.2775e-06, "loss": 6.8766, "step": 2555 }, { "epoch": 0.8012539184952978, "grad_norm": 11.213250160217285, "learning_rate": 1.278e-06, "loss": 8.3033, "step": 2556 }, { "epoch": 0.8015673981191223, "grad_norm": 10.06607723236084, "learning_rate": 1.2785e-06, "loss": 6.8206, "step": 2557 }, { "epoch": 0.8018808777429467, "grad_norm": 16.033727645874023, "learning_rate": 1.2790000000000002e-06, "loss": 10.1014, "step": 2558 }, { "epoch": 0.8021943573667711, "grad_norm": 10.360695838928223, "learning_rate": 1.2795000000000002e-06, "loss": 7.3716, "step": 2559 }, { "epoch": 0.8025078369905956, "grad_norm": 16.054412841796875, "learning_rate": 1.28e-06, "loss": 9.5552, "step": 2560 }, { "epoch": 0.80282131661442, "grad_norm": 10.377695083618164, "learning_rate": 1.2805e-06, "loss": 7.2903, "step": 2561 }, { "epoch": 0.8031347962382445, "grad_norm": 10.614144325256348, "learning_rate": 1.281e-06, "loss": 7.9967, "step": 2562 }, { "epoch": 0.803448275862069, "grad_norm": 10.514567375183105, "learning_rate": 1.2815e-06, "loss": 6.7008, "step": 2563 }, { "epoch": 0.8037617554858935, "grad_norm": 8.607209205627441, "learning_rate": 1.2820000000000002e-06, "loss": 5.8722, "step": 2564 }, { "epoch": 0.8040752351097179, "grad_norm": 10.558499336242676, "learning_rate": 1.2825000000000002e-06, "loss": 7.2417, "step": 2565 }, { "epoch": 0.8043887147335423, "grad_norm": 17.161361694335938, "learning_rate": 1.283e-06, "loss": 13.5513, "step": 2566 }, { "epoch": 0.8047021943573668, "grad_norm": 11.054380416870117, "learning_rate": 1.2835e-06, "loss": 8.4185, "step": 2567 }, { "epoch": 0.8050156739811912, "grad_norm": 15.804526329040527, "learning_rate": 1.284e-06, "loss": 12.1668, "step": 2568 }, { "epoch": 0.8053291536050157, "grad_norm": 15.922082901000977, "learning_rate": 1.2845000000000002e-06, "loss": 11.7197, "step": 2569 }, { "epoch": 0.8056426332288401, "grad_norm": 15.542136192321777, "learning_rate": 1.2850000000000002e-06, "loss": 5.9863, "step": 2570 }, { "epoch": 0.8059561128526646, "grad_norm": 13.482390403747559, "learning_rate": 1.2855e-06, "loss": 8.9265, "step": 2571 }, { "epoch": 0.806269592476489, "grad_norm": 15.561339378356934, "learning_rate": 1.286e-06, "loss": 8.5574, "step": 2572 }, { "epoch": 0.8065830721003134, "grad_norm": 14.538595199584961, "learning_rate": 1.2864999999999999e-06, "loss": 11.4279, "step": 2573 }, { "epoch": 0.8068965517241379, "grad_norm": 14.449934005737305, "learning_rate": 1.2870000000000001e-06, "loss": 12.0358, "step": 2574 }, { "epoch": 0.8072100313479624, "grad_norm": 13.869661331176758, "learning_rate": 1.2875000000000002e-06, "loss": 10.5568, "step": 2575 }, { "epoch": 0.8075235109717869, "grad_norm": 10.084548950195312, "learning_rate": 1.288e-06, "loss": 6.7925, "step": 2576 }, { "epoch": 0.8078369905956113, "grad_norm": 8.329049110412598, "learning_rate": 1.2885e-06, "loss": 7.129, "step": 2577 }, { "epoch": 0.8081504702194358, "grad_norm": 15.03673267364502, "learning_rate": 1.289e-06, "loss": 10.9492, "step": 2578 }, { "epoch": 0.8084639498432602, "grad_norm": 11.203147888183594, "learning_rate": 1.2895e-06, "loss": 6.868, "step": 2579 }, { "epoch": 0.8087774294670846, "grad_norm": 14.008033752441406, "learning_rate": 1.2900000000000001e-06, "loss": 7.7931, "step": 2580 }, { "epoch": 0.8090909090909091, "grad_norm": 11.999140739440918, "learning_rate": 1.2905000000000002e-06, "loss": 5.7815, "step": 2581 }, { "epoch": 0.8094043887147335, "grad_norm": 14.67009449005127, "learning_rate": 1.291e-06, "loss": 10.9111, "step": 2582 }, { "epoch": 0.809717868338558, "grad_norm": 10.367982864379883, "learning_rate": 1.2915e-06, "loss": 7.1978, "step": 2583 }, { "epoch": 0.8100313479623824, "grad_norm": 15.708711624145508, "learning_rate": 1.2920000000000003e-06, "loss": 8.7541, "step": 2584 }, { "epoch": 0.8103448275862069, "grad_norm": 21.055742263793945, "learning_rate": 1.2925000000000001e-06, "loss": 9.4178, "step": 2585 }, { "epoch": 0.8106583072100313, "grad_norm": 11.601629257202148, "learning_rate": 1.2930000000000002e-06, "loss": 7.239, "step": 2586 }, { "epoch": 0.8109717868338558, "grad_norm": 19.32367515563965, "learning_rate": 1.2935e-06, "loss": 12.215, "step": 2587 }, { "epoch": 0.8112852664576803, "grad_norm": 12.446346282958984, "learning_rate": 1.294e-06, "loss": 8.1503, "step": 2588 }, { "epoch": 0.8115987460815047, "grad_norm": 15.01457691192627, "learning_rate": 1.2945000000000003e-06, "loss": 9.3474, "step": 2589 }, { "epoch": 0.8119122257053292, "grad_norm": 8.85611629486084, "learning_rate": 1.295e-06, "loss": 6.0037, "step": 2590 }, { "epoch": 0.8122257053291536, "grad_norm": 12.6907320022583, "learning_rate": 1.2955000000000001e-06, "loss": 6.035, "step": 2591 }, { "epoch": 0.812539184952978, "grad_norm": 29.200088500976562, "learning_rate": 1.296e-06, "loss": 15.6593, "step": 2592 }, { "epoch": 0.8128526645768025, "grad_norm": 13.449995994567871, "learning_rate": 1.2965e-06, "loss": 7.4237, "step": 2593 }, { "epoch": 0.8131661442006269, "grad_norm": 12.185378074645996, "learning_rate": 1.2970000000000002e-06, "loss": 7.6051, "step": 2594 }, { "epoch": 0.8134796238244514, "grad_norm": 13.01950454711914, "learning_rate": 1.2975e-06, "loss": 9.8676, "step": 2595 }, { "epoch": 0.8137931034482758, "grad_norm": 12.636890411376953, "learning_rate": 1.2980000000000001e-06, "loss": 7.8251, "step": 2596 }, { "epoch": 0.8141065830721003, "grad_norm": 14.959136962890625, "learning_rate": 1.2985e-06, "loss": 8.8174, "step": 2597 }, { "epoch": 0.8144200626959248, "grad_norm": 12.237345695495605, "learning_rate": 1.299e-06, "loss": 6.9888, "step": 2598 }, { "epoch": 0.8147335423197493, "grad_norm": 9.535272598266602, "learning_rate": 1.2995000000000002e-06, "loss": 5.9116, "step": 2599 }, { "epoch": 0.8150470219435737, "grad_norm": 15.491859436035156, "learning_rate": 1.3e-06, "loss": 7.9926, "step": 2600 }, { "epoch": 0.8153605015673981, "grad_norm": 11.194720268249512, "learning_rate": 1.3005e-06, "loss": 7.7533, "step": 2601 }, { "epoch": 0.8156739811912226, "grad_norm": 18.05571746826172, "learning_rate": 1.3010000000000001e-06, "loss": 9.5349, "step": 2602 }, { "epoch": 0.815987460815047, "grad_norm": 9.782732009887695, "learning_rate": 1.3015e-06, "loss": 6.5899, "step": 2603 }, { "epoch": 0.8163009404388715, "grad_norm": 15.516654968261719, "learning_rate": 1.3020000000000002e-06, "loss": 7.107, "step": 2604 }, { "epoch": 0.8166144200626959, "grad_norm": 13.18638801574707, "learning_rate": 1.3025000000000002e-06, "loss": 6.44, "step": 2605 }, { "epoch": 0.8169278996865204, "grad_norm": 14.614501953125, "learning_rate": 1.303e-06, "loss": 7.5293, "step": 2606 }, { "epoch": 0.8172413793103448, "grad_norm": 11.227118492126465, "learning_rate": 1.3035e-06, "loss": 7.8937, "step": 2607 }, { "epoch": 0.8175548589341692, "grad_norm": 15.58757495880127, "learning_rate": 1.304e-06, "loss": 8.0596, "step": 2608 }, { "epoch": 0.8178683385579937, "grad_norm": 22.042787551879883, "learning_rate": 1.3045000000000002e-06, "loss": 8.2819, "step": 2609 }, { "epoch": 0.8181818181818182, "grad_norm": 21.173282623291016, "learning_rate": 1.3050000000000002e-06, "loss": 10.4288, "step": 2610 }, { "epoch": 0.8184952978056427, "grad_norm": 18.32832145690918, "learning_rate": 1.3055e-06, "loss": 12.1147, "step": 2611 }, { "epoch": 0.8188087774294671, "grad_norm": 15.920726776123047, "learning_rate": 1.306e-06, "loss": 9.0858, "step": 2612 }, { "epoch": 0.8191222570532916, "grad_norm": 12.705510139465332, "learning_rate": 1.3065e-06, "loss": 8.1776, "step": 2613 }, { "epoch": 0.819435736677116, "grad_norm": 13.502321243286133, "learning_rate": 1.3070000000000001e-06, "loss": 7.3692, "step": 2614 }, { "epoch": 0.8197492163009404, "grad_norm": 15.334294319152832, "learning_rate": 1.3075000000000002e-06, "loss": 8.4849, "step": 2615 }, { "epoch": 0.8200626959247649, "grad_norm": 13.238600730895996, "learning_rate": 1.308e-06, "loss": 6.9999, "step": 2616 }, { "epoch": 0.8203761755485893, "grad_norm": 20.93524742126465, "learning_rate": 1.3085e-06, "loss": 7.2483, "step": 2617 }, { "epoch": 0.8206896551724138, "grad_norm": 13.544681549072266, "learning_rate": 1.309e-06, "loss": 7.8575, "step": 2618 }, { "epoch": 0.8210031347962382, "grad_norm": 14.38907241821289, "learning_rate": 1.3095000000000001e-06, "loss": 6.2083, "step": 2619 }, { "epoch": 0.8213166144200627, "grad_norm": 20.6708927154541, "learning_rate": 1.3100000000000002e-06, "loss": 9.0823, "step": 2620 }, { "epoch": 0.8216300940438871, "grad_norm": 12.318846702575684, "learning_rate": 1.3105000000000002e-06, "loss": 6.4326, "step": 2621 }, { "epoch": 0.8219435736677116, "grad_norm": 40.32315444946289, "learning_rate": 1.311e-06, "loss": 14.9634, "step": 2622 }, { "epoch": 0.8222570532915361, "grad_norm": 15.944631576538086, "learning_rate": 1.3115e-06, "loss": 7.8842, "step": 2623 }, { "epoch": 0.8225705329153605, "grad_norm": 12.112171173095703, "learning_rate": 1.3120000000000003e-06, "loss": 5.5937, "step": 2624 }, { "epoch": 0.822884012539185, "grad_norm": 13.884896278381348, "learning_rate": 1.3125000000000001e-06, "loss": 6.3687, "step": 2625 }, { "epoch": 0.8231974921630094, "grad_norm": 22.745121002197266, "learning_rate": 1.3130000000000002e-06, "loss": 12.1536, "step": 2626 }, { "epoch": 0.8235109717868339, "grad_norm": 18.648685455322266, "learning_rate": 1.3135e-06, "loss": 7.5739, "step": 2627 }, { "epoch": 0.8238244514106583, "grad_norm": 18.585302352905273, "learning_rate": 1.314e-06, "loss": 8.4919, "step": 2628 }, { "epoch": 0.8241379310344827, "grad_norm": 26.631200790405273, "learning_rate": 1.3145000000000003e-06, "loss": 8.0334, "step": 2629 }, { "epoch": 0.8244514106583072, "grad_norm": 22.677764892578125, "learning_rate": 1.3150000000000001e-06, "loss": 8.3661, "step": 2630 }, { "epoch": 0.8247648902821316, "grad_norm": 27.903852462768555, "learning_rate": 1.3155000000000002e-06, "loss": 10.0715, "step": 2631 }, { "epoch": 0.8250783699059561, "grad_norm": 26.193275451660156, "learning_rate": 1.316e-06, "loss": 9.49, "step": 2632 }, { "epoch": 0.8253918495297806, "grad_norm": 15.721944808959961, "learning_rate": 1.3165e-06, "loss": 7.3057, "step": 2633 }, { "epoch": 0.8257053291536051, "grad_norm": 16.153507232666016, "learning_rate": 1.3170000000000003e-06, "loss": 6.6323, "step": 2634 }, { "epoch": 0.8260188087774295, "grad_norm": 17.67839813232422, "learning_rate": 1.3175e-06, "loss": 7.1916, "step": 2635 }, { "epoch": 0.826332288401254, "grad_norm": 18.60005760192871, "learning_rate": 1.3180000000000001e-06, "loss": 7.4737, "step": 2636 }, { "epoch": 0.8266457680250784, "grad_norm": 16.918989181518555, "learning_rate": 1.3185e-06, "loss": 9.5991, "step": 2637 }, { "epoch": 0.8269592476489028, "grad_norm": 28.17279624938965, "learning_rate": 1.319e-06, "loss": 9.0958, "step": 2638 }, { "epoch": 0.8272727272727273, "grad_norm": 22.73778533935547, "learning_rate": 1.3195000000000002e-06, "loss": 6.3333, "step": 2639 }, { "epoch": 0.8275862068965517, "grad_norm": 17.459001541137695, "learning_rate": 1.32e-06, "loss": 6.0962, "step": 2640 }, { "epoch": 0.8278996865203762, "grad_norm": 13.217561721801758, "learning_rate": 1.3205e-06, "loss": 6.563, "step": 2641 }, { "epoch": 0.8282131661442006, "grad_norm": 15.55752182006836, "learning_rate": 1.3210000000000001e-06, "loss": 5.5877, "step": 2642 }, { "epoch": 0.828526645768025, "grad_norm": 56.43824005126953, "learning_rate": 1.3215e-06, "loss": 7.7625, "step": 2643 }, { "epoch": 0.8288401253918495, "grad_norm": 21.975671768188477, "learning_rate": 1.3220000000000002e-06, "loss": 9.7735, "step": 2644 }, { "epoch": 0.829153605015674, "grad_norm": 19.311349868774414, "learning_rate": 1.3225000000000003e-06, "loss": 8.4531, "step": 2645 }, { "epoch": 0.8294670846394985, "grad_norm": 39.477603912353516, "learning_rate": 1.323e-06, "loss": 14.1747, "step": 2646 }, { "epoch": 0.8297805642633229, "grad_norm": 23.62912368774414, "learning_rate": 1.3235000000000001e-06, "loss": 9.9994, "step": 2647 }, { "epoch": 0.8300940438871474, "grad_norm": 17.039447784423828, "learning_rate": 1.324e-06, "loss": 6.8518, "step": 2648 }, { "epoch": 0.8304075235109718, "grad_norm": 21.798898696899414, "learning_rate": 1.3245000000000002e-06, "loss": 8.8252, "step": 2649 }, { "epoch": 0.8307210031347962, "grad_norm": 31.107290267944336, "learning_rate": 1.3250000000000002e-06, "loss": 9.1876, "step": 2650 }, { "epoch": 0.8310344827586207, "grad_norm": 21.375978469848633, "learning_rate": 1.3255e-06, "loss": 6.6534, "step": 2651 }, { "epoch": 0.8313479623824451, "grad_norm": 16.57238006591797, "learning_rate": 1.326e-06, "loss": 5.7842, "step": 2652 }, { "epoch": 0.8316614420062696, "grad_norm": 25.424558639526367, "learning_rate": 1.3265e-06, "loss": 8.5218, "step": 2653 }, { "epoch": 0.831974921630094, "grad_norm": 32.927066802978516, "learning_rate": 1.3270000000000002e-06, "loss": 9.3753, "step": 2654 }, { "epoch": 0.8322884012539185, "grad_norm": 18.978023529052734, "learning_rate": 1.3275000000000002e-06, "loss": 6.1247, "step": 2655 }, { "epoch": 0.8326018808777429, "grad_norm": 38.223243713378906, "learning_rate": 1.328e-06, "loss": 11.0785, "step": 2656 }, { "epoch": 0.8329153605015674, "grad_norm": 16.340456008911133, "learning_rate": 1.3285e-06, "loss": 6.4915, "step": 2657 }, { "epoch": 0.8332288401253919, "grad_norm": 18.187742233276367, "learning_rate": 1.3290000000000001e-06, "loss": 8.5854, "step": 2658 }, { "epoch": 0.8335423197492163, "grad_norm": 17.474607467651367, "learning_rate": 1.3295000000000001e-06, "loss": 7.5377, "step": 2659 }, { "epoch": 0.8338557993730408, "grad_norm": 19.971267700195312, "learning_rate": 1.3300000000000002e-06, "loss": 7.2454, "step": 2660 }, { "epoch": 0.8341692789968652, "grad_norm": 14.669326782226562, "learning_rate": 1.3305000000000002e-06, "loss": 6.1117, "step": 2661 }, { "epoch": 0.8344827586206897, "grad_norm": 18.929859161376953, "learning_rate": 1.331e-06, "loss": 7.8862, "step": 2662 }, { "epoch": 0.8347962382445141, "grad_norm": 29.54069709777832, "learning_rate": 1.3315e-06, "loss": 7.5074, "step": 2663 }, { "epoch": 0.8351097178683385, "grad_norm": 21.967472076416016, "learning_rate": 1.3320000000000003e-06, "loss": 7.0764, "step": 2664 }, { "epoch": 0.835423197492163, "grad_norm": 45.06687927246094, "learning_rate": 1.3325000000000002e-06, "loss": 9.701, "step": 2665 }, { "epoch": 0.8357366771159874, "grad_norm": 24.894472122192383, "learning_rate": 1.3330000000000002e-06, "loss": 8.708, "step": 2666 }, { "epoch": 0.8360501567398119, "grad_norm": 20.155099868774414, "learning_rate": 1.3335e-06, "loss": 6.1855, "step": 2667 }, { "epoch": 0.8363636363636363, "grad_norm": 20.541330337524414, "learning_rate": 1.334e-06, "loss": 6.6755, "step": 2668 }, { "epoch": 0.8366771159874609, "grad_norm": 22.628978729248047, "learning_rate": 1.3345000000000003e-06, "loss": 8.2543, "step": 2669 }, { "epoch": 0.8369905956112853, "grad_norm": 22.389394760131836, "learning_rate": 1.3350000000000001e-06, "loss": 7.0468, "step": 2670 }, { "epoch": 0.8373040752351097, "grad_norm": 23.417722702026367, "learning_rate": 1.3355000000000002e-06, "loss": 6.792, "step": 2671 }, { "epoch": 0.8376175548589342, "grad_norm": 21.459794998168945, "learning_rate": 1.336e-06, "loss": 6.9663, "step": 2672 }, { "epoch": 0.8379310344827586, "grad_norm": 32.657989501953125, "learning_rate": 1.3365e-06, "loss": 5.9438, "step": 2673 }, { "epoch": 0.8382445141065831, "grad_norm": 47.682586669921875, "learning_rate": 1.3370000000000003e-06, "loss": 10.6549, "step": 2674 }, { "epoch": 0.8385579937304075, "grad_norm": 20.36661720275879, "learning_rate": 1.3375000000000001e-06, "loss": 6.8448, "step": 2675 }, { "epoch": 0.838871473354232, "grad_norm": 14.008045196533203, "learning_rate": 1.3380000000000001e-06, "loss": 6.8451, "step": 2676 }, { "epoch": 0.8391849529780564, "grad_norm": 17.530170440673828, "learning_rate": 1.3385e-06, "loss": 5.8831, "step": 2677 }, { "epoch": 0.8394984326018808, "grad_norm": 17.58102798461914, "learning_rate": 1.339e-06, "loss": 6.3528, "step": 2678 }, { "epoch": 0.8398119122257053, "grad_norm": 23.83525848388672, "learning_rate": 1.3395000000000003e-06, "loss": 7.7174, "step": 2679 }, { "epoch": 0.8401253918495298, "grad_norm": 27.301860809326172, "learning_rate": 1.34e-06, "loss": 5.6273, "step": 2680 }, { "epoch": 0.8404388714733543, "grad_norm": 29.820091247558594, "learning_rate": 1.3405000000000001e-06, "loss": 7.9886, "step": 2681 }, { "epoch": 0.8407523510971787, "grad_norm": 31.963327407836914, "learning_rate": 1.3410000000000002e-06, "loss": 6.5933, "step": 2682 }, { "epoch": 0.8410658307210032, "grad_norm": 23.809494018554688, "learning_rate": 1.3415e-06, "loss": 7.4239, "step": 2683 }, { "epoch": 0.8413793103448276, "grad_norm": 26.722423553466797, "learning_rate": 1.3420000000000002e-06, "loss": 6.8609, "step": 2684 }, { "epoch": 0.841692789968652, "grad_norm": 20.260234832763672, "learning_rate": 1.3425000000000003e-06, "loss": 7.6539, "step": 2685 }, { "epoch": 0.8420062695924765, "grad_norm": 29.476564407348633, "learning_rate": 1.343e-06, "loss": 8.3136, "step": 2686 }, { "epoch": 0.8423197492163009, "grad_norm": 17.281204223632812, "learning_rate": 1.3435000000000001e-06, "loss": 5.9271, "step": 2687 }, { "epoch": 0.8426332288401254, "grad_norm": 39.53363037109375, "learning_rate": 1.344e-06, "loss": 8.3828, "step": 2688 }, { "epoch": 0.8429467084639498, "grad_norm": 19.412582397460938, "learning_rate": 1.3445e-06, "loss": 6.6062, "step": 2689 }, { "epoch": 0.8432601880877743, "grad_norm": 31.106475830078125, "learning_rate": 1.3450000000000003e-06, "loss": 7.9131, "step": 2690 }, { "epoch": 0.8435736677115987, "grad_norm": 37.35240936279297, "learning_rate": 1.3455e-06, "loss": 8.8415, "step": 2691 }, { "epoch": 0.8438871473354232, "grad_norm": 26.39192771911621, "learning_rate": 1.3460000000000001e-06, "loss": 6.637, "step": 2692 }, { "epoch": 0.8442006269592477, "grad_norm": 21.605411529541016, "learning_rate": 1.3465e-06, "loss": 6.3293, "step": 2693 }, { "epoch": 0.8445141065830721, "grad_norm": 21.46399688720703, "learning_rate": 1.347e-06, "loss": 6.8912, "step": 2694 }, { "epoch": 0.8448275862068966, "grad_norm": 26.33460807800293, "learning_rate": 1.3475000000000002e-06, "loss": 8.7624, "step": 2695 }, { "epoch": 0.845141065830721, "grad_norm": 21.527713775634766, "learning_rate": 1.348e-06, "loss": 5.8286, "step": 2696 }, { "epoch": 0.8454545454545455, "grad_norm": 30.383634567260742, "learning_rate": 1.3485e-06, "loss": 9.5837, "step": 2697 }, { "epoch": 0.8457680250783699, "grad_norm": 33.206356048583984, "learning_rate": 1.3490000000000001e-06, "loss": 7.8532, "step": 2698 }, { "epoch": 0.8460815047021943, "grad_norm": 36.5329475402832, "learning_rate": 1.3495e-06, "loss": 5.4797, "step": 2699 }, { "epoch": 0.8463949843260188, "grad_norm": 30.74308204650879, "learning_rate": 1.3500000000000002e-06, "loss": 7.289, "step": 2700 }, { "epoch": 0.8467084639498432, "grad_norm": 38.16331481933594, "learning_rate": 1.3505000000000002e-06, "loss": 7.9117, "step": 2701 }, { "epoch": 0.8470219435736677, "grad_norm": 26.741302490234375, "learning_rate": 1.351e-06, "loss": 5.9725, "step": 2702 }, { "epoch": 0.8473354231974921, "grad_norm": 34.44478988647461, "learning_rate": 1.3515e-06, "loss": 7.5997, "step": 2703 }, { "epoch": 0.8476489028213167, "grad_norm": 23.709253311157227, "learning_rate": 1.352e-06, "loss": 6.8915, "step": 2704 }, { "epoch": 0.8479623824451411, "grad_norm": 24.891733169555664, "learning_rate": 1.3525000000000002e-06, "loss": 5.9107, "step": 2705 }, { "epoch": 0.8482758620689655, "grad_norm": 24.73875617980957, "learning_rate": 1.3530000000000002e-06, "loss": 6.2778, "step": 2706 }, { "epoch": 0.84858934169279, "grad_norm": 18.066926956176758, "learning_rate": 1.3535e-06, "loss": 6.1856, "step": 2707 }, { "epoch": 0.8489028213166144, "grad_norm": 40.347476959228516, "learning_rate": 1.354e-06, "loss": 7.0503, "step": 2708 }, { "epoch": 0.8492163009404389, "grad_norm": 31.80996322631836, "learning_rate": 1.3545e-06, "loss": 8.8319, "step": 2709 }, { "epoch": 0.8495297805642633, "grad_norm": 27.470136642456055, "learning_rate": 1.3550000000000002e-06, "loss": 7.3765, "step": 2710 }, { "epoch": 0.8498432601880878, "grad_norm": 83.07367706298828, "learning_rate": 1.3555000000000002e-06, "loss": 20.241, "step": 2711 }, { "epoch": 0.8501567398119122, "grad_norm": 30.60701560974121, "learning_rate": 1.356e-06, "loss": 6.9115, "step": 2712 }, { "epoch": 0.8504702194357366, "grad_norm": 31.232662200927734, "learning_rate": 1.3565e-06, "loss": 7.2203, "step": 2713 }, { "epoch": 0.8507836990595611, "grad_norm": 29.009035110473633, "learning_rate": 1.3569999999999999e-06, "loss": 7.3232, "step": 2714 }, { "epoch": 0.8510971786833855, "grad_norm": 20.113876342773438, "learning_rate": 1.3575000000000001e-06, "loss": 5.1684, "step": 2715 }, { "epoch": 0.8514106583072101, "grad_norm": 21.815345764160156, "learning_rate": 1.3580000000000002e-06, "loss": 6.1103, "step": 2716 }, { "epoch": 0.8517241379310345, "grad_norm": 28.0091552734375, "learning_rate": 1.3585e-06, "loss": 7.3426, "step": 2717 }, { "epoch": 0.852037617554859, "grad_norm": 30.732309341430664, "learning_rate": 1.359e-06, "loss": 6.6737, "step": 2718 }, { "epoch": 0.8523510971786834, "grad_norm": 24.529449462890625, "learning_rate": 1.3595e-06, "loss": 6.6836, "step": 2719 }, { "epoch": 0.8526645768025078, "grad_norm": 21.116939544677734, "learning_rate": 1.3600000000000001e-06, "loss": 6.1635, "step": 2720 }, { "epoch": 0.8529780564263323, "grad_norm": 27.107351303100586, "learning_rate": 1.3605000000000001e-06, "loss": 6.0729, "step": 2721 }, { "epoch": 0.8532915360501567, "grad_norm": 19.72542953491211, "learning_rate": 1.3610000000000002e-06, "loss": 4.9322, "step": 2722 }, { "epoch": 0.8536050156739812, "grad_norm": 20.744781494140625, "learning_rate": 1.3615e-06, "loss": 6.2112, "step": 2723 }, { "epoch": 0.8539184952978056, "grad_norm": 23.486122131347656, "learning_rate": 1.362e-06, "loss": 6.0354, "step": 2724 }, { "epoch": 0.85423197492163, "grad_norm": 37.22897720336914, "learning_rate": 1.3625000000000003e-06, "loss": 6.7026, "step": 2725 }, { "epoch": 0.8545454545454545, "grad_norm": 28.69023323059082, "learning_rate": 1.3630000000000001e-06, "loss": 6.349, "step": 2726 }, { "epoch": 0.854858934169279, "grad_norm": 31.381017684936523, "learning_rate": 1.3635000000000002e-06, "loss": 6.7903, "step": 2727 }, { "epoch": 0.8551724137931035, "grad_norm": 23.358259201049805, "learning_rate": 1.364e-06, "loss": 5.7168, "step": 2728 }, { "epoch": 0.8554858934169279, "grad_norm": 29.059661865234375, "learning_rate": 1.3645e-06, "loss": 6.2689, "step": 2729 }, { "epoch": 0.8557993730407524, "grad_norm": 28.51305389404297, "learning_rate": 1.3650000000000003e-06, "loss": 5.9654, "step": 2730 }, { "epoch": 0.8561128526645768, "grad_norm": 24.55581283569336, "learning_rate": 1.3655e-06, "loss": 5.0936, "step": 2731 }, { "epoch": 0.8564263322884013, "grad_norm": 28.580116271972656, "learning_rate": 1.3660000000000001e-06, "loss": 6.2048, "step": 2732 }, { "epoch": 0.8567398119122257, "grad_norm": 37.072593688964844, "learning_rate": 1.3665e-06, "loss": 6.9439, "step": 2733 }, { "epoch": 0.8570532915360501, "grad_norm": 24.373992919921875, "learning_rate": 1.367e-06, "loss": 6.8616, "step": 2734 }, { "epoch": 0.8573667711598746, "grad_norm": 37.73335647583008, "learning_rate": 1.3675000000000002e-06, "loss": 6.2993, "step": 2735 }, { "epoch": 0.857680250783699, "grad_norm": 80.02848815917969, "learning_rate": 1.368e-06, "loss": 14.2041, "step": 2736 }, { "epoch": 0.8579937304075235, "grad_norm": 52.7184944152832, "learning_rate": 1.3685000000000001e-06, "loss": 8.9868, "step": 2737 }, { "epoch": 0.8583072100313479, "grad_norm": 31.193477630615234, "learning_rate": 1.3690000000000001e-06, "loss": 7.0294, "step": 2738 }, { "epoch": 0.8586206896551725, "grad_norm": 30.320158004760742, "learning_rate": 1.3695e-06, "loss": 6.4768, "step": 2739 }, { "epoch": 0.8589341692789969, "grad_norm": 30.57577896118164, "learning_rate": 1.3700000000000002e-06, "loss": 5.7312, "step": 2740 }, { "epoch": 0.8592476489028213, "grad_norm": 25.569894790649414, "learning_rate": 1.3705000000000003e-06, "loss": 5.5384, "step": 2741 }, { "epoch": 0.8595611285266458, "grad_norm": 40.88070297241211, "learning_rate": 1.371e-06, "loss": 7.8681, "step": 2742 }, { "epoch": 0.8598746081504702, "grad_norm": 32.0083122253418, "learning_rate": 1.3715000000000001e-06, "loss": 6.5132, "step": 2743 }, { "epoch": 0.8601880877742947, "grad_norm": 33.715633392333984, "learning_rate": 1.372e-06, "loss": 6.0128, "step": 2744 }, { "epoch": 0.8605015673981191, "grad_norm": 43.099884033203125, "learning_rate": 1.3725000000000002e-06, "loss": 6.8138, "step": 2745 }, { "epoch": 0.8608150470219436, "grad_norm": 47.59553909301758, "learning_rate": 1.3730000000000002e-06, "loss": 8.2016, "step": 2746 }, { "epoch": 0.861128526645768, "grad_norm": 40.0462532043457, "learning_rate": 1.3735e-06, "loss": 6.6381, "step": 2747 }, { "epoch": 0.8614420062695924, "grad_norm": 35.96194076538086, "learning_rate": 1.374e-06, "loss": 7.399, "step": 2748 }, { "epoch": 0.8617554858934169, "grad_norm": 39.26766586303711, "learning_rate": 1.3745e-06, "loss": 6.7996, "step": 2749 }, { "epoch": 0.8620689655172413, "grad_norm": 51.32083511352539, "learning_rate": 1.3750000000000002e-06, "loss": 8.2172, "step": 2750 }, { "epoch": 0.8623824451410659, "grad_norm": 18.551576614379883, "learning_rate": 1.3755000000000002e-06, "loss": 6.0155, "step": 2751 }, { "epoch": 0.8626959247648903, "grad_norm": 20.321107864379883, "learning_rate": 1.376e-06, "loss": 5.1774, "step": 2752 }, { "epoch": 0.8630094043887148, "grad_norm": 25.40889549255371, "learning_rate": 1.3765e-06, "loss": 5.933, "step": 2753 }, { "epoch": 0.8633228840125392, "grad_norm": 40.312984466552734, "learning_rate": 1.377e-06, "loss": 6.7176, "step": 2754 }, { "epoch": 0.8636363636363636, "grad_norm": 38.97128677368164, "learning_rate": 1.3775000000000002e-06, "loss": 5.2988, "step": 2755 }, { "epoch": 0.8639498432601881, "grad_norm": 39.15882110595703, "learning_rate": 1.3780000000000002e-06, "loss": 7.5587, "step": 2756 }, { "epoch": 0.8642633228840125, "grad_norm": 37.12295913696289, "learning_rate": 1.3785e-06, "loss": 5.3687, "step": 2757 }, { "epoch": 0.864576802507837, "grad_norm": 69.34944152832031, "learning_rate": 1.379e-06, "loss": 10.4011, "step": 2758 }, { "epoch": 0.8648902821316614, "grad_norm": 46.22438430786133, "learning_rate": 1.3795e-06, "loss": 7.7162, "step": 2759 }, { "epoch": 0.8652037617554859, "grad_norm": 21.386926651000977, "learning_rate": 1.3800000000000001e-06, "loss": 4.8139, "step": 2760 }, { "epoch": 0.8655172413793103, "grad_norm": 41.118896484375, "learning_rate": 1.3805000000000002e-06, "loss": 6.995, "step": 2761 }, { "epoch": 0.8658307210031349, "grad_norm": 29.520702362060547, "learning_rate": 1.3810000000000002e-06, "loss": 5.1273, "step": 2762 }, { "epoch": 0.8661442006269593, "grad_norm": 36.20766830444336, "learning_rate": 1.3815e-06, "loss": 6.9384, "step": 2763 }, { "epoch": 0.8664576802507837, "grad_norm": 26.937095642089844, "learning_rate": 1.382e-06, "loss": 6.1545, "step": 2764 }, { "epoch": 0.8667711598746082, "grad_norm": 61.82893753051758, "learning_rate": 1.3825000000000003e-06, "loss": 8.894, "step": 2765 }, { "epoch": 0.8670846394984326, "grad_norm": 27.976224899291992, "learning_rate": 1.3830000000000001e-06, "loss": 4.7557, "step": 2766 }, { "epoch": 0.8673981191222571, "grad_norm": 79.30841827392578, "learning_rate": 1.3835000000000002e-06, "loss": 8.8842, "step": 2767 }, { "epoch": 0.8677115987460815, "grad_norm": 32.45915985107422, "learning_rate": 1.384e-06, "loss": 5.9744, "step": 2768 }, { "epoch": 0.868025078369906, "grad_norm": 27.075918197631836, "learning_rate": 1.3845e-06, "loss": 4.6374, "step": 2769 }, { "epoch": 0.8683385579937304, "grad_norm": 60.21990203857422, "learning_rate": 1.3850000000000003e-06, "loss": 6.1392, "step": 2770 }, { "epoch": 0.8686520376175548, "grad_norm": 38.180335998535156, "learning_rate": 1.3855000000000001e-06, "loss": 5.6472, "step": 2771 }, { "epoch": 0.8689655172413793, "grad_norm": 44.778709411621094, "learning_rate": 1.3860000000000002e-06, "loss": 6.1988, "step": 2772 }, { "epoch": 0.8692789968652037, "grad_norm": 44.12003707885742, "learning_rate": 1.3865e-06, "loss": 6.2107, "step": 2773 }, { "epoch": 0.8695924764890283, "grad_norm": 115.1012954711914, "learning_rate": 1.387e-06, "loss": 10.4227, "step": 2774 }, { "epoch": 0.8699059561128527, "grad_norm": 47.72286605834961, "learning_rate": 1.3875000000000003e-06, "loss": 5.7027, "step": 2775 }, { "epoch": 0.8702194357366771, "grad_norm": 42.00877380371094, "learning_rate": 1.388e-06, "loss": 7.9404, "step": 2776 }, { "epoch": 0.8705329153605016, "grad_norm": 30.263233184814453, "learning_rate": 1.3885000000000001e-06, "loss": 5.9391, "step": 2777 }, { "epoch": 0.870846394984326, "grad_norm": 32.171573638916016, "learning_rate": 1.3890000000000002e-06, "loss": 5.6703, "step": 2778 }, { "epoch": 0.8711598746081505, "grad_norm": 29.78256607055664, "learning_rate": 1.3895e-06, "loss": 5.0913, "step": 2779 }, { "epoch": 0.8714733542319749, "grad_norm": 60.958702087402344, "learning_rate": 1.3900000000000002e-06, "loss": 8.0207, "step": 2780 }, { "epoch": 0.8717868338557994, "grad_norm": 43.16213607788086, "learning_rate": 1.3905000000000003e-06, "loss": 6.1787, "step": 2781 }, { "epoch": 0.8721003134796238, "grad_norm": 46.14605712890625, "learning_rate": 1.3910000000000001e-06, "loss": 5.719, "step": 2782 }, { "epoch": 0.8724137931034482, "grad_norm": 43.69407272338867, "learning_rate": 1.3915000000000001e-06, "loss": 7.7847, "step": 2783 }, { "epoch": 0.8727272727272727, "grad_norm": 41.964820861816406, "learning_rate": 1.392e-06, "loss": 5.3784, "step": 2784 }, { "epoch": 0.8730407523510971, "grad_norm": 33.25990295410156, "learning_rate": 1.3925000000000002e-06, "loss": 5.3528, "step": 2785 }, { "epoch": 0.8733542319749217, "grad_norm": 37.829498291015625, "learning_rate": 1.3930000000000003e-06, "loss": 5.1219, "step": 2786 }, { "epoch": 0.8736677115987461, "grad_norm": 42.88932418823242, "learning_rate": 1.3935e-06, "loss": 6.6543, "step": 2787 }, { "epoch": 0.8739811912225706, "grad_norm": 63.86898422241211, "learning_rate": 1.3940000000000001e-06, "loss": 9.7226, "step": 2788 }, { "epoch": 0.874294670846395, "grad_norm": 45.50909423828125, "learning_rate": 1.3945e-06, "loss": 7.3516, "step": 2789 }, { "epoch": 0.8746081504702194, "grad_norm": 40.1636962890625, "learning_rate": 1.3950000000000002e-06, "loss": 5.1074, "step": 2790 }, { "epoch": 0.8749216300940439, "grad_norm": 37.66518020629883, "learning_rate": 1.3955000000000002e-06, "loss": 5.9449, "step": 2791 }, { "epoch": 0.8752351097178683, "grad_norm": 77.52034759521484, "learning_rate": 1.396e-06, "loss": 8.5409, "step": 2792 }, { "epoch": 0.8755485893416928, "grad_norm": 45.52507019042969, "learning_rate": 1.3965e-06, "loss": 6.171, "step": 2793 }, { "epoch": 0.8758620689655172, "grad_norm": 41.04618835449219, "learning_rate": 1.397e-06, "loss": 5.7984, "step": 2794 }, { "epoch": 0.8761755485893417, "grad_norm": 57.51970672607422, "learning_rate": 1.3975000000000002e-06, "loss": 6.6904, "step": 2795 }, { "epoch": 0.8764890282131661, "grad_norm": 59.90044403076172, "learning_rate": 1.3980000000000002e-06, "loss": 6.9548, "step": 2796 }, { "epoch": 0.8768025078369905, "grad_norm": 37.544612884521484, "learning_rate": 1.3985e-06, "loss": 6.0155, "step": 2797 }, { "epoch": 0.8771159874608151, "grad_norm": 43.11286926269531, "learning_rate": 1.399e-06, "loss": 5.8831, "step": 2798 }, { "epoch": 0.8774294670846395, "grad_norm": 22.86078453063965, "learning_rate": 1.3995000000000001e-06, "loss": 3.8691, "step": 2799 }, { "epoch": 0.877742946708464, "grad_norm": 24.594881057739258, "learning_rate": 1.4000000000000001e-06, "loss": 4.9944, "step": 2800 }, { "epoch": 0.8780564263322884, "grad_norm": 32.79366683959961, "learning_rate": 1.4005000000000002e-06, "loss": 4.5094, "step": 2801 }, { "epoch": 0.8783699059561129, "grad_norm": 40.043540954589844, "learning_rate": 1.4010000000000002e-06, "loss": 5.0449, "step": 2802 }, { "epoch": 0.8786833855799373, "grad_norm": 39.7636604309082, "learning_rate": 1.4015e-06, "loss": 4.8512, "step": 2803 }, { "epoch": 0.8789968652037617, "grad_norm": 38.312355041503906, "learning_rate": 1.402e-06, "loss": 4.9508, "step": 2804 }, { "epoch": 0.8793103448275862, "grad_norm": 40.00060272216797, "learning_rate": 1.4025000000000003e-06, "loss": 6.2614, "step": 2805 }, { "epoch": 0.8796238244514106, "grad_norm": 38.401275634765625, "learning_rate": 1.4030000000000002e-06, "loss": 5.2398, "step": 2806 }, { "epoch": 0.8799373040752351, "grad_norm": 40.27410125732422, "learning_rate": 1.4035000000000002e-06, "loss": 5.9456, "step": 2807 }, { "epoch": 0.8802507836990595, "grad_norm": 28.530675888061523, "learning_rate": 1.404e-06, "loss": 4.5628, "step": 2808 }, { "epoch": 0.8805642633228841, "grad_norm": 39.466033935546875, "learning_rate": 1.4045e-06, "loss": 5.2629, "step": 2809 }, { "epoch": 0.8808777429467085, "grad_norm": 29.432931900024414, "learning_rate": 1.4050000000000003e-06, "loss": 4.6265, "step": 2810 }, { "epoch": 0.881191222570533, "grad_norm": 48.174129486083984, "learning_rate": 1.4055000000000001e-06, "loss": 4.853, "step": 2811 }, { "epoch": 0.8815047021943574, "grad_norm": 53.55119323730469, "learning_rate": 1.4060000000000002e-06, "loss": 7.2377, "step": 2812 }, { "epoch": 0.8818181818181818, "grad_norm": 36.38772964477539, "learning_rate": 1.4065e-06, "loss": 5.2177, "step": 2813 }, { "epoch": 0.8821316614420063, "grad_norm": 34.217830657958984, "learning_rate": 1.407e-06, "loss": 5.3052, "step": 2814 }, { "epoch": 0.8824451410658307, "grad_norm": 42.84861373901367, "learning_rate": 1.4075e-06, "loss": 5.6933, "step": 2815 }, { "epoch": 0.8827586206896552, "grad_norm": 53.72264862060547, "learning_rate": 1.4080000000000001e-06, "loss": 5.3901, "step": 2816 }, { "epoch": 0.8830721003134796, "grad_norm": 53.93974685668945, "learning_rate": 1.4085000000000002e-06, "loss": 5.2435, "step": 2817 }, { "epoch": 0.883385579937304, "grad_norm": 28.91839027404785, "learning_rate": 1.4090000000000002e-06, "loss": 3.7364, "step": 2818 }, { "epoch": 0.8836990595611285, "grad_norm": 36.89215850830078, "learning_rate": 1.4095e-06, "loss": 4.5822, "step": 2819 }, { "epoch": 0.8840125391849529, "grad_norm": 33.66713333129883, "learning_rate": 1.41e-06, "loss": 4.6832, "step": 2820 }, { "epoch": 0.8843260188087775, "grad_norm": 44.25896453857422, "learning_rate": 1.4105000000000003e-06, "loss": 6.3235, "step": 2821 }, { "epoch": 0.8846394984326019, "grad_norm": 34.358551025390625, "learning_rate": 1.4110000000000001e-06, "loss": 5.9712, "step": 2822 }, { "epoch": 0.8849529780564264, "grad_norm": 36.53178787231445, "learning_rate": 1.4115000000000002e-06, "loss": 5.4185, "step": 2823 }, { "epoch": 0.8852664576802508, "grad_norm": 61.49184036254883, "learning_rate": 1.412e-06, "loss": 6.4534, "step": 2824 }, { "epoch": 0.8855799373040752, "grad_norm": 96.07270812988281, "learning_rate": 1.4125e-06, "loss": 9.2125, "step": 2825 }, { "epoch": 0.8858934169278997, "grad_norm": 44.58479690551758, "learning_rate": 1.4130000000000003e-06, "loss": 4.9576, "step": 2826 }, { "epoch": 0.8862068965517241, "grad_norm": 52.979209899902344, "learning_rate": 1.4135e-06, "loss": 5.3499, "step": 2827 }, { "epoch": 0.8865203761755486, "grad_norm": 37.45901107788086, "learning_rate": 1.4140000000000001e-06, "loss": 4.4372, "step": 2828 }, { "epoch": 0.886833855799373, "grad_norm": 44.921173095703125, "learning_rate": 1.4145e-06, "loss": 5.9111, "step": 2829 }, { "epoch": 0.8871473354231975, "grad_norm": 38.52549362182617, "learning_rate": 1.415e-06, "loss": 4.8988, "step": 2830 }, { "epoch": 0.8874608150470219, "grad_norm": 48.08058166503906, "learning_rate": 1.4155000000000003e-06, "loss": 4.7856, "step": 2831 }, { "epoch": 0.8877742946708463, "grad_norm": 32.446327209472656, "learning_rate": 1.416e-06, "loss": 4.5528, "step": 2832 }, { "epoch": 0.8880877742946709, "grad_norm": 86.36289978027344, "learning_rate": 1.4165000000000001e-06, "loss": 7.3174, "step": 2833 }, { "epoch": 0.8884012539184953, "grad_norm": 64.09017944335938, "learning_rate": 1.417e-06, "loss": 6.8142, "step": 2834 }, { "epoch": 0.8887147335423198, "grad_norm": 52.754486083984375, "learning_rate": 1.4175e-06, "loss": 6.6618, "step": 2835 }, { "epoch": 0.8890282131661442, "grad_norm": 43.92601013183594, "learning_rate": 1.4180000000000002e-06, "loss": 5.7932, "step": 2836 }, { "epoch": 0.8893416927899687, "grad_norm": 39.43013000488281, "learning_rate": 1.4185e-06, "loss": 5.5289, "step": 2837 }, { "epoch": 0.8896551724137931, "grad_norm": 42.97591018676758, "learning_rate": 1.419e-06, "loss": 4.7558, "step": 2838 }, { "epoch": 0.8899686520376175, "grad_norm": 43.5150032043457, "learning_rate": 1.4195000000000001e-06, "loss": 4.532, "step": 2839 }, { "epoch": 0.890282131661442, "grad_norm": 40.31664276123047, "learning_rate": 1.42e-06, "loss": 5.3421, "step": 2840 }, { "epoch": 0.8905956112852664, "grad_norm": 34.56455612182617, "learning_rate": 1.4205000000000002e-06, "loss": 4.0237, "step": 2841 }, { "epoch": 0.8909090909090909, "grad_norm": 30.62310218811035, "learning_rate": 1.4210000000000002e-06, "loss": 4.2576, "step": 2842 }, { "epoch": 0.8912225705329153, "grad_norm": 40.746028900146484, "learning_rate": 1.4215e-06, "loss": 4.587, "step": 2843 }, { "epoch": 0.8915360501567398, "grad_norm": 31.952014923095703, "learning_rate": 1.4220000000000001e-06, "loss": 3.8408, "step": 2844 }, { "epoch": 0.8918495297805643, "grad_norm": 26.450883865356445, "learning_rate": 1.4225e-06, "loss": 4.4055, "step": 2845 }, { "epoch": 0.8921630094043888, "grad_norm": 47.278404235839844, "learning_rate": 1.4230000000000002e-06, "loss": 4.9155, "step": 2846 }, { "epoch": 0.8924764890282132, "grad_norm": 55.63157272338867, "learning_rate": 1.4235000000000002e-06, "loss": 5.8403, "step": 2847 }, { "epoch": 0.8927899686520376, "grad_norm": 30.85458755493164, "learning_rate": 1.424e-06, "loss": 4.5499, "step": 2848 }, { "epoch": 0.8931034482758621, "grad_norm": 36.83681106567383, "learning_rate": 1.4245e-06, "loss": 5.0289, "step": 2849 }, { "epoch": 0.8934169278996865, "grad_norm": 32.328121185302734, "learning_rate": 1.425e-06, "loss": 4.2722, "step": 2850 }, { "epoch": 0.893730407523511, "grad_norm": 44.29472732543945, "learning_rate": 1.4255000000000002e-06, "loss": 4.4436, "step": 2851 }, { "epoch": 0.8940438871473354, "grad_norm": 46.61187744140625, "learning_rate": 1.4260000000000002e-06, "loss": 5.3737, "step": 2852 }, { "epoch": 0.8943573667711598, "grad_norm": 35.942100524902344, "learning_rate": 1.4265e-06, "loss": 4.4256, "step": 2853 }, { "epoch": 0.8946708463949843, "grad_norm": 32.366519927978516, "learning_rate": 1.427e-06, "loss": 3.8425, "step": 2854 }, { "epoch": 0.8949843260188087, "grad_norm": 35.66120529174805, "learning_rate": 1.4275e-06, "loss": 4.4804, "step": 2855 }, { "epoch": 0.8952978056426333, "grad_norm": 37.548492431640625, "learning_rate": 1.4280000000000001e-06, "loss": 5.4055, "step": 2856 }, { "epoch": 0.8956112852664577, "grad_norm": 41.95212936401367, "learning_rate": 1.4285000000000002e-06, "loss": 4.8837, "step": 2857 }, { "epoch": 0.8959247648902822, "grad_norm": 34.09566879272461, "learning_rate": 1.4290000000000002e-06, "loss": 4.8557, "step": 2858 }, { "epoch": 0.8962382445141066, "grad_norm": 52.16231918334961, "learning_rate": 1.4295e-06, "loss": 4.9953, "step": 2859 }, { "epoch": 0.896551724137931, "grad_norm": 38.20033264160156, "learning_rate": 1.43e-06, "loss": 4.2838, "step": 2860 }, { "epoch": 0.8968652037617555, "grad_norm": 52.53467559814453, "learning_rate": 1.4305000000000003e-06, "loss": 5.1196, "step": 2861 }, { "epoch": 0.8971786833855799, "grad_norm": 68.67937469482422, "learning_rate": 1.4310000000000001e-06, "loss": 5.4295, "step": 2862 }, { "epoch": 0.8974921630094044, "grad_norm": 39.02657699584961, "learning_rate": 1.4315000000000002e-06, "loss": 3.6579, "step": 2863 }, { "epoch": 0.8978056426332288, "grad_norm": 27.187744140625, "learning_rate": 1.432e-06, "loss": 3.7283, "step": 2864 }, { "epoch": 0.8981191222570533, "grad_norm": 42.428802490234375, "learning_rate": 1.4325e-06, "loss": 4.022, "step": 2865 }, { "epoch": 0.8984326018808777, "grad_norm": 29.769405364990234, "learning_rate": 1.4330000000000003e-06, "loss": 3.7161, "step": 2866 }, { "epoch": 0.8987460815047021, "grad_norm": 40.81663513183594, "learning_rate": 1.4335000000000001e-06, "loss": 4.9217, "step": 2867 }, { "epoch": 0.8990595611285267, "grad_norm": 48.31073760986328, "learning_rate": 1.4340000000000002e-06, "loss": 4.4489, "step": 2868 }, { "epoch": 0.8993730407523511, "grad_norm": 32.35671615600586, "learning_rate": 1.4345e-06, "loss": 4.1166, "step": 2869 }, { "epoch": 0.8996865203761756, "grad_norm": 46.619041442871094, "learning_rate": 1.435e-06, "loss": 6.4497, "step": 2870 }, { "epoch": 0.9, "grad_norm": 56.73939514160156, "learning_rate": 1.4355000000000003e-06, "loss": 5.7911, "step": 2871 }, { "epoch": 0.9003134796238245, "grad_norm": 31.99533462524414, "learning_rate": 1.436e-06, "loss": 3.6651, "step": 2872 }, { "epoch": 0.9006269592476489, "grad_norm": 37.6666374206543, "learning_rate": 1.4365000000000001e-06, "loss": 3.6502, "step": 2873 }, { "epoch": 0.9009404388714733, "grad_norm": 23.286277770996094, "learning_rate": 1.437e-06, "loss": 3.7086, "step": 2874 }, { "epoch": 0.9012539184952978, "grad_norm": 42.88998794555664, "learning_rate": 1.4375e-06, "loss": 5.5071, "step": 2875 }, { "epoch": 0.9015673981191222, "grad_norm": 37.3940315246582, "learning_rate": 1.4380000000000003e-06, "loss": 5.1036, "step": 2876 }, { "epoch": 0.9018808777429467, "grad_norm": 48.386104583740234, "learning_rate": 1.4385e-06, "loss": 4.5443, "step": 2877 }, { "epoch": 0.9021943573667711, "grad_norm": 63.84465408325195, "learning_rate": 1.4390000000000001e-06, "loss": 6.2741, "step": 2878 }, { "epoch": 0.9025078369905956, "grad_norm": 32.616859436035156, "learning_rate": 1.4395000000000002e-06, "loss": 4.203, "step": 2879 }, { "epoch": 0.9028213166144201, "grad_norm": 42.289005279541016, "learning_rate": 1.44e-06, "loss": 5.2399, "step": 2880 }, { "epoch": 0.9031347962382446, "grad_norm": 62.790042877197266, "learning_rate": 1.4405000000000002e-06, "loss": 4.648, "step": 2881 }, { "epoch": 0.903448275862069, "grad_norm": 51.915889739990234, "learning_rate": 1.4410000000000003e-06, "loss": 5.2985, "step": 2882 }, { "epoch": 0.9037617554858934, "grad_norm": 75.9524154663086, "learning_rate": 1.4415e-06, "loss": 6.2667, "step": 2883 }, { "epoch": 0.9040752351097179, "grad_norm": 28.834205627441406, "learning_rate": 1.4420000000000001e-06, "loss": 3.4066, "step": 2884 }, { "epoch": 0.9043887147335423, "grad_norm": 51.43450927734375, "learning_rate": 1.4425e-06, "loss": 6.0204, "step": 2885 }, { "epoch": 0.9047021943573668, "grad_norm": 76.39936065673828, "learning_rate": 1.4430000000000002e-06, "loss": 5.3644, "step": 2886 }, { "epoch": 0.9050156739811912, "grad_norm": 39.322025299072266, "learning_rate": 1.4435000000000002e-06, "loss": 4.0316, "step": 2887 }, { "epoch": 0.9053291536050156, "grad_norm": 30.268177032470703, "learning_rate": 1.444e-06, "loss": 4.4225, "step": 2888 }, { "epoch": 0.9056426332288401, "grad_norm": 32.31449508666992, "learning_rate": 1.4445e-06, "loss": 4.538, "step": 2889 }, { "epoch": 0.9059561128526645, "grad_norm": 26.894195556640625, "learning_rate": 1.445e-06, "loss": 3.7654, "step": 2890 }, { "epoch": 0.9062695924764891, "grad_norm": 67.31766510009766, "learning_rate": 1.4455000000000002e-06, "loss": 6.158, "step": 2891 }, { "epoch": 0.9065830721003135, "grad_norm": 24.468050003051758, "learning_rate": 1.4460000000000002e-06, "loss": 3.9294, "step": 2892 }, { "epoch": 0.906896551724138, "grad_norm": 47.52812576293945, "learning_rate": 1.4465e-06, "loss": 5.1586, "step": 2893 }, { "epoch": 0.9072100313479624, "grad_norm": 35.4903450012207, "learning_rate": 1.447e-06, "loss": 4.418, "step": 2894 }, { "epoch": 0.9075235109717869, "grad_norm": 31.2788028717041, "learning_rate": 1.4475000000000001e-06, "loss": 3.6287, "step": 2895 }, { "epoch": 0.9078369905956113, "grad_norm": 31.898345947265625, "learning_rate": 1.4480000000000002e-06, "loss": 3.8173, "step": 2896 }, { "epoch": 0.9081504702194357, "grad_norm": 59.66981506347656, "learning_rate": 1.4485000000000002e-06, "loss": 5.177, "step": 2897 }, { "epoch": 0.9084639498432602, "grad_norm": 50.7946891784668, "learning_rate": 1.4490000000000002e-06, "loss": 5.3483, "step": 2898 }, { "epoch": 0.9087774294670846, "grad_norm": 27.841205596923828, "learning_rate": 1.4495e-06, "loss": 3.2566, "step": 2899 }, { "epoch": 0.9090909090909091, "grad_norm": 29.948871612548828, "learning_rate": 1.45e-06, "loss": 4.0416, "step": 2900 }, { "epoch": 0.9094043887147335, "grad_norm": 23.60964584350586, "learning_rate": 1.4505000000000003e-06, "loss": 3.1585, "step": 2901 }, { "epoch": 0.909717868338558, "grad_norm": 33.11689376831055, "learning_rate": 1.4510000000000002e-06, "loss": 4.4712, "step": 2902 }, { "epoch": 0.9100313479623825, "grad_norm": 62.05201721191406, "learning_rate": 1.4515000000000002e-06, "loss": 5.4037, "step": 2903 }, { "epoch": 0.9103448275862069, "grad_norm": 105.85570526123047, "learning_rate": 1.452e-06, "loss": 5.0095, "step": 2904 }, { "epoch": 0.9106583072100314, "grad_norm": 45.89641189575195, "learning_rate": 1.4525e-06, "loss": 3.9751, "step": 2905 }, { "epoch": 0.9109717868338558, "grad_norm": 52.051944732666016, "learning_rate": 1.4530000000000003e-06, "loss": 4.7415, "step": 2906 }, { "epoch": 0.9112852664576803, "grad_norm": 67.7939682006836, "learning_rate": 1.4535000000000001e-06, "loss": 4.7054, "step": 2907 }, { "epoch": 0.9115987460815047, "grad_norm": 41.19660949707031, "learning_rate": 1.4540000000000002e-06, "loss": 5.2859, "step": 2908 }, { "epoch": 0.9119122257053291, "grad_norm": 46.74116516113281, "learning_rate": 1.4545e-06, "loss": 4.0541, "step": 2909 }, { "epoch": 0.9122257053291536, "grad_norm": 84.27909088134766, "learning_rate": 1.455e-06, "loss": 5.2779, "step": 2910 }, { "epoch": 0.912539184952978, "grad_norm": 53.99125289916992, "learning_rate": 1.4555000000000003e-06, "loss": 4.3379, "step": 2911 }, { "epoch": 0.9128526645768025, "grad_norm": 52.822410583496094, "learning_rate": 1.4560000000000001e-06, "loss": 5.3916, "step": 2912 }, { "epoch": 0.9131661442006269, "grad_norm": 53.39067840576172, "learning_rate": 1.4565000000000002e-06, "loss": 5.6752, "step": 2913 }, { "epoch": 0.9134796238244514, "grad_norm": 42.46992111206055, "learning_rate": 1.457e-06, "loss": 3.9782, "step": 2914 }, { "epoch": 0.9137931034482759, "grad_norm": 31.098447799682617, "learning_rate": 1.4575e-06, "loss": 3.5816, "step": 2915 }, { "epoch": 0.9141065830721004, "grad_norm": 49.15898513793945, "learning_rate": 1.4580000000000003e-06, "loss": 3.5684, "step": 2916 }, { "epoch": 0.9144200626959248, "grad_norm": 49.75162124633789, "learning_rate": 1.4585e-06, "loss": 4.8455, "step": 2917 }, { "epoch": 0.9147335423197492, "grad_norm": 51.76414108276367, "learning_rate": 1.4590000000000001e-06, "loss": 4.1124, "step": 2918 }, { "epoch": 0.9150470219435737, "grad_norm": 41.26679229736328, "learning_rate": 1.4595000000000002e-06, "loss": 4.2793, "step": 2919 }, { "epoch": 0.9153605015673981, "grad_norm": 43.92702865600586, "learning_rate": 1.46e-06, "loss": 4.6872, "step": 2920 }, { "epoch": 0.9156739811912226, "grad_norm": 34.27779769897461, "learning_rate": 1.4605000000000002e-06, "loss": 3.8484, "step": 2921 }, { "epoch": 0.915987460815047, "grad_norm": 31.705305099487305, "learning_rate": 1.4610000000000003e-06, "loss": 4.3887, "step": 2922 }, { "epoch": 0.9163009404388714, "grad_norm": 53.87886428833008, "learning_rate": 1.4615000000000001e-06, "loss": 4.4851, "step": 2923 }, { "epoch": 0.9166144200626959, "grad_norm": 43.97981262207031, "learning_rate": 1.4620000000000001e-06, "loss": 3.952, "step": 2924 }, { "epoch": 0.9169278996865203, "grad_norm": 36.8157958984375, "learning_rate": 1.4625e-06, "loss": 3.4762, "step": 2925 }, { "epoch": 0.9172413793103448, "grad_norm": 35.28675079345703, "learning_rate": 1.4630000000000002e-06, "loss": 4.2781, "step": 2926 }, { "epoch": 0.9175548589341693, "grad_norm": 37.354148864746094, "learning_rate": 1.4635000000000003e-06, "loss": 3.6935, "step": 2927 }, { "epoch": 0.9178683385579938, "grad_norm": 61.448665618896484, "learning_rate": 1.464e-06, "loss": 5.5435, "step": 2928 }, { "epoch": 0.9181818181818182, "grad_norm": 33.78268814086914, "learning_rate": 1.4645000000000001e-06, "loss": 4.5201, "step": 2929 }, { "epoch": 0.9184952978056427, "grad_norm": 37.02507019042969, "learning_rate": 1.465e-06, "loss": 3.5038, "step": 2930 }, { "epoch": 0.9188087774294671, "grad_norm": 29.58689308166504, "learning_rate": 1.4655000000000002e-06, "loss": 3.4179, "step": 2931 }, { "epoch": 0.9191222570532915, "grad_norm": 56.62324142456055, "learning_rate": 1.4660000000000002e-06, "loss": 3.9863, "step": 2932 }, { "epoch": 0.919435736677116, "grad_norm": 30.186080932617188, "learning_rate": 1.4665e-06, "loss": 3.6394, "step": 2933 }, { "epoch": 0.9197492163009404, "grad_norm": 91.04447937011719, "learning_rate": 1.467e-06, "loss": 5.1371, "step": 2934 }, { "epoch": 0.9200626959247649, "grad_norm": 94.30821990966797, "learning_rate": 1.4675000000000001e-06, "loss": 4.2498, "step": 2935 }, { "epoch": 0.9203761755485893, "grad_norm": 42.256168365478516, "learning_rate": 1.4680000000000002e-06, "loss": 4.5557, "step": 2936 }, { "epoch": 0.9206896551724137, "grad_norm": 38.17436599731445, "learning_rate": 1.4685000000000002e-06, "loss": 4.0215, "step": 2937 }, { "epoch": 0.9210031347962383, "grad_norm": 38.624698638916016, "learning_rate": 1.4690000000000003e-06, "loss": 3.5599, "step": 2938 }, { "epoch": 0.9213166144200627, "grad_norm": 38.51246643066406, "learning_rate": 1.4695e-06, "loss": 3.5211, "step": 2939 }, { "epoch": 0.9216300940438872, "grad_norm": 37.080074310302734, "learning_rate": 1.4700000000000001e-06, "loss": 5.7069, "step": 2940 }, { "epoch": 0.9219435736677116, "grad_norm": 37.63076400756836, "learning_rate": 1.4705e-06, "loss": 4.2385, "step": 2941 }, { "epoch": 0.9222570532915361, "grad_norm": 82.26602935791016, "learning_rate": 1.4710000000000002e-06, "loss": 3.6196, "step": 2942 }, { "epoch": 0.9225705329153605, "grad_norm": 44.46656036376953, "learning_rate": 1.4715000000000002e-06, "loss": 4.0862, "step": 2943 }, { "epoch": 0.922884012539185, "grad_norm": 34.88580322265625, "learning_rate": 1.472e-06, "loss": 3.5532, "step": 2944 }, { "epoch": 0.9231974921630094, "grad_norm": 41.69892501831055, "learning_rate": 1.4725e-06, "loss": 4.2038, "step": 2945 }, { "epoch": 0.9235109717868338, "grad_norm": 48.94770431518555, "learning_rate": 1.473e-06, "loss": 4.1548, "step": 2946 }, { "epoch": 0.9238244514106583, "grad_norm": 30.48263168334961, "learning_rate": 1.4735000000000002e-06, "loss": 3.5391, "step": 2947 }, { "epoch": 0.9241379310344827, "grad_norm": 37.1695671081543, "learning_rate": 1.4740000000000002e-06, "loss": 3.9293, "step": 2948 }, { "epoch": 0.9244514106583072, "grad_norm": 132.91793823242188, "learning_rate": 1.4745e-06, "loss": 4.7049, "step": 2949 }, { "epoch": 0.9247648902821317, "grad_norm": 52.81797790527344, "learning_rate": 1.475e-06, "loss": 4.619, "step": 2950 }, { "epoch": 0.9250783699059562, "grad_norm": 27.379749298095703, "learning_rate": 1.4754999999999999e-06, "loss": 3.5159, "step": 2951 }, { "epoch": 0.9253918495297806, "grad_norm": 37.0819206237793, "learning_rate": 1.4760000000000001e-06, "loss": 5.0518, "step": 2952 }, { "epoch": 0.925705329153605, "grad_norm": 60.48858642578125, "learning_rate": 1.4765000000000002e-06, "loss": 4.3286, "step": 2953 }, { "epoch": 0.9260188087774295, "grad_norm": 60.34362030029297, "learning_rate": 1.477e-06, "loss": 4.1523, "step": 2954 }, { "epoch": 0.9263322884012539, "grad_norm": 47.12152862548828, "learning_rate": 1.4775e-06, "loss": 3.5215, "step": 2955 }, { "epoch": 0.9266457680250784, "grad_norm": 33.692054748535156, "learning_rate": 1.478e-06, "loss": 3.7666, "step": 2956 }, { "epoch": 0.9269592476489028, "grad_norm": 32.28064727783203, "learning_rate": 1.4785000000000001e-06, "loss": 3.2797, "step": 2957 }, { "epoch": 0.9272727272727272, "grad_norm": 47.644588470458984, "learning_rate": 1.4790000000000002e-06, "loss": 4.0666, "step": 2958 }, { "epoch": 0.9275862068965517, "grad_norm": 37.37916564941406, "learning_rate": 1.4795000000000002e-06, "loss": 3.9653, "step": 2959 }, { "epoch": 0.9278996865203761, "grad_norm": 34.87968444824219, "learning_rate": 1.48e-06, "loss": 3.4317, "step": 2960 }, { "epoch": 0.9282131661442006, "grad_norm": 75.75358581542969, "learning_rate": 1.4805e-06, "loss": 5.1771, "step": 2961 }, { "epoch": 0.9285266457680251, "grad_norm": 39.766578674316406, "learning_rate": 1.4810000000000003e-06, "loss": 3.8669, "step": 2962 }, { "epoch": 0.9288401253918496, "grad_norm": 69.24491119384766, "learning_rate": 1.4815000000000001e-06, "loss": 4.7225, "step": 2963 }, { "epoch": 0.929153605015674, "grad_norm": 84.43096923828125, "learning_rate": 1.4820000000000002e-06, "loss": 4.2543, "step": 2964 }, { "epoch": 0.9294670846394985, "grad_norm": 41.02848434448242, "learning_rate": 1.4825e-06, "loss": 4.1251, "step": 2965 }, { "epoch": 0.9297805642633229, "grad_norm": 41.528892517089844, "learning_rate": 1.483e-06, "loss": 3.6634, "step": 2966 }, { "epoch": 0.9300940438871473, "grad_norm": 28.742746353149414, "learning_rate": 1.4835000000000003e-06, "loss": 4.2559, "step": 2967 }, { "epoch": 0.9304075235109718, "grad_norm": 52.95411682128906, "learning_rate": 1.4840000000000001e-06, "loss": 3.8098, "step": 2968 }, { "epoch": 0.9307210031347962, "grad_norm": 48.885982513427734, "learning_rate": 1.4845000000000001e-06, "loss": 4.4849, "step": 2969 }, { "epoch": 0.9310344827586207, "grad_norm": 36.8331298828125, "learning_rate": 1.485e-06, "loss": 4.2251, "step": 2970 }, { "epoch": 0.9313479623824451, "grad_norm": 33.98138427734375, "learning_rate": 1.4855e-06, "loss": 3.6462, "step": 2971 }, { "epoch": 0.9316614420062695, "grad_norm": 33.16653823852539, "learning_rate": 1.4860000000000003e-06, "loss": 3.4936, "step": 2972 }, { "epoch": 0.931974921630094, "grad_norm": 23.719505310058594, "learning_rate": 1.4865e-06, "loss": 3.3029, "step": 2973 }, { "epoch": 0.9322884012539185, "grad_norm": 34.3863410949707, "learning_rate": 1.4870000000000001e-06, "loss": 4.0512, "step": 2974 }, { "epoch": 0.932601880877743, "grad_norm": 31.7304744720459, "learning_rate": 1.4875000000000002e-06, "loss": 3.8679, "step": 2975 }, { "epoch": 0.9329153605015674, "grad_norm": 36.177642822265625, "learning_rate": 1.488e-06, "loss": 3.5125, "step": 2976 }, { "epoch": 0.9332288401253919, "grad_norm": 48.70751953125, "learning_rate": 1.4885000000000002e-06, "loss": 4.5736, "step": 2977 }, { "epoch": 0.9335423197492163, "grad_norm": 47.91338348388672, "learning_rate": 1.4890000000000003e-06, "loss": 4.3447, "step": 2978 }, { "epoch": 0.9338557993730408, "grad_norm": 38.861026763916016, "learning_rate": 1.4895e-06, "loss": 3.2831, "step": 2979 }, { "epoch": 0.9341692789968652, "grad_norm": 31.002511978149414, "learning_rate": 1.4900000000000001e-06, "loss": 3.899, "step": 2980 }, { "epoch": 0.9344827586206896, "grad_norm": 45.859554290771484, "learning_rate": 1.4905e-06, "loss": 3.303, "step": 2981 }, { "epoch": 0.9347962382445141, "grad_norm": 37.03654479980469, "learning_rate": 1.4910000000000002e-06, "loss": 3.4903, "step": 2982 }, { "epoch": 0.9351097178683385, "grad_norm": 51.43019485473633, "learning_rate": 1.4915000000000002e-06, "loss": 3.4536, "step": 2983 }, { "epoch": 0.935423197492163, "grad_norm": 24.4971866607666, "learning_rate": 1.492e-06, "loss": 3.258, "step": 2984 }, { "epoch": 0.9357366771159875, "grad_norm": 35.39488220214844, "learning_rate": 1.4925000000000001e-06, "loss": 3.8771, "step": 2985 }, { "epoch": 0.936050156739812, "grad_norm": 33.246829986572266, "learning_rate": 1.493e-06, "loss": 3.5061, "step": 2986 }, { "epoch": 0.9363636363636364, "grad_norm": 80.51388549804688, "learning_rate": 1.4935000000000002e-06, "loss": 3.0297, "step": 2987 }, { "epoch": 0.9366771159874608, "grad_norm": 32.59467315673828, "learning_rate": 1.4940000000000002e-06, "loss": 3.5332, "step": 2988 }, { "epoch": 0.9369905956112853, "grad_norm": 35.27291488647461, "learning_rate": 1.4945e-06, "loss": 3.4688, "step": 2989 }, { "epoch": 0.9373040752351097, "grad_norm": 34.252323150634766, "learning_rate": 1.495e-06, "loss": 4.57, "step": 2990 }, { "epoch": 0.9376175548589342, "grad_norm": 38.402591705322266, "learning_rate": 1.4955e-06, "loss": 3.7867, "step": 2991 }, { "epoch": 0.9379310344827586, "grad_norm": 60.38703918457031, "learning_rate": 1.4960000000000002e-06, "loss": 3.5518, "step": 2992 }, { "epoch": 0.938244514106583, "grad_norm": 49.44574737548828, "learning_rate": 1.4965000000000002e-06, "loss": 4.4428, "step": 2993 }, { "epoch": 0.9385579937304075, "grad_norm": 36.240272521972656, "learning_rate": 1.497e-06, "loss": 3.9223, "step": 2994 }, { "epoch": 0.9388714733542319, "grad_norm": 43.761146545410156, "learning_rate": 1.4975e-06, "loss": 4.0446, "step": 2995 }, { "epoch": 0.9391849529780564, "grad_norm": 37.836631774902344, "learning_rate": 1.498e-06, "loss": 4.5072, "step": 2996 }, { "epoch": 0.9394984326018809, "grad_norm": 43.07557678222656, "learning_rate": 1.4985000000000001e-06, "loss": 3.9492, "step": 2997 }, { "epoch": 0.9398119122257054, "grad_norm": 45.68537521362305, "learning_rate": 1.4990000000000002e-06, "loss": 3.9818, "step": 2998 }, { "epoch": 0.9401253918495298, "grad_norm": 39.437374114990234, "learning_rate": 1.4995000000000002e-06, "loss": 3.8931, "step": 2999 }, { "epoch": 0.9404388714733543, "grad_norm": 43.97903823852539, "learning_rate": 1.5e-06, "loss": 3.9224, "step": 3000 }, { "epoch": 0.9407523510971787, "grad_norm": 39.799964904785156, "learning_rate": 1.5005e-06, "loss": 4.0903, "step": 3001 }, { "epoch": 0.9410658307210031, "grad_norm": 31.609031677246094, "learning_rate": 1.5010000000000003e-06, "loss": 3.1316, "step": 3002 }, { "epoch": 0.9413793103448276, "grad_norm": 44.23056411743164, "learning_rate": 1.5015000000000002e-06, "loss": 3.8267, "step": 3003 }, { "epoch": 0.941692789968652, "grad_norm": 54.30805206298828, "learning_rate": 1.5020000000000002e-06, "loss": 4.5096, "step": 3004 }, { "epoch": 0.9420062695924765, "grad_norm": 35.796661376953125, "learning_rate": 1.5025e-06, "loss": 3.364, "step": 3005 }, { "epoch": 0.9423197492163009, "grad_norm": 38.125091552734375, "learning_rate": 1.503e-06, "loss": 3.3353, "step": 3006 }, { "epoch": 0.9426332288401253, "grad_norm": 55.130775451660156, "learning_rate": 1.5035000000000003e-06, "loss": 3.269, "step": 3007 }, { "epoch": 0.9429467084639498, "grad_norm": 33.1065788269043, "learning_rate": 1.5040000000000001e-06, "loss": 3.8043, "step": 3008 }, { "epoch": 0.9432601880877743, "grad_norm": 32.7559814453125, "learning_rate": 1.5045000000000002e-06, "loss": 3.3734, "step": 3009 }, { "epoch": 0.9435736677115988, "grad_norm": 54.44166946411133, "learning_rate": 1.505e-06, "loss": 3.8441, "step": 3010 }, { "epoch": 0.9438871473354232, "grad_norm": 42.8200569152832, "learning_rate": 1.5055e-06, "loss": 3.2862, "step": 3011 }, { "epoch": 0.9442006269592477, "grad_norm": 33.59670639038086, "learning_rate": 1.5060000000000003e-06, "loss": 3.8924, "step": 3012 }, { "epoch": 0.9445141065830721, "grad_norm": 39.061279296875, "learning_rate": 1.5065e-06, "loss": 3.7316, "step": 3013 }, { "epoch": 0.9448275862068966, "grad_norm": 37.6108283996582, "learning_rate": 1.5070000000000001e-06, "loss": 4.0276, "step": 3014 }, { "epoch": 0.945141065830721, "grad_norm": 31.421833038330078, "learning_rate": 1.5075000000000002e-06, "loss": 3.2679, "step": 3015 }, { "epoch": 0.9454545454545454, "grad_norm": 34.2425422668457, "learning_rate": 1.508e-06, "loss": 3.7938, "step": 3016 }, { "epoch": 0.9457680250783699, "grad_norm": 58.670650482177734, "learning_rate": 1.5085000000000003e-06, "loss": 4.0145, "step": 3017 }, { "epoch": 0.9460815047021943, "grad_norm": 20.024412155151367, "learning_rate": 1.5090000000000003e-06, "loss": 3.1288, "step": 3018 }, { "epoch": 0.9463949843260188, "grad_norm": 58.533390045166016, "learning_rate": 1.5095000000000001e-06, "loss": 3.3225, "step": 3019 }, { "epoch": 0.9467084639498433, "grad_norm": 54.33335876464844, "learning_rate": 1.5100000000000002e-06, "loss": 6.4205, "step": 3020 }, { "epoch": 0.9470219435736678, "grad_norm": 48.962615966796875, "learning_rate": 1.5105e-06, "loss": 4.808, "step": 3021 }, { "epoch": 0.9473354231974922, "grad_norm": 45.55655288696289, "learning_rate": 1.5110000000000002e-06, "loss": 4.7946, "step": 3022 }, { "epoch": 0.9476489028213166, "grad_norm": 50.928199768066406, "learning_rate": 1.5115000000000003e-06, "loss": 2.9686, "step": 3023 }, { "epoch": 0.9479623824451411, "grad_norm": 48.291133880615234, "learning_rate": 1.512e-06, "loss": 4.4833, "step": 3024 }, { "epoch": 0.9482758620689655, "grad_norm": 53.17333221435547, "learning_rate": 1.5125000000000001e-06, "loss": 3.8591, "step": 3025 }, { "epoch": 0.94858934169279, "grad_norm": 39.857887268066406, "learning_rate": 1.513e-06, "loss": 3.6312, "step": 3026 }, { "epoch": 0.9489028213166144, "grad_norm": 54.0508918762207, "learning_rate": 1.5135000000000002e-06, "loss": 3.8041, "step": 3027 }, { "epoch": 0.9492163009404389, "grad_norm": 80.95004272460938, "learning_rate": 1.5140000000000002e-06, "loss": 5.0221, "step": 3028 }, { "epoch": 0.9495297805642633, "grad_norm": 49.28989791870117, "learning_rate": 1.5145e-06, "loss": 3.7739, "step": 3029 }, { "epoch": 0.9498432601880877, "grad_norm": 47.68058395385742, "learning_rate": 1.5150000000000001e-06, "loss": 4.321, "step": 3030 }, { "epoch": 0.9501567398119122, "grad_norm": 38.63204574584961, "learning_rate": 1.5155e-06, "loss": 3.6415, "step": 3031 }, { "epoch": 0.9504702194357367, "grad_norm": 71.45962524414062, "learning_rate": 1.5160000000000002e-06, "loss": 4.3374, "step": 3032 }, { "epoch": 0.9507836990595612, "grad_norm": 33.07627868652344, "learning_rate": 1.5165000000000002e-06, "loss": 4.8849, "step": 3033 }, { "epoch": 0.9510971786833856, "grad_norm": 43.36479187011719, "learning_rate": 1.517e-06, "loss": 4.4194, "step": 3034 }, { "epoch": 0.95141065830721, "grad_norm": 39.327003479003906, "learning_rate": 1.5175e-06, "loss": 3.8361, "step": 3035 }, { "epoch": 0.9517241379310345, "grad_norm": 42.822105407714844, "learning_rate": 1.5180000000000001e-06, "loss": 3.613, "step": 3036 }, { "epoch": 0.9520376175548589, "grad_norm": 36.2182502746582, "learning_rate": 1.5185000000000002e-06, "loss": 2.9997, "step": 3037 }, { "epoch": 0.9523510971786834, "grad_norm": 53.4159049987793, "learning_rate": 1.5190000000000002e-06, "loss": 3.312, "step": 3038 }, { "epoch": 0.9526645768025078, "grad_norm": 31.3487491607666, "learning_rate": 1.5195000000000002e-06, "loss": 3.2757, "step": 3039 }, { "epoch": 0.9529780564263323, "grad_norm": 30.916259765625, "learning_rate": 1.52e-06, "loss": 3.5675, "step": 3040 }, { "epoch": 0.9532915360501567, "grad_norm": 38.4372673034668, "learning_rate": 1.5205e-06, "loss": 3.7267, "step": 3041 }, { "epoch": 0.9536050156739811, "grad_norm": 33.22764587402344, "learning_rate": 1.5210000000000003e-06, "loss": 4.0307, "step": 3042 }, { "epoch": 0.9539184952978056, "grad_norm": 44.29248809814453, "learning_rate": 1.5215000000000002e-06, "loss": 4.2784, "step": 3043 }, { "epoch": 0.9542319749216301, "grad_norm": 28.41790199279785, "learning_rate": 1.5220000000000002e-06, "loss": 4.0886, "step": 3044 }, { "epoch": 0.9545454545454546, "grad_norm": 27.84803009033203, "learning_rate": 1.5225e-06, "loss": 3.059, "step": 3045 }, { "epoch": 0.954858934169279, "grad_norm": 64.40411376953125, "learning_rate": 1.523e-06, "loss": 4.3402, "step": 3046 }, { "epoch": 0.9551724137931035, "grad_norm": 41.377769470214844, "learning_rate": 1.5235000000000003e-06, "loss": 4.1004, "step": 3047 }, { "epoch": 0.9554858934169279, "grad_norm": 28.825105667114258, "learning_rate": 1.5240000000000001e-06, "loss": 3.5855, "step": 3048 }, { "epoch": 0.9557993730407524, "grad_norm": 36.527618408203125, "learning_rate": 1.5245000000000002e-06, "loss": 3.7634, "step": 3049 }, { "epoch": 0.9561128526645768, "grad_norm": 48.31051254272461, "learning_rate": 1.525e-06, "loss": 3.2032, "step": 3050 }, { "epoch": 0.9564263322884012, "grad_norm": 49.092552185058594, "learning_rate": 1.5255e-06, "loss": 3.2192, "step": 3051 }, { "epoch": 0.9567398119122257, "grad_norm": 28.44539451599121, "learning_rate": 1.5260000000000003e-06, "loss": 3.1739, "step": 3052 }, { "epoch": 0.9570532915360501, "grad_norm": 48.597190856933594, "learning_rate": 1.5265000000000001e-06, "loss": 5.3544, "step": 3053 }, { "epoch": 0.9573667711598746, "grad_norm": 37.88104248046875, "learning_rate": 1.5270000000000002e-06, "loss": 3.1668, "step": 3054 }, { "epoch": 0.957680250783699, "grad_norm": 88.76260375976562, "learning_rate": 1.5275000000000002e-06, "loss": 4.3589, "step": 3055 }, { "epoch": 0.9579937304075236, "grad_norm": 31.21636962890625, "learning_rate": 1.528e-06, "loss": 3.3685, "step": 3056 }, { "epoch": 0.958307210031348, "grad_norm": 44.11945343017578, "learning_rate": 1.5285000000000003e-06, "loss": 3.3106, "step": 3057 }, { "epoch": 0.9586206896551724, "grad_norm": 46.66257095336914, "learning_rate": 1.5290000000000003e-06, "loss": 3.5772, "step": 3058 }, { "epoch": 0.9589341692789969, "grad_norm": 45.46147537231445, "learning_rate": 1.5295000000000001e-06, "loss": 6.1372, "step": 3059 }, { "epoch": 0.9592476489028213, "grad_norm": 37.89000701904297, "learning_rate": 1.5300000000000002e-06, "loss": 3.3852, "step": 3060 }, { "epoch": 0.9595611285266458, "grad_norm": 38.45157241821289, "learning_rate": 1.5305e-06, "loss": 3.4452, "step": 3061 }, { "epoch": 0.9598746081504702, "grad_norm": 29.158227920532227, "learning_rate": 1.5310000000000002e-06, "loss": 3.0486, "step": 3062 }, { "epoch": 0.9601880877742947, "grad_norm": 97.31979370117188, "learning_rate": 1.5315000000000003e-06, "loss": 4.3573, "step": 3063 }, { "epoch": 0.9605015673981191, "grad_norm": 63.818050384521484, "learning_rate": 1.5320000000000001e-06, "loss": 4.6664, "step": 3064 }, { "epoch": 0.9608150470219435, "grad_norm": 28.385637283325195, "learning_rate": 1.5325000000000002e-06, "loss": 3.4442, "step": 3065 }, { "epoch": 0.961128526645768, "grad_norm": 40.3345947265625, "learning_rate": 1.533e-06, "loss": 3.7686, "step": 3066 }, { "epoch": 0.9614420062695925, "grad_norm": 35.8496208190918, "learning_rate": 1.5335e-06, "loss": 3.5075, "step": 3067 }, { "epoch": 0.961755485893417, "grad_norm": 31.015871047973633, "learning_rate": 1.5340000000000003e-06, "loss": 3.5269, "step": 3068 }, { "epoch": 0.9620689655172414, "grad_norm": 26.467029571533203, "learning_rate": 1.5345e-06, "loss": 3.0488, "step": 3069 }, { "epoch": 0.9623824451410659, "grad_norm": 57.80925369262695, "learning_rate": 1.5350000000000001e-06, "loss": 4.5774, "step": 3070 }, { "epoch": 0.9626959247648903, "grad_norm": 68.83757019042969, "learning_rate": 1.5355e-06, "loss": 4.7216, "step": 3071 }, { "epoch": 0.9630094043887147, "grad_norm": 44.15302658081055, "learning_rate": 1.536e-06, "loss": 4.4129, "step": 3072 }, { "epoch": 0.9633228840125392, "grad_norm": 48.796966552734375, "learning_rate": 1.5365000000000002e-06, "loss": 3.7593, "step": 3073 }, { "epoch": 0.9636363636363636, "grad_norm": 52.55888748168945, "learning_rate": 1.537e-06, "loss": 3.7959, "step": 3074 }, { "epoch": 0.9639498432601881, "grad_norm": 31.702449798583984, "learning_rate": 1.5375e-06, "loss": 3.7677, "step": 3075 }, { "epoch": 0.9642633228840125, "grad_norm": 47.94102096557617, "learning_rate": 1.5380000000000001e-06, "loss": 3.5232, "step": 3076 }, { "epoch": 0.964576802507837, "grad_norm": 46.60993957519531, "learning_rate": 1.5385e-06, "loss": 4.0139, "step": 3077 }, { "epoch": 0.9648902821316614, "grad_norm": 28.9523983001709, "learning_rate": 1.5390000000000002e-06, "loss": 2.7694, "step": 3078 }, { "epoch": 0.965203761755486, "grad_norm": 44.399925231933594, "learning_rate": 1.5395000000000003e-06, "loss": 3.5384, "step": 3079 }, { "epoch": 0.9655172413793104, "grad_norm": 42.65142822265625, "learning_rate": 1.54e-06, "loss": 3.7671, "step": 3080 }, { "epoch": 0.9658307210031348, "grad_norm": 51.35871887207031, "learning_rate": 1.5405000000000001e-06, "loss": 3.4669, "step": 3081 }, { "epoch": 0.9661442006269593, "grad_norm": 41.16212844848633, "learning_rate": 1.541e-06, "loss": 3.4601, "step": 3082 }, { "epoch": 0.9664576802507837, "grad_norm": 36.21253204345703, "learning_rate": 1.5415000000000002e-06, "loss": 3.063, "step": 3083 }, { "epoch": 0.9667711598746082, "grad_norm": 29.120792388916016, "learning_rate": 1.5420000000000002e-06, "loss": 3.5839, "step": 3084 }, { "epoch": 0.9670846394984326, "grad_norm": 32.0816535949707, "learning_rate": 1.5425e-06, "loss": 3.7305, "step": 3085 }, { "epoch": 0.967398119122257, "grad_norm": 30.316551208496094, "learning_rate": 1.543e-06, "loss": 3.2055, "step": 3086 }, { "epoch": 0.9677115987460815, "grad_norm": 33.83134841918945, "learning_rate": 1.5435e-06, "loss": 3.3693, "step": 3087 }, { "epoch": 0.9680250783699059, "grad_norm": 47.991493225097656, "learning_rate": 1.5440000000000002e-06, "loss": 4.0049, "step": 3088 }, { "epoch": 0.9683385579937304, "grad_norm": 24.988670349121094, "learning_rate": 1.5445000000000002e-06, "loss": 3.81, "step": 3089 }, { "epoch": 0.9686520376175548, "grad_norm": 39.11276626586914, "learning_rate": 1.545e-06, "loss": 3.7266, "step": 3090 }, { "epoch": 0.9689655172413794, "grad_norm": 69.33980560302734, "learning_rate": 1.5455e-06, "loss": 4.3037, "step": 3091 }, { "epoch": 0.9692789968652038, "grad_norm": 43.25937271118164, "learning_rate": 1.546e-06, "loss": 3.6816, "step": 3092 }, { "epoch": 0.9695924764890282, "grad_norm": 56.0028190612793, "learning_rate": 1.5465000000000001e-06, "loss": 3.5904, "step": 3093 }, { "epoch": 0.9699059561128527, "grad_norm": 62.90016555786133, "learning_rate": 1.5470000000000002e-06, "loss": 4.1878, "step": 3094 }, { "epoch": 0.9702194357366771, "grad_norm": 27.32510757446289, "learning_rate": 1.5475000000000002e-06, "loss": 2.9088, "step": 3095 }, { "epoch": 0.9705329153605016, "grad_norm": 29.192907333374023, "learning_rate": 1.548e-06, "loss": 3.6891, "step": 3096 }, { "epoch": 0.970846394984326, "grad_norm": 29.538455963134766, "learning_rate": 1.5485e-06, "loss": 3.5678, "step": 3097 }, { "epoch": 0.9711598746081505, "grad_norm": 39.13743209838867, "learning_rate": 1.5490000000000003e-06, "loss": 3.5836, "step": 3098 }, { "epoch": 0.9714733542319749, "grad_norm": 35.45713424682617, "learning_rate": 1.5495000000000002e-06, "loss": 3.5822, "step": 3099 }, { "epoch": 0.9717868338557993, "grad_norm": 37.34980392456055, "learning_rate": 1.5500000000000002e-06, "loss": 3.4961, "step": 3100 }, { "epoch": 0.9721003134796238, "grad_norm": 52.05093765258789, "learning_rate": 1.5505e-06, "loss": 4.2753, "step": 3101 }, { "epoch": 0.9724137931034482, "grad_norm": 43.123226165771484, "learning_rate": 1.551e-06, "loss": 4.7245, "step": 3102 }, { "epoch": 0.9727272727272728, "grad_norm": 30.87041473388672, "learning_rate": 1.5515000000000003e-06, "loss": 3.187, "step": 3103 }, { "epoch": 0.9730407523510972, "grad_norm": 61.1220588684082, "learning_rate": 1.5520000000000001e-06, "loss": 3.7726, "step": 3104 }, { "epoch": 0.9733542319749217, "grad_norm": 24.472267150878906, "learning_rate": 1.5525000000000002e-06, "loss": 3.2109, "step": 3105 }, { "epoch": 0.9736677115987461, "grad_norm": 41.3580436706543, "learning_rate": 1.553e-06, "loss": 4.827, "step": 3106 }, { "epoch": 0.9739811912225705, "grad_norm": 23.06800079345703, "learning_rate": 1.5535e-06, "loss": 3.6523, "step": 3107 }, { "epoch": 0.974294670846395, "grad_norm": 45.498695373535156, "learning_rate": 1.5540000000000003e-06, "loss": 4.1928, "step": 3108 }, { "epoch": 0.9746081504702194, "grad_norm": 32.11305618286133, "learning_rate": 1.5545000000000001e-06, "loss": 3.395, "step": 3109 }, { "epoch": 0.9749216300940439, "grad_norm": 22.9804744720459, "learning_rate": 1.5550000000000001e-06, "loss": 2.9892, "step": 3110 }, { "epoch": 0.9752351097178683, "grad_norm": 33.303001403808594, "learning_rate": 1.5555e-06, "loss": 3.2164, "step": 3111 }, { "epoch": 0.9755485893416928, "grad_norm": 35.61023712158203, "learning_rate": 1.556e-06, "loss": 3.2413, "step": 3112 }, { "epoch": 0.9758620689655172, "grad_norm": 32.667476654052734, "learning_rate": 1.5565000000000003e-06, "loss": 4.2483, "step": 3113 }, { "epoch": 0.9761755485893417, "grad_norm": 30.902109146118164, "learning_rate": 1.557e-06, "loss": 3.5198, "step": 3114 }, { "epoch": 0.9764890282131662, "grad_norm": 29.37360191345215, "learning_rate": 1.5575000000000001e-06, "loss": 3.6189, "step": 3115 }, { "epoch": 0.9768025078369906, "grad_norm": 41.751956939697266, "learning_rate": 1.5580000000000002e-06, "loss": 3.3907, "step": 3116 }, { "epoch": 0.9771159874608151, "grad_norm": 48.0351676940918, "learning_rate": 1.5585e-06, "loss": 3.2474, "step": 3117 }, { "epoch": 0.9774294670846395, "grad_norm": 33.41267395019531, "learning_rate": 1.5590000000000002e-06, "loss": 3.4043, "step": 3118 }, { "epoch": 0.977742946708464, "grad_norm": 26.803760528564453, "learning_rate": 1.5595000000000003e-06, "loss": 3.016, "step": 3119 }, { "epoch": 0.9780564263322884, "grad_norm": 59.20711135864258, "learning_rate": 1.56e-06, "loss": 3.8982, "step": 3120 }, { "epoch": 0.9783699059561128, "grad_norm": 25.695276260375977, "learning_rate": 1.5605000000000001e-06, "loss": 2.929, "step": 3121 }, { "epoch": 0.9786833855799373, "grad_norm": 41.91707229614258, "learning_rate": 1.561e-06, "loss": 3.3785, "step": 3122 }, { "epoch": 0.9789968652037617, "grad_norm": 26.014942169189453, "learning_rate": 1.5615000000000002e-06, "loss": 4.3426, "step": 3123 }, { "epoch": 0.9793103448275862, "grad_norm": 43.355995178222656, "learning_rate": 1.5620000000000002e-06, "loss": 3.191, "step": 3124 }, { "epoch": 0.9796238244514106, "grad_norm": 52.087005615234375, "learning_rate": 1.5625e-06, "loss": 3.2585, "step": 3125 }, { "epoch": 0.9799373040752352, "grad_norm": 61.68992233276367, "learning_rate": 1.5630000000000001e-06, "loss": 3.7206, "step": 3126 }, { "epoch": 0.9802507836990596, "grad_norm": 34.2900390625, "learning_rate": 1.5635e-06, "loss": 3.2416, "step": 3127 }, { "epoch": 0.980564263322884, "grad_norm": 51.79576873779297, "learning_rate": 1.5640000000000002e-06, "loss": 5.8384, "step": 3128 }, { "epoch": 0.9808777429467085, "grad_norm": 26.661314010620117, "learning_rate": 1.5645000000000002e-06, "loss": 3.4689, "step": 3129 }, { "epoch": 0.9811912225705329, "grad_norm": 35.80608367919922, "learning_rate": 1.565e-06, "loss": 3.2023, "step": 3130 }, { "epoch": 0.9815047021943574, "grad_norm": 33.74555587768555, "learning_rate": 1.5655e-06, "loss": 3.3833, "step": 3131 }, { "epoch": 0.9818181818181818, "grad_norm": 40.17083740234375, "learning_rate": 1.566e-06, "loss": 3.5556, "step": 3132 }, { "epoch": 0.9821316614420063, "grad_norm": 39.58924865722656, "learning_rate": 1.5665000000000002e-06, "loss": 3.0832, "step": 3133 }, { "epoch": 0.9824451410658307, "grad_norm": 45.79573440551758, "learning_rate": 1.5670000000000002e-06, "loss": 3.5993, "step": 3134 }, { "epoch": 0.9827586206896551, "grad_norm": 30.124801635742188, "learning_rate": 1.5675e-06, "loss": 3.0965, "step": 3135 }, { "epoch": 0.9830721003134796, "grad_norm": 50.938419342041016, "learning_rate": 1.568e-06, "loss": 3.1747, "step": 3136 }, { "epoch": 0.983385579937304, "grad_norm": 37.803470611572266, "learning_rate": 1.5685e-06, "loss": 3.724, "step": 3137 }, { "epoch": 0.9836990595611286, "grad_norm": 45.604373931884766, "learning_rate": 1.5690000000000001e-06, "loss": 2.8654, "step": 3138 }, { "epoch": 0.984012539184953, "grad_norm": 29.251733779907227, "learning_rate": 1.5695000000000002e-06, "loss": 3.0197, "step": 3139 }, { "epoch": 0.9843260188087775, "grad_norm": 40.991294860839844, "learning_rate": 1.5700000000000002e-06, "loss": 3.0759, "step": 3140 }, { "epoch": 0.9846394984326019, "grad_norm": 39.70516586303711, "learning_rate": 1.5705e-06, "loss": 3.2381, "step": 3141 }, { "epoch": 0.9849529780564263, "grad_norm": 104.91124725341797, "learning_rate": 1.571e-06, "loss": 3.9789, "step": 3142 }, { "epoch": 0.9852664576802508, "grad_norm": 69.99864959716797, "learning_rate": 1.5715000000000003e-06, "loss": 5.4289, "step": 3143 }, { "epoch": 0.9855799373040752, "grad_norm": 47.31104278564453, "learning_rate": 1.5720000000000002e-06, "loss": 3.5801, "step": 3144 }, { "epoch": 0.9858934169278997, "grad_norm": 23.58669090270996, "learning_rate": 1.5725000000000002e-06, "loss": 3.2141, "step": 3145 }, { "epoch": 0.9862068965517241, "grad_norm": 60.75740432739258, "learning_rate": 1.573e-06, "loss": 3.1711, "step": 3146 }, { "epoch": 0.9865203761755486, "grad_norm": 35.55626678466797, "learning_rate": 1.5735e-06, "loss": 3.0043, "step": 3147 }, { "epoch": 0.986833855799373, "grad_norm": 56.99102783203125, "learning_rate": 1.5740000000000003e-06, "loss": 4.1672, "step": 3148 }, { "epoch": 0.9871473354231975, "grad_norm": 25.71376609802246, "learning_rate": 1.5745000000000001e-06, "loss": 2.92, "step": 3149 }, { "epoch": 0.987460815047022, "grad_norm": 31.668563842773438, "learning_rate": 1.5750000000000002e-06, "loss": 3.2671, "step": 3150 }, { "epoch": 0.9877742946708464, "grad_norm": 32.40145492553711, "learning_rate": 1.5755e-06, "loss": 3.1882, "step": 3151 }, { "epoch": 0.9880877742946709, "grad_norm": 33.56290054321289, "learning_rate": 1.576e-06, "loss": 2.97, "step": 3152 }, { "epoch": 0.9884012539184953, "grad_norm": 23.0459041595459, "learning_rate": 1.5765000000000003e-06, "loss": 2.8921, "step": 3153 }, { "epoch": 0.9887147335423198, "grad_norm": 31.948535919189453, "learning_rate": 1.577e-06, "loss": 3.348, "step": 3154 }, { "epoch": 0.9890282131661442, "grad_norm": 39.63069534301758, "learning_rate": 1.5775000000000001e-06, "loss": 2.9237, "step": 3155 }, { "epoch": 0.9893416927899686, "grad_norm": 32.34980392456055, "learning_rate": 1.5780000000000002e-06, "loss": 3.245, "step": 3156 }, { "epoch": 0.9896551724137931, "grad_norm": 41.89958953857422, "learning_rate": 1.5785e-06, "loss": 3.4284, "step": 3157 }, { "epoch": 0.9899686520376175, "grad_norm": 51.7044792175293, "learning_rate": 1.5790000000000003e-06, "loss": 3.0146, "step": 3158 }, { "epoch": 0.990282131661442, "grad_norm": 110.62555694580078, "learning_rate": 1.5795000000000003e-06, "loss": 3.7889, "step": 3159 }, { "epoch": 0.9905956112852664, "grad_norm": 42.60169982910156, "learning_rate": 1.5800000000000001e-06, "loss": 4.1996, "step": 3160 }, { "epoch": 0.990909090909091, "grad_norm": 45.62926483154297, "learning_rate": 1.5805000000000002e-06, "loss": 2.8792, "step": 3161 }, { "epoch": 0.9912225705329154, "grad_norm": 29.770870208740234, "learning_rate": 1.581e-06, "loss": 3.162, "step": 3162 }, { "epoch": 0.9915360501567398, "grad_norm": 40.25620651245117, "learning_rate": 1.5815000000000002e-06, "loss": 3.2687, "step": 3163 }, { "epoch": 0.9918495297805643, "grad_norm": 51.3452262878418, "learning_rate": 1.5820000000000003e-06, "loss": 3.6536, "step": 3164 }, { "epoch": 0.9921630094043887, "grad_norm": 42.561981201171875, "learning_rate": 1.5825e-06, "loss": 3.2372, "step": 3165 }, { "epoch": 0.9924764890282132, "grad_norm": 104.1058578491211, "learning_rate": 1.5830000000000001e-06, "loss": 3.6081, "step": 3166 }, { "epoch": 0.9927899686520376, "grad_norm": 33.43682098388672, "learning_rate": 1.5835e-06, "loss": 3.0076, "step": 3167 }, { "epoch": 0.993103448275862, "grad_norm": 31.840662002563477, "learning_rate": 1.5840000000000002e-06, "loss": 3.4222, "step": 3168 }, { "epoch": 0.9934169278996865, "grad_norm": 39.76411437988281, "learning_rate": 1.5845000000000002e-06, "loss": 3.4354, "step": 3169 }, { "epoch": 0.9937304075235109, "grad_norm": 45.577274322509766, "learning_rate": 1.585e-06, "loss": 4.2827, "step": 3170 }, { "epoch": 0.9940438871473354, "grad_norm": 40.2696647644043, "learning_rate": 1.5855000000000001e-06, "loss": 3.1225, "step": 3171 }, { "epoch": 0.9943573667711598, "grad_norm": 37.35997009277344, "learning_rate": 1.586e-06, "loss": 3.626, "step": 3172 }, { "epoch": 0.9946708463949844, "grad_norm": 29.363245010375977, "learning_rate": 1.5865000000000002e-06, "loss": 3.437, "step": 3173 }, { "epoch": 0.9949843260188088, "grad_norm": 30.479536056518555, "learning_rate": 1.5870000000000002e-06, "loss": 3.4304, "step": 3174 }, { "epoch": 0.9952978056426333, "grad_norm": 27.580543518066406, "learning_rate": 1.5875e-06, "loss": 3.2106, "step": 3175 }, { "epoch": 0.9956112852664577, "grad_norm": 39.4871711730957, "learning_rate": 1.588e-06, "loss": 3.465, "step": 3176 }, { "epoch": 0.9959247648902821, "grad_norm": 20.305816650390625, "learning_rate": 1.5885000000000001e-06, "loss": 2.8586, "step": 3177 }, { "epoch": 0.9962382445141066, "grad_norm": 54.459625244140625, "learning_rate": 1.5890000000000002e-06, "loss": 3.8607, "step": 3178 }, { "epoch": 0.996551724137931, "grad_norm": 45.21859359741211, "learning_rate": 1.5895000000000002e-06, "loss": 4.0433, "step": 3179 }, { "epoch": 0.9968652037617555, "grad_norm": 30.275541305541992, "learning_rate": 1.5900000000000002e-06, "loss": 3.1236, "step": 3180 }, { "epoch": 0.9971786833855799, "grad_norm": 24.845712661743164, "learning_rate": 1.5905e-06, "loss": 3.052, "step": 3181 }, { "epoch": 0.9974921630094044, "grad_norm": 47.068382263183594, "learning_rate": 1.591e-06, "loss": 4.1035, "step": 3182 }, { "epoch": 0.9978056426332288, "grad_norm": 86.9577865600586, "learning_rate": 1.5915000000000003e-06, "loss": 3.6548, "step": 3183 }, { "epoch": 0.9981191222570532, "grad_norm": 45.81187057495117, "learning_rate": 1.5920000000000002e-06, "loss": 4.0105, "step": 3184 }, { "epoch": 0.9984326018808778, "grad_norm": 54.567386627197266, "learning_rate": 1.5925000000000002e-06, "loss": 3.7583, "step": 3185 }, { "epoch": 0.9987460815047022, "grad_norm": 45.75572967529297, "learning_rate": 1.593e-06, "loss": 3.3029, "step": 3186 }, { "epoch": 0.9990595611285267, "grad_norm": 56.43427658081055, "learning_rate": 1.5935e-06, "loss": 2.8933, "step": 3187 }, { "epoch": 0.9993730407523511, "grad_norm": 41.16676712036133, "learning_rate": 1.594e-06, "loss": 3.0468, "step": 3188 }, { "epoch": 0.9996865203761756, "grad_norm": 38.499820709228516, "learning_rate": 1.5945000000000001e-06, "loss": 3.9996, "step": 3189 }, { "epoch": 1.0, "grad_norm": 20.5953426361084, "learning_rate": 1.5950000000000002e-06, "loss": 2.7745, "step": 3190 }, { "epoch": 1.0003134796238244, "grad_norm": 34.326446533203125, "learning_rate": 1.5955e-06, "loss": 3.4488, "step": 3191 }, { "epoch": 1.0006269592476489, "grad_norm": 25.23051643371582, "learning_rate": 1.596e-06, "loss": 2.7374, "step": 3192 }, { "epoch": 1.0006269592476489, "eval_loss": 5.7285990715026855, "eval_runtime": 21.0416, "eval_samples_per_second": 127.7, "eval_steps_per_second": 7.984, "step": 3192 }, { "epoch": 1.0009404388714733, "grad_norm": 27.973400115966797, "learning_rate": 1.5965e-06, "loss": 3.0561, "step": 3193 }, { "epoch": 1.0012539184952978, "grad_norm": 28.687578201293945, "learning_rate": 1.5970000000000001e-06, "loss": 3.1456, "step": 3194 }, { "epoch": 1.0015673981191222, "grad_norm": 54.51884078979492, "learning_rate": 1.5975000000000002e-06, "loss": 4.3574, "step": 3195 }, { "epoch": 1.0018808777429467, "grad_norm": 22.76805877685547, "learning_rate": 1.5980000000000002e-06, "loss": 3.5964, "step": 3196 }, { "epoch": 1.002194357366771, "grad_norm": 42.05739974975586, "learning_rate": 1.5985e-06, "loss": 2.9885, "step": 3197 }, { "epoch": 1.0025078369905955, "grad_norm": 55.0859260559082, "learning_rate": 1.599e-06, "loss": 3.6646, "step": 3198 }, { "epoch": 1.00282131661442, "grad_norm": 48.74159240722656, "learning_rate": 1.5995000000000003e-06, "loss": 3.5255, "step": 3199 }, { "epoch": 1.0031347962382444, "grad_norm": 28.012622833251953, "learning_rate": 1.6000000000000001e-06, "loss": 3.4958, "step": 3200 }, { "epoch": 1.0034482758620689, "grad_norm": 38.07377624511719, "learning_rate": 1.6005000000000002e-06, "loss": 3.3704, "step": 3201 }, { "epoch": 1.0037617554858935, "grad_norm": 28.49401092529297, "learning_rate": 1.601e-06, "loss": 4.2952, "step": 3202 }, { "epoch": 1.004075235109718, "grad_norm": 94.79960632324219, "learning_rate": 1.6015e-06, "loss": 5.6727, "step": 3203 }, { "epoch": 1.0043887147335424, "grad_norm": 21.47426414489746, "learning_rate": 1.6020000000000003e-06, "loss": 3.5125, "step": 3204 }, { "epoch": 1.0047021943573669, "grad_norm": 36.450836181640625, "learning_rate": 1.6025000000000001e-06, "loss": 2.8938, "step": 3205 }, { "epoch": 1.0050156739811913, "grad_norm": 38.39944076538086, "learning_rate": 1.6030000000000002e-06, "loss": 3.4479, "step": 3206 }, { "epoch": 1.0053291536050157, "grad_norm": 26.999618530273438, "learning_rate": 1.6035e-06, "loss": 3.1897, "step": 3207 }, { "epoch": 1.0056426332288402, "grad_norm": 38.76042938232422, "learning_rate": 1.604e-06, "loss": 3.3169, "step": 3208 }, { "epoch": 1.0059561128526646, "grad_norm": 34.40362548828125, "learning_rate": 1.6045000000000003e-06, "loss": 3.5853, "step": 3209 }, { "epoch": 1.006269592476489, "grad_norm": 29.437618255615234, "learning_rate": 1.605e-06, "loss": 2.737, "step": 3210 }, { "epoch": 1.0065830721003135, "grad_norm": 31.488759994506836, "learning_rate": 1.6055000000000001e-06, "loss": 4.0049, "step": 3211 }, { "epoch": 1.006896551724138, "grad_norm": 32.4052619934082, "learning_rate": 1.606e-06, "loss": 3.9704, "step": 3212 }, { "epoch": 1.0072100313479624, "grad_norm": 41.16117858886719, "learning_rate": 1.6065e-06, "loss": 3.4824, "step": 3213 }, { "epoch": 1.0075235109717868, "grad_norm": 39.67842102050781, "learning_rate": 1.6070000000000002e-06, "loss": 3.2149, "step": 3214 }, { "epoch": 1.0078369905956113, "grad_norm": 34.92125701904297, "learning_rate": 1.6075e-06, "loss": 3.4013, "step": 3215 }, { "epoch": 1.0081504702194357, "grad_norm": 35.19260787963867, "learning_rate": 1.608e-06, "loss": 3.0669, "step": 3216 }, { "epoch": 1.0084639498432602, "grad_norm": 30.965526580810547, "learning_rate": 1.6085000000000001e-06, "loss": 3.5108, "step": 3217 }, { "epoch": 1.0087774294670846, "grad_norm": 55.61679458618164, "learning_rate": 1.609e-06, "loss": 3.2511, "step": 3218 }, { "epoch": 1.009090909090909, "grad_norm": 52.900753021240234, "learning_rate": 1.6095000000000002e-06, "loss": 4.8786, "step": 3219 }, { "epoch": 1.0094043887147335, "grad_norm": 31.876253128051758, "learning_rate": 1.6100000000000003e-06, "loss": 3.5526, "step": 3220 }, { "epoch": 1.009717868338558, "grad_norm": 70.21955871582031, "learning_rate": 1.6105e-06, "loss": 3.2994, "step": 3221 }, { "epoch": 1.0100313479623824, "grad_norm": 57.42060470581055, "learning_rate": 1.6110000000000001e-06, "loss": 3.6681, "step": 3222 }, { "epoch": 1.0103448275862068, "grad_norm": 20.363557815551758, "learning_rate": 1.6115e-06, "loss": 2.969, "step": 3223 }, { "epoch": 1.0106583072100312, "grad_norm": 46.56084442138672, "learning_rate": 1.6120000000000002e-06, "loss": 3.2407, "step": 3224 }, { "epoch": 1.0109717868338557, "grad_norm": 26.832813262939453, "learning_rate": 1.6125000000000002e-06, "loss": 2.859, "step": 3225 }, { "epoch": 1.0112852664576804, "grad_norm": 39.61360549926758, "learning_rate": 1.613e-06, "loss": 2.8402, "step": 3226 }, { "epoch": 1.0115987460815048, "grad_norm": 59.345333099365234, "learning_rate": 1.6135e-06, "loss": 3.2726, "step": 3227 }, { "epoch": 1.0119122257053292, "grad_norm": 34.26584243774414, "learning_rate": 1.614e-06, "loss": 3.2933, "step": 3228 }, { "epoch": 1.0122257053291537, "grad_norm": 31.62544822692871, "learning_rate": 1.6145000000000002e-06, "loss": 2.8915, "step": 3229 }, { "epoch": 1.0125391849529781, "grad_norm": 24.0006160736084, "learning_rate": 1.6150000000000002e-06, "loss": 2.4861, "step": 3230 }, { "epoch": 1.0128526645768026, "grad_norm": 23.52138328552246, "learning_rate": 1.6155e-06, "loss": 3.1468, "step": 3231 }, { "epoch": 1.013166144200627, "grad_norm": 48.56825256347656, "learning_rate": 1.616e-06, "loss": 3.2097, "step": 3232 }, { "epoch": 1.0134796238244514, "grad_norm": 40.454193115234375, "learning_rate": 1.6165000000000001e-06, "loss": 3.646, "step": 3233 }, { "epoch": 1.013793103448276, "grad_norm": 27.93584632873535, "learning_rate": 1.6170000000000001e-06, "loss": 3.0358, "step": 3234 }, { "epoch": 1.0141065830721003, "grad_norm": 37.334739685058594, "learning_rate": 1.6175000000000002e-06, "loss": 3.0387, "step": 3235 }, { "epoch": 1.0144200626959248, "grad_norm": 29.4758358001709, "learning_rate": 1.6180000000000002e-06, "loss": 3.141, "step": 3236 }, { "epoch": 1.0147335423197492, "grad_norm": 19.978744506835938, "learning_rate": 1.6185e-06, "loss": 2.5931, "step": 3237 }, { "epoch": 1.0150470219435737, "grad_norm": 57.191749572753906, "learning_rate": 1.619e-06, "loss": 3.0585, "step": 3238 }, { "epoch": 1.015360501567398, "grad_norm": 36.548648834228516, "learning_rate": 1.6195000000000003e-06, "loss": 3.1298, "step": 3239 }, { "epoch": 1.0156739811912225, "grad_norm": 32.01398849487305, "learning_rate": 1.6200000000000002e-06, "loss": 3.2455, "step": 3240 }, { "epoch": 1.015987460815047, "grad_norm": 34.381431579589844, "learning_rate": 1.6205000000000002e-06, "loss": 3.4494, "step": 3241 }, { "epoch": 1.0163009404388714, "grad_norm": 59.548458099365234, "learning_rate": 1.621e-06, "loss": 4.1441, "step": 3242 }, { "epoch": 1.0166144200626959, "grad_norm": 59.294395446777344, "learning_rate": 1.6215e-06, "loss": 4.3494, "step": 3243 }, { "epoch": 1.0169278996865203, "grad_norm": 31.622285842895508, "learning_rate": 1.6220000000000003e-06, "loss": 3.1537, "step": 3244 }, { "epoch": 1.0172413793103448, "grad_norm": 33.20928955078125, "learning_rate": 1.6225000000000001e-06, "loss": 2.9726, "step": 3245 }, { "epoch": 1.0175548589341692, "grad_norm": 75.9956283569336, "learning_rate": 1.6230000000000002e-06, "loss": 3.3001, "step": 3246 }, { "epoch": 1.0178683385579936, "grad_norm": 44.67719650268555, "learning_rate": 1.6235e-06, "loss": 3.0272, "step": 3247 }, { "epoch": 1.018181818181818, "grad_norm": 27.121854782104492, "learning_rate": 1.624e-06, "loss": 3.3043, "step": 3248 }, { "epoch": 1.0184952978056427, "grad_norm": 34.486473083496094, "learning_rate": 1.6245000000000003e-06, "loss": 2.8611, "step": 3249 }, { "epoch": 1.0188087774294672, "grad_norm": 70.01057434082031, "learning_rate": 1.6250000000000001e-06, "loss": 3.9254, "step": 3250 }, { "epoch": 1.0191222570532916, "grad_norm": 35.10891342163086, "learning_rate": 1.6255000000000001e-06, "loss": 3.2911, "step": 3251 }, { "epoch": 1.019435736677116, "grad_norm": 45.17849349975586, "learning_rate": 1.626e-06, "loss": 3.5813, "step": 3252 }, { "epoch": 1.0197492163009405, "grad_norm": 36.457950592041016, "learning_rate": 1.6265e-06, "loss": 3.16, "step": 3253 }, { "epoch": 1.020062695924765, "grad_norm": 49.345062255859375, "learning_rate": 1.6270000000000003e-06, "loss": 4.0663, "step": 3254 }, { "epoch": 1.0203761755485894, "grad_norm": 22.22081756591797, "learning_rate": 1.6275e-06, "loss": 2.8179, "step": 3255 }, { "epoch": 1.0206896551724138, "grad_norm": 44.726043701171875, "learning_rate": 1.6280000000000001e-06, "loss": 3.1774, "step": 3256 }, { "epoch": 1.0210031347962383, "grad_norm": 63.41642379760742, "learning_rate": 1.6285000000000002e-06, "loss": 3.1283, "step": 3257 }, { "epoch": 1.0213166144200627, "grad_norm": 25.33333396911621, "learning_rate": 1.629e-06, "loss": 2.8745, "step": 3258 }, { "epoch": 1.0216300940438872, "grad_norm": 26.45008659362793, "learning_rate": 1.6295000000000002e-06, "loss": 3.3849, "step": 3259 }, { "epoch": 1.0219435736677116, "grad_norm": 29.69173812866211, "learning_rate": 1.6300000000000003e-06, "loss": 2.9018, "step": 3260 }, { "epoch": 1.022257053291536, "grad_norm": 29.18816375732422, "learning_rate": 1.6305e-06, "loss": 3.2486, "step": 3261 }, { "epoch": 1.0225705329153605, "grad_norm": 29.251667022705078, "learning_rate": 1.6310000000000001e-06, "loss": 2.9324, "step": 3262 }, { "epoch": 1.022884012539185, "grad_norm": 28.868215560913086, "learning_rate": 1.6315e-06, "loss": 2.872, "step": 3263 }, { "epoch": 1.0231974921630094, "grad_norm": 31.325496673583984, "learning_rate": 1.6320000000000002e-06, "loss": 3.2128, "step": 3264 }, { "epoch": 1.0235109717868338, "grad_norm": 42.07441711425781, "learning_rate": 1.6325000000000003e-06, "loss": 2.9007, "step": 3265 }, { "epoch": 1.0238244514106583, "grad_norm": 68.76583099365234, "learning_rate": 1.633e-06, "loss": 4.2267, "step": 3266 }, { "epoch": 1.0241379310344827, "grad_norm": 48.943477630615234, "learning_rate": 1.6335000000000001e-06, "loss": 3.0372, "step": 3267 }, { "epoch": 1.0244514106583071, "grad_norm": 40.075984954833984, "learning_rate": 1.634e-06, "loss": 3.9948, "step": 3268 }, { "epoch": 1.0247648902821316, "grad_norm": 21.297189712524414, "learning_rate": 1.6345000000000002e-06, "loss": 2.7858, "step": 3269 }, { "epoch": 1.025078369905956, "grad_norm": 40.05287551879883, "learning_rate": 1.6350000000000002e-06, "loss": 3.8012, "step": 3270 }, { "epoch": 1.0253918495297805, "grad_norm": 68.54923248291016, "learning_rate": 1.6355e-06, "loss": 3.6058, "step": 3271 }, { "epoch": 1.0257053291536051, "grad_norm": 37.00851058959961, "learning_rate": 1.636e-06, "loss": 3.347, "step": 3272 }, { "epoch": 1.0260188087774296, "grad_norm": 50.256134033203125, "learning_rate": 1.6365000000000001e-06, "loss": 3.2308, "step": 3273 }, { "epoch": 1.026332288401254, "grad_norm": 19.08588218688965, "learning_rate": 1.6370000000000002e-06, "loss": 2.9165, "step": 3274 }, { "epoch": 1.0266457680250785, "grad_norm": 26.314943313598633, "learning_rate": 1.6375000000000002e-06, "loss": 3.1059, "step": 3275 }, { "epoch": 1.026959247648903, "grad_norm": 74.63359069824219, "learning_rate": 1.6380000000000002e-06, "loss": 4.7147, "step": 3276 }, { "epoch": 1.0272727272727273, "grad_norm": 38.578792572021484, "learning_rate": 1.6385e-06, "loss": 3.2019, "step": 3277 }, { "epoch": 1.0275862068965518, "grad_norm": 40.135738372802734, "learning_rate": 1.639e-06, "loss": 2.7686, "step": 3278 }, { "epoch": 1.0278996865203762, "grad_norm": 23.660158157348633, "learning_rate": 1.6395000000000004e-06, "loss": 2.9382, "step": 3279 }, { "epoch": 1.0282131661442007, "grad_norm": 39.85804748535156, "learning_rate": 1.6400000000000002e-06, "loss": 2.8816, "step": 3280 }, { "epoch": 1.028526645768025, "grad_norm": 46.418426513671875, "learning_rate": 1.6405000000000002e-06, "loss": 3.0901, "step": 3281 }, { "epoch": 1.0288401253918495, "grad_norm": 190.2622528076172, "learning_rate": 1.641e-06, "loss": 2.9474, "step": 3282 }, { "epoch": 1.029153605015674, "grad_norm": 43.696956634521484, "learning_rate": 1.6415e-06, "loss": 3.5476, "step": 3283 }, { "epoch": 1.0294670846394984, "grad_norm": 22.224021911621094, "learning_rate": 1.6420000000000003e-06, "loss": 3.1479, "step": 3284 }, { "epoch": 1.0297805642633229, "grad_norm": 29.27781867980957, "learning_rate": 1.6425000000000002e-06, "loss": 2.6123, "step": 3285 }, { "epoch": 1.0300940438871473, "grad_norm": 24.40782356262207, "learning_rate": 1.6430000000000002e-06, "loss": 2.6373, "step": 3286 }, { "epoch": 1.0304075235109718, "grad_norm": 27.865283966064453, "learning_rate": 1.6435e-06, "loss": 3.657, "step": 3287 }, { "epoch": 1.0307210031347962, "grad_norm": 33.429283142089844, "learning_rate": 1.644e-06, "loss": 3.7763, "step": 3288 }, { "epoch": 1.0310344827586206, "grad_norm": 20.046314239501953, "learning_rate": 1.6445000000000003e-06, "loss": 3.4536, "step": 3289 }, { "epoch": 1.031347962382445, "grad_norm": 26.210359573364258, "learning_rate": 1.6450000000000001e-06, "loss": 2.8382, "step": 3290 }, { "epoch": 1.0316614420062695, "grad_norm": 27.057741165161133, "learning_rate": 1.6455000000000002e-06, "loss": 2.6236, "step": 3291 }, { "epoch": 1.031974921630094, "grad_norm": 60.8143310546875, "learning_rate": 1.646e-06, "loss": 4.561, "step": 3292 }, { "epoch": 1.0322884012539184, "grad_norm": 22.38360023498535, "learning_rate": 1.6465e-06, "loss": 2.4903, "step": 3293 }, { "epoch": 1.0326018808777429, "grad_norm": 32.65427780151367, "learning_rate": 1.6470000000000003e-06, "loss": 3.5099, "step": 3294 }, { "epoch": 1.0329153605015673, "grad_norm": 32.22104263305664, "learning_rate": 1.6475000000000001e-06, "loss": 2.9126, "step": 3295 }, { "epoch": 1.033228840125392, "grad_norm": 18.776548385620117, "learning_rate": 1.6480000000000001e-06, "loss": 2.7612, "step": 3296 }, { "epoch": 1.0335423197492164, "grad_norm": 25.23478126525879, "learning_rate": 1.6485000000000002e-06, "loss": 2.8399, "step": 3297 }, { "epoch": 1.0338557993730408, "grad_norm": 60.96337890625, "learning_rate": 1.649e-06, "loss": 2.9111, "step": 3298 }, { "epoch": 1.0341692789968653, "grad_norm": 26.42510414123535, "learning_rate": 1.6495000000000003e-06, "loss": 3.0101, "step": 3299 }, { "epoch": 1.0344827586206897, "grad_norm": 31.34357261657715, "learning_rate": 1.6500000000000003e-06, "loss": 2.8724, "step": 3300 }, { "epoch": 1.0347962382445142, "grad_norm": 64.86882781982422, "learning_rate": 1.6505000000000001e-06, "loss": 3.3274, "step": 3301 }, { "epoch": 1.0351097178683386, "grad_norm": 21.959516525268555, "learning_rate": 1.6510000000000002e-06, "loss": 3.0008, "step": 3302 }, { "epoch": 1.035423197492163, "grad_norm": 32.650943756103516, "learning_rate": 1.6515e-06, "loss": 3.1787, "step": 3303 }, { "epoch": 1.0357366771159875, "grad_norm": 36.1429328918457, "learning_rate": 1.6520000000000002e-06, "loss": 3.0022, "step": 3304 }, { "epoch": 1.036050156739812, "grad_norm": 48.24531936645508, "learning_rate": 1.6525000000000003e-06, "loss": 2.9519, "step": 3305 }, { "epoch": 1.0363636363636364, "grad_norm": 34.19054412841797, "learning_rate": 1.653e-06, "loss": 3.331, "step": 3306 }, { "epoch": 1.0366771159874608, "grad_norm": 18.031879425048828, "learning_rate": 1.6535000000000001e-06, "loss": 3.028, "step": 3307 }, { "epoch": 1.0369905956112853, "grad_norm": 29.253751754760742, "learning_rate": 1.654e-06, "loss": 2.7076, "step": 3308 }, { "epoch": 1.0373040752351097, "grad_norm": 43.910423278808594, "learning_rate": 1.6545000000000002e-06, "loss": 3.3512, "step": 3309 }, { "epoch": 1.0376175548589341, "grad_norm": 25.297222137451172, "learning_rate": 1.6550000000000002e-06, "loss": 3.3843, "step": 3310 }, { "epoch": 1.0379310344827586, "grad_norm": 27.861984252929688, "learning_rate": 1.6555e-06, "loss": 2.8931, "step": 3311 }, { "epoch": 1.038244514106583, "grad_norm": 29.09917640686035, "learning_rate": 1.6560000000000001e-06, "loss": 2.8308, "step": 3312 }, { "epoch": 1.0385579937304075, "grad_norm": 23.79581642150879, "learning_rate": 1.6565000000000001e-06, "loss": 2.7549, "step": 3313 }, { "epoch": 1.038871473354232, "grad_norm": 23.295839309692383, "learning_rate": 1.657e-06, "loss": 2.7089, "step": 3314 }, { "epoch": 1.0391849529780564, "grad_norm": 20.08068084716797, "learning_rate": 1.6575000000000002e-06, "loss": 2.7575, "step": 3315 }, { "epoch": 1.0394984326018808, "grad_norm": 28.336292266845703, "learning_rate": 1.6580000000000003e-06, "loss": 2.9939, "step": 3316 }, { "epoch": 1.0398119122257052, "grad_norm": 26.962081909179688, "learning_rate": 1.6585e-06, "loss": 2.9673, "step": 3317 }, { "epoch": 1.0401253918495297, "grad_norm": 71.06791687011719, "learning_rate": 1.6590000000000001e-06, "loss": 5.612, "step": 3318 }, { "epoch": 1.0404388714733543, "grad_norm": 22.833223342895508, "learning_rate": 1.6595e-06, "loss": 2.9429, "step": 3319 }, { "epoch": 1.0407523510971788, "grad_norm": 73.84259033203125, "learning_rate": 1.6600000000000002e-06, "loss": 4.2648, "step": 3320 }, { "epoch": 1.0410658307210032, "grad_norm": 38.24066925048828, "learning_rate": 1.6605000000000002e-06, "loss": 3.4204, "step": 3321 }, { "epoch": 1.0413793103448277, "grad_norm": 85.08003997802734, "learning_rate": 1.661e-06, "loss": 3.7887, "step": 3322 }, { "epoch": 1.041692789968652, "grad_norm": 36.931583404541016, "learning_rate": 1.6615e-06, "loss": 2.9921, "step": 3323 }, { "epoch": 1.0420062695924766, "grad_norm": 64.08250427246094, "learning_rate": 1.662e-06, "loss": 3.7738, "step": 3324 }, { "epoch": 1.042319749216301, "grad_norm": 32.0848388671875, "learning_rate": 1.6625000000000002e-06, "loss": 3.3746, "step": 3325 }, { "epoch": 1.0426332288401254, "grad_norm": 19.127561569213867, "learning_rate": 1.6630000000000002e-06, "loss": 2.7594, "step": 3326 }, { "epoch": 1.0429467084639499, "grad_norm": 97.25066375732422, "learning_rate": 1.6635e-06, "loss": 3.3868, "step": 3327 }, { "epoch": 1.0432601880877743, "grad_norm": 63.51421356201172, "learning_rate": 1.664e-06, "loss": 3.442, "step": 3328 }, { "epoch": 1.0435736677115988, "grad_norm": 38.333282470703125, "learning_rate": 1.6645e-06, "loss": 3.4402, "step": 3329 }, { "epoch": 1.0438871473354232, "grad_norm": 39.50896072387695, "learning_rate": 1.6650000000000002e-06, "loss": 3.0081, "step": 3330 }, { "epoch": 1.0442006269592476, "grad_norm": 43.86787033081055, "learning_rate": 1.6655000000000002e-06, "loss": 2.6378, "step": 3331 }, { "epoch": 1.044514106583072, "grad_norm": 25.9218692779541, "learning_rate": 1.666e-06, "loss": 2.7456, "step": 3332 }, { "epoch": 1.0448275862068965, "grad_norm": 41.19293212890625, "learning_rate": 1.6665e-06, "loss": 3.3948, "step": 3333 }, { "epoch": 1.045141065830721, "grad_norm": 27.02749252319336, "learning_rate": 1.667e-06, "loss": 3.3631, "step": 3334 }, { "epoch": 1.0454545454545454, "grad_norm": 48.45111083984375, "learning_rate": 1.6675000000000001e-06, "loss": 3.0751, "step": 3335 }, { "epoch": 1.0457680250783699, "grad_norm": 27.78561782836914, "learning_rate": 1.6680000000000002e-06, "loss": 3.616, "step": 3336 }, { "epoch": 1.0460815047021943, "grad_norm": 37.05852127075195, "learning_rate": 1.6685000000000002e-06, "loss": 2.9797, "step": 3337 }, { "epoch": 1.0463949843260187, "grad_norm": 33.00432205200195, "learning_rate": 1.669e-06, "loss": 3.8552, "step": 3338 }, { "epoch": 1.0467084639498432, "grad_norm": 55.03670120239258, "learning_rate": 1.6695e-06, "loss": 3.2324, "step": 3339 }, { "epoch": 1.0470219435736676, "grad_norm": 40.15440368652344, "learning_rate": 1.6700000000000003e-06, "loss": 3.3079, "step": 3340 }, { "epoch": 1.047335423197492, "grad_norm": 57.2756233215332, "learning_rate": 1.6705000000000001e-06, "loss": 3.9613, "step": 3341 }, { "epoch": 1.0476489028213165, "grad_norm": 44.24393844604492, "learning_rate": 1.6710000000000002e-06, "loss": 3.7156, "step": 3342 }, { "epoch": 1.0479623824451412, "grad_norm": 72.37336730957031, "learning_rate": 1.6715e-06, "loss": 2.5366, "step": 3343 }, { "epoch": 1.0482758620689656, "grad_norm": 44.30312728881836, "learning_rate": 1.672e-06, "loss": 3.3909, "step": 3344 }, { "epoch": 1.04858934169279, "grad_norm": 34.76454162597656, "learning_rate": 1.6725000000000003e-06, "loss": 3.5469, "step": 3345 }, { "epoch": 1.0489028213166145, "grad_norm": 50.73592758178711, "learning_rate": 1.6730000000000001e-06, "loss": 2.8253, "step": 3346 }, { "epoch": 1.049216300940439, "grad_norm": 33.41584396362305, "learning_rate": 1.6735000000000002e-06, "loss": 3.7759, "step": 3347 }, { "epoch": 1.0495297805642634, "grad_norm": 92.3950424194336, "learning_rate": 1.674e-06, "loss": 3.4374, "step": 3348 }, { "epoch": 1.0498432601880878, "grad_norm": 18.99493408203125, "learning_rate": 1.6745e-06, "loss": 2.7457, "step": 3349 }, { "epoch": 1.0501567398119123, "grad_norm": 65.04867553710938, "learning_rate": 1.6750000000000003e-06, "loss": 3.2615, "step": 3350 }, { "epoch": 1.0504702194357367, "grad_norm": 19.495983123779297, "learning_rate": 1.6755e-06, "loss": 2.5306, "step": 3351 }, { "epoch": 1.0507836990595611, "grad_norm": 70.9605712890625, "learning_rate": 1.6760000000000001e-06, "loss": 3.8023, "step": 3352 }, { "epoch": 1.0510971786833856, "grad_norm": 35.70189666748047, "learning_rate": 1.6765000000000002e-06, "loss": 2.8319, "step": 3353 }, { "epoch": 1.05141065830721, "grad_norm": 50.487388610839844, "learning_rate": 1.677e-06, "loss": 3.9747, "step": 3354 }, { "epoch": 1.0517241379310345, "grad_norm": 31.325843811035156, "learning_rate": 1.6775000000000002e-06, "loss": 3.1756, "step": 3355 }, { "epoch": 1.052037617554859, "grad_norm": 32.679195404052734, "learning_rate": 1.6780000000000003e-06, "loss": 2.9121, "step": 3356 }, { "epoch": 1.0523510971786834, "grad_norm": 40.90104675292969, "learning_rate": 1.6785000000000001e-06, "loss": 3.2628, "step": 3357 }, { "epoch": 1.0526645768025078, "grad_norm": 33.8775634765625, "learning_rate": 1.6790000000000001e-06, "loss": 2.3563, "step": 3358 }, { "epoch": 1.0529780564263322, "grad_norm": 25.625640869140625, "learning_rate": 1.6795e-06, "loss": 2.7023, "step": 3359 }, { "epoch": 1.0532915360501567, "grad_norm": 29.225643157958984, "learning_rate": 1.6800000000000002e-06, "loss": 2.8436, "step": 3360 }, { "epoch": 1.0536050156739811, "grad_norm": 24.310447692871094, "learning_rate": 1.6805000000000003e-06, "loss": 2.8803, "step": 3361 }, { "epoch": 1.0539184952978056, "grad_norm": 17.30785369873047, "learning_rate": 1.681e-06, "loss": 2.3881, "step": 3362 }, { "epoch": 1.05423197492163, "grad_norm": 21.964794158935547, "learning_rate": 1.6815000000000001e-06, "loss": 2.4327, "step": 3363 }, { "epoch": 1.0545454545454545, "grad_norm": 38.354225158691406, "learning_rate": 1.682e-06, "loss": 3.6111, "step": 3364 }, { "epoch": 1.054858934169279, "grad_norm": 27.03491973876953, "learning_rate": 1.6825000000000002e-06, "loss": 3.5078, "step": 3365 }, { "epoch": 1.0551724137931036, "grad_norm": 34.087646484375, "learning_rate": 1.6830000000000002e-06, "loss": 2.872, "step": 3366 }, { "epoch": 1.055485893416928, "grad_norm": 57.53993225097656, "learning_rate": 1.6835e-06, "loss": 3.4415, "step": 3367 }, { "epoch": 1.0557993730407524, "grad_norm": 119.94190216064453, "learning_rate": 1.684e-06, "loss": 3.5462, "step": 3368 }, { "epoch": 1.0561128526645769, "grad_norm": 46.50600051879883, "learning_rate": 1.6845e-06, "loss": 2.8154, "step": 3369 }, { "epoch": 1.0564263322884013, "grad_norm": 19.846092224121094, "learning_rate": 1.6850000000000002e-06, "loss": 2.6868, "step": 3370 }, { "epoch": 1.0567398119122258, "grad_norm": 42.88930892944336, "learning_rate": 1.6855000000000002e-06, "loss": 3.3179, "step": 3371 }, { "epoch": 1.0570532915360502, "grad_norm": 26.39129638671875, "learning_rate": 1.686e-06, "loss": 3.0197, "step": 3372 }, { "epoch": 1.0573667711598747, "grad_norm": 23.650426864624023, "learning_rate": 1.6865e-06, "loss": 3.1575, "step": 3373 }, { "epoch": 1.057680250783699, "grad_norm": 43.16667556762695, "learning_rate": 1.6870000000000001e-06, "loss": 3.9021, "step": 3374 }, { "epoch": 1.0579937304075235, "grad_norm": 27.793106079101562, "learning_rate": 1.6875000000000001e-06, "loss": 3.1221, "step": 3375 }, { "epoch": 1.058307210031348, "grad_norm": 29.778823852539062, "learning_rate": 1.6880000000000002e-06, "loss": 3.1776, "step": 3376 }, { "epoch": 1.0586206896551724, "grad_norm": 59.28425979614258, "learning_rate": 1.6885000000000002e-06, "loss": 3.1223, "step": 3377 }, { "epoch": 1.0589341692789969, "grad_norm": 26.89908218383789, "learning_rate": 1.689e-06, "loss": 2.7911, "step": 3378 }, { "epoch": 1.0592476489028213, "grad_norm": 33.18085479736328, "learning_rate": 1.6895e-06, "loss": 2.8164, "step": 3379 }, { "epoch": 1.0595611285266457, "grad_norm": 25.427278518676758, "learning_rate": 1.6900000000000003e-06, "loss": 2.6253, "step": 3380 }, { "epoch": 1.0598746081504702, "grad_norm": 67.79798126220703, "learning_rate": 1.6905000000000002e-06, "loss": 3.7484, "step": 3381 }, { "epoch": 1.0601880877742946, "grad_norm": 41.082977294921875, "learning_rate": 1.6910000000000002e-06, "loss": 3.2418, "step": 3382 }, { "epoch": 1.060501567398119, "grad_norm": 113.1292953491211, "learning_rate": 1.6915e-06, "loss": 3.2591, "step": 3383 }, { "epoch": 1.0608150470219435, "grad_norm": 29.729745864868164, "learning_rate": 1.692e-06, "loss": 3.3393, "step": 3384 }, { "epoch": 1.061128526645768, "grad_norm": 39.9203987121582, "learning_rate": 1.6925000000000003e-06, "loss": 3.3071, "step": 3385 }, { "epoch": 1.0614420062695924, "grad_norm": 51.5232048034668, "learning_rate": 1.6930000000000001e-06, "loss": 3.2022, "step": 3386 }, { "epoch": 1.0617554858934168, "grad_norm": 28.027128219604492, "learning_rate": 1.6935000000000002e-06, "loss": 2.9686, "step": 3387 }, { "epoch": 1.0620689655172413, "grad_norm": 17.445575714111328, "learning_rate": 1.694e-06, "loss": 2.3405, "step": 3388 }, { "epoch": 1.0623824451410657, "grad_norm": 31.220802307128906, "learning_rate": 1.6945e-06, "loss": 3.3004, "step": 3389 }, { "epoch": 1.0626959247648902, "grad_norm": 36.90548324584961, "learning_rate": 1.6950000000000003e-06, "loss": 3.3895, "step": 3390 }, { "epoch": 1.0630094043887148, "grad_norm": 19.605175018310547, "learning_rate": 1.6955000000000001e-06, "loss": 3.3256, "step": 3391 }, { "epoch": 1.0633228840125393, "grad_norm": 25.22102928161621, "learning_rate": 1.6960000000000002e-06, "loss": 3.0132, "step": 3392 }, { "epoch": 1.0636363636363637, "grad_norm": 31.76527976989746, "learning_rate": 1.6965000000000002e-06, "loss": 2.8584, "step": 3393 }, { "epoch": 1.0639498432601882, "grad_norm": 27.360450744628906, "learning_rate": 1.697e-06, "loss": 3.5555, "step": 3394 }, { "epoch": 1.0642633228840126, "grad_norm": 39.99030303955078, "learning_rate": 1.6975000000000003e-06, "loss": 3.9428, "step": 3395 }, { "epoch": 1.064576802507837, "grad_norm": 20.094100952148438, "learning_rate": 1.6980000000000003e-06, "loss": 2.9556, "step": 3396 }, { "epoch": 1.0648902821316615, "grad_norm": 27.564455032348633, "learning_rate": 1.6985000000000001e-06, "loss": 2.6065, "step": 3397 }, { "epoch": 1.065203761755486, "grad_norm": 23.814006805419922, "learning_rate": 1.6990000000000002e-06, "loss": 3.0498, "step": 3398 }, { "epoch": 1.0655172413793104, "grad_norm": 22.892719268798828, "learning_rate": 1.6995e-06, "loss": 3.1003, "step": 3399 }, { "epoch": 1.0658307210031348, "grad_norm": 37.36176681518555, "learning_rate": 1.7000000000000002e-06, "loss": 2.8643, "step": 3400 }, { "epoch": 1.0661442006269592, "grad_norm": 27.134132385253906, "learning_rate": 1.7005000000000003e-06, "loss": 3.4379, "step": 3401 }, { "epoch": 1.0664576802507837, "grad_norm": 45.84774398803711, "learning_rate": 1.701e-06, "loss": 3.2277, "step": 3402 }, { "epoch": 1.0667711598746081, "grad_norm": 40.72565841674805, "learning_rate": 1.7015000000000001e-06, "loss": 2.9327, "step": 3403 }, { "epoch": 1.0670846394984326, "grad_norm": 32.90715789794922, "learning_rate": 1.702e-06, "loss": 4.1643, "step": 3404 }, { "epoch": 1.067398119122257, "grad_norm": 29.4326229095459, "learning_rate": 1.7025000000000002e-06, "loss": 3.0146, "step": 3405 }, { "epoch": 1.0677115987460815, "grad_norm": 32.895835876464844, "learning_rate": 1.7030000000000003e-06, "loss": 2.61, "step": 3406 }, { "epoch": 1.068025078369906, "grad_norm": 26.282489776611328, "learning_rate": 1.7035e-06, "loss": 2.6263, "step": 3407 }, { "epoch": 1.0683385579937303, "grad_norm": 50.594940185546875, "learning_rate": 1.7040000000000001e-06, "loss": 2.5367, "step": 3408 }, { "epoch": 1.0686520376175548, "grad_norm": 23.735204696655273, "learning_rate": 1.7045e-06, "loss": 3.0807, "step": 3409 }, { "epoch": 1.0689655172413792, "grad_norm": 29.3944034576416, "learning_rate": 1.7050000000000002e-06, "loss": 2.8559, "step": 3410 }, { "epoch": 1.0692789968652037, "grad_norm": 101.75618743896484, "learning_rate": 1.7055000000000002e-06, "loss": 2.8017, "step": 3411 }, { "epoch": 1.069592476489028, "grad_norm": 32.16104507446289, "learning_rate": 1.706e-06, "loss": 3.0168, "step": 3412 }, { "epoch": 1.0699059561128528, "grad_norm": 33.96107864379883, "learning_rate": 1.7065e-06, "loss": 3.183, "step": 3413 }, { "epoch": 1.0702194357366772, "grad_norm": 24.979026794433594, "learning_rate": 1.7070000000000001e-06, "loss": 2.7527, "step": 3414 }, { "epoch": 1.0705329153605017, "grad_norm": 61.97425842285156, "learning_rate": 1.7075000000000002e-06, "loss": 3.9153, "step": 3415 }, { "epoch": 1.070846394984326, "grad_norm": 33.35090637207031, "learning_rate": 1.7080000000000002e-06, "loss": 3.1761, "step": 3416 }, { "epoch": 1.0711598746081505, "grad_norm": 24.40055274963379, "learning_rate": 1.7085000000000002e-06, "loss": 2.7182, "step": 3417 }, { "epoch": 1.071473354231975, "grad_norm": 36.82908630371094, "learning_rate": 1.709e-06, "loss": 3.2681, "step": 3418 }, { "epoch": 1.0717868338557994, "grad_norm": 75.78736114501953, "learning_rate": 1.7095000000000001e-06, "loss": 3.3688, "step": 3419 }, { "epoch": 1.0721003134796239, "grad_norm": 53.95949935913086, "learning_rate": 1.7100000000000004e-06, "loss": 3.3434, "step": 3420 }, { "epoch": 1.0724137931034483, "grad_norm": 38.91935348510742, "learning_rate": 1.7105000000000002e-06, "loss": 2.6255, "step": 3421 }, { "epoch": 1.0727272727272728, "grad_norm": 31.636262893676758, "learning_rate": 1.7110000000000002e-06, "loss": 3.2822, "step": 3422 }, { "epoch": 1.0730407523510972, "grad_norm": 49.99222183227539, "learning_rate": 1.7115e-06, "loss": 3.5524, "step": 3423 }, { "epoch": 1.0733542319749216, "grad_norm": 28.117033004760742, "learning_rate": 1.712e-06, "loss": 3.115, "step": 3424 }, { "epoch": 1.073667711598746, "grad_norm": 20.26447105407715, "learning_rate": 1.7125000000000003e-06, "loss": 2.6173, "step": 3425 }, { "epoch": 1.0739811912225705, "grad_norm": 17.405311584472656, "learning_rate": 1.7130000000000002e-06, "loss": 3.3576, "step": 3426 }, { "epoch": 1.074294670846395, "grad_norm": 42.87114715576172, "learning_rate": 1.7135000000000002e-06, "loss": 3.0969, "step": 3427 }, { "epoch": 1.0746081504702194, "grad_norm": 20.595788955688477, "learning_rate": 1.714e-06, "loss": 2.6237, "step": 3428 }, { "epoch": 1.0749216300940438, "grad_norm": 36.38020706176758, "learning_rate": 1.7145e-06, "loss": 3.5374, "step": 3429 }, { "epoch": 1.0752351097178683, "grad_norm": 31.2501163482666, "learning_rate": 1.7150000000000003e-06, "loss": 2.8522, "step": 3430 }, { "epoch": 1.0755485893416927, "grad_norm": 29.131214141845703, "learning_rate": 1.7155000000000001e-06, "loss": 2.7789, "step": 3431 }, { "epoch": 1.0758620689655172, "grad_norm": 26.90103530883789, "learning_rate": 1.7160000000000002e-06, "loss": 3.4374, "step": 3432 }, { "epoch": 1.0761755485893416, "grad_norm": 34.345088958740234, "learning_rate": 1.7165000000000002e-06, "loss": 2.7771, "step": 3433 }, { "epoch": 1.076489028213166, "grad_norm": 18.91362190246582, "learning_rate": 1.717e-06, "loss": 2.7293, "step": 3434 }, { "epoch": 1.0768025078369905, "grad_norm": 58.237918853759766, "learning_rate": 1.7175000000000003e-06, "loss": 2.9041, "step": 3435 }, { "epoch": 1.077115987460815, "grad_norm": 45.42815017700195, "learning_rate": 1.7180000000000003e-06, "loss": 3.1195, "step": 3436 }, { "epoch": 1.0774294670846394, "grad_norm": 20.961240768432617, "learning_rate": 1.7185000000000001e-06, "loss": 3.0985, "step": 3437 }, { "epoch": 1.077742946708464, "grad_norm": 31.146696090698242, "learning_rate": 1.7190000000000002e-06, "loss": 2.8074, "step": 3438 }, { "epoch": 1.0780564263322885, "grad_norm": 50.56280517578125, "learning_rate": 1.7195e-06, "loss": 3.9483, "step": 3439 }, { "epoch": 1.078369905956113, "grad_norm": 30.90845489501953, "learning_rate": 1.72e-06, "loss": 2.9276, "step": 3440 }, { "epoch": 1.0786833855799374, "grad_norm": 36.373931884765625, "learning_rate": 1.7205000000000003e-06, "loss": 2.7398, "step": 3441 }, { "epoch": 1.0789968652037618, "grad_norm": 38.69664764404297, "learning_rate": 1.7210000000000001e-06, "loss": 3.1982, "step": 3442 }, { "epoch": 1.0793103448275863, "grad_norm": 52.687503814697266, "learning_rate": 1.7215000000000002e-06, "loss": 2.7719, "step": 3443 }, { "epoch": 1.0796238244514107, "grad_norm": 28.561540603637695, "learning_rate": 1.722e-06, "loss": 2.7896, "step": 3444 }, { "epoch": 1.0799373040752351, "grad_norm": 104.28028869628906, "learning_rate": 1.7225e-06, "loss": 3.9382, "step": 3445 }, { "epoch": 1.0802507836990596, "grad_norm": 35.79922103881836, "learning_rate": 1.7230000000000003e-06, "loss": 3.2333, "step": 3446 }, { "epoch": 1.080564263322884, "grad_norm": 28.640287399291992, "learning_rate": 1.7235e-06, "loss": 2.6655, "step": 3447 }, { "epoch": 1.0808777429467085, "grad_norm": 28.324596405029297, "learning_rate": 1.7240000000000001e-06, "loss": 3.4944, "step": 3448 }, { "epoch": 1.081191222570533, "grad_norm": 35.781715393066406, "learning_rate": 1.7245e-06, "loss": 3.1321, "step": 3449 }, { "epoch": 1.0815047021943573, "grad_norm": 39.355499267578125, "learning_rate": 1.725e-06, "loss": 3.2125, "step": 3450 }, { "epoch": 1.0818181818181818, "grad_norm": 43.08556365966797, "learning_rate": 1.7255000000000003e-06, "loss": 3.4494, "step": 3451 }, { "epoch": 1.0821316614420062, "grad_norm": 49.84754943847656, "learning_rate": 1.726e-06, "loss": 3.232, "step": 3452 }, { "epoch": 1.0824451410658307, "grad_norm": 72.55428314208984, "learning_rate": 1.7265000000000001e-06, "loss": 3.0765, "step": 3453 }, { "epoch": 1.0827586206896551, "grad_norm": 43.84035110473633, "learning_rate": 1.7270000000000002e-06, "loss": 3.2667, "step": 3454 }, { "epoch": 1.0830721003134796, "grad_norm": 21.691329956054688, "learning_rate": 1.7275e-06, "loss": 2.6102, "step": 3455 }, { "epoch": 1.083385579937304, "grad_norm": 28.518117904663086, "learning_rate": 1.7280000000000002e-06, "loss": 3.7609, "step": 3456 }, { "epoch": 1.0836990595611284, "grad_norm": 20.661483764648438, "learning_rate": 1.7285000000000003e-06, "loss": 2.8076, "step": 3457 }, { "epoch": 1.0840125391849529, "grad_norm": 21.91556167602539, "learning_rate": 1.729e-06, "loss": 3.0376, "step": 3458 }, { "epoch": 1.0843260188087773, "grad_norm": 31.704559326171875, "learning_rate": 1.7295000000000001e-06, "loss": 2.8547, "step": 3459 }, { "epoch": 1.084639498432602, "grad_norm": 103.53755187988281, "learning_rate": 1.73e-06, "loss": 3.1362, "step": 3460 }, { "epoch": 1.0849529780564264, "grad_norm": 21.896881103515625, "learning_rate": 1.7305000000000002e-06, "loss": 2.8094, "step": 3461 }, { "epoch": 1.0852664576802509, "grad_norm": 26.890310287475586, "learning_rate": 1.7310000000000002e-06, "loss": 2.6375, "step": 3462 }, { "epoch": 1.0855799373040753, "grad_norm": 21.702211380004883, "learning_rate": 1.7315e-06, "loss": 2.839, "step": 3463 }, { "epoch": 1.0858934169278998, "grad_norm": 17.826766967773438, "learning_rate": 1.732e-06, "loss": 2.754, "step": 3464 }, { "epoch": 1.0862068965517242, "grad_norm": 27.036876678466797, "learning_rate": 1.7325e-06, "loss": 3.0524, "step": 3465 }, { "epoch": 1.0865203761755486, "grad_norm": 50.12677764892578, "learning_rate": 1.7330000000000002e-06, "loss": 4.3457, "step": 3466 }, { "epoch": 1.086833855799373, "grad_norm": 36.31730651855469, "learning_rate": 1.7335000000000002e-06, "loss": 2.7942, "step": 3467 }, { "epoch": 1.0871473354231975, "grad_norm": 29.601634979248047, "learning_rate": 1.734e-06, "loss": 2.8078, "step": 3468 }, { "epoch": 1.087460815047022, "grad_norm": 33.91842269897461, "learning_rate": 1.7345e-06, "loss": 2.7449, "step": 3469 }, { "epoch": 1.0877742946708464, "grad_norm": 27.274818420410156, "learning_rate": 1.7350000000000001e-06, "loss": 2.7981, "step": 3470 }, { "epoch": 1.0880877742946709, "grad_norm": 23.352020263671875, "learning_rate": 1.7355000000000002e-06, "loss": 2.7459, "step": 3471 }, { "epoch": 1.0884012539184953, "grad_norm": 39.16693115234375, "learning_rate": 1.7360000000000002e-06, "loss": 2.7875, "step": 3472 }, { "epoch": 1.0887147335423197, "grad_norm": 32.53330993652344, "learning_rate": 1.7365000000000002e-06, "loss": 3.6786, "step": 3473 }, { "epoch": 1.0890282131661442, "grad_norm": 124.53848266601562, "learning_rate": 1.737e-06, "loss": 3.1399, "step": 3474 }, { "epoch": 1.0893416927899686, "grad_norm": 42.1822395324707, "learning_rate": 1.7375e-06, "loss": 3.3776, "step": 3475 }, { "epoch": 1.089655172413793, "grad_norm": 19.54660987854004, "learning_rate": 1.7380000000000003e-06, "loss": 2.7341, "step": 3476 }, { "epoch": 1.0899686520376175, "grad_norm": 25.96428108215332, "learning_rate": 1.7385000000000002e-06, "loss": 2.8165, "step": 3477 }, { "epoch": 1.090282131661442, "grad_norm": 46.42351150512695, "learning_rate": 1.7390000000000002e-06, "loss": 2.9387, "step": 3478 }, { "epoch": 1.0905956112852664, "grad_norm": 16.924457550048828, "learning_rate": 1.7395e-06, "loss": 2.653, "step": 3479 }, { "epoch": 1.0909090909090908, "grad_norm": 25.124286651611328, "learning_rate": 1.74e-06, "loss": 2.5537, "step": 3480 }, { "epoch": 1.0912225705329153, "grad_norm": 64.10070037841797, "learning_rate": 1.7405000000000003e-06, "loss": 3.6643, "step": 3481 }, { "epoch": 1.0915360501567397, "grad_norm": 24.5253963470459, "learning_rate": 1.7410000000000001e-06, "loss": 2.676, "step": 3482 }, { "epoch": 1.0918495297805642, "grad_norm": 32.09825897216797, "learning_rate": 1.7415000000000002e-06, "loss": 3.3551, "step": 3483 }, { "epoch": 1.0921630094043888, "grad_norm": 35.84313201904297, "learning_rate": 1.742e-06, "loss": 3.4427, "step": 3484 }, { "epoch": 1.0924764890282133, "grad_norm": 51.0280647277832, "learning_rate": 1.7425e-06, "loss": 3.3035, "step": 3485 }, { "epoch": 1.0927899686520377, "grad_norm": 23.819229125976562, "learning_rate": 1.7430000000000003e-06, "loss": 2.9676, "step": 3486 }, { "epoch": 1.0931034482758621, "grad_norm": 38.78548812866211, "learning_rate": 1.7435000000000001e-06, "loss": 3.3918, "step": 3487 }, { "epoch": 1.0934169278996866, "grad_norm": 38.34880065917969, "learning_rate": 1.7440000000000002e-06, "loss": 3.0882, "step": 3488 }, { "epoch": 1.093730407523511, "grad_norm": 32.39979553222656, "learning_rate": 1.7445e-06, "loss": 2.9143, "step": 3489 }, { "epoch": 1.0940438871473355, "grad_norm": 51.53925704956055, "learning_rate": 1.745e-06, "loss": 2.7739, "step": 3490 }, { "epoch": 1.09435736677116, "grad_norm": 47.33768081665039, "learning_rate": 1.7455000000000003e-06, "loss": 3.2417, "step": 3491 }, { "epoch": 1.0946708463949844, "grad_norm": 23.84410285949707, "learning_rate": 1.746e-06, "loss": 2.9263, "step": 3492 }, { "epoch": 1.0949843260188088, "grad_norm": 27.154666900634766, "learning_rate": 1.7465000000000001e-06, "loss": 2.8824, "step": 3493 }, { "epoch": 1.0952978056426332, "grad_norm": 19.792095184326172, "learning_rate": 1.7470000000000002e-06, "loss": 2.5153, "step": 3494 }, { "epoch": 1.0956112852664577, "grad_norm": 29.83961296081543, "learning_rate": 1.7475e-06, "loss": 2.6682, "step": 3495 }, { "epoch": 1.0959247648902821, "grad_norm": 22.432218551635742, "learning_rate": 1.7480000000000002e-06, "loss": 2.6332, "step": 3496 }, { "epoch": 1.0962382445141066, "grad_norm": 34.51369857788086, "learning_rate": 1.7485000000000003e-06, "loss": 2.8702, "step": 3497 }, { "epoch": 1.096551724137931, "grad_norm": 22.819372177124023, "learning_rate": 1.7490000000000001e-06, "loss": 2.5255, "step": 3498 }, { "epoch": 1.0968652037617554, "grad_norm": 22.45680809020996, "learning_rate": 1.7495000000000001e-06, "loss": 3.092, "step": 3499 }, { "epoch": 1.09717868338558, "grad_norm": 47.499847412109375, "learning_rate": 1.75e-06, "loss": 4.2197, "step": 3500 }, { "epoch": 1.0974921630094043, "grad_norm": 75.1513442993164, "learning_rate": 1.7505000000000002e-06, "loss": 3.1316, "step": 3501 }, { "epoch": 1.0978056426332288, "grad_norm": 18.71675682067871, "learning_rate": 1.7510000000000003e-06, "loss": 2.5032, "step": 3502 }, { "epoch": 1.0981191222570532, "grad_norm": 51.24578094482422, "learning_rate": 1.7515e-06, "loss": 3.402, "step": 3503 }, { "epoch": 1.0984326018808777, "grad_norm": 16.40824317932129, "learning_rate": 1.7520000000000001e-06, "loss": 2.4963, "step": 3504 }, { "epoch": 1.098746081504702, "grad_norm": 87.81695556640625, "learning_rate": 1.7525e-06, "loss": 4.5785, "step": 3505 }, { "epoch": 1.0990595611285268, "grad_norm": 46.15744400024414, "learning_rate": 1.7530000000000002e-06, "loss": 3.1064, "step": 3506 }, { "epoch": 1.0993730407523512, "grad_norm": 34.15064239501953, "learning_rate": 1.7535000000000002e-06, "loss": 3.1321, "step": 3507 }, { "epoch": 1.0996865203761756, "grad_norm": 25.923816680908203, "learning_rate": 1.754e-06, "loss": 3.0893, "step": 3508 }, { "epoch": 1.1, "grad_norm": 32.922027587890625, "learning_rate": 1.7545e-06, "loss": 2.608, "step": 3509 }, { "epoch": 1.1003134796238245, "grad_norm": 37.616695404052734, "learning_rate": 1.7550000000000001e-06, "loss": 3.7035, "step": 3510 }, { "epoch": 1.100626959247649, "grad_norm": 35.45209884643555, "learning_rate": 1.7555000000000002e-06, "loss": 2.9419, "step": 3511 }, { "epoch": 1.1009404388714734, "grad_norm": 35.20869064331055, "learning_rate": 1.7560000000000002e-06, "loss": 3.0732, "step": 3512 }, { "epoch": 1.1012539184952979, "grad_norm": 44.86465835571289, "learning_rate": 1.7565000000000003e-06, "loss": 3.3239, "step": 3513 }, { "epoch": 1.1015673981191223, "grad_norm": 22.87040901184082, "learning_rate": 1.757e-06, "loss": 2.6935, "step": 3514 }, { "epoch": 1.1018808777429467, "grad_norm": 33.13298034667969, "learning_rate": 1.7575000000000001e-06, "loss": 2.7612, "step": 3515 }, { "epoch": 1.1021943573667712, "grad_norm": 18.739044189453125, "learning_rate": 1.7580000000000004e-06, "loss": 2.7834, "step": 3516 }, { "epoch": 1.1025078369905956, "grad_norm": 40.461647033691406, "learning_rate": 1.7585000000000002e-06, "loss": 3.0305, "step": 3517 }, { "epoch": 1.10282131661442, "grad_norm": 27.588546752929688, "learning_rate": 1.7590000000000002e-06, "loss": 2.4332, "step": 3518 }, { "epoch": 1.1031347962382445, "grad_norm": 26.194929122924805, "learning_rate": 1.7595e-06, "loss": 2.8379, "step": 3519 }, { "epoch": 1.103448275862069, "grad_norm": 34.12715530395508, "learning_rate": 1.76e-06, "loss": 2.7955, "step": 3520 }, { "epoch": 1.1037617554858934, "grad_norm": 24.855409622192383, "learning_rate": 1.7605000000000003e-06, "loss": 2.7204, "step": 3521 }, { "epoch": 1.1040752351097178, "grad_norm": 22.457441329956055, "learning_rate": 1.7610000000000002e-06, "loss": 3.2287, "step": 3522 }, { "epoch": 1.1043887147335423, "grad_norm": 21.55569839477539, "learning_rate": 1.7615000000000002e-06, "loss": 3.0252, "step": 3523 }, { "epoch": 1.1047021943573667, "grad_norm": 22.316848754882812, "learning_rate": 1.762e-06, "loss": 2.7789, "step": 3524 }, { "epoch": 1.1050156739811912, "grad_norm": 51.03260803222656, "learning_rate": 1.7625e-06, "loss": 3.6778, "step": 3525 }, { "epoch": 1.1053291536050156, "grad_norm": 20.623655319213867, "learning_rate": 1.7630000000000003e-06, "loss": 2.9086, "step": 3526 }, { "epoch": 1.10564263322884, "grad_norm": 22.799224853515625, "learning_rate": 1.7635000000000001e-06, "loss": 2.9854, "step": 3527 }, { "epoch": 1.1059561128526645, "grad_norm": 104.5829086303711, "learning_rate": 1.7640000000000002e-06, "loss": 3.4163, "step": 3528 }, { "epoch": 1.106269592476489, "grad_norm": 46.149879455566406, "learning_rate": 1.7645e-06, "loss": 2.5803, "step": 3529 }, { "epoch": 1.1065830721003134, "grad_norm": 21.742412567138672, "learning_rate": 1.765e-06, "loss": 2.3909, "step": 3530 }, { "epoch": 1.106896551724138, "grad_norm": 21.718233108520508, "learning_rate": 1.7655000000000003e-06, "loss": 2.7341, "step": 3531 }, { "epoch": 1.1072100313479625, "grad_norm": 19.40610122680664, "learning_rate": 1.7660000000000001e-06, "loss": 2.5342, "step": 3532 }, { "epoch": 1.107523510971787, "grad_norm": 29.3079891204834, "learning_rate": 1.7665000000000002e-06, "loss": 4.2854, "step": 3533 }, { "epoch": 1.1078369905956114, "grad_norm": 31.97117805480957, "learning_rate": 1.7670000000000002e-06, "loss": 2.8468, "step": 3534 }, { "epoch": 1.1081504702194358, "grad_norm": 32.589317321777344, "learning_rate": 1.7675e-06, "loss": 2.9813, "step": 3535 }, { "epoch": 1.1084639498432602, "grad_norm": 24.950849533081055, "learning_rate": 1.7680000000000003e-06, "loss": 2.7851, "step": 3536 }, { "epoch": 1.1087774294670847, "grad_norm": 29.49528694152832, "learning_rate": 1.7685000000000003e-06, "loss": 3.3423, "step": 3537 }, { "epoch": 1.1090909090909091, "grad_norm": 23.108978271484375, "learning_rate": 1.7690000000000001e-06, "loss": 2.9801, "step": 3538 }, { "epoch": 1.1094043887147336, "grad_norm": 26.269826889038086, "learning_rate": 1.7695000000000002e-06, "loss": 2.9439, "step": 3539 }, { "epoch": 1.109717868338558, "grad_norm": 25.8447322845459, "learning_rate": 1.77e-06, "loss": 2.7837, "step": 3540 }, { "epoch": 1.1100313479623825, "grad_norm": 25.49906349182129, "learning_rate": 1.7705000000000002e-06, "loss": 2.6496, "step": 3541 }, { "epoch": 1.110344827586207, "grad_norm": 31.025354385375977, "learning_rate": 1.7710000000000003e-06, "loss": 2.546, "step": 3542 }, { "epoch": 1.1106583072100313, "grad_norm": 34.72405242919922, "learning_rate": 1.7715000000000001e-06, "loss": 2.8026, "step": 3543 }, { "epoch": 1.1109717868338558, "grad_norm": 36.77132034301758, "learning_rate": 1.7720000000000001e-06, "loss": 3.3022, "step": 3544 }, { "epoch": 1.1112852664576802, "grad_norm": 32.44343566894531, "learning_rate": 1.7725e-06, "loss": 3.0009, "step": 3545 }, { "epoch": 1.1115987460815047, "grad_norm": 18.060049057006836, "learning_rate": 1.7730000000000002e-06, "loss": 2.5424, "step": 3546 }, { "epoch": 1.111912225705329, "grad_norm": 33.717308044433594, "learning_rate": 1.7735000000000003e-06, "loss": 3.7549, "step": 3547 }, { "epoch": 1.1122257053291535, "grad_norm": 18.197097778320312, "learning_rate": 1.774e-06, "loss": 2.7566, "step": 3548 }, { "epoch": 1.112539184952978, "grad_norm": 31.980981826782227, "learning_rate": 1.7745000000000001e-06, "loss": 3.3021, "step": 3549 }, { "epoch": 1.1128526645768024, "grad_norm": 47.19995880126953, "learning_rate": 1.7750000000000002e-06, "loss": 2.9785, "step": 3550 }, { "epoch": 1.1131661442006269, "grad_norm": 23.19911766052246, "learning_rate": 1.7755000000000002e-06, "loss": 3.0185, "step": 3551 }, { "epoch": 1.1134796238244513, "grad_norm": 35.693580627441406, "learning_rate": 1.7760000000000002e-06, "loss": 3.4506, "step": 3552 }, { "epoch": 1.113793103448276, "grad_norm": 36.20033645629883, "learning_rate": 1.7765000000000003e-06, "loss": 2.6537, "step": 3553 }, { "epoch": 1.1141065830721004, "grad_norm": 40.76347732543945, "learning_rate": 1.777e-06, "loss": 2.6187, "step": 3554 }, { "epoch": 1.1144200626959249, "grad_norm": 33.449562072753906, "learning_rate": 1.7775000000000001e-06, "loss": 3.0086, "step": 3555 }, { "epoch": 1.1147335423197493, "grad_norm": 70.72737884521484, "learning_rate": 1.7780000000000004e-06, "loss": 3.4293, "step": 3556 }, { "epoch": 1.1150470219435737, "grad_norm": 24.169923782348633, "learning_rate": 1.7785000000000002e-06, "loss": 2.5885, "step": 3557 }, { "epoch": 1.1153605015673982, "grad_norm": 22.239110946655273, "learning_rate": 1.7790000000000002e-06, "loss": 2.6941, "step": 3558 }, { "epoch": 1.1156739811912226, "grad_norm": 31.481800079345703, "learning_rate": 1.7795e-06, "loss": 2.9983, "step": 3559 }, { "epoch": 1.115987460815047, "grad_norm": 32.51259994506836, "learning_rate": 1.7800000000000001e-06, "loss": 2.5775, "step": 3560 }, { "epoch": 1.1163009404388715, "grad_norm": 23.58956527709961, "learning_rate": 1.7805000000000004e-06, "loss": 3.3801, "step": 3561 }, { "epoch": 1.116614420062696, "grad_norm": 22.67048454284668, "learning_rate": 1.7810000000000002e-06, "loss": 2.5815, "step": 3562 }, { "epoch": 1.1169278996865204, "grad_norm": 28.150209426879883, "learning_rate": 1.7815000000000002e-06, "loss": 2.6133, "step": 3563 }, { "epoch": 1.1172413793103448, "grad_norm": 63.03923416137695, "learning_rate": 1.782e-06, "loss": 2.9811, "step": 3564 }, { "epoch": 1.1175548589341693, "grad_norm": 101.25431060791016, "learning_rate": 1.7825e-06, "loss": 3.5238, "step": 3565 }, { "epoch": 1.1178683385579937, "grad_norm": 70.68958282470703, "learning_rate": 1.783e-06, "loss": 3.4413, "step": 3566 }, { "epoch": 1.1181818181818182, "grad_norm": 18.009140014648438, "learning_rate": 1.7835000000000002e-06, "loss": 2.4277, "step": 3567 }, { "epoch": 1.1184952978056426, "grad_norm": 41.71234893798828, "learning_rate": 1.7840000000000002e-06, "loss": 3.7431, "step": 3568 }, { "epoch": 1.118808777429467, "grad_norm": 62.48421096801758, "learning_rate": 1.7845e-06, "loss": 3.1696, "step": 3569 }, { "epoch": 1.1191222570532915, "grad_norm": 36.058467864990234, "learning_rate": 1.785e-06, "loss": 4.098, "step": 3570 }, { "epoch": 1.119435736677116, "grad_norm": 45.1207160949707, "learning_rate": 1.7855e-06, "loss": 2.8022, "step": 3571 }, { "epoch": 1.1197492163009404, "grad_norm": 200.1226348876953, "learning_rate": 1.7860000000000001e-06, "loss": 3.0398, "step": 3572 }, { "epoch": 1.1200626959247648, "grad_norm": 20.55527114868164, "learning_rate": 1.7865000000000002e-06, "loss": 2.7789, "step": 3573 }, { "epoch": 1.1203761755485893, "grad_norm": 42.2872428894043, "learning_rate": 1.7870000000000002e-06, "loss": 2.9237, "step": 3574 }, { "epoch": 1.1206896551724137, "grad_norm": 35.486915588378906, "learning_rate": 1.7875e-06, "loss": 3.1821, "step": 3575 }, { "epoch": 1.1210031347962381, "grad_norm": 40.49298095703125, "learning_rate": 1.788e-06, "loss": 3.2534, "step": 3576 }, { "epoch": 1.1213166144200626, "grad_norm": 24.555992126464844, "learning_rate": 1.7885000000000003e-06, "loss": 2.5797, "step": 3577 }, { "epoch": 1.1216300940438872, "grad_norm": 35.7850341796875, "learning_rate": 1.7890000000000002e-06, "loss": 3.2042, "step": 3578 }, { "epoch": 1.1219435736677117, "grad_norm": 47.83175277709961, "learning_rate": 1.7895000000000002e-06, "loss": 2.9337, "step": 3579 }, { "epoch": 1.1222570532915361, "grad_norm": 29.239980697631836, "learning_rate": 1.79e-06, "loss": 2.608, "step": 3580 }, { "epoch": 1.1225705329153606, "grad_norm": 22.09000015258789, "learning_rate": 1.7905e-06, "loss": 2.3577, "step": 3581 }, { "epoch": 1.122884012539185, "grad_norm": 33.30189514160156, "learning_rate": 1.7910000000000003e-06, "loss": 2.899, "step": 3582 }, { "epoch": 1.1231974921630095, "grad_norm": 18.214956283569336, "learning_rate": 1.7915000000000001e-06, "loss": 2.6793, "step": 3583 }, { "epoch": 1.123510971786834, "grad_norm": 29.422252655029297, "learning_rate": 1.7920000000000002e-06, "loss": 2.7343, "step": 3584 }, { "epoch": 1.1238244514106583, "grad_norm": 54.86771011352539, "learning_rate": 1.7925e-06, "loss": 2.9922, "step": 3585 }, { "epoch": 1.1241379310344828, "grad_norm": 33.0071907043457, "learning_rate": 1.793e-06, "loss": 2.7256, "step": 3586 }, { "epoch": 1.1244514106583072, "grad_norm": 38.235755920410156, "learning_rate": 1.7935000000000003e-06, "loss": 3.1273, "step": 3587 }, { "epoch": 1.1247648902821317, "grad_norm": 34.35319900512695, "learning_rate": 1.794e-06, "loss": 2.7182, "step": 3588 }, { "epoch": 1.125078369905956, "grad_norm": 24.03645896911621, "learning_rate": 1.7945000000000001e-06, "loss": 3.0134, "step": 3589 }, { "epoch": 1.1253918495297806, "grad_norm": 22.386899948120117, "learning_rate": 1.7950000000000002e-06, "loss": 2.56, "step": 3590 }, { "epoch": 1.125705329153605, "grad_norm": 21.036725997924805, "learning_rate": 1.7955e-06, "loss": 3.0022, "step": 3591 }, { "epoch": 1.1260188087774294, "grad_norm": 37.754215240478516, "learning_rate": 1.7960000000000003e-06, "loss": 2.8063, "step": 3592 }, { "epoch": 1.1263322884012539, "grad_norm": 34.971317291259766, "learning_rate": 1.7965000000000003e-06, "loss": 2.4451, "step": 3593 }, { "epoch": 1.1266457680250783, "grad_norm": 27.26007080078125, "learning_rate": 1.7970000000000001e-06, "loss": 2.6376, "step": 3594 }, { "epoch": 1.1269592476489028, "grad_norm": 31.517719268798828, "learning_rate": 1.7975000000000002e-06, "loss": 2.7497, "step": 3595 }, { "epoch": 1.1272727272727272, "grad_norm": 32.8683967590332, "learning_rate": 1.798e-06, "loss": 2.6048, "step": 3596 }, { "epoch": 1.1275862068965516, "grad_norm": 25.371549606323242, "learning_rate": 1.7985000000000002e-06, "loss": 2.5012, "step": 3597 }, { "epoch": 1.127899686520376, "grad_norm": 41.98431396484375, "learning_rate": 1.7990000000000003e-06, "loss": 3.0868, "step": 3598 }, { "epoch": 1.1282131661442005, "grad_norm": 32.28398513793945, "learning_rate": 1.7995e-06, "loss": 2.5616, "step": 3599 }, { "epoch": 1.1285266457680252, "grad_norm": 41.10795974731445, "learning_rate": 1.8000000000000001e-06, "loss": 2.9102, "step": 3600 }, { "epoch": 1.1288401253918496, "grad_norm": 40.71234130859375, "learning_rate": 1.8005e-06, "loss": 3.9757, "step": 3601 }, { "epoch": 1.129153605015674, "grad_norm": 29.76866340637207, "learning_rate": 1.8010000000000002e-06, "loss": 3.0691, "step": 3602 }, { "epoch": 1.1294670846394985, "grad_norm": 43.9595832824707, "learning_rate": 1.8015000000000002e-06, "loss": 2.918, "step": 3603 }, { "epoch": 1.129780564263323, "grad_norm": 34.53847122192383, "learning_rate": 1.802e-06, "loss": 3.1386, "step": 3604 }, { "epoch": 1.1300940438871474, "grad_norm": 29.814184188842773, "learning_rate": 1.8025000000000001e-06, "loss": 3.0358, "step": 3605 }, { "epoch": 1.1304075235109718, "grad_norm": 36.51237106323242, "learning_rate": 1.803e-06, "loss": 3.2394, "step": 3606 }, { "epoch": 1.1307210031347963, "grad_norm": 21.537824630737305, "learning_rate": 1.8035000000000002e-06, "loss": 2.8301, "step": 3607 }, { "epoch": 1.1310344827586207, "grad_norm": 53.89097213745117, "learning_rate": 1.8040000000000002e-06, "loss": 3.1312, "step": 3608 }, { "epoch": 1.1313479623824452, "grad_norm": 25.06736946105957, "learning_rate": 1.8045e-06, "loss": 2.9153, "step": 3609 }, { "epoch": 1.1316614420062696, "grad_norm": 21.754596710205078, "learning_rate": 1.805e-06, "loss": 3.1249, "step": 3610 }, { "epoch": 1.131974921630094, "grad_norm": 71.8990707397461, "learning_rate": 1.8055000000000001e-06, "loss": 2.648, "step": 3611 }, { "epoch": 1.1322884012539185, "grad_norm": 29.600143432617188, "learning_rate": 1.8060000000000002e-06, "loss": 2.9847, "step": 3612 }, { "epoch": 1.132601880877743, "grad_norm": 20.959943771362305, "learning_rate": 1.8065000000000002e-06, "loss": 2.438, "step": 3613 }, { "epoch": 1.1329153605015674, "grad_norm": 31.19559669494629, "learning_rate": 1.8070000000000002e-06, "loss": 3.2214, "step": 3614 }, { "epoch": 1.1332288401253918, "grad_norm": 42.91053009033203, "learning_rate": 1.8075e-06, "loss": 2.8965, "step": 3615 }, { "epoch": 1.1335423197492163, "grad_norm": 82.3578872680664, "learning_rate": 1.808e-06, "loss": 3.4786, "step": 3616 }, { "epoch": 1.1338557993730407, "grad_norm": 33.798587799072266, "learning_rate": 1.8085000000000003e-06, "loss": 3.9894, "step": 3617 }, { "epoch": 1.1341692789968651, "grad_norm": 23.663040161132812, "learning_rate": 1.8090000000000002e-06, "loss": 2.8058, "step": 3618 }, { "epoch": 1.1344827586206896, "grad_norm": 25.969789505004883, "learning_rate": 1.8095000000000002e-06, "loss": 2.981, "step": 3619 }, { "epoch": 1.134796238244514, "grad_norm": 34.255165100097656, "learning_rate": 1.81e-06, "loss": 3.3658, "step": 3620 }, { "epoch": 1.1351097178683385, "grad_norm": 25.960176467895508, "learning_rate": 1.8105e-06, "loss": 2.7904, "step": 3621 }, { "epoch": 1.135423197492163, "grad_norm": 23.85843849182129, "learning_rate": 1.8110000000000003e-06, "loss": 3.1468, "step": 3622 }, { "epoch": 1.1357366771159874, "grad_norm": 40.975677490234375, "learning_rate": 1.8115000000000001e-06, "loss": 2.6138, "step": 3623 }, { "epoch": 1.1360501567398118, "grad_norm": 53.12677001953125, "learning_rate": 1.8120000000000002e-06, "loss": 3.0403, "step": 3624 }, { "epoch": 1.1363636363636362, "grad_norm": 28.290157318115234, "learning_rate": 1.8125e-06, "loss": 2.918, "step": 3625 }, { "epoch": 1.136677115987461, "grad_norm": 25.967588424682617, "learning_rate": 1.813e-06, "loss": 3.2999, "step": 3626 }, { "epoch": 1.1369905956112853, "grad_norm": 28.83611297607422, "learning_rate": 1.8135000000000003e-06, "loss": 2.4556, "step": 3627 }, { "epoch": 1.1373040752351098, "grad_norm": 33.06849670410156, "learning_rate": 1.8140000000000001e-06, "loss": 2.8766, "step": 3628 }, { "epoch": 1.1376175548589342, "grad_norm": 109.74273681640625, "learning_rate": 1.8145000000000002e-06, "loss": 2.9064, "step": 3629 }, { "epoch": 1.1379310344827587, "grad_norm": 29.229536056518555, "learning_rate": 1.8150000000000002e-06, "loss": 2.6592, "step": 3630 }, { "epoch": 1.1382445141065831, "grad_norm": 55.26179885864258, "learning_rate": 1.8155e-06, "loss": 2.8912, "step": 3631 }, { "epoch": 1.1385579937304076, "grad_norm": 23.847789764404297, "learning_rate": 1.8160000000000003e-06, "loss": 2.6741, "step": 3632 }, { "epoch": 1.138871473354232, "grad_norm": 93.24641418457031, "learning_rate": 1.8165000000000003e-06, "loss": 3.6647, "step": 3633 }, { "epoch": 1.1391849529780564, "grad_norm": 24.152320861816406, "learning_rate": 1.8170000000000001e-06, "loss": 2.5677, "step": 3634 }, { "epoch": 1.1394984326018809, "grad_norm": 37.18745040893555, "learning_rate": 1.8175000000000002e-06, "loss": 2.7119, "step": 3635 }, { "epoch": 1.1398119122257053, "grad_norm": 18.85127067565918, "learning_rate": 1.818e-06, "loss": 2.6048, "step": 3636 }, { "epoch": 1.1401253918495298, "grad_norm": 64.41525268554688, "learning_rate": 1.8185000000000002e-06, "loss": 3.0811, "step": 3637 }, { "epoch": 1.1404388714733542, "grad_norm": 47.21731185913086, "learning_rate": 1.8190000000000003e-06, "loss": 2.8125, "step": 3638 }, { "epoch": 1.1407523510971787, "grad_norm": 87.95037078857422, "learning_rate": 1.8195000000000001e-06, "loss": 3.9969, "step": 3639 }, { "epoch": 1.141065830721003, "grad_norm": 69.32103729248047, "learning_rate": 1.8200000000000002e-06, "loss": 3.752, "step": 3640 }, { "epoch": 1.1413793103448275, "grad_norm": 42.07493591308594, "learning_rate": 1.8205e-06, "loss": 3.4211, "step": 3641 }, { "epoch": 1.141692789968652, "grad_norm": 27.793582916259766, "learning_rate": 1.8210000000000002e-06, "loss": 2.6521, "step": 3642 }, { "epoch": 1.1420062695924764, "grad_norm": 25.326906204223633, "learning_rate": 1.8215000000000003e-06, "loss": 2.6708, "step": 3643 }, { "epoch": 1.1423197492163009, "grad_norm": 52.35178756713867, "learning_rate": 1.822e-06, "loss": 3.6085, "step": 3644 }, { "epoch": 1.1426332288401253, "grad_norm": 26.3311824798584, "learning_rate": 1.8225000000000001e-06, "loss": 3.0224, "step": 3645 }, { "epoch": 1.14294670846395, "grad_norm": 23.436771392822266, "learning_rate": 1.823e-06, "loss": 3.0373, "step": 3646 }, { "epoch": 1.1432601880877744, "grad_norm": 21.673803329467773, "learning_rate": 1.8235000000000002e-06, "loss": 2.7107, "step": 3647 }, { "epoch": 1.1435736677115989, "grad_norm": 18.09606170654297, "learning_rate": 1.8240000000000002e-06, "loss": 2.8579, "step": 3648 }, { "epoch": 1.1438871473354233, "grad_norm": 15.521315574645996, "learning_rate": 1.8245e-06, "loss": 2.6568, "step": 3649 }, { "epoch": 1.1442006269592477, "grad_norm": 22.498708724975586, "learning_rate": 1.825e-06, "loss": 2.571, "step": 3650 }, { "epoch": 1.1445141065830722, "grad_norm": 40.23341369628906, "learning_rate": 1.8255000000000001e-06, "loss": 2.7546, "step": 3651 }, { "epoch": 1.1448275862068966, "grad_norm": 20.80241584777832, "learning_rate": 1.8260000000000002e-06, "loss": 2.4507, "step": 3652 }, { "epoch": 1.145141065830721, "grad_norm": 16.20147705078125, "learning_rate": 1.8265000000000002e-06, "loss": 2.4255, "step": 3653 }, { "epoch": 1.1454545454545455, "grad_norm": 25.723865509033203, "learning_rate": 1.8270000000000003e-06, "loss": 2.5392, "step": 3654 }, { "epoch": 1.14576802507837, "grad_norm": 35.9820671081543, "learning_rate": 1.8275e-06, "loss": 3.0196, "step": 3655 }, { "epoch": 1.1460815047021944, "grad_norm": 22.180850982666016, "learning_rate": 1.8280000000000001e-06, "loss": 2.7761, "step": 3656 }, { "epoch": 1.1463949843260188, "grad_norm": 29.314672470092773, "learning_rate": 1.8285000000000004e-06, "loss": 3.6485, "step": 3657 }, { "epoch": 1.1467084639498433, "grad_norm": 48.099334716796875, "learning_rate": 1.8290000000000002e-06, "loss": 2.9849, "step": 3658 }, { "epoch": 1.1470219435736677, "grad_norm": 28.2821102142334, "learning_rate": 1.8295000000000002e-06, "loss": 2.592, "step": 3659 }, { "epoch": 1.1473354231974922, "grad_norm": 25.157413482666016, "learning_rate": 1.83e-06, "loss": 2.9226, "step": 3660 }, { "epoch": 1.1476489028213166, "grad_norm": 57.659461975097656, "learning_rate": 1.8305e-06, "loss": 2.345, "step": 3661 }, { "epoch": 1.147962382445141, "grad_norm": 38.85525894165039, "learning_rate": 1.8310000000000003e-06, "loss": 3.045, "step": 3662 }, { "epoch": 1.1482758620689655, "grad_norm": 60.336917877197266, "learning_rate": 1.8315000000000002e-06, "loss": 3.0008, "step": 3663 }, { "epoch": 1.14858934169279, "grad_norm": 30.43072509765625, "learning_rate": 1.8320000000000002e-06, "loss": 3.0494, "step": 3664 }, { "epoch": 1.1489028213166144, "grad_norm": 70.96421813964844, "learning_rate": 1.8325e-06, "loss": 2.9899, "step": 3665 }, { "epoch": 1.1492163009404388, "grad_norm": 24.374061584472656, "learning_rate": 1.833e-06, "loss": 2.9697, "step": 3666 }, { "epoch": 1.1495297805642632, "grad_norm": 23.166545867919922, "learning_rate": 1.8335000000000003e-06, "loss": 2.4767, "step": 3667 }, { "epoch": 1.1498432601880877, "grad_norm": 23.386566162109375, "learning_rate": 1.8340000000000001e-06, "loss": 2.6403, "step": 3668 }, { "epoch": 1.1501567398119121, "grad_norm": 46.93557357788086, "learning_rate": 1.8345000000000002e-06, "loss": 3.3012, "step": 3669 }, { "epoch": 1.1504702194357366, "grad_norm": 73.08238220214844, "learning_rate": 1.8350000000000002e-06, "loss": 3.0202, "step": 3670 }, { "epoch": 1.150783699059561, "grad_norm": 28.248533248901367, "learning_rate": 1.8355e-06, "loss": 3.7443, "step": 3671 }, { "epoch": 1.1510971786833855, "grad_norm": 27.702686309814453, "learning_rate": 1.8360000000000003e-06, "loss": 3.2577, "step": 3672 }, { "epoch": 1.1514106583072101, "grad_norm": 27.311302185058594, "learning_rate": 1.8365000000000003e-06, "loss": 2.8966, "step": 3673 }, { "epoch": 1.1517241379310346, "grad_norm": 51.159976959228516, "learning_rate": 1.8370000000000002e-06, "loss": 3.3169, "step": 3674 }, { "epoch": 1.152037617554859, "grad_norm": 32.074832916259766, "learning_rate": 1.8375000000000002e-06, "loss": 2.2569, "step": 3675 }, { "epoch": 1.1523510971786834, "grad_norm": 35.85694885253906, "learning_rate": 1.838e-06, "loss": 2.7525, "step": 3676 }, { "epoch": 1.152664576802508, "grad_norm": 24.524778366088867, "learning_rate": 1.8385000000000003e-06, "loss": 3.047, "step": 3677 }, { "epoch": 1.1529780564263323, "grad_norm": 38.84641647338867, "learning_rate": 1.8390000000000003e-06, "loss": 3.4556, "step": 3678 }, { "epoch": 1.1532915360501568, "grad_norm": 22.57017707824707, "learning_rate": 1.8395000000000001e-06, "loss": 2.5072, "step": 3679 }, { "epoch": 1.1536050156739812, "grad_norm": 62.06796646118164, "learning_rate": 1.8400000000000002e-06, "loss": 3.7502, "step": 3680 }, { "epoch": 1.1539184952978057, "grad_norm": 19.97160530090332, "learning_rate": 1.8405e-06, "loss": 2.7028, "step": 3681 }, { "epoch": 1.15423197492163, "grad_norm": 34.08852005004883, "learning_rate": 1.8410000000000002e-06, "loss": 2.8553, "step": 3682 }, { "epoch": 1.1545454545454545, "grad_norm": 24.2499942779541, "learning_rate": 1.8415000000000003e-06, "loss": 3.4699, "step": 3683 }, { "epoch": 1.154858934169279, "grad_norm": 26.6270694732666, "learning_rate": 1.8420000000000001e-06, "loss": 2.8316, "step": 3684 }, { "epoch": 1.1551724137931034, "grad_norm": 18.375274658203125, "learning_rate": 1.8425000000000001e-06, "loss": 2.711, "step": 3685 }, { "epoch": 1.1554858934169279, "grad_norm": 18.933826446533203, "learning_rate": 1.843e-06, "loss": 2.9843, "step": 3686 }, { "epoch": 1.1557993730407523, "grad_norm": 19.643722534179688, "learning_rate": 1.8435000000000002e-06, "loss": 2.8432, "step": 3687 }, { "epoch": 1.1561128526645768, "grad_norm": 103.77330017089844, "learning_rate": 1.8440000000000003e-06, "loss": 2.9164, "step": 3688 }, { "epoch": 1.1564263322884012, "grad_norm": 103.93336486816406, "learning_rate": 1.8445e-06, "loss": 2.6389, "step": 3689 }, { "epoch": 1.1567398119122256, "grad_norm": 16.40753746032715, "learning_rate": 1.8450000000000001e-06, "loss": 2.5037, "step": 3690 }, { "epoch": 1.15705329153605, "grad_norm": 31.478145599365234, "learning_rate": 1.8455000000000002e-06, "loss": 2.0934, "step": 3691 }, { "epoch": 1.1573667711598745, "grad_norm": 17.865964889526367, "learning_rate": 1.846e-06, "loss": 2.6648, "step": 3692 }, { "epoch": 1.1576802507836992, "grad_norm": 22.187891006469727, "learning_rate": 1.8465000000000002e-06, "loss": 2.4895, "step": 3693 }, { "epoch": 1.1579937304075236, "grad_norm": 22.099071502685547, "learning_rate": 1.8470000000000003e-06, "loss": 2.8543, "step": 3694 }, { "epoch": 1.158307210031348, "grad_norm": 67.45195007324219, "learning_rate": 1.8475e-06, "loss": 2.8602, "step": 3695 }, { "epoch": 1.1586206896551725, "grad_norm": 29.56635284423828, "learning_rate": 1.8480000000000001e-06, "loss": 2.5955, "step": 3696 }, { "epoch": 1.158934169278997, "grad_norm": 23.719898223876953, "learning_rate": 1.8485e-06, "loss": 2.8614, "step": 3697 }, { "epoch": 1.1592476489028214, "grad_norm": 27.395971298217773, "learning_rate": 1.8490000000000002e-06, "loss": 2.8527, "step": 3698 }, { "epoch": 1.1595611285266458, "grad_norm": 18.854639053344727, "learning_rate": 1.8495000000000002e-06, "loss": 2.5318, "step": 3699 }, { "epoch": 1.1598746081504703, "grad_norm": 17.002197265625, "learning_rate": 1.85e-06, "loss": 2.5792, "step": 3700 }, { "epoch": 1.1601880877742947, "grad_norm": 22.917043685913086, "learning_rate": 1.8505000000000001e-06, "loss": 2.4931, "step": 3701 }, { "epoch": 1.1605015673981192, "grad_norm": 45.23145294189453, "learning_rate": 1.851e-06, "loss": 2.9302, "step": 3702 }, { "epoch": 1.1608150470219436, "grad_norm": 34.68376922607422, "learning_rate": 1.8515000000000002e-06, "loss": 2.8426, "step": 3703 }, { "epoch": 1.161128526645768, "grad_norm": 30.177770614624023, "learning_rate": 1.8520000000000002e-06, "loss": 2.4237, "step": 3704 }, { "epoch": 1.1614420062695925, "grad_norm": 62.9578971862793, "learning_rate": 1.8525e-06, "loss": 4.0864, "step": 3705 }, { "epoch": 1.161755485893417, "grad_norm": 28.61890411376953, "learning_rate": 1.853e-06, "loss": 2.6907, "step": 3706 }, { "epoch": 1.1620689655172414, "grad_norm": 32.607295989990234, "learning_rate": 1.8535000000000001e-06, "loss": 2.6239, "step": 3707 }, { "epoch": 1.1623824451410658, "grad_norm": 32.68761444091797, "learning_rate": 1.8540000000000002e-06, "loss": 2.8229, "step": 3708 }, { "epoch": 1.1626959247648903, "grad_norm": 26.804277420043945, "learning_rate": 1.8545000000000002e-06, "loss": 2.6434, "step": 3709 }, { "epoch": 1.1630094043887147, "grad_norm": 30.005191802978516, "learning_rate": 1.8550000000000002e-06, "loss": 3.8045, "step": 3710 }, { "epoch": 1.1633228840125391, "grad_norm": 91.55377960205078, "learning_rate": 1.8555e-06, "loss": 2.93, "step": 3711 }, { "epoch": 1.1636363636363636, "grad_norm": 40.16615295410156, "learning_rate": 1.856e-06, "loss": 2.7624, "step": 3712 }, { "epoch": 1.163949843260188, "grad_norm": 29.836450576782227, "learning_rate": 1.8565000000000004e-06, "loss": 2.6629, "step": 3713 }, { "epoch": 1.1642633228840125, "grad_norm": 46.64818572998047, "learning_rate": 1.8570000000000002e-06, "loss": 3.0849, "step": 3714 }, { "epoch": 1.164576802507837, "grad_norm": 27.5681095123291, "learning_rate": 1.8575000000000002e-06, "loss": 2.522, "step": 3715 }, { "epoch": 1.1648902821316613, "grad_norm": 35.278045654296875, "learning_rate": 1.858e-06, "loss": 2.6238, "step": 3716 }, { "epoch": 1.1652037617554858, "grad_norm": 18.06938934326172, "learning_rate": 1.8585e-06, "loss": 2.4448, "step": 3717 }, { "epoch": 1.1655172413793102, "grad_norm": 25.87566566467285, "learning_rate": 1.8590000000000003e-06, "loss": 2.6019, "step": 3718 }, { "epoch": 1.1658307210031347, "grad_norm": 36.46488952636719, "learning_rate": 1.8595000000000002e-06, "loss": 3.7691, "step": 3719 }, { "epoch": 1.1661442006269593, "grad_norm": 31.678668975830078, "learning_rate": 1.8600000000000002e-06, "loss": 2.6131, "step": 3720 }, { "epoch": 1.1664576802507838, "grad_norm": 26.029878616333008, "learning_rate": 1.8605e-06, "loss": 3.1041, "step": 3721 }, { "epoch": 1.1667711598746082, "grad_norm": 31.820192337036133, "learning_rate": 1.861e-06, "loss": 2.592, "step": 3722 }, { "epoch": 1.1670846394984327, "grad_norm": 39.595542907714844, "learning_rate": 1.8615000000000003e-06, "loss": 2.8279, "step": 3723 }, { "epoch": 1.167398119122257, "grad_norm": 42.314151763916016, "learning_rate": 1.8620000000000001e-06, "loss": 3.0745, "step": 3724 }, { "epoch": 1.1677115987460815, "grad_norm": 38.720455169677734, "learning_rate": 1.8625000000000002e-06, "loss": 2.8476, "step": 3725 }, { "epoch": 1.168025078369906, "grad_norm": 65.40015411376953, "learning_rate": 1.863e-06, "loss": 4.3198, "step": 3726 }, { "epoch": 1.1683385579937304, "grad_norm": 139.53359985351562, "learning_rate": 1.8635e-06, "loss": 2.9609, "step": 3727 }, { "epoch": 1.1686520376175549, "grad_norm": 28.293638229370117, "learning_rate": 1.8640000000000003e-06, "loss": 2.9129, "step": 3728 }, { "epoch": 1.1689655172413793, "grad_norm": 27.988492965698242, "learning_rate": 1.8645e-06, "loss": 2.3799, "step": 3729 }, { "epoch": 1.1692789968652038, "grad_norm": 14.3427095413208, "learning_rate": 1.8650000000000001e-06, "loss": 2.2311, "step": 3730 }, { "epoch": 1.1695924764890282, "grad_norm": 40.683189392089844, "learning_rate": 1.8655000000000002e-06, "loss": 2.8328, "step": 3731 }, { "epoch": 1.1699059561128526, "grad_norm": 39.50870132446289, "learning_rate": 1.866e-06, "loss": 2.3849, "step": 3732 }, { "epoch": 1.170219435736677, "grad_norm": 39.7574462890625, "learning_rate": 1.8665000000000003e-06, "loss": 2.6584, "step": 3733 }, { "epoch": 1.1705329153605015, "grad_norm": 31.841501235961914, "learning_rate": 1.8670000000000003e-06, "loss": 3.718, "step": 3734 }, { "epoch": 1.170846394984326, "grad_norm": 57.89259338378906, "learning_rate": 1.8675000000000001e-06, "loss": 2.6181, "step": 3735 }, { "epoch": 1.1711598746081504, "grad_norm": 44.258182525634766, "learning_rate": 1.8680000000000002e-06, "loss": 2.5498, "step": 3736 }, { "epoch": 1.1714733542319749, "grad_norm": 50.148067474365234, "learning_rate": 1.8685e-06, "loss": 2.8004, "step": 3737 }, { "epoch": 1.1717868338557993, "grad_norm": 38.96084976196289, "learning_rate": 1.8690000000000002e-06, "loss": 3.2472, "step": 3738 }, { "epoch": 1.1721003134796237, "grad_norm": 13.656538963317871, "learning_rate": 1.8695000000000003e-06, "loss": 2.5985, "step": 3739 }, { "epoch": 1.1724137931034484, "grad_norm": 32.31047821044922, "learning_rate": 1.87e-06, "loss": 2.8887, "step": 3740 }, { "epoch": 1.1727272727272728, "grad_norm": 14.074935913085938, "learning_rate": 1.8705000000000001e-06, "loss": 2.8068, "step": 3741 }, { "epoch": 1.1730407523510973, "grad_norm": 23.464031219482422, "learning_rate": 1.871e-06, "loss": 3.1187, "step": 3742 }, { "epoch": 1.1733542319749217, "grad_norm": 26.261005401611328, "learning_rate": 1.8715000000000002e-06, "loss": 2.7901, "step": 3743 }, { "epoch": 1.1736677115987462, "grad_norm": 28.60563850402832, "learning_rate": 1.8720000000000002e-06, "loss": 3.1748, "step": 3744 }, { "epoch": 1.1739811912225706, "grad_norm": 38.07712936401367, "learning_rate": 1.8725e-06, "loss": 2.8271, "step": 3745 }, { "epoch": 1.174294670846395, "grad_norm": 114.45543670654297, "learning_rate": 1.8730000000000001e-06, "loss": 2.6053, "step": 3746 }, { "epoch": 1.1746081504702195, "grad_norm": 22.89457893371582, "learning_rate": 1.8735000000000001e-06, "loss": 2.4069, "step": 3747 }, { "epoch": 1.174921630094044, "grad_norm": 38.383277893066406, "learning_rate": 1.8740000000000002e-06, "loss": 3.9307, "step": 3748 }, { "epoch": 1.1752351097178684, "grad_norm": 42.46405029296875, "learning_rate": 1.8745000000000002e-06, "loss": 3.3047, "step": 3749 }, { "epoch": 1.1755485893416928, "grad_norm": 17.158899307250977, "learning_rate": 1.8750000000000003e-06, "loss": 2.7314, "step": 3750 }, { "epoch": 1.1758620689655173, "grad_norm": 28.514450073242188, "learning_rate": 1.8755e-06, "loss": 3.8771, "step": 3751 }, { "epoch": 1.1761755485893417, "grad_norm": 33.93141174316406, "learning_rate": 1.8760000000000001e-06, "loss": 2.4683, "step": 3752 }, { "epoch": 1.1764890282131661, "grad_norm": 17.947032928466797, "learning_rate": 1.8765000000000002e-06, "loss": 2.5114, "step": 3753 }, { "epoch": 1.1768025078369906, "grad_norm": 39.59684371948242, "learning_rate": 1.8770000000000002e-06, "loss": 2.6302, "step": 3754 }, { "epoch": 1.177115987460815, "grad_norm": 41.95842361450195, "learning_rate": 1.8775000000000002e-06, "loss": 3.1639, "step": 3755 }, { "epoch": 1.1774294670846395, "grad_norm": 36.54000473022461, "learning_rate": 1.878e-06, "loss": 3.0288, "step": 3756 }, { "epoch": 1.177742946708464, "grad_norm": 33.35289764404297, "learning_rate": 1.8785e-06, "loss": 2.664, "step": 3757 }, { "epoch": 1.1780564263322884, "grad_norm": 24.627397537231445, "learning_rate": 1.8790000000000003e-06, "loss": 2.712, "step": 3758 }, { "epoch": 1.1783699059561128, "grad_norm": 28.08515167236328, "learning_rate": 1.8795000000000002e-06, "loss": 2.8725, "step": 3759 }, { "epoch": 1.1786833855799372, "grad_norm": 61.68031311035156, "learning_rate": 1.8800000000000002e-06, "loss": 2.816, "step": 3760 }, { "epoch": 1.1789968652037617, "grad_norm": 87.2820053100586, "learning_rate": 1.8805e-06, "loss": 3.6346, "step": 3761 }, { "epoch": 1.1793103448275861, "grad_norm": 71.9816665649414, "learning_rate": 1.881e-06, "loss": 3.4425, "step": 3762 }, { "epoch": 1.1796238244514106, "grad_norm": 20.74867057800293, "learning_rate": 1.8815000000000003e-06, "loss": 2.4513, "step": 3763 }, { "epoch": 1.179937304075235, "grad_norm": 35.601165771484375, "learning_rate": 1.8820000000000001e-06, "loss": 3.3946, "step": 3764 }, { "epoch": 1.1802507836990594, "grad_norm": 55.310768127441406, "learning_rate": 1.8825000000000002e-06, "loss": 2.894, "step": 3765 }, { "epoch": 1.1805642633228839, "grad_norm": 41.345428466796875, "learning_rate": 1.883e-06, "loss": 2.7336, "step": 3766 }, { "epoch": 1.1808777429467086, "grad_norm": 23.19663429260254, "learning_rate": 1.8835e-06, "loss": 2.5383, "step": 3767 }, { "epoch": 1.181191222570533, "grad_norm": 35.46568298339844, "learning_rate": 1.8840000000000003e-06, "loss": 3.1715, "step": 3768 }, { "epoch": 1.1815047021943574, "grad_norm": 20.419200897216797, "learning_rate": 1.8845000000000001e-06, "loss": 2.8013, "step": 3769 }, { "epoch": 1.1818181818181819, "grad_norm": 19.463045120239258, "learning_rate": 1.8850000000000002e-06, "loss": 2.4863, "step": 3770 }, { "epoch": 1.1821316614420063, "grad_norm": 34.91453552246094, "learning_rate": 1.8855000000000002e-06, "loss": 2.868, "step": 3771 }, { "epoch": 1.1824451410658308, "grad_norm": 16.39441680908203, "learning_rate": 1.886e-06, "loss": 2.2358, "step": 3772 }, { "epoch": 1.1827586206896552, "grad_norm": 29.26861000061035, "learning_rate": 1.8865000000000003e-06, "loss": 2.2603, "step": 3773 }, { "epoch": 1.1830721003134796, "grad_norm": 38.201534271240234, "learning_rate": 1.8870000000000003e-06, "loss": 3.9458, "step": 3774 }, { "epoch": 1.183385579937304, "grad_norm": 37.52971649169922, "learning_rate": 1.8875000000000001e-06, "loss": 2.4286, "step": 3775 }, { "epoch": 1.1836990595611285, "grad_norm": 36.337181091308594, "learning_rate": 1.8880000000000002e-06, "loss": 3.3092, "step": 3776 }, { "epoch": 1.184012539184953, "grad_norm": 44.544219970703125, "learning_rate": 1.8885e-06, "loss": 2.5326, "step": 3777 }, { "epoch": 1.1843260188087774, "grad_norm": 60.01102066040039, "learning_rate": 1.8890000000000003e-06, "loss": 2.7836, "step": 3778 }, { "epoch": 1.1846394984326019, "grad_norm": 26.756061553955078, "learning_rate": 1.8895000000000003e-06, "loss": 2.7295, "step": 3779 }, { "epoch": 1.1849529780564263, "grad_norm": 51.751670837402344, "learning_rate": 1.8900000000000001e-06, "loss": 2.6031, "step": 3780 }, { "epoch": 1.1852664576802507, "grad_norm": 37.55747985839844, "learning_rate": 1.8905000000000002e-06, "loss": 3.4953, "step": 3781 }, { "epoch": 1.1855799373040752, "grad_norm": 17.792362213134766, "learning_rate": 1.891e-06, "loss": 2.438, "step": 3782 }, { "epoch": 1.1858934169278996, "grad_norm": 42.68111801147461, "learning_rate": 1.8915000000000002e-06, "loss": 2.534, "step": 3783 }, { "epoch": 1.186206896551724, "grad_norm": 20.284029006958008, "learning_rate": 1.8920000000000003e-06, "loss": 2.8496, "step": 3784 }, { "epoch": 1.1865203761755485, "grad_norm": 22.83557891845703, "learning_rate": 1.8925e-06, "loss": 2.8779, "step": 3785 }, { "epoch": 1.186833855799373, "grad_norm": 20.988258361816406, "learning_rate": 1.8930000000000001e-06, "loss": 3.2298, "step": 3786 }, { "epoch": 1.1871473354231976, "grad_norm": 52.38770294189453, "learning_rate": 1.8935e-06, "loss": 2.6163, "step": 3787 }, { "epoch": 1.187460815047022, "grad_norm": 65.69499969482422, "learning_rate": 1.8940000000000002e-06, "loss": 3.5213, "step": 3788 }, { "epoch": 1.1877742946708465, "grad_norm": 30.15526008605957, "learning_rate": 1.8945000000000002e-06, "loss": 2.7002, "step": 3789 }, { "epoch": 1.188087774294671, "grad_norm": 52.1644401550293, "learning_rate": 1.895e-06, "loss": 3.1846, "step": 3790 }, { "epoch": 1.1884012539184954, "grad_norm": 24.36962890625, "learning_rate": 1.8955e-06, "loss": 2.4365, "step": 3791 }, { "epoch": 1.1887147335423198, "grad_norm": 19.295175552368164, "learning_rate": 1.8960000000000001e-06, "loss": 2.8495, "step": 3792 }, { "epoch": 1.1890282131661443, "grad_norm": 26.62535285949707, "learning_rate": 1.8965000000000002e-06, "loss": 2.7075, "step": 3793 }, { "epoch": 1.1893416927899687, "grad_norm": 40.80854034423828, "learning_rate": 1.8970000000000002e-06, "loss": 2.7312, "step": 3794 }, { "epoch": 1.1896551724137931, "grad_norm": 57.89705276489258, "learning_rate": 1.8975000000000003e-06, "loss": 2.5595, "step": 3795 }, { "epoch": 1.1899686520376176, "grad_norm": 26.473291397094727, "learning_rate": 1.898e-06, "loss": 2.5633, "step": 3796 }, { "epoch": 1.190282131661442, "grad_norm": 44.334529876708984, "learning_rate": 1.8985000000000001e-06, "loss": 2.4269, "step": 3797 }, { "epoch": 1.1905956112852665, "grad_norm": 27.930118560791016, "learning_rate": 1.8990000000000004e-06, "loss": 2.9725, "step": 3798 }, { "epoch": 1.190909090909091, "grad_norm": 45.654109954833984, "learning_rate": 1.8995000000000002e-06, "loss": 3.2554, "step": 3799 }, { "epoch": 1.1912225705329154, "grad_norm": 16.584321975708008, "learning_rate": 1.9000000000000002e-06, "loss": 2.5039, "step": 3800 }, { "epoch": 1.1915360501567398, "grad_norm": 30.959943771362305, "learning_rate": 1.9005e-06, "loss": 2.9912, "step": 3801 }, { "epoch": 1.1918495297805642, "grad_norm": 31.72456169128418, "learning_rate": 1.901e-06, "loss": 2.6606, "step": 3802 }, { "epoch": 1.1921630094043887, "grad_norm": 23.852758407592773, "learning_rate": 1.9015000000000003e-06, "loss": 2.865, "step": 3803 }, { "epoch": 1.1924764890282131, "grad_norm": 34.01231002807617, "learning_rate": 1.9020000000000002e-06, "loss": 2.3327, "step": 3804 }, { "epoch": 1.1927899686520376, "grad_norm": 57.21983337402344, "learning_rate": 1.9025000000000002e-06, "loss": 2.9505, "step": 3805 }, { "epoch": 1.193103448275862, "grad_norm": 129.81326293945312, "learning_rate": 1.903e-06, "loss": 2.698, "step": 3806 }, { "epoch": 1.1934169278996865, "grad_norm": 51.74315643310547, "learning_rate": 1.9035e-06, "loss": 3.5234, "step": 3807 }, { "epoch": 1.193730407523511, "grad_norm": 22.179800033569336, "learning_rate": 1.9040000000000003e-06, "loss": 2.6641, "step": 3808 }, { "epoch": 1.1940438871473353, "grad_norm": 15.864103317260742, "learning_rate": 1.9045000000000001e-06, "loss": 2.7566, "step": 3809 }, { "epoch": 1.1943573667711598, "grad_norm": 29.442760467529297, "learning_rate": 1.9050000000000002e-06, "loss": 3.1718, "step": 3810 }, { "epoch": 1.1946708463949842, "grad_norm": 67.95341491699219, "learning_rate": 1.9055000000000002e-06, "loss": 3.1144, "step": 3811 }, { "epoch": 1.1949843260188087, "grad_norm": 27.336172103881836, "learning_rate": 1.906e-06, "loss": 2.7236, "step": 3812 }, { "epoch": 1.195297805642633, "grad_norm": 76.92893981933594, "learning_rate": 1.9065e-06, "loss": 2.2788, "step": 3813 }, { "epoch": 1.1956112852664578, "grad_norm": 28.78301239013672, "learning_rate": 1.9070000000000003e-06, "loss": 2.6817, "step": 3814 }, { "epoch": 1.1959247648902822, "grad_norm": 57.42724609375, "learning_rate": 1.9075000000000004e-06, "loss": 3.3968, "step": 3815 }, { "epoch": 1.1962382445141067, "grad_norm": 25.262216567993164, "learning_rate": 1.908e-06, "loss": 2.6335, "step": 3816 }, { "epoch": 1.196551724137931, "grad_norm": 19.571935653686523, "learning_rate": 1.9085e-06, "loss": 2.5299, "step": 3817 }, { "epoch": 1.1968652037617555, "grad_norm": 23.045181274414062, "learning_rate": 1.909e-06, "loss": 2.6344, "step": 3818 }, { "epoch": 1.19717868338558, "grad_norm": 53.79544448852539, "learning_rate": 1.9095e-06, "loss": 3.178, "step": 3819 }, { "epoch": 1.1974921630094044, "grad_norm": 25.499238967895508, "learning_rate": 1.9100000000000003e-06, "loss": 3.3268, "step": 3820 }, { "epoch": 1.1978056426332289, "grad_norm": 44.6558837890625, "learning_rate": 1.9105e-06, "loss": 3.7756, "step": 3821 }, { "epoch": 1.1981191222570533, "grad_norm": 33.91236114501953, "learning_rate": 1.911e-06, "loss": 2.9224, "step": 3822 }, { "epoch": 1.1984326018808777, "grad_norm": 40.67926788330078, "learning_rate": 1.9115e-06, "loss": 2.646, "step": 3823 }, { "epoch": 1.1987460815047022, "grad_norm": 28.280128479003906, "learning_rate": 1.912e-06, "loss": 2.5554, "step": 3824 }, { "epoch": 1.1990595611285266, "grad_norm": 31.250951766967773, "learning_rate": 1.9125000000000003e-06, "loss": 2.6868, "step": 3825 }, { "epoch": 1.199373040752351, "grad_norm": 53.51694107055664, "learning_rate": 1.913e-06, "loss": 3.5194, "step": 3826 }, { "epoch": 1.1996865203761755, "grad_norm": 44.22115707397461, "learning_rate": 1.9135e-06, "loss": 3.2924, "step": 3827 }, { "epoch": 1.2, "grad_norm": 27.570470809936523, "learning_rate": 1.9140000000000002e-06, "loss": 2.7071, "step": 3828 }, { "epoch": 1.2003134796238244, "grad_norm": 29.043636322021484, "learning_rate": 1.9145e-06, "loss": 2.9722, "step": 3829 }, { "epoch": 1.2006269592476488, "grad_norm": 48.598636627197266, "learning_rate": 1.9150000000000003e-06, "loss": 2.6605, "step": 3830 }, { "epoch": 1.2009404388714733, "grad_norm": 25.80156898498535, "learning_rate": 1.9155e-06, "loss": 2.7321, "step": 3831 }, { "epoch": 1.2012539184952977, "grad_norm": 36.27217483520508, "learning_rate": 1.916e-06, "loss": 2.3933, "step": 3832 }, { "epoch": 1.2015673981191222, "grad_norm": 50.53173065185547, "learning_rate": 1.9165e-06, "loss": 2.8351, "step": 3833 }, { "epoch": 1.2018808777429468, "grad_norm": 33.122230529785156, "learning_rate": 1.9170000000000005e-06, "loss": 3.0608, "step": 3834 }, { "epoch": 1.2021943573667713, "grad_norm": 66.05673217773438, "learning_rate": 1.9175000000000003e-06, "loss": 3.3519, "step": 3835 }, { "epoch": 1.2025078369905957, "grad_norm": 27.911020278930664, "learning_rate": 1.918e-06, "loss": 2.6034, "step": 3836 }, { "epoch": 1.2028213166144202, "grad_norm": 43.20354080200195, "learning_rate": 1.9185e-06, "loss": 2.8307, "step": 3837 }, { "epoch": 1.2031347962382446, "grad_norm": 23.005508422851562, "learning_rate": 1.919e-06, "loss": 2.6978, "step": 3838 }, { "epoch": 1.203448275862069, "grad_norm": 67.49423217773438, "learning_rate": 1.9195000000000004e-06, "loss": 3.424, "step": 3839 }, { "epoch": 1.2037617554858935, "grad_norm": 30.004865646362305, "learning_rate": 1.9200000000000003e-06, "loss": 2.5142, "step": 3840 }, { "epoch": 1.204075235109718, "grad_norm": 33.494834899902344, "learning_rate": 1.9205e-06, "loss": 3.2469, "step": 3841 }, { "epoch": 1.2043887147335424, "grad_norm": 27.38806915283203, "learning_rate": 1.921e-06, "loss": 2.7793, "step": 3842 }, { "epoch": 1.2047021943573668, "grad_norm": 52.8104133605957, "learning_rate": 1.9215e-06, "loss": 2.6053, "step": 3843 }, { "epoch": 1.2050156739811912, "grad_norm": 18.836341857910156, "learning_rate": 1.9220000000000004e-06, "loss": 2.3581, "step": 3844 }, { "epoch": 1.2053291536050157, "grad_norm": 28.024110794067383, "learning_rate": 1.9225000000000002e-06, "loss": 2.6733, "step": 3845 }, { "epoch": 1.2056426332288401, "grad_norm": 13.282366752624512, "learning_rate": 1.923e-06, "loss": 2.4797, "step": 3846 }, { "epoch": 1.2059561128526646, "grad_norm": 16.323963165283203, "learning_rate": 1.9235e-06, "loss": 2.3768, "step": 3847 }, { "epoch": 1.206269592476489, "grad_norm": 27.969438552856445, "learning_rate": 1.924e-06, "loss": 2.2672, "step": 3848 }, { "epoch": 1.2065830721003135, "grad_norm": 27.688213348388672, "learning_rate": 1.9245000000000004e-06, "loss": 2.5915, "step": 3849 }, { "epoch": 1.206896551724138, "grad_norm": 26.980260848999023, "learning_rate": 1.925e-06, "loss": 2.6703, "step": 3850 }, { "epoch": 1.2072100313479623, "grad_norm": 25.653573989868164, "learning_rate": 1.9255e-06, "loss": 2.3697, "step": 3851 }, { "epoch": 1.2075235109717868, "grad_norm": 16.955244064331055, "learning_rate": 1.9260000000000003e-06, "loss": 2.5476, "step": 3852 }, { "epoch": 1.2078369905956112, "grad_norm": 51.82334518432617, "learning_rate": 1.9265e-06, "loss": 2.7162, "step": 3853 }, { "epoch": 1.2081504702194357, "grad_norm": 19.893192291259766, "learning_rate": 1.9270000000000004e-06, "loss": 2.3514, "step": 3854 }, { "epoch": 1.20846394984326, "grad_norm": 19.577396392822266, "learning_rate": 1.9275e-06, "loss": 2.6409, "step": 3855 }, { "epoch": 1.2087774294670846, "grad_norm": 15.481810569763184, "learning_rate": 1.928e-06, "loss": 2.5522, "step": 3856 }, { "epoch": 1.209090909090909, "grad_norm": 20.681058883666992, "learning_rate": 1.9285000000000003e-06, "loss": 2.7032, "step": 3857 }, { "epoch": 1.2094043887147334, "grad_norm": 36.33036804199219, "learning_rate": 1.929e-06, "loss": 3.2028, "step": 3858 }, { "epoch": 1.2097178683385579, "grad_norm": 20.78473472595215, "learning_rate": 1.9295000000000003e-06, "loss": 2.487, "step": 3859 }, { "epoch": 1.2100313479623825, "grad_norm": 89.45281982421875, "learning_rate": 1.93e-06, "loss": 5.2864, "step": 3860 }, { "epoch": 1.210344827586207, "grad_norm": 32.773216247558594, "learning_rate": 1.9305e-06, "loss": 2.6761, "step": 3861 }, { "epoch": 1.2106583072100314, "grad_norm": 17.46109962463379, "learning_rate": 1.9310000000000002e-06, "loss": 2.4264, "step": 3862 }, { "epoch": 1.2109717868338559, "grad_norm": 28.945659637451172, "learning_rate": 1.9315e-06, "loss": 2.73, "step": 3863 }, { "epoch": 1.2112852664576803, "grad_norm": 150.18414306640625, "learning_rate": 1.9320000000000003e-06, "loss": 2.8617, "step": 3864 }, { "epoch": 1.2115987460815048, "grad_norm": 27.77288055419922, "learning_rate": 1.9325e-06, "loss": 2.7205, "step": 3865 }, { "epoch": 1.2119122257053292, "grad_norm": 23.874618530273438, "learning_rate": 1.933e-06, "loss": 2.5498, "step": 3866 }, { "epoch": 1.2122257053291536, "grad_norm": 118.2887191772461, "learning_rate": 1.9335e-06, "loss": 2.8452, "step": 3867 }, { "epoch": 1.212539184952978, "grad_norm": 44.203243255615234, "learning_rate": 1.934e-06, "loss": 3.3568, "step": 3868 }, { "epoch": 1.2128526645768025, "grad_norm": 22.49334144592285, "learning_rate": 1.9345000000000003e-06, "loss": 3.0688, "step": 3869 }, { "epoch": 1.213166144200627, "grad_norm": 20.258432388305664, "learning_rate": 1.935e-06, "loss": 2.781, "step": 3870 }, { "epoch": 1.2134796238244514, "grad_norm": 26.8780517578125, "learning_rate": 1.9355000000000004e-06, "loss": 3.2032, "step": 3871 }, { "epoch": 1.2137931034482758, "grad_norm": 27.263999938964844, "learning_rate": 1.936e-06, "loss": 2.3119, "step": 3872 }, { "epoch": 1.2141065830721003, "grad_norm": 29.607118606567383, "learning_rate": 1.9365e-06, "loss": 2.9999, "step": 3873 }, { "epoch": 1.2144200626959247, "grad_norm": 36.53831481933594, "learning_rate": 1.9370000000000003e-06, "loss": 3.1496, "step": 3874 }, { "epoch": 1.2147335423197492, "grad_norm": 18.061384201049805, "learning_rate": 1.9375e-06, "loss": 2.596, "step": 3875 }, { "epoch": 1.2150470219435736, "grad_norm": 41.549800872802734, "learning_rate": 1.9380000000000003e-06, "loss": 3.3622, "step": 3876 }, { "epoch": 1.215360501567398, "grad_norm": 26.330778121948242, "learning_rate": 1.9385e-06, "loss": 2.5022, "step": 3877 }, { "epoch": 1.2156739811912225, "grad_norm": 41.321956634521484, "learning_rate": 1.939e-06, "loss": 3.0533, "step": 3878 }, { "epoch": 1.215987460815047, "grad_norm": 16.474641799926758, "learning_rate": 1.9395000000000002e-06, "loss": 2.7315, "step": 3879 }, { "epoch": 1.2163009404388714, "grad_norm": 46.212467193603516, "learning_rate": 1.94e-06, "loss": 2.9196, "step": 3880 }, { "epoch": 1.216614420062696, "grad_norm": 31.109722137451172, "learning_rate": 1.9405000000000003e-06, "loss": 2.6945, "step": 3881 }, { "epoch": 1.2169278996865205, "grad_norm": 18.186649322509766, "learning_rate": 1.941e-06, "loss": 2.4777, "step": 3882 }, { "epoch": 1.217241379310345, "grad_norm": 28.388580322265625, "learning_rate": 1.9415e-06, "loss": 2.8091, "step": 3883 }, { "epoch": 1.2175548589341694, "grad_norm": 16.219085693359375, "learning_rate": 1.942e-06, "loss": 2.3407, "step": 3884 }, { "epoch": 1.2178683385579938, "grad_norm": 25.237590789794922, "learning_rate": 1.9425e-06, "loss": 2.5794, "step": 3885 }, { "epoch": 1.2181818181818183, "grad_norm": 17.25516700744629, "learning_rate": 1.9430000000000003e-06, "loss": 2.3214, "step": 3886 }, { "epoch": 1.2184952978056427, "grad_norm": 20.798234939575195, "learning_rate": 1.9435e-06, "loss": 2.4373, "step": 3887 }, { "epoch": 1.2188087774294671, "grad_norm": 18.155517578125, "learning_rate": 1.944e-06, "loss": 2.3363, "step": 3888 }, { "epoch": 1.2191222570532916, "grad_norm": 27.072912216186523, "learning_rate": 1.9445e-06, "loss": 2.6745, "step": 3889 }, { "epoch": 1.219435736677116, "grad_norm": 34.35905075073242, "learning_rate": 1.945e-06, "loss": 2.7749, "step": 3890 }, { "epoch": 1.2197492163009405, "grad_norm": 19.70022964477539, "learning_rate": 1.9455000000000003e-06, "loss": 2.7745, "step": 3891 }, { "epoch": 1.220062695924765, "grad_norm": 13.153515815734863, "learning_rate": 1.946e-06, "loss": 2.3028, "step": 3892 }, { "epoch": 1.2203761755485893, "grad_norm": 42.72065734863281, "learning_rate": 1.9465e-06, "loss": 2.7218, "step": 3893 }, { "epoch": 1.2206896551724138, "grad_norm": 38.443115234375, "learning_rate": 1.947e-06, "loss": 2.8484, "step": 3894 }, { "epoch": 1.2210031347962382, "grad_norm": 20.11655044555664, "learning_rate": 1.9475000000000004e-06, "loss": 2.6273, "step": 3895 }, { "epoch": 1.2213166144200627, "grad_norm": 46.73358917236328, "learning_rate": 1.9480000000000002e-06, "loss": 2.4925, "step": 3896 }, { "epoch": 1.2216300940438871, "grad_norm": 82.11383056640625, "learning_rate": 1.9485e-06, "loss": 3.4626, "step": 3897 }, { "epoch": 1.2219435736677116, "grad_norm": 22.496259689331055, "learning_rate": 1.949e-06, "loss": 2.4596, "step": 3898 }, { "epoch": 1.222257053291536, "grad_norm": 21.611501693725586, "learning_rate": 1.9495e-06, "loss": 2.3298, "step": 3899 }, { "epoch": 1.2225705329153604, "grad_norm": 16.648698806762695, "learning_rate": 1.9500000000000004e-06, "loss": 2.5577, "step": 3900 }, { "epoch": 1.2228840125391849, "grad_norm": 17.495756149291992, "learning_rate": 1.9505000000000002e-06, "loss": 2.3889, "step": 3901 }, { "epoch": 1.2231974921630093, "grad_norm": 25.892221450805664, "learning_rate": 1.951e-06, "loss": 2.4928, "step": 3902 }, { "epoch": 1.2235109717868338, "grad_norm": 27.336132049560547, "learning_rate": 1.9515e-06, "loss": 2.8628, "step": 3903 }, { "epoch": 1.2238244514106582, "grad_norm": 21.546873092651367, "learning_rate": 1.952e-06, "loss": 2.6961, "step": 3904 }, { "epoch": 1.2241379310344827, "grad_norm": 24.31972312927246, "learning_rate": 1.9525000000000004e-06, "loss": 2.6013, "step": 3905 }, { "epoch": 1.224451410658307, "grad_norm": 39.16084671020508, "learning_rate": 1.953e-06, "loss": 2.5389, "step": 3906 }, { "epoch": 1.2247648902821318, "grad_norm": 218.3396453857422, "learning_rate": 1.9535e-06, "loss": 3.3436, "step": 3907 }, { "epoch": 1.2250783699059562, "grad_norm": 27.832767486572266, "learning_rate": 1.9540000000000003e-06, "loss": 2.7819, "step": 3908 }, { "epoch": 1.2253918495297806, "grad_norm": 13.001688003540039, "learning_rate": 1.9545e-06, "loss": 2.5069, "step": 3909 }, { "epoch": 1.225705329153605, "grad_norm": 14.492598533630371, "learning_rate": 1.9550000000000003e-06, "loss": 2.1693, "step": 3910 }, { "epoch": 1.2260188087774295, "grad_norm": 60.928680419921875, "learning_rate": 1.9555e-06, "loss": 2.4301, "step": 3911 }, { "epoch": 1.226332288401254, "grad_norm": 64.4627914428711, "learning_rate": 1.956e-06, "loss": 3.0376, "step": 3912 }, { "epoch": 1.2266457680250784, "grad_norm": 61.340904235839844, "learning_rate": 1.9565000000000002e-06, "loss": 2.5944, "step": 3913 }, { "epoch": 1.2269592476489029, "grad_norm": 33.53535842895508, "learning_rate": 1.9570000000000005e-06, "loss": 3.2278, "step": 3914 }, { "epoch": 1.2272727272727273, "grad_norm": 59.37514114379883, "learning_rate": 1.9575000000000003e-06, "loss": 2.9019, "step": 3915 }, { "epoch": 1.2275862068965517, "grad_norm": 18.784873962402344, "learning_rate": 1.958e-06, "loss": 2.5877, "step": 3916 }, { "epoch": 1.2278996865203762, "grad_norm": 30.17689323425293, "learning_rate": 1.9585e-06, "loss": 2.6973, "step": 3917 }, { "epoch": 1.2282131661442006, "grad_norm": 29.341289520263672, "learning_rate": 1.9590000000000002e-06, "loss": 2.8073, "step": 3918 }, { "epoch": 1.228526645768025, "grad_norm": 46.65104293823242, "learning_rate": 1.9595000000000005e-06, "loss": 2.9391, "step": 3919 }, { "epoch": 1.2288401253918495, "grad_norm": 31.895326614379883, "learning_rate": 1.9600000000000003e-06, "loss": 2.5958, "step": 3920 }, { "epoch": 1.229153605015674, "grad_norm": 48.45664596557617, "learning_rate": 1.9605e-06, "loss": 2.5412, "step": 3921 }, { "epoch": 1.2294670846394984, "grad_norm": 13.672981262207031, "learning_rate": 1.961e-06, "loss": 2.3096, "step": 3922 }, { "epoch": 1.2297805642633228, "grad_norm": 30.28753662109375, "learning_rate": 1.9615e-06, "loss": 2.4116, "step": 3923 }, { "epoch": 1.2300940438871473, "grad_norm": 39.5670280456543, "learning_rate": 1.9620000000000004e-06, "loss": 3.0972, "step": 3924 }, { "epoch": 1.2304075235109717, "grad_norm": 17.293338775634766, "learning_rate": 1.9625000000000003e-06, "loss": 2.3708, "step": 3925 }, { "epoch": 1.2307210031347962, "grad_norm": 25.217784881591797, "learning_rate": 1.963e-06, "loss": 2.3573, "step": 3926 }, { "epoch": 1.2310344827586206, "grad_norm": 29.8195743560791, "learning_rate": 1.9635e-06, "loss": 2.9449, "step": 3927 }, { "epoch": 1.2313479623824453, "grad_norm": 55.18169403076172, "learning_rate": 1.964e-06, "loss": 2.5012, "step": 3928 }, { "epoch": 1.2316614420062697, "grad_norm": 28.454805374145508, "learning_rate": 1.9645000000000004e-06, "loss": 2.8373, "step": 3929 }, { "epoch": 1.2319749216300941, "grad_norm": 24.07086944580078, "learning_rate": 1.9650000000000002e-06, "loss": 2.5167, "step": 3930 }, { "epoch": 1.2322884012539186, "grad_norm": 26.33777618408203, "learning_rate": 1.9655e-06, "loss": 3.0033, "step": 3931 }, { "epoch": 1.232601880877743, "grad_norm": 26.97733497619629, "learning_rate": 1.9660000000000003e-06, "loss": 2.8334, "step": 3932 }, { "epoch": 1.2329153605015675, "grad_norm": 38.315311431884766, "learning_rate": 1.9665e-06, "loss": 3.6005, "step": 3933 }, { "epoch": 1.233228840125392, "grad_norm": 37.27223587036133, "learning_rate": 1.9670000000000004e-06, "loss": 3.188, "step": 3934 }, { "epoch": 1.2335423197492164, "grad_norm": 45.02766036987305, "learning_rate": 1.9675000000000002e-06, "loss": 3.2214, "step": 3935 }, { "epoch": 1.2338557993730408, "grad_norm": 24.59618377685547, "learning_rate": 1.968e-06, "loss": 2.6329, "step": 3936 }, { "epoch": 1.2341692789968652, "grad_norm": 30.170249938964844, "learning_rate": 1.9685000000000003e-06, "loss": 2.9202, "step": 3937 }, { "epoch": 1.2344827586206897, "grad_norm": 20.308732986450195, "learning_rate": 1.969e-06, "loss": 2.4924, "step": 3938 }, { "epoch": 1.2347962382445141, "grad_norm": 35.08584213256836, "learning_rate": 1.9695e-06, "loss": 2.1855, "step": 3939 }, { "epoch": 1.2351097178683386, "grad_norm": 24.181821823120117, "learning_rate": 1.97e-06, "loss": 2.4697, "step": 3940 }, { "epoch": 1.235423197492163, "grad_norm": 16.770347595214844, "learning_rate": 1.9705e-06, "loss": 2.365, "step": 3941 }, { "epoch": 1.2357366771159874, "grad_norm": 23.461381912231445, "learning_rate": 1.9710000000000003e-06, "loss": 2.9199, "step": 3942 }, { "epoch": 1.236050156739812, "grad_norm": 20.003427505493164, "learning_rate": 1.9715e-06, "loss": 2.3127, "step": 3943 }, { "epoch": 1.2363636363636363, "grad_norm": 33.976295471191406, "learning_rate": 1.972e-06, "loss": 2.6393, "step": 3944 }, { "epoch": 1.2366771159874608, "grad_norm": 61.14518737792969, "learning_rate": 1.9725e-06, "loss": 2.6706, "step": 3945 }, { "epoch": 1.2369905956112852, "grad_norm": 56.58893585205078, "learning_rate": 1.973e-06, "loss": 2.5611, "step": 3946 }, { "epoch": 1.2373040752351097, "grad_norm": 51.902347564697266, "learning_rate": 1.9735000000000002e-06, "loss": 2.9941, "step": 3947 }, { "epoch": 1.237617554858934, "grad_norm": 21.652944564819336, "learning_rate": 1.974e-06, "loss": 2.7382, "step": 3948 }, { "epoch": 1.2379310344827585, "grad_norm": 44.5388298034668, "learning_rate": 1.9745e-06, "loss": 3.0546, "step": 3949 }, { "epoch": 1.238244514106583, "grad_norm": 24.932762145996094, "learning_rate": 1.975e-06, "loss": 2.3333, "step": 3950 }, { "epoch": 1.2385579937304074, "grad_norm": 27.233287811279297, "learning_rate": 1.9755000000000004e-06, "loss": 2.8792, "step": 3951 }, { "epoch": 1.2388714733542319, "grad_norm": 24.123865127563477, "learning_rate": 1.9760000000000002e-06, "loss": 2.5114, "step": 3952 }, { "epoch": 1.2391849529780563, "grad_norm": 25.245868682861328, "learning_rate": 1.9765e-06, "loss": 2.279, "step": 3953 }, { "epoch": 1.239498432601881, "grad_norm": 38.38374710083008, "learning_rate": 1.977e-06, "loss": 2.7466, "step": 3954 }, { "epoch": 1.2398119122257054, "grad_norm": 18.390085220336914, "learning_rate": 1.9775e-06, "loss": 2.734, "step": 3955 }, { "epoch": 1.2401253918495299, "grad_norm": 36.78880310058594, "learning_rate": 1.9780000000000004e-06, "loss": 2.7813, "step": 3956 }, { "epoch": 1.2404388714733543, "grad_norm": 66.53557586669922, "learning_rate": 1.9785e-06, "loss": 2.5124, "step": 3957 }, { "epoch": 1.2407523510971787, "grad_norm": 103.23516082763672, "learning_rate": 1.979e-06, "loss": 2.4264, "step": 3958 }, { "epoch": 1.2410658307210032, "grad_norm": 49.23095703125, "learning_rate": 1.9795e-06, "loss": 2.6846, "step": 3959 }, { "epoch": 1.2413793103448276, "grad_norm": 15.24638557434082, "learning_rate": 1.98e-06, "loss": 2.175, "step": 3960 }, { "epoch": 1.241692789968652, "grad_norm": 39.227378845214844, "learning_rate": 1.9805000000000004e-06, "loss": 3.1209, "step": 3961 }, { "epoch": 1.2420062695924765, "grad_norm": 32.122440338134766, "learning_rate": 1.981e-06, "loss": 2.5471, "step": 3962 }, { "epoch": 1.242319749216301, "grad_norm": 22.874025344848633, "learning_rate": 1.9815e-06, "loss": 3.025, "step": 3963 }, { "epoch": 1.2426332288401254, "grad_norm": 25.451332092285156, "learning_rate": 1.982e-06, "loss": 2.7846, "step": 3964 }, { "epoch": 1.2429467084639498, "grad_norm": 65.10897064208984, "learning_rate": 1.9825e-06, "loss": 2.6536, "step": 3965 }, { "epoch": 1.2432601880877743, "grad_norm": 41.65166091918945, "learning_rate": 1.9830000000000003e-06, "loss": 3.1561, "step": 3966 }, { "epoch": 1.2435736677115987, "grad_norm": 60.49795150756836, "learning_rate": 1.9835e-06, "loss": 3.6986, "step": 3967 }, { "epoch": 1.2438871473354232, "grad_norm": 21.87533187866211, "learning_rate": 1.984e-06, "loss": 2.6943, "step": 3968 }, { "epoch": 1.2442006269592476, "grad_norm": 14.572546005249023, "learning_rate": 1.9845000000000002e-06, "loss": 2.5391, "step": 3969 }, { "epoch": 1.244514106583072, "grad_norm": 24.31571388244629, "learning_rate": 1.985e-06, "loss": 2.9149, "step": 3970 }, { "epoch": 1.2448275862068965, "grad_norm": 24.90671730041504, "learning_rate": 1.9855000000000003e-06, "loss": 2.8762, "step": 3971 }, { "epoch": 1.245141065830721, "grad_norm": 24.37377166748047, "learning_rate": 1.986e-06, "loss": 2.6314, "step": 3972 }, { "epoch": 1.2454545454545454, "grad_norm": 19.115659713745117, "learning_rate": 1.9865e-06, "loss": 2.46, "step": 3973 }, { "epoch": 1.2457680250783698, "grad_norm": 21.906410217285156, "learning_rate": 1.987e-06, "loss": 2.7101, "step": 3974 }, { "epoch": 1.2460815047021945, "grad_norm": 30.39508819580078, "learning_rate": 1.9875000000000005e-06, "loss": 2.9468, "step": 3975 }, { "epoch": 1.246394984326019, "grad_norm": 40.24324417114258, "learning_rate": 1.9880000000000003e-06, "loss": 2.492, "step": 3976 }, { "epoch": 1.2467084639498434, "grad_norm": 18.688854217529297, "learning_rate": 1.9885e-06, "loss": 2.4316, "step": 3977 }, { "epoch": 1.2470219435736678, "grad_norm": 114.26399993896484, "learning_rate": 1.989e-06, "loss": 3.3366, "step": 3978 }, { "epoch": 1.2473354231974922, "grad_norm": 13.459341049194336, "learning_rate": 1.9895e-06, "loss": 2.2597, "step": 3979 }, { "epoch": 1.2476489028213167, "grad_norm": 24.026880264282227, "learning_rate": 1.9900000000000004e-06, "loss": 2.5099, "step": 3980 }, { "epoch": 1.2479623824451411, "grad_norm": 30.318620681762695, "learning_rate": 1.9905000000000003e-06, "loss": 2.4107, "step": 3981 }, { "epoch": 1.2482758620689656, "grad_norm": 14.148162841796875, "learning_rate": 1.991e-06, "loss": 2.4989, "step": 3982 }, { "epoch": 1.24858934169279, "grad_norm": 34.825653076171875, "learning_rate": 1.9915e-06, "loss": 3.2823, "step": 3983 }, { "epoch": 1.2489028213166145, "grad_norm": 23.375911712646484, "learning_rate": 1.992e-06, "loss": 2.8601, "step": 3984 }, { "epoch": 1.249216300940439, "grad_norm": 75.13655853271484, "learning_rate": 1.9925000000000004e-06, "loss": 2.4779, "step": 3985 }, { "epoch": 1.2495297805642633, "grad_norm": 12.734591484069824, "learning_rate": 1.9930000000000002e-06, "loss": 2.1784, "step": 3986 }, { "epoch": 1.2498432601880878, "grad_norm": 34.51366424560547, "learning_rate": 1.9935e-06, "loss": 2.566, "step": 3987 }, { "epoch": 1.2501567398119122, "grad_norm": 16.85667610168457, "learning_rate": 1.9940000000000003e-06, "loss": 2.4532, "step": 3988 }, { "epoch": 1.2504702194357367, "grad_norm": 66.03433227539062, "learning_rate": 1.9945e-06, "loss": 2.7693, "step": 3989 }, { "epoch": 1.250783699059561, "grad_norm": 49.91973114013672, "learning_rate": 1.9950000000000004e-06, "loss": 3.3225, "step": 3990 }, { "epoch": 1.250783699059561, "eval_loss": 3.8327951431274414, "eval_runtime": 21.0158, "eval_samples_per_second": 127.856, "eval_steps_per_second": 7.994, "step": 3990 }, { "epoch": 1.2510971786833855, "grad_norm": 28.950902938842773, "learning_rate": 1.9955e-06, "loss": 2.7085, "step": 3991 }, { "epoch": 1.25141065830721, "grad_norm": 48.79054641723633, "learning_rate": 1.996e-06, "loss": 2.7858, "step": 3992 }, { "epoch": 1.2517241379310344, "grad_norm": 25.618690490722656, "learning_rate": 1.9965000000000003e-06, "loss": 2.6399, "step": 3993 }, { "epoch": 1.2520376175548589, "grad_norm": 13.908455848693848, "learning_rate": 1.997e-06, "loss": 2.2832, "step": 3994 }, { "epoch": 1.2523510971786833, "grad_norm": 16.477970123291016, "learning_rate": 1.9975000000000004e-06, "loss": 2.6561, "step": 3995 }, { "epoch": 1.2526645768025078, "grad_norm": 39.627376556396484, "learning_rate": 1.998e-06, "loss": 2.8566, "step": 3996 }, { "epoch": 1.2529780564263322, "grad_norm": 15.530925750732422, "learning_rate": 1.9985e-06, "loss": 2.701, "step": 3997 }, { "epoch": 1.2532915360501566, "grad_norm": 71.57562255859375, "learning_rate": 1.9990000000000003e-06, "loss": 3.97, "step": 3998 }, { "epoch": 1.253605015673981, "grad_norm": 50.008113861083984, "learning_rate": 1.9995e-06, "loss": 2.8347, "step": 3999 }, { "epoch": 1.2539184952978055, "grad_norm": 58.43327331542969, "learning_rate": 2.0000000000000003e-06, "loss": 3.1059, "step": 4000 }, { "epoch": 1.25423197492163, "grad_norm": 27.29250144958496, "learning_rate": 2.0005e-06, "loss": 2.5445, "step": 4001 }, { "epoch": 1.2545454545454544, "grad_norm": 14.431353569030762, "learning_rate": 2.001e-06, "loss": 2.4125, "step": 4002 }, { "epoch": 1.254858934169279, "grad_norm": 34.21233367919922, "learning_rate": 2.0015000000000002e-06, "loss": 3.0746, "step": 4003 }, { "epoch": 1.2551724137931035, "grad_norm": 35.75080871582031, "learning_rate": 2.002e-06, "loss": 2.6628, "step": 4004 }, { "epoch": 1.255485893416928, "grad_norm": 39.31830978393555, "learning_rate": 2.0025000000000003e-06, "loss": 2.8678, "step": 4005 }, { "epoch": 1.2557993730407524, "grad_norm": 25.93036651611328, "learning_rate": 2.003e-06, "loss": 2.7643, "step": 4006 }, { "epoch": 1.2561128526645768, "grad_norm": 16.806570053100586, "learning_rate": 2.0035e-06, "loss": 2.3947, "step": 4007 }, { "epoch": 1.2564263322884013, "grad_norm": 101.331787109375, "learning_rate": 2.004e-06, "loss": 3.1663, "step": 4008 }, { "epoch": 1.2567398119122257, "grad_norm": 141.9791717529297, "learning_rate": 2.0045e-06, "loss": 3.1777, "step": 4009 }, { "epoch": 1.2570532915360502, "grad_norm": 18.34634017944336, "learning_rate": 2.0050000000000003e-06, "loss": 2.7777, "step": 4010 }, { "epoch": 1.2573667711598746, "grad_norm": 19.01302719116211, "learning_rate": 2.0055e-06, "loss": 2.5083, "step": 4011 }, { "epoch": 1.257680250783699, "grad_norm": 39.72523880004883, "learning_rate": 2.0060000000000004e-06, "loss": 2.4291, "step": 4012 }, { "epoch": 1.2579937304075235, "grad_norm": 24.59177017211914, "learning_rate": 2.0065e-06, "loss": 2.7645, "step": 4013 }, { "epoch": 1.258307210031348, "grad_norm": 28.052520751953125, "learning_rate": 2.007e-06, "loss": 2.4558, "step": 4014 }, { "epoch": 1.2586206896551724, "grad_norm": 18.298568725585938, "learning_rate": 2.0075000000000003e-06, "loss": 2.3395, "step": 4015 }, { "epoch": 1.2589341692789968, "grad_norm": 19.31409454345703, "learning_rate": 2.008e-06, "loss": 2.6477, "step": 4016 }, { "epoch": 1.2592476489028213, "grad_norm": 61.653419494628906, "learning_rate": 2.0085000000000003e-06, "loss": 3.8697, "step": 4017 }, { "epoch": 1.2595611285266457, "grad_norm": 33.65704345703125, "learning_rate": 2.009e-06, "loss": 2.7283, "step": 4018 }, { "epoch": 1.2598746081504701, "grad_norm": 49.89202117919922, "learning_rate": 2.0095e-06, "loss": 2.9251, "step": 4019 }, { "epoch": 1.2601880877742948, "grad_norm": 25.42795181274414, "learning_rate": 2.0100000000000002e-06, "loss": 2.4845, "step": 4020 }, { "epoch": 1.2605015673981192, "grad_norm": 19.237651824951172, "learning_rate": 2.0105e-06, "loss": 2.5299, "step": 4021 }, { "epoch": 1.2608150470219437, "grad_norm": 20.69957160949707, "learning_rate": 2.0110000000000003e-06, "loss": 2.7474, "step": 4022 }, { "epoch": 1.2611285266457681, "grad_norm": 47.299766540527344, "learning_rate": 2.0115e-06, "loss": 2.7008, "step": 4023 }, { "epoch": 1.2614420062695926, "grad_norm": 26.17997169494629, "learning_rate": 2.012e-06, "loss": 2.9662, "step": 4024 }, { "epoch": 1.261755485893417, "grad_norm": 14.28285026550293, "learning_rate": 2.0125000000000002e-06, "loss": 2.4881, "step": 4025 }, { "epoch": 1.2620689655172415, "grad_norm": 27.487199783325195, "learning_rate": 2.013e-06, "loss": 2.4464, "step": 4026 }, { "epoch": 1.262382445141066, "grad_norm": 20.70392608642578, "learning_rate": 2.0135000000000003e-06, "loss": 2.4101, "step": 4027 }, { "epoch": 1.2626959247648903, "grad_norm": 19.474803924560547, "learning_rate": 2.014e-06, "loss": 2.0501, "step": 4028 }, { "epoch": 1.2630094043887148, "grad_norm": 39.99258804321289, "learning_rate": 2.0145e-06, "loss": 2.5471, "step": 4029 }, { "epoch": 1.2633228840125392, "grad_norm": 26.0811767578125, "learning_rate": 2.015e-06, "loss": 2.3593, "step": 4030 }, { "epoch": 1.2636363636363637, "grad_norm": 19.96682357788086, "learning_rate": 2.0155000000000004e-06, "loss": 2.6898, "step": 4031 }, { "epoch": 1.263949843260188, "grad_norm": 16.560874938964844, "learning_rate": 2.0160000000000003e-06, "loss": 2.4135, "step": 4032 }, { "epoch": 1.2642633228840126, "grad_norm": 44.8605842590332, "learning_rate": 2.0165e-06, "loss": 3.0805, "step": 4033 }, { "epoch": 1.264576802507837, "grad_norm": 14.197938919067383, "learning_rate": 2.017e-06, "loss": 2.8498, "step": 4034 }, { "epoch": 1.2648902821316614, "grad_norm": 28.6807861328125, "learning_rate": 2.0175e-06, "loss": 2.584, "step": 4035 }, { "epoch": 1.2652037617554859, "grad_norm": 69.7178726196289, "learning_rate": 2.0180000000000004e-06, "loss": 2.8437, "step": 4036 }, { "epoch": 1.2655172413793103, "grad_norm": 19.459928512573242, "learning_rate": 2.0185000000000002e-06, "loss": 2.813, "step": 4037 }, { "epoch": 1.2658307210031348, "grad_norm": 12.748106956481934, "learning_rate": 2.019e-06, "loss": 2.5893, "step": 4038 }, { "epoch": 1.2661442006269592, "grad_norm": 58.045040130615234, "learning_rate": 2.0195e-06, "loss": 2.9038, "step": 4039 }, { "epoch": 1.2664576802507836, "grad_norm": 23.788843154907227, "learning_rate": 2.02e-06, "loss": 2.1679, "step": 4040 }, { "epoch": 1.266771159874608, "grad_norm": 29.40966033935547, "learning_rate": 2.0205000000000004e-06, "loss": 2.2847, "step": 4041 }, { "epoch": 1.2670846394984325, "grad_norm": 19.985910415649414, "learning_rate": 2.0210000000000002e-06, "loss": 2.9521, "step": 4042 }, { "epoch": 1.267398119122257, "grad_norm": 91.1566390991211, "learning_rate": 2.0215e-06, "loss": 3.4109, "step": 4043 }, { "epoch": 1.2677115987460814, "grad_norm": 34.98796463012695, "learning_rate": 2.022e-06, "loss": 2.7613, "step": 4044 }, { "epoch": 1.2680250783699059, "grad_norm": 24.939620971679688, "learning_rate": 2.0225e-06, "loss": 2.2957, "step": 4045 }, { "epoch": 1.2683385579937303, "grad_norm": 66.59795379638672, "learning_rate": 2.0230000000000004e-06, "loss": 2.8673, "step": 4046 }, { "epoch": 1.2686520376175547, "grad_norm": 14.904571533203125, "learning_rate": 2.0235e-06, "loss": 2.5483, "step": 4047 }, { "epoch": 1.2689655172413792, "grad_norm": 24.920495986938477, "learning_rate": 2.024e-06, "loss": 2.8943, "step": 4048 }, { "epoch": 1.2692789968652036, "grad_norm": 26.274133682250977, "learning_rate": 2.0245000000000003e-06, "loss": 2.5536, "step": 4049 }, { "epoch": 1.2695924764890283, "grad_norm": 30.83713722229004, "learning_rate": 2.025e-06, "loss": 2.3649, "step": 4050 }, { "epoch": 1.2699059561128527, "grad_norm": 25.9537296295166, "learning_rate": 2.0255000000000003e-06, "loss": 2.3121, "step": 4051 }, { "epoch": 1.2702194357366772, "grad_norm": 16.14348602294922, "learning_rate": 2.026e-06, "loss": 2.421, "step": 4052 }, { "epoch": 1.2705329153605016, "grad_norm": 54.47565841674805, "learning_rate": 2.0265e-06, "loss": 3.325, "step": 4053 }, { "epoch": 1.270846394984326, "grad_norm": 42.51778793334961, "learning_rate": 2.0270000000000002e-06, "loss": 3.1028, "step": 4054 }, { "epoch": 1.2711598746081505, "grad_norm": 28.14337921142578, "learning_rate": 2.0275000000000005e-06, "loss": 2.5186, "step": 4055 }, { "epoch": 1.271473354231975, "grad_norm": 14.081459999084473, "learning_rate": 2.0280000000000003e-06, "loss": 2.3879, "step": 4056 }, { "epoch": 1.2717868338557994, "grad_norm": 32.00901412963867, "learning_rate": 2.0285e-06, "loss": 2.5596, "step": 4057 }, { "epoch": 1.2721003134796238, "grad_norm": 15.976505279541016, "learning_rate": 2.029e-06, "loss": 2.8214, "step": 4058 }, { "epoch": 1.2724137931034483, "grad_norm": 16.56611442565918, "learning_rate": 2.0295000000000002e-06, "loss": 2.4008, "step": 4059 }, { "epoch": 1.2727272727272727, "grad_norm": 37.53776931762695, "learning_rate": 2.0300000000000005e-06, "loss": 2.5785, "step": 4060 }, { "epoch": 1.2730407523510971, "grad_norm": 27.125139236450195, "learning_rate": 2.0305000000000003e-06, "loss": 2.687, "step": 4061 }, { "epoch": 1.2733542319749216, "grad_norm": 21.245471954345703, "learning_rate": 2.031e-06, "loss": 2.6162, "step": 4062 }, { "epoch": 1.273667711598746, "grad_norm": 20.019229888916016, "learning_rate": 2.0315e-06, "loss": 2.2253, "step": 4063 }, { "epoch": 1.2739811912225705, "grad_norm": 14.50991153717041, "learning_rate": 2.032e-06, "loss": 2.3581, "step": 4064 }, { "epoch": 1.274294670846395, "grad_norm": 19.146757125854492, "learning_rate": 2.0325e-06, "loss": 3.0703, "step": 4065 }, { "epoch": 1.2746081504702194, "grad_norm": 55.10849380493164, "learning_rate": 2.0330000000000003e-06, "loss": 2.585, "step": 4066 }, { "epoch": 1.274921630094044, "grad_norm": 15.288493156433105, "learning_rate": 2.0335e-06, "loss": 2.4462, "step": 4067 }, { "epoch": 1.2752351097178685, "grad_norm": 27.375669479370117, "learning_rate": 2.0340000000000003e-06, "loss": 2.4656, "step": 4068 }, { "epoch": 1.275548589341693, "grad_norm": 21.695545196533203, "learning_rate": 2.0345e-06, "loss": 2.659, "step": 4069 }, { "epoch": 1.2758620689655173, "grad_norm": 16.515893936157227, "learning_rate": 2.035e-06, "loss": 2.5209, "step": 4070 }, { "epoch": 1.2761755485893418, "grad_norm": 22.51224708557129, "learning_rate": 2.0355000000000002e-06, "loss": 2.46, "step": 4071 }, { "epoch": 1.2764890282131662, "grad_norm": 19.671100616455078, "learning_rate": 2.036e-06, "loss": 2.6473, "step": 4072 }, { "epoch": 1.2768025078369907, "grad_norm": 21.77365493774414, "learning_rate": 2.0365000000000003e-06, "loss": 2.7338, "step": 4073 }, { "epoch": 1.2771159874608151, "grad_norm": 19.614160537719727, "learning_rate": 2.037e-06, "loss": 2.583, "step": 4074 }, { "epoch": 1.2774294670846396, "grad_norm": 28.54057502746582, "learning_rate": 2.0375e-06, "loss": 3.0976, "step": 4075 }, { "epoch": 1.277742946708464, "grad_norm": 16.560401916503906, "learning_rate": 2.0380000000000002e-06, "loss": 2.4488, "step": 4076 }, { "epoch": 1.2780564263322884, "grad_norm": 22.638416290283203, "learning_rate": 2.0385e-06, "loss": 2.3756, "step": 4077 }, { "epoch": 1.2783699059561129, "grad_norm": 29.103925704956055, "learning_rate": 2.0390000000000003e-06, "loss": 2.5986, "step": 4078 }, { "epoch": 1.2786833855799373, "grad_norm": 30.395748138427734, "learning_rate": 2.0395e-06, "loss": 3.4387, "step": 4079 }, { "epoch": 1.2789968652037618, "grad_norm": 33.306697845458984, "learning_rate": 2.04e-06, "loss": 2.4712, "step": 4080 }, { "epoch": 1.2793103448275862, "grad_norm": 29.530553817749023, "learning_rate": 2.0405e-06, "loss": 2.4654, "step": 4081 }, { "epoch": 1.2796238244514107, "grad_norm": 21.296916961669922, "learning_rate": 2.041e-06, "loss": 2.4743, "step": 4082 }, { "epoch": 1.279937304075235, "grad_norm": 25.111255645751953, "learning_rate": 2.0415000000000003e-06, "loss": 2.4727, "step": 4083 }, { "epoch": 1.2802507836990595, "grad_norm": 22.278154373168945, "learning_rate": 2.042e-06, "loss": 2.2619, "step": 4084 }, { "epoch": 1.280564263322884, "grad_norm": 33.86940383911133, "learning_rate": 2.0425e-06, "loss": 2.5844, "step": 4085 }, { "epoch": 1.2808777429467084, "grad_norm": 22.564645767211914, "learning_rate": 2.043e-06, "loss": 2.6423, "step": 4086 }, { "epoch": 1.2811912225705329, "grad_norm": 24.984695434570312, "learning_rate": 2.0435e-06, "loss": 2.3474, "step": 4087 }, { "epoch": 1.2815047021943573, "grad_norm": 73.06382751464844, "learning_rate": 2.0440000000000003e-06, "loss": 2.3279, "step": 4088 }, { "epoch": 1.2818181818181817, "grad_norm": 24.00354766845703, "learning_rate": 2.0445e-06, "loss": 2.4972, "step": 4089 }, { "epoch": 1.2821316614420062, "grad_norm": 28.285858154296875, "learning_rate": 2.045e-06, "loss": 2.4998, "step": 4090 }, { "epoch": 1.2824451410658306, "grad_norm": 52.50784683227539, "learning_rate": 2.0455e-06, "loss": 3.6619, "step": 4091 }, { "epoch": 1.282758620689655, "grad_norm": 26.97268295288086, "learning_rate": 2.0460000000000004e-06, "loss": 2.7269, "step": 4092 }, { "epoch": 1.2830721003134795, "grad_norm": 26.74242401123047, "learning_rate": 2.0465000000000002e-06, "loss": 2.3844, "step": 4093 }, { "epoch": 1.283385579937304, "grad_norm": 35.4241943359375, "learning_rate": 2.047e-06, "loss": 2.8735, "step": 4094 }, { "epoch": 1.2836990595611284, "grad_norm": 26.71773910522461, "learning_rate": 2.0475e-06, "loss": 2.6532, "step": 4095 }, { "epoch": 1.284012539184953, "grad_norm": 41.195072174072266, "learning_rate": 2.048e-06, "loss": 2.6725, "step": 4096 }, { "epoch": 1.2843260188087775, "grad_norm": 63.87744903564453, "learning_rate": 2.0485000000000004e-06, "loss": 3.0468, "step": 4097 }, { "epoch": 1.284639498432602, "grad_norm": 26.91925048828125, "learning_rate": 2.049e-06, "loss": 2.7028, "step": 4098 }, { "epoch": 1.2849529780564264, "grad_norm": 36.127159118652344, "learning_rate": 2.0495e-06, "loss": 2.801, "step": 4099 }, { "epoch": 1.2852664576802508, "grad_norm": 45.2885856628418, "learning_rate": 2.05e-06, "loss": 2.5757, "step": 4100 }, { "epoch": 1.2855799373040753, "grad_norm": 21.48421859741211, "learning_rate": 2.0505e-06, "loss": 2.5982, "step": 4101 }, { "epoch": 1.2858934169278997, "grad_norm": 21.25076675415039, "learning_rate": 2.0510000000000004e-06, "loss": 2.4623, "step": 4102 }, { "epoch": 1.2862068965517242, "grad_norm": 22.256900787353516, "learning_rate": 2.0515e-06, "loss": 2.4937, "step": 4103 }, { "epoch": 1.2865203761755486, "grad_norm": 35.315425872802734, "learning_rate": 2.052e-06, "loss": 2.5487, "step": 4104 }, { "epoch": 1.286833855799373, "grad_norm": 21.41267204284668, "learning_rate": 2.0525000000000003e-06, "loss": 2.6247, "step": 4105 }, { "epoch": 1.2871473354231975, "grad_norm": 23.571125030517578, "learning_rate": 2.053e-06, "loss": 2.4657, "step": 4106 }, { "epoch": 1.287460815047022, "grad_norm": 20.08584976196289, "learning_rate": 2.0535000000000003e-06, "loss": 2.6455, "step": 4107 }, { "epoch": 1.2877742946708464, "grad_norm": 26.432762145996094, "learning_rate": 2.054e-06, "loss": 2.3619, "step": 4108 }, { "epoch": 1.2880877742946708, "grad_norm": 28.0866641998291, "learning_rate": 2.0545e-06, "loss": 2.5186, "step": 4109 }, { "epoch": 1.2884012539184952, "grad_norm": 58.62767028808594, "learning_rate": 2.0550000000000002e-06, "loss": 2.4484, "step": 4110 }, { "epoch": 1.2887147335423197, "grad_norm": 19.969892501831055, "learning_rate": 2.0555000000000005e-06, "loss": 2.268, "step": 4111 }, { "epoch": 1.2890282131661441, "grad_norm": 20.90323257446289, "learning_rate": 2.0560000000000003e-06, "loss": 2.4765, "step": 4112 }, { "epoch": 1.2893416927899686, "grad_norm": 33.036624908447266, "learning_rate": 2.0565e-06, "loss": 2.888, "step": 4113 }, { "epoch": 1.2896551724137932, "grad_norm": 24.629642486572266, "learning_rate": 2.057e-06, "loss": 2.5645, "step": 4114 }, { "epoch": 1.2899686520376177, "grad_norm": 12.752878189086914, "learning_rate": 2.0575e-06, "loss": 2.5795, "step": 4115 }, { "epoch": 1.2902821316614421, "grad_norm": 33.927825927734375, "learning_rate": 2.0580000000000005e-06, "loss": 2.7409, "step": 4116 }, { "epoch": 1.2905956112852666, "grad_norm": 12.693510055541992, "learning_rate": 2.0585000000000003e-06, "loss": 2.3876, "step": 4117 }, { "epoch": 1.290909090909091, "grad_norm": 37.0070915222168, "learning_rate": 2.059e-06, "loss": 2.8024, "step": 4118 }, { "epoch": 1.2912225705329154, "grad_norm": 24.57074546813965, "learning_rate": 2.0595e-06, "loss": 2.6066, "step": 4119 }, { "epoch": 1.29153605015674, "grad_norm": 28.160541534423828, "learning_rate": 2.06e-06, "loss": 2.6939, "step": 4120 }, { "epoch": 1.2918495297805643, "grad_norm": 37.937007904052734, "learning_rate": 2.0605000000000004e-06, "loss": 2.9812, "step": 4121 }, { "epoch": 1.2921630094043888, "grad_norm": 16.546600341796875, "learning_rate": 2.0610000000000003e-06, "loss": 2.44, "step": 4122 }, { "epoch": 1.2924764890282132, "grad_norm": 71.4891128540039, "learning_rate": 2.0615e-06, "loss": 2.7788, "step": 4123 }, { "epoch": 1.2927899686520377, "grad_norm": 33.58329772949219, "learning_rate": 2.062e-06, "loss": 2.6808, "step": 4124 }, { "epoch": 1.293103448275862, "grad_norm": 46.48095703125, "learning_rate": 2.0625e-06, "loss": 2.614, "step": 4125 }, { "epoch": 1.2934169278996865, "grad_norm": 26.5600643157959, "learning_rate": 2.0630000000000004e-06, "loss": 2.6591, "step": 4126 }, { "epoch": 1.293730407523511, "grad_norm": 26.45882225036621, "learning_rate": 2.0635000000000002e-06, "loss": 2.9518, "step": 4127 }, { "epoch": 1.2940438871473354, "grad_norm": 34.29823684692383, "learning_rate": 2.064e-06, "loss": 2.7566, "step": 4128 }, { "epoch": 1.2943573667711599, "grad_norm": 20.29995346069336, "learning_rate": 2.0645000000000003e-06, "loss": 2.8827, "step": 4129 }, { "epoch": 1.2946708463949843, "grad_norm": 14.124770164489746, "learning_rate": 2.065e-06, "loss": 2.2836, "step": 4130 }, { "epoch": 1.2949843260188088, "grad_norm": 20.10064697265625, "learning_rate": 2.0655000000000004e-06, "loss": 3.0029, "step": 4131 }, { "epoch": 1.2952978056426332, "grad_norm": 41.575111389160156, "learning_rate": 2.066e-06, "loss": 2.2947, "step": 4132 }, { "epoch": 1.2956112852664576, "grad_norm": 15.761857032775879, "learning_rate": 2.0665e-06, "loss": 2.3627, "step": 4133 }, { "epoch": 1.295924764890282, "grad_norm": 15.230303764343262, "learning_rate": 2.0670000000000003e-06, "loss": 2.9609, "step": 4134 }, { "epoch": 1.2962382445141065, "grad_norm": 25.583053588867188, "learning_rate": 2.0675e-06, "loss": 2.3677, "step": 4135 }, { "epoch": 1.296551724137931, "grad_norm": 22.213247299194336, "learning_rate": 2.0680000000000004e-06, "loss": 2.5241, "step": 4136 }, { "epoch": 1.2968652037617554, "grad_norm": 25.54466438293457, "learning_rate": 2.0685e-06, "loss": 2.6116, "step": 4137 }, { "epoch": 1.2971786833855798, "grad_norm": 23.40324592590332, "learning_rate": 2.069e-06, "loss": 2.124, "step": 4138 }, { "epoch": 1.2974921630094043, "grad_norm": 41.573062896728516, "learning_rate": 2.0695000000000003e-06, "loss": 3.7863, "step": 4139 }, { "epoch": 1.2978056426332287, "grad_norm": 13.942071914672852, "learning_rate": 2.07e-06, "loss": 2.3136, "step": 4140 }, { "epoch": 1.2981191222570532, "grad_norm": 15.588929176330566, "learning_rate": 2.0705000000000003e-06, "loss": 2.3499, "step": 4141 }, { "epoch": 1.2984326018808776, "grad_norm": 117.79044342041016, "learning_rate": 2.071e-06, "loss": 2.8299, "step": 4142 }, { "epoch": 1.2987460815047023, "grad_norm": 33.719791412353516, "learning_rate": 2.0715e-06, "loss": 2.3125, "step": 4143 }, { "epoch": 1.2990595611285267, "grad_norm": 53.90446472167969, "learning_rate": 2.0720000000000002e-06, "loss": 2.7984, "step": 4144 }, { "epoch": 1.2993730407523512, "grad_norm": 61.47092819213867, "learning_rate": 2.0725e-06, "loss": 3.0994, "step": 4145 }, { "epoch": 1.2996865203761756, "grad_norm": 33.92884063720703, "learning_rate": 2.0730000000000003e-06, "loss": 3.1014, "step": 4146 }, { "epoch": 1.3, "grad_norm": 29.12774085998535, "learning_rate": 2.0735e-06, "loss": 2.4252, "step": 4147 }, { "epoch": 1.3003134796238245, "grad_norm": 20.700414657592773, "learning_rate": 2.0740000000000004e-06, "loss": 3.0268, "step": 4148 }, { "epoch": 1.300626959247649, "grad_norm": 48.68048095703125, "learning_rate": 2.0745000000000002e-06, "loss": 2.9331, "step": 4149 }, { "epoch": 1.3009404388714734, "grad_norm": 17.82050323486328, "learning_rate": 2.075e-06, "loss": 2.3952, "step": 4150 }, { "epoch": 1.3012539184952978, "grad_norm": 20.30829429626465, "learning_rate": 2.0755000000000003e-06, "loss": 2.4276, "step": 4151 }, { "epoch": 1.3015673981191223, "grad_norm": 28.34781265258789, "learning_rate": 2.076e-06, "loss": 2.5247, "step": 4152 }, { "epoch": 1.3018808777429467, "grad_norm": 15.627279281616211, "learning_rate": 2.0765000000000004e-06, "loss": 2.5008, "step": 4153 }, { "epoch": 1.3021943573667711, "grad_norm": 24.798925399780273, "learning_rate": 2.077e-06, "loss": 2.7212, "step": 4154 }, { "epoch": 1.3025078369905956, "grad_norm": 27.888092041015625, "learning_rate": 2.0775e-06, "loss": 2.6418, "step": 4155 }, { "epoch": 1.30282131661442, "grad_norm": 15.778170585632324, "learning_rate": 2.0780000000000003e-06, "loss": 2.2588, "step": 4156 }, { "epoch": 1.3031347962382445, "grad_norm": 21.949968338012695, "learning_rate": 2.0785e-06, "loss": 2.2574, "step": 4157 }, { "epoch": 1.303448275862069, "grad_norm": 25.079078674316406, "learning_rate": 2.0790000000000003e-06, "loss": 2.8032, "step": 4158 }, { "epoch": 1.3037617554858933, "grad_norm": 25.341110229492188, "learning_rate": 2.0795e-06, "loss": 2.5002, "step": 4159 }, { "epoch": 1.3040752351097178, "grad_norm": 44.27704620361328, "learning_rate": 2.08e-06, "loss": 3.2607, "step": 4160 }, { "epoch": 1.3043887147335425, "grad_norm": 47.27009582519531, "learning_rate": 2.0805000000000002e-06, "loss": 2.3683, "step": 4161 }, { "epoch": 1.304702194357367, "grad_norm": 29.553010940551758, "learning_rate": 2.081e-06, "loss": 2.5803, "step": 4162 }, { "epoch": 1.3050156739811913, "grad_norm": 25.054651260375977, "learning_rate": 2.0815000000000003e-06, "loss": 2.265, "step": 4163 }, { "epoch": 1.3053291536050158, "grad_norm": 83.10655212402344, "learning_rate": 2.082e-06, "loss": 2.4293, "step": 4164 }, { "epoch": 1.3056426332288402, "grad_norm": 20.534961700439453, "learning_rate": 2.0825e-06, "loss": 2.5412, "step": 4165 }, { "epoch": 1.3059561128526647, "grad_norm": 27.032581329345703, "learning_rate": 2.0830000000000002e-06, "loss": 2.8988, "step": 4166 }, { "epoch": 1.306269592476489, "grad_norm": 19.306673049926758, "learning_rate": 2.0835e-06, "loss": 2.5186, "step": 4167 }, { "epoch": 1.3065830721003135, "grad_norm": 34.469871520996094, "learning_rate": 2.0840000000000003e-06, "loss": 2.4477, "step": 4168 }, { "epoch": 1.306896551724138, "grad_norm": 23.936790466308594, "learning_rate": 2.0845e-06, "loss": 2.4572, "step": 4169 }, { "epoch": 1.3072100313479624, "grad_norm": 26.452882766723633, "learning_rate": 2.085e-06, "loss": 2.33, "step": 4170 }, { "epoch": 1.3075235109717869, "grad_norm": 36.12260437011719, "learning_rate": 2.0855e-06, "loss": 2.9178, "step": 4171 }, { "epoch": 1.3078369905956113, "grad_norm": 19.62932777404785, "learning_rate": 2.0860000000000004e-06, "loss": 2.5279, "step": 4172 }, { "epoch": 1.3081504702194358, "grad_norm": 21.609783172607422, "learning_rate": 2.0865000000000003e-06, "loss": 2.7524, "step": 4173 }, { "epoch": 1.3084639498432602, "grad_norm": 62.739112854003906, "learning_rate": 2.087e-06, "loss": 2.985, "step": 4174 }, { "epoch": 1.3087774294670846, "grad_norm": 34.94454574584961, "learning_rate": 2.0875e-06, "loss": 3.0326, "step": 4175 }, { "epoch": 1.309090909090909, "grad_norm": 54.120731353759766, "learning_rate": 2.088e-06, "loss": 3.2535, "step": 4176 }, { "epoch": 1.3094043887147335, "grad_norm": 35.5368537902832, "learning_rate": 2.0885000000000004e-06, "loss": 2.9125, "step": 4177 }, { "epoch": 1.309717868338558, "grad_norm": 38.99037170410156, "learning_rate": 2.0890000000000002e-06, "loss": 2.7989, "step": 4178 }, { "epoch": 1.3100313479623824, "grad_norm": 14.405024528503418, "learning_rate": 2.0895e-06, "loss": 2.4821, "step": 4179 }, { "epoch": 1.3103448275862069, "grad_norm": 25.834924697875977, "learning_rate": 2.09e-06, "loss": 2.5991, "step": 4180 }, { "epoch": 1.3106583072100313, "grad_norm": 23.313091278076172, "learning_rate": 2.0905e-06, "loss": 2.4887, "step": 4181 }, { "epoch": 1.3109717868338557, "grad_norm": 23.875391006469727, "learning_rate": 2.0910000000000004e-06, "loss": 2.8547, "step": 4182 }, { "epoch": 1.3112852664576802, "grad_norm": 22.093515396118164, "learning_rate": 2.0915000000000002e-06, "loss": 2.6691, "step": 4183 }, { "epoch": 1.3115987460815046, "grad_norm": 14.514626502990723, "learning_rate": 2.092e-06, "loss": 2.4244, "step": 4184 }, { "epoch": 1.311912225705329, "grad_norm": 50.77438735961914, "learning_rate": 2.0925000000000003e-06, "loss": 2.8881, "step": 4185 }, { "epoch": 1.3122257053291535, "grad_norm": 34.015419006347656, "learning_rate": 2.093e-06, "loss": 2.848, "step": 4186 }, { "epoch": 1.312539184952978, "grad_norm": 32.5399284362793, "learning_rate": 2.0935000000000004e-06, "loss": 2.7736, "step": 4187 }, { "epoch": 1.3128526645768024, "grad_norm": 22.101634979248047, "learning_rate": 2.094e-06, "loss": 2.662, "step": 4188 }, { "epoch": 1.3131661442006268, "grad_norm": 22.13960838317871, "learning_rate": 2.0945e-06, "loss": 2.4422, "step": 4189 }, { "epoch": 1.3134796238244515, "grad_norm": 34.24423599243164, "learning_rate": 2.0950000000000003e-06, "loss": 2.9224, "step": 4190 }, { "epoch": 1.313793103448276, "grad_norm": 26.192934036254883, "learning_rate": 2.0955e-06, "loss": 2.556, "step": 4191 }, { "epoch": 1.3141065830721004, "grad_norm": 22.4783992767334, "learning_rate": 2.0960000000000003e-06, "loss": 3.2724, "step": 4192 }, { "epoch": 1.3144200626959248, "grad_norm": 30.03993034362793, "learning_rate": 2.0965e-06, "loss": 2.2862, "step": 4193 }, { "epoch": 1.3147335423197493, "grad_norm": 44.098018646240234, "learning_rate": 2.097e-06, "loss": 2.8181, "step": 4194 }, { "epoch": 1.3150470219435737, "grad_norm": 34.14385223388672, "learning_rate": 2.0975000000000002e-06, "loss": 2.5418, "step": 4195 }, { "epoch": 1.3153605015673981, "grad_norm": 25.11463165283203, "learning_rate": 2.098e-06, "loss": 2.5286, "step": 4196 }, { "epoch": 1.3156739811912226, "grad_norm": 28.286510467529297, "learning_rate": 2.0985000000000003e-06, "loss": 2.2723, "step": 4197 }, { "epoch": 1.315987460815047, "grad_norm": 29.696142196655273, "learning_rate": 2.099e-06, "loss": 2.5886, "step": 4198 }, { "epoch": 1.3163009404388715, "grad_norm": 16.753244400024414, "learning_rate": 2.0995e-06, "loss": 2.6589, "step": 4199 }, { "epoch": 1.316614420062696, "grad_norm": 23.720779418945312, "learning_rate": 2.1000000000000002e-06, "loss": 2.5827, "step": 4200 }, { "epoch": 1.3169278996865204, "grad_norm": 15.066795349121094, "learning_rate": 2.1005e-06, "loss": 2.4665, "step": 4201 }, { "epoch": 1.3172413793103448, "grad_norm": 26.113821029663086, "learning_rate": 2.1010000000000003e-06, "loss": 3.0411, "step": 4202 }, { "epoch": 1.3175548589341692, "grad_norm": 43.554443359375, "learning_rate": 2.1015e-06, "loss": 2.5904, "step": 4203 }, { "epoch": 1.3178683385579937, "grad_norm": 20.175739288330078, "learning_rate": 2.102e-06, "loss": 2.5329, "step": 4204 }, { "epoch": 1.3181818181818181, "grad_norm": 15.298727035522461, "learning_rate": 2.1025e-06, "loss": 2.4321, "step": 4205 }, { "epoch": 1.3184952978056426, "grad_norm": 48.11604309082031, "learning_rate": 2.103e-06, "loss": 2.5175, "step": 4206 }, { "epoch": 1.318808777429467, "grad_norm": 73.75093841552734, "learning_rate": 2.1035000000000003e-06, "loss": 2.6158, "step": 4207 }, { "epoch": 1.3191222570532917, "grad_norm": 49.25511932373047, "learning_rate": 2.104e-06, "loss": 2.5209, "step": 4208 }, { "epoch": 1.3194357366771161, "grad_norm": 19.584362030029297, "learning_rate": 2.1045000000000003e-06, "loss": 2.7566, "step": 4209 }, { "epoch": 1.3197492163009406, "grad_norm": 31.413415908813477, "learning_rate": 2.105e-06, "loss": 2.6506, "step": 4210 }, { "epoch": 1.320062695924765, "grad_norm": 24.068866729736328, "learning_rate": 2.1055e-06, "loss": 2.6964, "step": 4211 }, { "epoch": 1.3203761755485894, "grad_norm": 18.091983795166016, "learning_rate": 2.1060000000000002e-06, "loss": 2.4337, "step": 4212 }, { "epoch": 1.3206896551724139, "grad_norm": 21.102827072143555, "learning_rate": 2.1065e-06, "loss": 2.5642, "step": 4213 }, { "epoch": 1.3210031347962383, "grad_norm": 36.744781494140625, "learning_rate": 2.1070000000000003e-06, "loss": 2.3418, "step": 4214 }, { "epoch": 1.3213166144200628, "grad_norm": 113.0008316040039, "learning_rate": 2.1075e-06, "loss": 3.1146, "step": 4215 }, { "epoch": 1.3216300940438872, "grad_norm": 28.350322723388672, "learning_rate": 2.108e-06, "loss": 2.5306, "step": 4216 }, { "epoch": 1.3219435736677116, "grad_norm": 15.95138931274414, "learning_rate": 2.1085000000000002e-06, "loss": 2.3577, "step": 4217 }, { "epoch": 1.322257053291536, "grad_norm": 54.72137451171875, "learning_rate": 2.109e-06, "loss": 2.5567, "step": 4218 }, { "epoch": 1.3225705329153605, "grad_norm": 14.051758766174316, "learning_rate": 2.1095000000000003e-06, "loss": 2.2394, "step": 4219 }, { "epoch": 1.322884012539185, "grad_norm": 27.43135643005371, "learning_rate": 2.11e-06, "loss": 2.2638, "step": 4220 }, { "epoch": 1.3231974921630094, "grad_norm": 19.857425689697266, "learning_rate": 2.1105e-06, "loss": 2.7347, "step": 4221 }, { "epoch": 1.3235109717868339, "grad_norm": 16.808263778686523, "learning_rate": 2.111e-06, "loss": 2.3323, "step": 4222 }, { "epoch": 1.3238244514106583, "grad_norm": 31.731477737426758, "learning_rate": 2.1115e-06, "loss": 2.594, "step": 4223 }, { "epoch": 1.3241379310344827, "grad_norm": 42.76530075073242, "learning_rate": 2.1120000000000003e-06, "loss": 2.4346, "step": 4224 }, { "epoch": 1.3244514106583072, "grad_norm": 13.989803314208984, "learning_rate": 2.1125e-06, "loss": 2.5349, "step": 4225 }, { "epoch": 1.3247648902821316, "grad_norm": 18.851530075073242, "learning_rate": 2.113e-06, "loss": 2.4989, "step": 4226 }, { "epoch": 1.325078369905956, "grad_norm": 57.27273178100586, "learning_rate": 2.1135e-06, "loss": 2.3327, "step": 4227 }, { "epoch": 1.3253918495297805, "grad_norm": 21.600278854370117, "learning_rate": 2.1140000000000004e-06, "loss": 2.6823, "step": 4228 }, { "epoch": 1.325705329153605, "grad_norm": 18.557979583740234, "learning_rate": 2.1145000000000003e-06, "loss": 2.8352, "step": 4229 }, { "epoch": 1.3260188087774294, "grad_norm": 10.887141227722168, "learning_rate": 2.115e-06, "loss": 2.4158, "step": 4230 }, { "epoch": 1.3263322884012538, "grad_norm": 29.551910400390625, "learning_rate": 2.1155e-06, "loss": 2.6161, "step": 4231 }, { "epoch": 1.3266457680250783, "grad_norm": 46.62107849121094, "learning_rate": 2.116e-06, "loss": 3.0583, "step": 4232 }, { "epoch": 1.3269592476489027, "grad_norm": 22.17835235595703, "learning_rate": 2.1165000000000004e-06, "loss": 2.5605, "step": 4233 }, { "epoch": 1.3272727272727272, "grad_norm": 15.31783390045166, "learning_rate": 2.1170000000000002e-06, "loss": 2.3408, "step": 4234 }, { "epoch": 1.3275862068965516, "grad_norm": 32.893497467041016, "learning_rate": 2.1175e-06, "loss": 2.7138, "step": 4235 }, { "epoch": 1.327899686520376, "grad_norm": 37.14299774169922, "learning_rate": 2.118e-06, "loss": 2.4061, "step": 4236 }, { "epoch": 1.3282131661442007, "grad_norm": 45.42509841918945, "learning_rate": 2.1185e-06, "loss": 3.1221, "step": 4237 }, { "epoch": 1.3285266457680251, "grad_norm": 34.94936752319336, "learning_rate": 2.1190000000000004e-06, "loss": 2.9343, "step": 4238 }, { "epoch": 1.3288401253918496, "grad_norm": 41.42539978027344, "learning_rate": 2.1195e-06, "loss": 2.8523, "step": 4239 }, { "epoch": 1.329153605015674, "grad_norm": 25.758621215820312, "learning_rate": 2.12e-06, "loss": 2.2589, "step": 4240 }, { "epoch": 1.3294670846394985, "grad_norm": 37.20207595825195, "learning_rate": 2.1205e-06, "loss": 2.4667, "step": 4241 }, { "epoch": 1.329780564263323, "grad_norm": 14.378911018371582, "learning_rate": 2.121e-06, "loss": 2.6144, "step": 4242 }, { "epoch": 1.3300940438871474, "grad_norm": 30.05830955505371, "learning_rate": 2.1215000000000004e-06, "loss": 2.7819, "step": 4243 }, { "epoch": 1.3304075235109718, "grad_norm": 24.931983947753906, "learning_rate": 2.122e-06, "loss": 3.0302, "step": 4244 }, { "epoch": 1.3307210031347962, "grad_norm": 34.86129379272461, "learning_rate": 2.1225e-06, "loss": 3.0939, "step": 4245 }, { "epoch": 1.3310344827586207, "grad_norm": 30.962717056274414, "learning_rate": 2.1230000000000003e-06, "loss": 2.6051, "step": 4246 }, { "epoch": 1.3313479623824451, "grad_norm": 23.41077423095703, "learning_rate": 2.1235e-06, "loss": 2.5413, "step": 4247 }, { "epoch": 1.3316614420062696, "grad_norm": 16.523805618286133, "learning_rate": 2.1240000000000003e-06, "loss": 2.4509, "step": 4248 }, { "epoch": 1.331974921630094, "grad_norm": 25.446996688842773, "learning_rate": 2.1245e-06, "loss": 2.3192, "step": 4249 }, { "epoch": 1.3322884012539185, "grad_norm": 50.319541931152344, "learning_rate": 2.125e-06, "loss": 2.6523, "step": 4250 }, { "epoch": 1.332601880877743, "grad_norm": 30.700286865234375, "learning_rate": 2.1255000000000002e-06, "loss": 2.7076, "step": 4251 }, { "epoch": 1.3329153605015673, "grad_norm": 25.385419845581055, "learning_rate": 2.1260000000000005e-06, "loss": 2.541, "step": 4252 }, { "epoch": 1.3332288401253918, "grad_norm": 21.421798706054688, "learning_rate": 2.1265000000000003e-06, "loss": 3.12, "step": 4253 }, { "epoch": 1.3335423197492162, "grad_norm": 33.206661224365234, "learning_rate": 2.127e-06, "loss": 2.5614, "step": 4254 }, { "epoch": 1.3338557993730409, "grad_norm": 31.233118057250977, "learning_rate": 2.1275e-06, "loss": 2.4797, "step": 4255 }, { "epoch": 1.3341692789968653, "grad_norm": 21.808687210083008, "learning_rate": 2.128e-06, "loss": 2.3788, "step": 4256 }, { "epoch": 1.3344827586206898, "grad_norm": 17.938234329223633, "learning_rate": 2.1285000000000005e-06, "loss": 2.4703, "step": 4257 }, { "epoch": 1.3347962382445142, "grad_norm": 43.056495666503906, "learning_rate": 2.1290000000000003e-06, "loss": 2.7721, "step": 4258 }, { "epoch": 1.3351097178683387, "grad_norm": 27.85002899169922, "learning_rate": 2.1295e-06, "loss": 2.2889, "step": 4259 }, { "epoch": 1.335423197492163, "grad_norm": 14.552286148071289, "learning_rate": 2.13e-06, "loss": 2.3216, "step": 4260 }, { "epoch": 1.3357366771159875, "grad_norm": 53.797794342041016, "learning_rate": 2.1305e-06, "loss": 2.9491, "step": 4261 }, { "epoch": 1.336050156739812, "grad_norm": 28.714412689208984, "learning_rate": 2.1310000000000004e-06, "loss": 2.4772, "step": 4262 }, { "epoch": 1.3363636363636364, "grad_norm": 21.757843017578125, "learning_rate": 2.1315000000000003e-06, "loss": 2.6578, "step": 4263 }, { "epoch": 1.3366771159874609, "grad_norm": 18.476587295532227, "learning_rate": 2.132e-06, "loss": 2.4202, "step": 4264 }, { "epoch": 1.3369905956112853, "grad_norm": 13.806644439697266, "learning_rate": 2.1325000000000003e-06, "loss": 2.2175, "step": 4265 }, { "epoch": 1.3373040752351097, "grad_norm": 18.443056106567383, "learning_rate": 2.133e-06, "loss": 2.6481, "step": 4266 }, { "epoch": 1.3376175548589342, "grad_norm": 50.59961700439453, "learning_rate": 2.1335000000000004e-06, "loss": 3.1419, "step": 4267 }, { "epoch": 1.3379310344827586, "grad_norm": 94.30250549316406, "learning_rate": 2.1340000000000002e-06, "loss": 3.4595, "step": 4268 }, { "epoch": 1.338244514106583, "grad_norm": 28.462663650512695, "learning_rate": 2.1345e-06, "loss": 2.2085, "step": 4269 }, { "epoch": 1.3385579937304075, "grad_norm": 43.41497802734375, "learning_rate": 2.1350000000000003e-06, "loss": 2.5531, "step": 4270 }, { "epoch": 1.338871473354232, "grad_norm": 14.940423965454102, "learning_rate": 2.1355e-06, "loss": 2.3722, "step": 4271 }, { "epoch": 1.3391849529780564, "grad_norm": 18.490428924560547, "learning_rate": 2.1360000000000004e-06, "loss": 2.3419, "step": 4272 }, { "epoch": 1.3394984326018808, "grad_norm": 30.41261100769043, "learning_rate": 2.1365000000000002e-06, "loss": 2.837, "step": 4273 }, { "epoch": 1.3398119122257053, "grad_norm": 31.712522506713867, "learning_rate": 2.137e-06, "loss": 3.1905, "step": 4274 }, { "epoch": 1.3401253918495297, "grad_norm": 16.07421875, "learning_rate": 2.1375000000000003e-06, "loss": 2.1409, "step": 4275 }, { "epoch": 1.3404388714733542, "grad_norm": 21.069381713867188, "learning_rate": 2.138e-06, "loss": 2.3961, "step": 4276 }, { "epoch": 1.3407523510971786, "grad_norm": 36.013614654541016, "learning_rate": 2.1385000000000004e-06, "loss": 2.3364, "step": 4277 }, { "epoch": 1.341065830721003, "grad_norm": 17.197298049926758, "learning_rate": 2.139e-06, "loss": 2.506, "step": 4278 }, { "epoch": 1.3413793103448275, "grad_norm": 17.188724517822266, "learning_rate": 2.1395e-06, "loss": 2.6044, "step": 4279 }, { "epoch": 1.341692789968652, "grad_norm": 25.85085105895996, "learning_rate": 2.1400000000000003e-06, "loss": 3.0611, "step": 4280 }, { "epoch": 1.3420062695924764, "grad_norm": 18.796472549438477, "learning_rate": 2.1405e-06, "loss": 2.241, "step": 4281 }, { "epoch": 1.3423197492163008, "grad_norm": 16.755332946777344, "learning_rate": 2.1410000000000003e-06, "loss": 2.2393, "step": 4282 }, { "epoch": 1.3426332288401253, "grad_norm": 19.352550506591797, "learning_rate": 2.1415e-06, "loss": 2.61, "step": 4283 }, { "epoch": 1.34294670846395, "grad_norm": 29.244054794311523, "learning_rate": 2.142e-06, "loss": 2.7959, "step": 4284 }, { "epoch": 1.3432601880877744, "grad_norm": 40.46537780761719, "learning_rate": 2.1425000000000002e-06, "loss": 3.0749, "step": 4285 }, { "epoch": 1.3435736677115988, "grad_norm": 17.840002059936523, "learning_rate": 2.143e-06, "loss": 2.266, "step": 4286 }, { "epoch": 1.3438871473354232, "grad_norm": 47.583038330078125, "learning_rate": 2.1435000000000003e-06, "loss": 2.668, "step": 4287 }, { "epoch": 1.3442006269592477, "grad_norm": 23.055110931396484, "learning_rate": 2.144e-06, "loss": 2.6395, "step": 4288 }, { "epoch": 1.3445141065830721, "grad_norm": 38.746795654296875, "learning_rate": 2.1445000000000004e-06, "loss": 3.8204, "step": 4289 }, { "epoch": 1.3448275862068966, "grad_norm": 19.897480010986328, "learning_rate": 2.1450000000000002e-06, "loss": 2.3241, "step": 4290 }, { "epoch": 1.345141065830721, "grad_norm": 257.05810546875, "learning_rate": 2.1455e-06, "loss": 3.1263, "step": 4291 }, { "epoch": 1.3454545454545455, "grad_norm": 61.294918060302734, "learning_rate": 2.1460000000000003e-06, "loss": 2.7713, "step": 4292 }, { "epoch": 1.34576802507837, "grad_norm": 21.1854305267334, "learning_rate": 2.1465e-06, "loss": 2.2955, "step": 4293 }, { "epoch": 1.3460815047021943, "grad_norm": 21.163742065429688, "learning_rate": 2.1470000000000004e-06, "loss": 2.4182, "step": 4294 }, { "epoch": 1.3463949843260188, "grad_norm": 18.480819702148438, "learning_rate": 2.1475e-06, "loss": 2.6374, "step": 4295 }, { "epoch": 1.3467084639498432, "grad_norm": 13.95515251159668, "learning_rate": 2.148e-06, "loss": 2.2971, "step": 4296 }, { "epoch": 1.3470219435736677, "grad_norm": 40.44957733154297, "learning_rate": 2.1485000000000003e-06, "loss": 2.5437, "step": 4297 }, { "epoch": 1.347335423197492, "grad_norm": 63.295997619628906, "learning_rate": 2.149e-06, "loss": 2.7892, "step": 4298 }, { "epoch": 1.3476489028213166, "grad_norm": 33.74834060668945, "learning_rate": 2.1495000000000003e-06, "loss": 2.737, "step": 4299 }, { "epoch": 1.347962382445141, "grad_norm": 28.50554656982422, "learning_rate": 2.15e-06, "loss": 2.3759, "step": 4300 }, { "epoch": 1.3482758620689654, "grad_norm": 22.857704162597656, "learning_rate": 2.1505e-06, "loss": 2.9179, "step": 4301 }, { "epoch": 1.34858934169279, "grad_norm": 15.061685562133789, "learning_rate": 2.1510000000000002e-06, "loss": 2.2522, "step": 4302 }, { "epoch": 1.3489028213166145, "grad_norm": 19.26007080078125, "learning_rate": 2.1515e-06, "loss": 2.4863, "step": 4303 }, { "epoch": 1.349216300940439, "grad_norm": 62.05062484741211, "learning_rate": 2.1520000000000003e-06, "loss": 3.1491, "step": 4304 }, { "epoch": 1.3495297805642634, "grad_norm": 21.983184814453125, "learning_rate": 2.1525e-06, "loss": 2.4744, "step": 4305 }, { "epoch": 1.3498432601880879, "grad_norm": 60.55722427368164, "learning_rate": 2.153e-06, "loss": 2.5645, "step": 4306 }, { "epoch": 1.3501567398119123, "grad_norm": 14.09309196472168, "learning_rate": 2.1535000000000002e-06, "loss": 2.71, "step": 4307 }, { "epoch": 1.3504702194357368, "grad_norm": 43.35044860839844, "learning_rate": 2.1540000000000005e-06, "loss": 2.8857, "step": 4308 }, { "epoch": 1.3507836990595612, "grad_norm": 23.708900451660156, "learning_rate": 2.1545000000000003e-06, "loss": 2.5606, "step": 4309 }, { "epoch": 1.3510971786833856, "grad_norm": 20.863204956054688, "learning_rate": 2.155e-06, "loss": 2.6164, "step": 4310 }, { "epoch": 1.35141065830721, "grad_norm": 16.67975616455078, "learning_rate": 2.1555e-06, "loss": 2.3722, "step": 4311 }, { "epoch": 1.3517241379310345, "grad_norm": 16.18927001953125, "learning_rate": 2.156e-06, "loss": 2.2388, "step": 4312 }, { "epoch": 1.352037617554859, "grad_norm": 21.1960506439209, "learning_rate": 2.1565000000000004e-06, "loss": 2.4443, "step": 4313 }, { "epoch": 1.3523510971786834, "grad_norm": 50.9581184387207, "learning_rate": 2.1570000000000003e-06, "loss": 2.742, "step": 4314 }, { "epoch": 1.3526645768025078, "grad_norm": 29.12017250061035, "learning_rate": 2.1575e-06, "loss": 2.7771, "step": 4315 }, { "epoch": 1.3529780564263323, "grad_norm": 16.846023559570312, "learning_rate": 2.158e-06, "loss": 2.5052, "step": 4316 }, { "epoch": 1.3532915360501567, "grad_norm": 50.6660270690918, "learning_rate": 2.1585e-06, "loss": 2.8763, "step": 4317 }, { "epoch": 1.3536050156739812, "grad_norm": 15.816275596618652, "learning_rate": 2.1590000000000004e-06, "loss": 2.5491, "step": 4318 }, { "epoch": 1.3539184952978056, "grad_norm": 52.377525329589844, "learning_rate": 2.1595000000000002e-06, "loss": 2.6742, "step": 4319 }, { "epoch": 1.35423197492163, "grad_norm": 370.2469482421875, "learning_rate": 2.16e-06, "loss": 2.6247, "step": 4320 }, { "epoch": 1.3545454545454545, "grad_norm": 46.37929153442383, "learning_rate": 2.1605e-06, "loss": 3.0429, "step": 4321 }, { "epoch": 1.354858934169279, "grad_norm": 44.83418655395508, "learning_rate": 2.161e-06, "loss": 2.7189, "step": 4322 }, { "epoch": 1.3551724137931034, "grad_norm": 12.778990745544434, "learning_rate": 2.1615000000000004e-06, "loss": 2.3526, "step": 4323 }, { "epoch": 1.3554858934169278, "grad_norm": 21.996225357055664, "learning_rate": 2.1620000000000002e-06, "loss": 2.9304, "step": 4324 }, { "epoch": 1.3557993730407523, "grad_norm": 16.98282814025879, "learning_rate": 2.1625e-06, "loss": 2.3532, "step": 4325 }, { "epoch": 1.3561128526645767, "grad_norm": 21.168886184692383, "learning_rate": 2.1630000000000003e-06, "loss": 2.7953, "step": 4326 }, { "epoch": 1.3564263322884011, "grad_norm": 23.252321243286133, "learning_rate": 2.1635e-06, "loss": 2.3939, "step": 4327 }, { "epoch": 1.3567398119122256, "grad_norm": 24.73003578186035, "learning_rate": 2.1640000000000004e-06, "loss": 2.3674, "step": 4328 }, { "epoch": 1.35705329153605, "grad_norm": 40.02139663696289, "learning_rate": 2.1645e-06, "loss": 2.5032, "step": 4329 }, { "epoch": 1.3573667711598745, "grad_norm": 19.943395614624023, "learning_rate": 2.165e-06, "loss": 2.5056, "step": 4330 }, { "epoch": 1.3576802507836991, "grad_norm": 18.23369789123535, "learning_rate": 2.1655000000000003e-06, "loss": 2.3793, "step": 4331 }, { "epoch": 1.3579937304075236, "grad_norm": 222.7714385986328, "learning_rate": 2.166e-06, "loss": 2.5105, "step": 4332 }, { "epoch": 1.358307210031348, "grad_norm": 18.434926986694336, "learning_rate": 2.1665000000000003e-06, "loss": 2.3635, "step": 4333 }, { "epoch": 1.3586206896551725, "grad_norm": 48.5857048034668, "learning_rate": 2.167e-06, "loss": 2.5115, "step": 4334 }, { "epoch": 1.358934169278997, "grad_norm": 24.345428466796875, "learning_rate": 2.1675e-06, "loss": 3.1558, "step": 4335 }, { "epoch": 1.3592476489028213, "grad_norm": 24.408594131469727, "learning_rate": 2.1680000000000002e-06, "loss": 2.7029, "step": 4336 }, { "epoch": 1.3595611285266458, "grad_norm": 52.189109802246094, "learning_rate": 2.1685e-06, "loss": 2.9846, "step": 4337 }, { "epoch": 1.3598746081504702, "grad_norm": 58.49778366088867, "learning_rate": 2.1690000000000003e-06, "loss": 2.9688, "step": 4338 }, { "epoch": 1.3601880877742947, "grad_norm": 49.80043029785156, "learning_rate": 2.1695e-06, "loss": 2.7229, "step": 4339 }, { "epoch": 1.3605015673981191, "grad_norm": 29.659666061401367, "learning_rate": 2.17e-06, "loss": 3.1569, "step": 4340 }, { "epoch": 1.3608150470219436, "grad_norm": 16.015514373779297, "learning_rate": 2.1705000000000002e-06, "loss": 2.3724, "step": 4341 }, { "epoch": 1.361128526645768, "grad_norm": 26.102127075195312, "learning_rate": 2.171e-06, "loss": 2.4503, "step": 4342 }, { "epoch": 1.3614420062695924, "grad_norm": 17.50678062438965, "learning_rate": 2.1715000000000003e-06, "loss": 2.0142, "step": 4343 }, { "epoch": 1.3617554858934169, "grad_norm": 19.81712532043457, "learning_rate": 2.172e-06, "loss": 2.5277, "step": 4344 }, { "epoch": 1.3620689655172413, "grad_norm": 40.197269439697266, "learning_rate": 2.1725000000000004e-06, "loss": 2.5226, "step": 4345 }, { "epoch": 1.3623824451410658, "grad_norm": 83.39940643310547, "learning_rate": 2.173e-06, "loss": 3.6011, "step": 4346 }, { "epoch": 1.3626959247648902, "grad_norm": 27.205394744873047, "learning_rate": 2.1735e-06, "loss": 2.4416, "step": 4347 }, { "epoch": 1.3630094043887147, "grad_norm": 80.09081268310547, "learning_rate": 2.1740000000000003e-06, "loss": 2.7729, "step": 4348 }, { "epoch": 1.3633228840125393, "grad_norm": 11.565630912780762, "learning_rate": 2.1745e-06, "loss": 2.2525, "step": 4349 }, { "epoch": 1.3636363636363638, "grad_norm": 18.49755859375, "learning_rate": 2.1750000000000004e-06, "loss": 2.3291, "step": 4350 }, { "epoch": 1.3639498432601882, "grad_norm": 27.100292205810547, "learning_rate": 2.1755e-06, "loss": 2.4243, "step": 4351 }, { "epoch": 1.3642633228840126, "grad_norm": 18.38869285583496, "learning_rate": 2.176e-06, "loss": 2.7035, "step": 4352 }, { "epoch": 1.364576802507837, "grad_norm": 37.79602813720703, "learning_rate": 2.1765000000000003e-06, "loss": 2.428, "step": 4353 }, { "epoch": 1.3648902821316615, "grad_norm": 12.350529670715332, "learning_rate": 2.177e-06, "loss": 2.4816, "step": 4354 }, { "epoch": 1.365203761755486, "grad_norm": 33.481956481933594, "learning_rate": 2.1775000000000003e-06, "loss": 2.5288, "step": 4355 }, { "epoch": 1.3655172413793104, "grad_norm": 54.41046905517578, "learning_rate": 2.178e-06, "loss": 2.6805, "step": 4356 }, { "epoch": 1.3658307210031349, "grad_norm": 32.40873336791992, "learning_rate": 2.1785e-06, "loss": 2.4722, "step": 4357 }, { "epoch": 1.3661442006269593, "grad_norm": 13.064126014709473, "learning_rate": 2.1790000000000002e-06, "loss": 2.1733, "step": 4358 }, { "epoch": 1.3664576802507837, "grad_norm": 26.094358444213867, "learning_rate": 2.1795e-06, "loss": 2.9938, "step": 4359 }, { "epoch": 1.3667711598746082, "grad_norm": 15.258416175842285, "learning_rate": 2.1800000000000003e-06, "loss": 2.3782, "step": 4360 }, { "epoch": 1.3670846394984326, "grad_norm": 21.28451919555664, "learning_rate": 2.1805e-06, "loss": 2.8561, "step": 4361 }, { "epoch": 1.367398119122257, "grad_norm": 32.01693344116211, "learning_rate": 2.181e-06, "loss": 2.6646, "step": 4362 }, { "epoch": 1.3677115987460815, "grad_norm": 22.742752075195312, "learning_rate": 2.1815e-06, "loss": 2.3205, "step": 4363 }, { "epoch": 1.368025078369906, "grad_norm": 17.927749633789062, "learning_rate": 2.182e-06, "loss": 2.4711, "step": 4364 }, { "epoch": 1.3683385579937304, "grad_norm": 11.641703605651855, "learning_rate": 2.1825000000000003e-06, "loss": 2.9432, "step": 4365 }, { "epoch": 1.3686520376175548, "grad_norm": 19.79996109008789, "learning_rate": 2.183e-06, "loss": 2.511, "step": 4366 }, { "epoch": 1.3689655172413793, "grad_norm": 15.667484283447266, "learning_rate": 2.1835e-06, "loss": 2.2901, "step": 4367 }, { "epoch": 1.3692789968652037, "grad_norm": 20.27992820739746, "learning_rate": 2.184e-06, "loss": 2.6813, "step": 4368 }, { "epoch": 1.3695924764890282, "grad_norm": 13.770139694213867, "learning_rate": 2.1845000000000004e-06, "loss": 2.275, "step": 4369 }, { "epoch": 1.3699059561128526, "grad_norm": 50.38872146606445, "learning_rate": 2.1850000000000003e-06, "loss": 3.0291, "step": 4370 }, { "epoch": 1.370219435736677, "grad_norm": 23.927005767822266, "learning_rate": 2.1855e-06, "loss": 2.3998, "step": 4371 }, { "epoch": 1.3705329153605015, "grad_norm": 20.26904296875, "learning_rate": 2.186e-06, "loss": 2.5815, "step": 4372 }, { "epoch": 1.370846394984326, "grad_norm": 16.98995018005371, "learning_rate": 2.1865e-06, "loss": 2.415, "step": 4373 }, { "epoch": 1.3711598746081504, "grad_norm": 52.3552131652832, "learning_rate": 2.1870000000000004e-06, "loss": 2.8687, "step": 4374 }, { "epoch": 1.3714733542319748, "grad_norm": 40.32173538208008, "learning_rate": 2.1875000000000002e-06, "loss": 2.2187, "step": 4375 }, { "epoch": 1.3717868338557992, "grad_norm": 25.02348518371582, "learning_rate": 2.188e-06, "loss": 2.166, "step": 4376 }, { "epoch": 1.3721003134796237, "grad_norm": 25.96660614013672, "learning_rate": 2.1885e-06, "loss": 2.5841, "step": 4377 }, { "epoch": 1.3724137931034484, "grad_norm": 37.6526985168457, "learning_rate": 2.189e-06, "loss": 2.7103, "step": 4378 }, { "epoch": 1.3727272727272728, "grad_norm": 21.919849395751953, "learning_rate": 2.1895000000000004e-06, "loss": 2.9718, "step": 4379 }, { "epoch": 1.3730407523510972, "grad_norm": 29.23079490661621, "learning_rate": 2.19e-06, "loss": 2.2765, "step": 4380 }, { "epoch": 1.3733542319749217, "grad_norm": 37.13560104370117, "learning_rate": 2.1905e-06, "loss": 2.5976, "step": 4381 }, { "epoch": 1.3736677115987461, "grad_norm": 18.35036277770996, "learning_rate": 2.191e-06, "loss": 2.4564, "step": 4382 }, { "epoch": 1.3739811912225706, "grad_norm": 25.74932861328125, "learning_rate": 2.1915e-06, "loss": 2.43, "step": 4383 }, { "epoch": 1.374294670846395, "grad_norm": 22.727750778198242, "learning_rate": 2.1920000000000004e-06, "loss": 1.9894, "step": 4384 }, { "epoch": 1.3746081504702194, "grad_norm": 32.51811599731445, "learning_rate": 2.1925e-06, "loss": 2.3064, "step": 4385 }, { "epoch": 1.374921630094044, "grad_norm": 60.5367317199707, "learning_rate": 2.193e-06, "loss": 2.453, "step": 4386 }, { "epoch": 1.3752351097178683, "grad_norm": 16.200775146484375, "learning_rate": 2.1935000000000003e-06, "loss": 2.3594, "step": 4387 }, { "epoch": 1.3755485893416928, "grad_norm": 27.528079986572266, "learning_rate": 2.194e-06, "loss": 3.6015, "step": 4388 }, { "epoch": 1.3758620689655172, "grad_norm": 32.33339309692383, "learning_rate": 2.1945000000000003e-06, "loss": 2.6115, "step": 4389 }, { "epoch": 1.3761755485893417, "grad_norm": 52.41084289550781, "learning_rate": 2.195e-06, "loss": 3.0907, "step": 4390 }, { "epoch": 1.376489028213166, "grad_norm": 15.238964080810547, "learning_rate": 2.1955e-06, "loss": 2.2765, "step": 4391 }, { "epoch": 1.3768025078369905, "grad_norm": 42.972206115722656, "learning_rate": 2.1960000000000002e-06, "loss": 3.603, "step": 4392 }, { "epoch": 1.377115987460815, "grad_norm": 13.6290922164917, "learning_rate": 2.1965000000000005e-06, "loss": 2.302, "step": 4393 }, { "epoch": 1.3774294670846394, "grad_norm": 11.237236022949219, "learning_rate": 2.1970000000000003e-06, "loss": 2.3066, "step": 4394 }, { "epoch": 1.3777429467084639, "grad_norm": 18.652162551879883, "learning_rate": 2.1975e-06, "loss": 2.8458, "step": 4395 }, { "epoch": 1.3780564263322885, "grad_norm": 20.605178833007812, "learning_rate": 2.198e-06, "loss": 2.1595, "step": 4396 }, { "epoch": 1.378369905956113, "grad_norm": 24.695146560668945, "learning_rate": 2.1985e-06, "loss": 2.3621, "step": 4397 }, { "epoch": 1.3786833855799374, "grad_norm": 19.619272232055664, "learning_rate": 2.1990000000000005e-06, "loss": 2.431, "step": 4398 }, { "epoch": 1.3789968652037619, "grad_norm": 22.981626510620117, "learning_rate": 2.1995000000000003e-06, "loss": 2.3365, "step": 4399 }, { "epoch": 1.3793103448275863, "grad_norm": 19.587995529174805, "learning_rate": 2.2e-06, "loss": 2.4834, "step": 4400 }, { "epoch": 1.3796238244514107, "grad_norm": 25.917057037353516, "learning_rate": 2.2005e-06, "loss": 2.5633, "step": 4401 }, { "epoch": 1.3799373040752352, "grad_norm": 46.26815414428711, "learning_rate": 2.201e-06, "loss": 2.897, "step": 4402 }, { "epoch": 1.3802507836990596, "grad_norm": 91.74821472167969, "learning_rate": 2.2015000000000004e-06, "loss": 2.8106, "step": 4403 }, { "epoch": 1.380564263322884, "grad_norm": 26.963945388793945, "learning_rate": 2.2020000000000003e-06, "loss": 2.4099, "step": 4404 }, { "epoch": 1.3808777429467085, "grad_norm": 29.442838668823242, "learning_rate": 2.2025e-06, "loss": 2.529, "step": 4405 }, { "epoch": 1.381191222570533, "grad_norm": 23.21820831298828, "learning_rate": 2.2030000000000003e-06, "loss": 2.5125, "step": 4406 }, { "epoch": 1.3815047021943574, "grad_norm": 38.09263610839844, "learning_rate": 2.2035e-06, "loss": 2.4609, "step": 4407 }, { "epoch": 1.3818181818181818, "grad_norm": 17.69852638244629, "learning_rate": 2.2040000000000004e-06, "loss": 3.0537, "step": 4408 }, { "epoch": 1.3821316614420063, "grad_norm": 12.266457557678223, "learning_rate": 2.2045000000000002e-06, "loss": 2.3415, "step": 4409 }, { "epoch": 1.3824451410658307, "grad_norm": 19.07096290588379, "learning_rate": 2.205e-06, "loss": 2.213, "step": 4410 }, { "epoch": 1.3827586206896552, "grad_norm": 19.074338912963867, "learning_rate": 2.2055000000000003e-06, "loss": 2.6777, "step": 4411 }, { "epoch": 1.3830721003134796, "grad_norm": 36.32673263549805, "learning_rate": 2.206e-06, "loss": 2.8512, "step": 4412 }, { "epoch": 1.383385579937304, "grad_norm": 11.210381507873535, "learning_rate": 2.2065000000000004e-06, "loss": 2.3242, "step": 4413 }, { "epoch": 1.3836990595611285, "grad_norm": 21.469053268432617, "learning_rate": 2.2070000000000002e-06, "loss": 2.3212, "step": 4414 }, { "epoch": 1.384012539184953, "grad_norm": 31.558183670043945, "learning_rate": 2.2075e-06, "loss": 2.4386, "step": 4415 }, { "epoch": 1.3843260188087774, "grad_norm": 13.501238822937012, "learning_rate": 2.2080000000000003e-06, "loss": 2.2854, "step": 4416 }, { "epoch": 1.3846394984326018, "grad_norm": 22.445682525634766, "learning_rate": 2.2085e-06, "loss": 2.6778, "step": 4417 }, { "epoch": 1.3849529780564263, "grad_norm": 28.20890235900879, "learning_rate": 2.2090000000000004e-06, "loss": 2.5873, "step": 4418 }, { "epoch": 1.3852664576802507, "grad_norm": 19.3082275390625, "learning_rate": 2.2095e-06, "loss": 2.2456, "step": 4419 }, { "epoch": 1.3855799373040751, "grad_norm": 19.401933670043945, "learning_rate": 2.21e-06, "loss": 2.4951, "step": 4420 }, { "epoch": 1.3858934169278996, "grad_norm": 38.82876968383789, "learning_rate": 2.2105000000000003e-06, "loss": 2.6545, "step": 4421 }, { "epoch": 1.386206896551724, "grad_norm": 28.123075485229492, "learning_rate": 2.211e-06, "loss": 2.5914, "step": 4422 }, { "epoch": 1.3865203761755485, "grad_norm": 29.51925277709961, "learning_rate": 2.2115000000000003e-06, "loss": 2.3974, "step": 4423 }, { "epoch": 1.386833855799373, "grad_norm": 19.382112503051758, "learning_rate": 2.212e-06, "loss": 2.4665, "step": 4424 }, { "epoch": 1.3871473354231976, "grad_norm": 43.75849533081055, "learning_rate": 2.2125e-06, "loss": 2.5744, "step": 4425 }, { "epoch": 1.387460815047022, "grad_norm": 58.4948616027832, "learning_rate": 2.2130000000000002e-06, "loss": 2.1157, "step": 4426 }, { "epoch": 1.3877742946708465, "grad_norm": 46.00054168701172, "learning_rate": 2.2135e-06, "loss": 2.3246, "step": 4427 }, { "epoch": 1.388087774294671, "grad_norm": 26.322551727294922, "learning_rate": 2.2140000000000003e-06, "loss": 2.5217, "step": 4428 }, { "epoch": 1.3884012539184953, "grad_norm": 163.227783203125, "learning_rate": 2.2145e-06, "loss": 3.6124, "step": 4429 }, { "epoch": 1.3887147335423198, "grad_norm": 47.88997268676758, "learning_rate": 2.2150000000000004e-06, "loss": 2.552, "step": 4430 }, { "epoch": 1.3890282131661442, "grad_norm": 26.8756103515625, "learning_rate": 2.2155000000000002e-06, "loss": 2.6725, "step": 4431 }, { "epoch": 1.3893416927899687, "grad_norm": 17.241147994995117, "learning_rate": 2.216e-06, "loss": 2.2438, "step": 4432 }, { "epoch": 1.389655172413793, "grad_norm": 40.995697021484375, "learning_rate": 2.2165000000000003e-06, "loss": 2.6698, "step": 4433 }, { "epoch": 1.3899686520376175, "grad_norm": 17.262331008911133, "learning_rate": 2.217e-06, "loss": 2.5749, "step": 4434 }, { "epoch": 1.390282131661442, "grad_norm": 18.621685028076172, "learning_rate": 2.2175000000000004e-06, "loss": 2.4253, "step": 4435 }, { "epoch": 1.3905956112852664, "grad_norm": 30.165283203125, "learning_rate": 2.218e-06, "loss": 2.3773, "step": 4436 }, { "epoch": 1.3909090909090909, "grad_norm": 17.12919807434082, "learning_rate": 2.2185e-06, "loss": 2.3988, "step": 4437 }, { "epoch": 1.3912225705329153, "grad_norm": 63.46233367919922, "learning_rate": 2.219e-06, "loss": 2.7529, "step": 4438 }, { "epoch": 1.3915360501567398, "grad_norm": 30.933561325073242, "learning_rate": 2.2195e-06, "loss": 2.4569, "step": 4439 }, { "epoch": 1.3918495297805642, "grad_norm": 12.137598991394043, "learning_rate": 2.2200000000000003e-06, "loss": 2.2861, "step": 4440 }, { "epoch": 1.3921630094043886, "grad_norm": 38.616188049316406, "learning_rate": 2.2205e-06, "loss": 2.258, "step": 4441 }, { "epoch": 1.3924764890282133, "grad_norm": 23.51974105834961, "learning_rate": 2.221e-06, "loss": 2.4517, "step": 4442 }, { "epoch": 1.3927899686520377, "grad_norm": 15.76685619354248, "learning_rate": 2.2215000000000002e-06, "loss": 2.4993, "step": 4443 }, { "epoch": 1.3931034482758622, "grad_norm": 18.977575302124023, "learning_rate": 2.222e-06, "loss": 2.4115, "step": 4444 }, { "epoch": 1.3934169278996866, "grad_norm": 32.84709548950195, "learning_rate": 2.2225000000000003e-06, "loss": 2.6831, "step": 4445 }, { "epoch": 1.393730407523511, "grad_norm": 41.960670471191406, "learning_rate": 2.223e-06, "loss": 2.5226, "step": 4446 }, { "epoch": 1.3940438871473355, "grad_norm": 14.326679229736328, "learning_rate": 2.2235e-06, "loss": 2.8367, "step": 4447 }, { "epoch": 1.39435736677116, "grad_norm": 17.317045211791992, "learning_rate": 2.2240000000000002e-06, "loss": 2.5747, "step": 4448 }, { "epoch": 1.3946708463949844, "grad_norm": 27.15635108947754, "learning_rate": 2.2245000000000005e-06, "loss": 2.6648, "step": 4449 }, { "epoch": 1.3949843260188088, "grad_norm": 22.993974685668945, "learning_rate": 2.2250000000000003e-06, "loss": 2.3917, "step": 4450 }, { "epoch": 1.3952978056426333, "grad_norm": 15.782670021057129, "learning_rate": 2.2255e-06, "loss": 2.229, "step": 4451 }, { "epoch": 1.3956112852664577, "grad_norm": 20.578027725219727, "learning_rate": 2.226e-06, "loss": 2.5032, "step": 4452 }, { "epoch": 1.3959247648902822, "grad_norm": 26.298973083496094, "learning_rate": 2.2265e-06, "loss": 2.4797, "step": 4453 }, { "epoch": 1.3962382445141066, "grad_norm": 17.40005111694336, "learning_rate": 2.2270000000000004e-06, "loss": 2.354, "step": 4454 }, { "epoch": 1.396551724137931, "grad_norm": 22.421411514282227, "learning_rate": 2.2275000000000003e-06, "loss": 2.2626, "step": 4455 }, { "epoch": 1.3968652037617555, "grad_norm": 35.079627990722656, "learning_rate": 2.228e-06, "loss": 2.4148, "step": 4456 }, { "epoch": 1.39717868338558, "grad_norm": 22.324115753173828, "learning_rate": 2.2285e-06, "loss": 2.5046, "step": 4457 }, { "epoch": 1.3974921630094044, "grad_norm": 47.19550323486328, "learning_rate": 2.229e-06, "loss": 2.8975, "step": 4458 }, { "epoch": 1.3978056426332288, "grad_norm": 13.401718139648438, "learning_rate": 2.2295000000000004e-06, "loss": 2.3232, "step": 4459 }, { "epoch": 1.3981191222570533, "grad_norm": 32.29413986206055, "learning_rate": 2.2300000000000002e-06, "loss": 2.6065, "step": 4460 }, { "epoch": 1.3984326018808777, "grad_norm": 28.262371063232422, "learning_rate": 2.2305e-06, "loss": 2.1882, "step": 4461 }, { "epoch": 1.3987460815047021, "grad_norm": 17.66546058654785, "learning_rate": 2.231e-06, "loss": 2.3207, "step": 4462 }, { "epoch": 1.3990595611285266, "grad_norm": 32.400699615478516, "learning_rate": 2.2315e-06, "loss": 2.1721, "step": 4463 }, { "epoch": 1.399373040752351, "grad_norm": 18.58637237548828, "learning_rate": 2.2320000000000004e-06, "loss": 2.2821, "step": 4464 }, { "epoch": 1.3996865203761755, "grad_norm": 19.35753059387207, "learning_rate": 2.2325000000000002e-06, "loss": 3.3655, "step": 4465 }, { "epoch": 1.4, "grad_norm": 28.945085525512695, "learning_rate": 2.233e-06, "loss": 3.0592, "step": 4466 }, { "epoch": 1.4003134796238244, "grad_norm": 10.396805763244629, "learning_rate": 2.2335000000000003e-06, "loss": 2.4453, "step": 4467 }, { "epoch": 1.4006269592476488, "grad_norm": 49.911705017089844, "learning_rate": 2.234e-06, "loss": 2.4485, "step": 4468 }, { "epoch": 1.4009404388714732, "grad_norm": 79.0918960571289, "learning_rate": 2.2345000000000004e-06, "loss": 1.9819, "step": 4469 }, { "epoch": 1.4012539184952977, "grad_norm": 22.501100540161133, "learning_rate": 2.235e-06, "loss": 2.4589, "step": 4470 }, { "epoch": 1.4015673981191221, "grad_norm": 36.03977966308594, "learning_rate": 2.2355e-06, "loss": 2.4418, "step": 4471 }, { "epoch": 1.4018808777429468, "grad_norm": 16.972667694091797, "learning_rate": 2.2360000000000003e-06, "loss": 2.3223, "step": 4472 }, { "epoch": 1.4021943573667712, "grad_norm": 14.537110328674316, "learning_rate": 2.2365e-06, "loss": 2.3733, "step": 4473 }, { "epoch": 1.4025078369905957, "grad_norm": 53.65971755981445, "learning_rate": 2.2370000000000004e-06, "loss": 2.5804, "step": 4474 }, { "epoch": 1.40282131661442, "grad_norm": 19.41697883605957, "learning_rate": 2.2375e-06, "loss": 2.681, "step": 4475 }, { "epoch": 1.4031347962382446, "grad_norm": 15.1221923828125, "learning_rate": 2.238e-06, "loss": 2.1564, "step": 4476 }, { "epoch": 1.403448275862069, "grad_norm": 19.52741241455078, "learning_rate": 2.2385000000000003e-06, "loss": 2.1644, "step": 4477 }, { "epoch": 1.4037617554858934, "grad_norm": 48.018367767333984, "learning_rate": 2.239e-06, "loss": 2.6797, "step": 4478 }, { "epoch": 1.4040752351097179, "grad_norm": 41.445098876953125, "learning_rate": 2.2395000000000003e-06, "loss": 2.8707, "step": 4479 }, { "epoch": 1.4043887147335423, "grad_norm": 53.0460319519043, "learning_rate": 2.24e-06, "loss": 3.1522, "step": 4480 }, { "epoch": 1.4047021943573668, "grad_norm": 17.2395076751709, "learning_rate": 2.2405e-06, "loss": 2.2166, "step": 4481 }, { "epoch": 1.4050156739811912, "grad_norm": 22.412208557128906, "learning_rate": 2.2410000000000002e-06, "loss": 2.4682, "step": 4482 }, { "epoch": 1.4053291536050156, "grad_norm": 54.72515106201172, "learning_rate": 2.2415e-06, "loss": 2.521, "step": 4483 }, { "epoch": 1.40564263322884, "grad_norm": 26.304935455322266, "learning_rate": 2.2420000000000003e-06, "loss": 2.0913, "step": 4484 }, { "epoch": 1.4059561128526645, "grad_norm": 16.559471130371094, "learning_rate": 2.2425e-06, "loss": 2.1316, "step": 4485 }, { "epoch": 1.406269592476489, "grad_norm": 24.916353225708008, "learning_rate": 2.2430000000000004e-06, "loss": 2.594, "step": 4486 }, { "epoch": 1.4065830721003134, "grad_norm": 31.964988708496094, "learning_rate": 2.2435e-06, "loss": 2.5372, "step": 4487 }, { "epoch": 1.4068965517241379, "grad_norm": 60.68818283081055, "learning_rate": 2.244e-06, "loss": 2.9889, "step": 4488 }, { "epoch": 1.4072100313479625, "grad_norm": 16.230152130126953, "learning_rate": 2.2445000000000003e-06, "loss": 2.4079, "step": 4489 }, { "epoch": 1.407523510971787, "grad_norm": 19.387493133544922, "learning_rate": 2.245e-06, "loss": 2.5174, "step": 4490 }, { "epoch": 1.4078369905956114, "grad_norm": 14.607769012451172, "learning_rate": 2.2455000000000004e-06, "loss": 2.1988, "step": 4491 }, { "epoch": 1.4081504702194358, "grad_norm": 23.40793228149414, "learning_rate": 2.246e-06, "loss": 2.7084, "step": 4492 }, { "epoch": 1.4084639498432603, "grad_norm": 75.68866729736328, "learning_rate": 2.2465e-06, "loss": 2.672, "step": 4493 }, { "epoch": 1.4087774294670847, "grad_norm": 66.80253601074219, "learning_rate": 2.2470000000000003e-06, "loss": 2.8853, "step": 4494 }, { "epoch": 1.4090909090909092, "grad_norm": 35.71215057373047, "learning_rate": 2.2475e-06, "loss": 2.5797, "step": 4495 }, { "epoch": 1.4094043887147336, "grad_norm": 119.49893188476562, "learning_rate": 2.2480000000000003e-06, "loss": 2.1623, "step": 4496 }, { "epoch": 1.409717868338558, "grad_norm": 20.179807662963867, "learning_rate": 2.2485e-06, "loss": 2.3686, "step": 4497 }, { "epoch": 1.4100313479623825, "grad_norm": 33.06385040283203, "learning_rate": 2.249e-06, "loss": 3.224, "step": 4498 }, { "epoch": 1.410344827586207, "grad_norm": 28.012863159179688, "learning_rate": 2.2495000000000002e-06, "loss": 2.5953, "step": 4499 }, { "epoch": 1.4106583072100314, "grad_norm": 23.963817596435547, "learning_rate": 2.25e-06, "loss": 2.0826, "step": 4500 }, { "epoch": 1.4109717868338558, "grad_norm": 18.82379913330078, "learning_rate": 2.2505000000000003e-06, "loss": 2.5676, "step": 4501 }, { "epoch": 1.4112852664576803, "grad_norm": 29.09153175354004, "learning_rate": 2.251e-06, "loss": 2.5424, "step": 4502 }, { "epoch": 1.4115987460815047, "grad_norm": 24.60426902770996, "learning_rate": 2.2515e-06, "loss": 2.4428, "step": 4503 }, { "epoch": 1.4119122257053291, "grad_norm": 12.828309059143066, "learning_rate": 2.252e-06, "loss": 2.4397, "step": 4504 }, { "epoch": 1.4122257053291536, "grad_norm": 10.570173263549805, "learning_rate": 2.2525e-06, "loss": 2.289, "step": 4505 }, { "epoch": 1.412539184952978, "grad_norm": 11.049269676208496, "learning_rate": 2.2530000000000003e-06, "loss": 2.133, "step": 4506 }, { "epoch": 1.4128526645768025, "grad_norm": 77.35066986083984, "learning_rate": 2.2535e-06, "loss": 2.4723, "step": 4507 }, { "epoch": 1.413166144200627, "grad_norm": 56.341064453125, "learning_rate": 2.254e-06, "loss": 3.4043, "step": 4508 }, { "epoch": 1.4134796238244514, "grad_norm": 18.001556396484375, "learning_rate": 2.2545e-06, "loss": 2.5181, "step": 4509 }, { "epoch": 1.4137931034482758, "grad_norm": 11.483774185180664, "learning_rate": 2.2550000000000004e-06, "loss": 2.4675, "step": 4510 }, { "epoch": 1.4141065830721002, "grad_norm": 92.04100799560547, "learning_rate": 2.2555000000000003e-06, "loss": 3.0268, "step": 4511 }, { "epoch": 1.4144200626959247, "grad_norm": 25.606027603149414, "learning_rate": 2.256e-06, "loss": 2.5514, "step": 4512 }, { "epoch": 1.4147335423197491, "grad_norm": 25.288551330566406, "learning_rate": 2.2565e-06, "loss": 2.4726, "step": 4513 }, { "epoch": 1.4150470219435736, "grad_norm": 30.17818260192871, "learning_rate": 2.257e-06, "loss": 2.3311, "step": 4514 }, { "epoch": 1.415360501567398, "grad_norm": 42.18940734863281, "learning_rate": 2.2575000000000004e-06, "loss": 2.6688, "step": 4515 }, { "epoch": 1.4156739811912225, "grad_norm": 18.35502815246582, "learning_rate": 2.2580000000000002e-06, "loss": 2.3161, "step": 4516 }, { "epoch": 1.415987460815047, "grad_norm": 20.00831413269043, "learning_rate": 2.2585e-06, "loss": 2.3968, "step": 4517 }, { "epoch": 1.4163009404388713, "grad_norm": 21.05620574951172, "learning_rate": 2.259e-06, "loss": 2.4714, "step": 4518 }, { "epoch": 1.416614420062696, "grad_norm": 21.366777420043945, "learning_rate": 2.2595e-06, "loss": 2.7164, "step": 4519 }, { "epoch": 1.4169278996865204, "grad_norm": 21.714614868164062, "learning_rate": 2.2600000000000004e-06, "loss": 2.5011, "step": 4520 }, { "epoch": 1.4172413793103449, "grad_norm": 29.17438316345215, "learning_rate": 2.2605e-06, "loss": 2.6016, "step": 4521 }, { "epoch": 1.4175548589341693, "grad_norm": 21.83228874206543, "learning_rate": 2.261e-06, "loss": 2.3649, "step": 4522 }, { "epoch": 1.4178683385579938, "grad_norm": 36.300941467285156, "learning_rate": 2.2615000000000003e-06, "loss": 3.4691, "step": 4523 }, { "epoch": 1.4181818181818182, "grad_norm": 52.76793670654297, "learning_rate": 2.262e-06, "loss": 2.8731, "step": 4524 }, { "epoch": 1.4184952978056427, "grad_norm": 44.001461029052734, "learning_rate": 2.2625000000000004e-06, "loss": 2.4197, "step": 4525 }, { "epoch": 1.418808777429467, "grad_norm": 14.982115745544434, "learning_rate": 2.263e-06, "loss": 2.2569, "step": 4526 }, { "epoch": 1.4191222570532915, "grad_norm": 14.590595245361328, "learning_rate": 2.2635e-06, "loss": 2.3871, "step": 4527 }, { "epoch": 1.419435736677116, "grad_norm": 23.147096633911133, "learning_rate": 2.2640000000000003e-06, "loss": 2.2429, "step": 4528 }, { "epoch": 1.4197492163009404, "grad_norm": 25.500978469848633, "learning_rate": 2.2645000000000005e-06, "loss": 2.4662, "step": 4529 }, { "epoch": 1.4200626959247649, "grad_norm": 56.13792037963867, "learning_rate": 2.2650000000000003e-06, "loss": 3.5036, "step": 4530 }, { "epoch": 1.4203761755485893, "grad_norm": 27.729318618774414, "learning_rate": 2.2655e-06, "loss": 2.3546, "step": 4531 }, { "epoch": 1.4206896551724137, "grad_norm": 19.810440063476562, "learning_rate": 2.266e-06, "loss": 2.6285, "step": 4532 }, { "epoch": 1.4210031347962382, "grad_norm": 19.486492156982422, "learning_rate": 2.2665000000000002e-06, "loss": 2.3256, "step": 4533 }, { "epoch": 1.4213166144200626, "grad_norm": 32.0871696472168, "learning_rate": 2.2670000000000005e-06, "loss": 2.2786, "step": 4534 }, { "epoch": 1.421630094043887, "grad_norm": 10.243577003479004, "learning_rate": 2.2675000000000003e-06, "loss": 2.8419, "step": 4535 }, { "epoch": 1.4219435736677117, "grad_norm": 15.97329044342041, "learning_rate": 2.268e-06, "loss": 2.5527, "step": 4536 }, { "epoch": 1.4222570532915362, "grad_norm": 24.04090118408203, "learning_rate": 2.2685e-06, "loss": 2.4916, "step": 4537 }, { "epoch": 1.4225705329153606, "grad_norm": 33.413352966308594, "learning_rate": 2.2690000000000002e-06, "loss": 2.8125, "step": 4538 }, { "epoch": 1.422884012539185, "grad_norm": 42.97076416015625, "learning_rate": 2.2695000000000005e-06, "loss": 3.0929, "step": 4539 }, { "epoch": 1.4231974921630095, "grad_norm": 50.45700454711914, "learning_rate": 2.2700000000000003e-06, "loss": 2.6702, "step": 4540 }, { "epoch": 1.423510971786834, "grad_norm": 13.318603515625, "learning_rate": 2.2705e-06, "loss": 2.333, "step": 4541 }, { "epoch": 1.4238244514106584, "grad_norm": 30.421119689941406, "learning_rate": 2.271e-06, "loss": 2.4485, "step": 4542 }, { "epoch": 1.4241379310344828, "grad_norm": 31.187414169311523, "learning_rate": 2.2715e-06, "loss": 2.3791, "step": 4543 }, { "epoch": 1.4244514106583073, "grad_norm": 14.010133743286133, "learning_rate": 2.2720000000000004e-06, "loss": 2.2676, "step": 4544 }, { "epoch": 1.4247648902821317, "grad_norm": 18.06087875366211, "learning_rate": 2.2725000000000003e-06, "loss": 2.3697, "step": 4545 }, { "epoch": 1.4250783699059562, "grad_norm": 28.26776123046875, "learning_rate": 2.273e-06, "loss": 2.2602, "step": 4546 }, { "epoch": 1.4253918495297806, "grad_norm": 78.11388397216797, "learning_rate": 2.2735000000000003e-06, "loss": 3.5192, "step": 4547 }, { "epoch": 1.425705329153605, "grad_norm": 30.041149139404297, "learning_rate": 2.274e-06, "loss": 2.425, "step": 4548 }, { "epoch": 1.4260188087774295, "grad_norm": 35.09251403808594, "learning_rate": 2.2745000000000004e-06, "loss": 2.3432, "step": 4549 }, { "epoch": 1.426332288401254, "grad_norm": 20.941200256347656, "learning_rate": 2.2750000000000002e-06, "loss": 2.6254, "step": 4550 }, { "epoch": 1.4266457680250784, "grad_norm": 25.721513748168945, "learning_rate": 2.2755e-06, "loss": 3.0662, "step": 4551 }, { "epoch": 1.4269592476489028, "grad_norm": 19.826330184936523, "learning_rate": 2.2760000000000003e-06, "loss": 2.2957, "step": 4552 }, { "epoch": 1.4272727272727272, "grad_norm": 27.101993560791016, "learning_rate": 2.2765e-06, "loss": 2.7257, "step": 4553 }, { "epoch": 1.4275862068965517, "grad_norm": 30.94487190246582, "learning_rate": 2.2770000000000004e-06, "loss": 2.3578, "step": 4554 }, { "epoch": 1.4278996865203761, "grad_norm": 20.627241134643555, "learning_rate": 2.2775000000000002e-06, "loss": 2.4182, "step": 4555 }, { "epoch": 1.4282131661442006, "grad_norm": 41.094146728515625, "learning_rate": 2.278e-06, "loss": 2.4636, "step": 4556 }, { "epoch": 1.428526645768025, "grad_norm": 22.273983001708984, "learning_rate": 2.2785000000000003e-06, "loss": 2.4823, "step": 4557 }, { "epoch": 1.4288401253918495, "grad_norm": 15.536066055297852, "learning_rate": 2.279e-06, "loss": 2.3666, "step": 4558 }, { "epoch": 1.429153605015674, "grad_norm": 19.999664306640625, "learning_rate": 2.2795000000000004e-06, "loss": 2.3434, "step": 4559 }, { "epoch": 1.4294670846394983, "grad_norm": 29.738088607788086, "learning_rate": 2.28e-06, "loss": 3.1658, "step": 4560 }, { "epoch": 1.4297805642633228, "grad_norm": 28.52236557006836, "learning_rate": 2.2805e-06, "loss": 2.5009, "step": 4561 }, { "epoch": 1.4300940438871472, "grad_norm": 58.76250457763672, "learning_rate": 2.2810000000000003e-06, "loss": 2.8935, "step": 4562 }, { "epoch": 1.4304075235109717, "grad_norm": 13.827445030212402, "learning_rate": 2.2815e-06, "loss": 2.5779, "step": 4563 }, { "epoch": 1.430721003134796, "grad_norm": 24.913976669311523, "learning_rate": 2.282e-06, "loss": 2.5309, "step": 4564 }, { "epoch": 1.4310344827586206, "grad_norm": 14.8967924118042, "learning_rate": 2.2825e-06, "loss": 2.3282, "step": 4565 }, { "epoch": 1.4313479623824452, "grad_norm": 27.908395767211914, "learning_rate": 2.2830000000000004e-06, "loss": 2.1359, "step": 4566 }, { "epoch": 1.4316614420062697, "grad_norm": 44.49998474121094, "learning_rate": 2.2835000000000002e-06, "loss": 2.3364, "step": 4567 }, { "epoch": 1.431974921630094, "grad_norm": 78.90461730957031, "learning_rate": 2.284e-06, "loss": 2.8442, "step": 4568 }, { "epoch": 1.4322884012539185, "grad_norm": 61.37258529663086, "learning_rate": 2.2845e-06, "loss": 2.2926, "step": 4569 }, { "epoch": 1.432601880877743, "grad_norm": 17.646041870117188, "learning_rate": 2.285e-06, "loss": 2.453, "step": 4570 }, { "epoch": 1.4329153605015674, "grad_norm": 26.07595443725586, "learning_rate": 2.2855000000000004e-06, "loss": 2.5075, "step": 4571 }, { "epoch": 1.4332288401253919, "grad_norm": 30.174352645874023, "learning_rate": 2.2860000000000002e-06, "loss": 2.6651, "step": 4572 }, { "epoch": 1.4335423197492163, "grad_norm": 12.280838012695312, "learning_rate": 2.2865e-06, "loss": 2.3135, "step": 4573 }, { "epoch": 1.4338557993730408, "grad_norm": 17.326416015625, "learning_rate": 2.287e-06, "loss": 2.3963, "step": 4574 }, { "epoch": 1.4341692789968652, "grad_norm": 48.17243957519531, "learning_rate": 2.2875e-06, "loss": 2.5341, "step": 4575 }, { "epoch": 1.4344827586206896, "grad_norm": 75.73432922363281, "learning_rate": 2.2880000000000004e-06, "loss": 2.8398, "step": 4576 }, { "epoch": 1.434796238244514, "grad_norm": 280.7005920410156, "learning_rate": 2.2885e-06, "loss": 3.0092, "step": 4577 }, { "epoch": 1.4351097178683385, "grad_norm": 15.257850646972656, "learning_rate": 2.289e-06, "loss": 2.2889, "step": 4578 }, { "epoch": 1.435423197492163, "grad_norm": 31.832490921020508, "learning_rate": 2.2895e-06, "loss": 2.4703, "step": 4579 }, { "epoch": 1.4357366771159874, "grad_norm": 24.959474563598633, "learning_rate": 2.29e-06, "loss": 1.9712, "step": 4580 }, { "epoch": 1.4360501567398118, "grad_norm": 30.74005126953125, "learning_rate": 2.2905000000000003e-06, "loss": 2.8657, "step": 4581 }, { "epoch": 1.4363636363636363, "grad_norm": 36.41959762573242, "learning_rate": 2.291e-06, "loss": 2.5452, "step": 4582 }, { "epoch": 1.436677115987461, "grad_norm": 28.88225746154785, "learning_rate": 2.2915e-06, "loss": 2.789, "step": 4583 }, { "epoch": 1.4369905956112854, "grad_norm": 22.670028686523438, "learning_rate": 2.2920000000000002e-06, "loss": 2.5401, "step": 4584 }, { "epoch": 1.4373040752351098, "grad_norm": 17.49390411376953, "learning_rate": 2.2925e-06, "loss": 2.2482, "step": 4585 }, { "epoch": 1.4376175548589343, "grad_norm": 43.223487854003906, "learning_rate": 2.2930000000000003e-06, "loss": 2.6491, "step": 4586 }, { "epoch": 1.4379310344827587, "grad_norm": 28.049163818359375, "learning_rate": 2.2935e-06, "loss": 2.4151, "step": 4587 }, { "epoch": 1.4382445141065832, "grad_norm": 23.431224822998047, "learning_rate": 2.294e-06, "loss": 2.3691, "step": 4588 }, { "epoch": 1.4385579937304076, "grad_norm": 12.071837425231934, "learning_rate": 2.2945000000000002e-06, "loss": 2.3541, "step": 4589 }, { "epoch": 1.438871473354232, "grad_norm": 28.91056251525879, "learning_rate": 2.2950000000000005e-06, "loss": 2.4894, "step": 4590 }, { "epoch": 1.4391849529780565, "grad_norm": 22.10399055480957, "learning_rate": 2.2955000000000003e-06, "loss": 2.3852, "step": 4591 }, { "epoch": 1.439498432601881, "grad_norm": 23.183076858520508, "learning_rate": 2.296e-06, "loss": 2.2229, "step": 4592 }, { "epoch": 1.4398119122257054, "grad_norm": 15.805237770080566, "learning_rate": 2.2965e-06, "loss": 2.3327, "step": 4593 }, { "epoch": 1.4401253918495298, "grad_norm": 24.32073211669922, "learning_rate": 2.297e-06, "loss": 2.4777, "step": 4594 }, { "epoch": 1.4404388714733543, "grad_norm": 29.6636905670166, "learning_rate": 2.2975000000000004e-06, "loss": 2.4994, "step": 4595 }, { "epoch": 1.4407523510971787, "grad_norm": 21.071327209472656, "learning_rate": 2.2980000000000003e-06, "loss": 2.5011, "step": 4596 }, { "epoch": 1.4410658307210031, "grad_norm": 30.68376922607422, "learning_rate": 2.2985e-06, "loss": 2.6823, "step": 4597 }, { "epoch": 1.4413793103448276, "grad_norm": 49.39833068847656, "learning_rate": 2.299e-06, "loss": 3.1115, "step": 4598 }, { "epoch": 1.441692789968652, "grad_norm": 13.447183609008789, "learning_rate": 2.2995e-06, "loss": 2.4562, "step": 4599 }, { "epoch": 1.4420062695924765, "grad_norm": 27.433025360107422, "learning_rate": 2.3000000000000004e-06, "loss": 2.4388, "step": 4600 }, { "epoch": 1.442319749216301, "grad_norm": 26.882158279418945, "learning_rate": 2.3005000000000003e-06, "loss": 2.3435, "step": 4601 }, { "epoch": 1.4426332288401253, "grad_norm": 24.261220932006836, "learning_rate": 2.301e-06, "loss": 2.383, "step": 4602 }, { "epoch": 1.4429467084639498, "grad_norm": 18.171157836914062, "learning_rate": 2.3015000000000003e-06, "loss": 2.3532, "step": 4603 }, { "epoch": 1.4432601880877742, "grad_norm": 27.184085845947266, "learning_rate": 2.302e-06, "loss": 2.3824, "step": 4604 }, { "epoch": 1.4435736677115987, "grad_norm": 25.665733337402344, "learning_rate": 2.3025000000000004e-06, "loss": 2.2047, "step": 4605 }, { "epoch": 1.4438871473354231, "grad_norm": 29.451372146606445, "learning_rate": 2.3030000000000002e-06, "loss": 2.244, "step": 4606 }, { "epoch": 1.4442006269592476, "grad_norm": 16.427772521972656, "learning_rate": 2.3035e-06, "loss": 2.5831, "step": 4607 }, { "epoch": 1.444514106583072, "grad_norm": 30.14790916442871, "learning_rate": 2.3040000000000003e-06, "loss": 2.8414, "step": 4608 }, { "epoch": 1.4448275862068964, "grad_norm": 16.03703498840332, "learning_rate": 2.3045e-06, "loss": 2.26, "step": 4609 }, { "epoch": 1.4451410658307209, "grad_norm": 126.84292602539062, "learning_rate": 2.3050000000000004e-06, "loss": 2.7126, "step": 4610 }, { "epoch": 1.4454545454545453, "grad_norm": 18.969295501708984, "learning_rate": 2.3055e-06, "loss": 2.0886, "step": 4611 }, { "epoch": 1.4457680250783698, "grad_norm": 42.71048355102539, "learning_rate": 2.306e-06, "loss": 3.2491, "step": 4612 }, { "epoch": 1.4460815047021944, "grad_norm": 18.34791374206543, "learning_rate": 2.3065000000000003e-06, "loss": 2.4025, "step": 4613 }, { "epoch": 1.4463949843260189, "grad_norm": 23.1646728515625, "learning_rate": 2.307e-06, "loss": 2.9714, "step": 4614 }, { "epoch": 1.4467084639498433, "grad_norm": 39.93225860595703, "learning_rate": 2.3075000000000004e-06, "loss": 2.9443, "step": 4615 }, { "epoch": 1.4470219435736678, "grad_norm": 17.18674087524414, "learning_rate": 2.308e-06, "loss": 2.4679, "step": 4616 }, { "epoch": 1.4473354231974922, "grad_norm": 40.53770065307617, "learning_rate": 2.3085e-06, "loss": 2.764, "step": 4617 }, { "epoch": 1.4476489028213166, "grad_norm": 36.780208587646484, "learning_rate": 2.3090000000000003e-06, "loss": 2.8383, "step": 4618 }, { "epoch": 1.447962382445141, "grad_norm": 14.106476783752441, "learning_rate": 2.3095e-06, "loss": 2.404, "step": 4619 }, { "epoch": 1.4482758620689655, "grad_norm": 13.64006233215332, "learning_rate": 2.3100000000000003e-06, "loss": 2.2731, "step": 4620 }, { "epoch": 1.44858934169279, "grad_norm": 26.457382202148438, "learning_rate": 2.3105e-06, "loss": 2.7478, "step": 4621 }, { "epoch": 1.4489028213166144, "grad_norm": 38.400936126708984, "learning_rate": 2.311e-06, "loss": 2.5102, "step": 4622 }, { "epoch": 1.4492163009404389, "grad_norm": 41.096771240234375, "learning_rate": 2.3115000000000002e-06, "loss": 2.3325, "step": 4623 }, { "epoch": 1.4495297805642633, "grad_norm": 37.16647720336914, "learning_rate": 2.312e-06, "loss": 2.8084, "step": 4624 }, { "epoch": 1.4498432601880877, "grad_norm": 12.608713150024414, "learning_rate": 2.3125000000000003e-06, "loss": 2.474, "step": 4625 }, { "epoch": 1.4501567398119122, "grad_norm": 11.61148452758789, "learning_rate": 2.313e-06, "loss": 2.4906, "step": 4626 }, { "epoch": 1.4504702194357366, "grad_norm": 152.3636474609375, "learning_rate": 2.3135000000000004e-06, "loss": 2.8927, "step": 4627 }, { "epoch": 1.450783699059561, "grad_norm": 22.41962242126465, "learning_rate": 2.314e-06, "loss": 2.3821, "step": 4628 }, { "epoch": 1.4510971786833855, "grad_norm": 12.883655548095703, "learning_rate": 2.3145e-06, "loss": 2.2177, "step": 4629 }, { "epoch": 1.4514106583072102, "grad_norm": 85.4560317993164, "learning_rate": 2.3150000000000003e-06, "loss": 3.2228, "step": 4630 }, { "epoch": 1.4517241379310346, "grad_norm": 26.63489532470703, "learning_rate": 2.3155e-06, "loss": 2.3603, "step": 4631 }, { "epoch": 1.452037617554859, "grad_norm": 32.63912582397461, "learning_rate": 2.3160000000000004e-06, "loss": 2.5887, "step": 4632 }, { "epoch": 1.4523510971786835, "grad_norm": 27.428691864013672, "learning_rate": 2.3165e-06, "loss": 2.6201, "step": 4633 }, { "epoch": 1.452664576802508, "grad_norm": 12.853342056274414, "learning_rate": 2.317e-06, "loss": 2.4724, "step": 4634 }, { "epoch": 1.4529780564263324, "grad_norm": 31.954179763793945, "learning_rate": 2.3175000000000003e-06, "loss": 2.8328, "step": 4635 }, { "epoch": 1.4532915360501568, "grad_norm": 17.225984573364258, "learning_rate": 2.318e-06, "loss": 2.4483, "step": 4636 }, { "epoch": 1.4536050156739813, "grad_norm": 39.024723052978516, "learning_rate": 2.3185000000000003e-06, "loss": 2.6955, "step": 4637 }, { "epoch": 1.4539184952978057, "grad_norm": 13.877724647521973, "learning_rate": 2.319e-06, "loss": 2.5151, "step": 4638 }, { "epoch": 1.4542319749216301, "grad_norm": 20.05229377746582, "learning_rate": 2.3195e-06, "loss": 2.4511, "step": 4639 }, { "epoch": 1.4545454545454546, "grad_norm": 40.79631805419922, "learning_rate": 2.3200000000000002e-06, "loss": 3.2781, "step": 4640 }, { "epoch": 1.454858934169279, "grad_norm": 26.610153198242188, "learning_rate": 2.3205e-06, "loss": 2.237, "step": 4641 }, { "epoch": 1.4551724137931035, "grad_norm": 43.11115264892578, "learning_rate": 2.3210000000000003e-06, "loss": 1.9168, "step": 4642 }, { "epoch": 1.455485893416928, "grad_norm": 31.909210205078125, "learning_rate": 2.3215e-06, "loss": 2.3807, "step": 4643 }, { "epoch": 1.4557993730407524, "grad_norm": 34.96866989135742, "learning_rate": 2.322e-06, "loss": 2.5393, "step": 4644 }, { "epoch": 1.4561128526645768, "grad_norm": 16.582456588745117, "learning_rate": 2.3225e-06, "loss": 2.4451, "step": 4645 }, { "epoch": 1.4564263322884012, "grad_norm": 13.87610912322998, "learning_rate": 2.3230000000000005e-06, "loss": 2.4433, "step": 4646 }, { "epoch": 1.4567398119122257, "grad_norm": 23.799659729003906, "learning_rate": 2.3235000000000003e-06, "loss": 2.5115, "step": 4647 }, { "epoch": 1.4570532915360501, "grad_norm": 27.039108276367188, "learning_rate": 2.324e-06, "loss": 2.2724, "step": 4648 }, { "epoch": 1.4573667711598746, "grad_norm": 21.01291275024414, "learning_rate": 2.3245e-06, "loss": 2.5922, "step": 4649 }, { "epoch": 1.457680250783699, "grad_norm": 25.368284225463867, "learning_rate": 2.325e-06, "loss": 3.2784, "step": 4650 }, { "epoch": 1.4579937304075234, "grad_norm": 21.040912628173828, "learning_rate": 2.3255000000000004e-06, "loss": 2.2421, "step": 4651 }, { "epoch": 1.458307210031348, "grad_norm": 27.654041290283203, "learning_rate": 2.3260000000000003e-06, "loss": 2.5671, "step": 4652 }, { "epoch": 1.4586206896551723, "grad_norm": 14.569658279418945, "learning_rate": 2.3265e-06, "loss": 2.3534, "step": 4653 }, { "epoch": 1.4589341692789968, "grad_norm": 27.91167640686035, "learning_rate": 2.327e-06, "loss": 2.7539, "step": 4654 }, { "epoch": 1.4592476489028212, "grad_norm": 17.407148361206055, "learning_rate": 2.3275e-06, "loss": 2.5176, "step": 4655 }, { "epoch": 1.4595611285266457, "grad_norm": 20.95456886291504, "learning_rate": 2.3280000000000004e-06, "loss": 2.2847, "step": 4656 }, { "epoch": 1.45987460815047, "grad_norm": 15.59076976776123, "learning_rate": 2.3285000000000002e-06, "loss": 2.2954, "step": 4657 }, { "epoch": 1.4601880877742945, "grad_norm": 18.56783676147461, "learning_rate": 2.329e-06, "loss": 2.4062, "step": 4658 }, { "epoch": 1.4605015673981192, "grad_norm": 18.09410858154297, "learning_rate": 2.3295e-06, "loss": 2.2867, "step": 4659 }, { "epoch": 1.4608150470219436, "grad_norm": 10.945845603942871, "learning_rate": 2.33e-06, "loss": 2.2486, "step": 4660 }, { "epoch": 1.461128526645768, "grad_norm": 19.199756622314453, "learning_rate": 2.3305000000000004e-06, "loss": 2.6845, "step": 4661 }, { "epoch": 1.4614420062695925, "grad_norm": 15.891563415527344, "learning_rate": 2.3310000000000002e-06, "loss": 2.6765, "step": 4662 }, { "epoch": 1.461755485893417, "grad_norm": 31.088024139404297, "learning_rate": 2.3315e-06, "loss": 2.3015, "step": 4663 }, { "epoch": 1.4620689655172414, "grad_norm": 72.7515869140625, "learning_rate": 2.3320000000000003e-06, "loss": 2.8665, "step": 4664 }, { "epoch": 1.4623824451410659, "grad_norm": 22.520038604736328, "learning_rate": 2.3325e-06, "loss": 2.517, "step": 4665 }, { "epoch": 1.4626959247648903, "grad_norm": 19.68512725830078, "learning_rate": 2.3330000000000004e-06, "loss": 2.5447, "step": 4666 }, { "epoch": 1.4630094043887147, "grad_norm": 55.96286392211914, "learning_rate": 2.3335e-06, "loss": 2.6517, "step": 4667 }, { "epoch": 1.4633228840125392, "grad_norm": 38.48992919921875, "learning_rate": 2.334e-06, "loss": 2.9361, "step": 4668 }, { "epoch": 1.4636363636363636, "grad_norm": 25.144210815429688, "learning_rate": 2.3345000000000003e-06, "loss": 2.3563, "step": 4669 }, { "epoch": 1.463949843260188, "grad_norm": 16.994159698486328, "learning_rate": 2.3350000000000005e-06, "loss": 2.4947, "step": 4670 }, { "epoch": 1.4642633228840125, "grad_norm": 30.933141708374023, "learning_rate": 2.3355000000000003e-06, "loss": 2.469, "step": 4671 }, { "epoch": 1.464576802507837, "grad_norm": 33.76378631591797, "learning_rate": 2.336e-06, "loss": 2.8162, "step": 4672 }, { "epoch": 1.4648902821316614, "grad_norm": 13.684147834777832, "learning_rate": 2.3365e-06, "loss": 2.3385, "step": 4673 }, { "epoch": 1.4652037617554858, "grad_norm": 33.54969787597656, "learning_rate": 2.3370000000000002e-06, "loss": 2.4732, "step": 4674 }, { "epoch": 1.4655172413793103, "grad_norm": 21.35070037841797, "learning_rate": 2.3375000000000005e-06, "loss": 2.8555, "step": 4675 }, { "epoch": 1.4658307210031347, "grad_norm": 20.555835723876953, "learning_rate": 2.3380000000000003e-06, "loss": 2.4961, "step": 4676 }, { "epoch": 1.4661442006269594, "grad_norm": 26.707595825195312, "learning_rate": 2.3385e-06, "loss": 2.4832, "step": 4677 }, { "epoch": 1.4664576802507838, "grad_norm": 18.20785140991211, "learning_rate": 2.339e-06, "loss": 2.9194, "step": 4678 }, { "epoch": 1.4667711598746083, "grad_norm": 94.4541244506836, "learning_rate": 2.3395000000000002e-06, "loss": 2.3439, "step": 4679 }, { "epoch": 1.4670846394984327, "grad_norm": 25.442766189575195, "learning_rate": 2.3400000000000005e-06, "loss": 2.6061, "step": 4680 }, { "epoch": 1.4673981191222571, "grad_norm": 33.864078521728516, "learning_rate": 2.3405000000000003e-06, "loss": 2.5939, "step": 4681 }, { "epoch": 1.4677115987460816, "grad_norm": 41.32418441772461, "learning_rate": 2.341e-06, "loss": 3.0855, "step": 4682 }, { "epoch": 1.468025078369906, "grad_norm": 20.600738525390625, "learning_rate": 2.3415000000000004e-06, "loss": 2.4121, "step": 4683 }, { "epoch": 1.4683385579937305, "grad_norm": 21.15034294128418, "learning_rate": 2.342e-06, "loss": 2.4212, "step": 4684 }, { "epoch": 1.468652037617555, "grad_norm": 16.126068115234375, "learning_rate": 2.3425000000000004e-06, "loss": 2.5768, "step": 4685 }, { "epoch": 1.4689655172413794, "grad_norm": 18.89388084411621, "learning_rate": 2.3430000000000003e-06, "loss": 2.1556, "step": 4686 }, { "epoch": 1.4692789968652038, "grad_norm": 49.33967208862305, "learning_rate": 2.3435e-06, "loss": 2.4046, "step": 4687 }, { "epoch": 1.4695924764890282, "grad_norm": 18.668262481689453, "learning_rate": 2.3440000000000003e-06, "loss": 2.4608, "step": 4688 }, { "epoch": 1.4699059561128527, "grad_norm": 15.122355461120605, "learning_rate": 2.3445e-06, "loss": 2.412, "step": 4689 }, { "epoch": 1.4702194357366771, "grad_norm": 9.459912300109863, "learning_rate": 2.345e-06, "loss": 2.3784, "step": 4690 }, { "epoch": 1.4705329153605016, "grad_norm": 13.91557788848877, "learning_rate": 2.3455000000000002e-06, "loss": 2.6558, "step": 4691 }, { "epoch": 1.470846394984326, "grad_norm": 53.33099365234375, "learning_rate": 2.346e-06, "loss": 2.5759, "step": 4692 }, { "epoch": 1.4711598746081505, "grad_norm": 26.81603240966797, "learning_rate": 2.3465000000000003e-06, "loss": 2.4183, "step": 4693 }, { "epoch": 1.471473354231975, "grad_norm": 15.679391860961914, "learning_rate": 2.347e-06, "loss": 2.8987, "step": 4694 }, { "epoch": 1.4717868338557993, "grad_norm": 16.509136199951172, "learning_rate": 2.3475e-06, "loss": 2.4195, "step": 4695 }, { "epoch": 1.4721003134796238, "grad_norm": 30.074337005615234, "learning_rate": 2.3480000000000002e-06, "loss": 2.5757, "step": 4696 }, { "epoch": 1.4724137931034482, "grad_norm": 19.880718231201172, "learning_rate": 2.3485e-06, "loss": 2.3326, "step": 4697 }, { "epoch": 1.4727272727272727, "grad_norm": 37.2962646484375, "learning_rate": 2.3490000000000003e-06, "loss": 2.1726, "step": 4698 }, { "epoch": 1.473040752351097, "grad_norm": 29.554244995117188, "learning_rate": 2.3495e-06, "loss": 2.377, "step": 4699 }, { "epoch": 1.4733542319749215, "grad_norm": 24.98297691345215, "learning_rate": 2.35e-06, "loss": 2.597, "step": 4700 }, { "epoch": 1.473667711598746, "grad_norm": 13.754597663879395, "learning_rate": 2.3505e-06, "loss": 2.4814, "step": 4701 }, { "epoch": 1.4739811912225704, "grad_norm": 54.246803283691406, "learning_rate": 2.351e-06, "loss": 2.4151, "step": 4702 }, { "epoch": 1.4742946708463949, "grad_norm": 9.66529369354248, "learning_rate": 2.3515000000000003e-06, "loss": 2.3911, "step": 4703 }, { "epoch": 1.4746081504702193, "grad_norm": 16.990436553955078, "learning_rate": 2.352e-06, "loss": 2.7343, "step": 4704 }, { "epoch": 1.4749216300940438, "grad_norm": 18.184396743774414, "learning_rate": 2.3525e-06, "loss": 2.3241, "step": 4705 }, { "epoch": 1.4752351097178684, "grad_norm": 16.607677459716797, "learning_rate": 2.353e-06, "loss": 2.5718, "step": 4706 }, { "epoch": 1.4755485893416929, "grad_norm": 25.31608772277832, "learning_rate": 2.3535000000000004e-06, "loss": 2.612, "step": 4707 }, { "epoch": 1.4758620689655173, "grad_norm": 9.628955841064453, "learning_rate": 2.3540000000000002e-06, "loss": 2.3875, "step": 4708 }, { "epoch": 1.4761755485893417, "grad_norm": 20.810176849365234, "learning_rate": 2.3545e-06, "loss": 2.4388, "step": 4709 }, { "epoch": 1.4764890282131662, "grad_norm": 47.8430061340332, "learning_rate": 2.355e-06, "loss": 2.5104, "step": 4710 }, { "epoch": 1.4768025078369906, "grad_norm": 44.49211502075195, "learning_rate": 2.3555e-06, "loss": 2.1969, "step": 4711 }, { "epoch": 1.477115987460815, "grad_norm": 16.758508682250977, "learning_rate": 2.3560000000000004e-06, "loss": 2.4783, "step": 4712 }, { "epoch": 1.4774294670846395, "grad_norm": 20.397388458251953, "learning_rate": 2.3565000000000002e-06, "loss": 2.361, "step": 4713 }, { "epoch": 1.477742946708464, "grad_norm": 26.805633544921875, "learning_rate": 2.357e-06, "loss": 2.5757, "step": 4714 }, { "epoch": 1.4780564263322884, "grad_norm": 44.96598815917969, "learning_rate": 2.3575e-06, "loss": 2.8678, "step": 4715 }, { "epoch": 1.4783699059561128, "grad_norm": 23.84420394897461, "learning_rate": 2.358e-06, "loss": 2.7157, "step": 4716 }, { "epoch": 1.4786833855799373, "grad_norm": 14.02918529510498, "learning_rate": 2.3585000000000004e-06, "loss": 2.2836, "step": 4717 }, { "epoch": 1.4789968652037617, "grad_norm": 16.57590675354004, "learning_rate": 2.359e-06, "loss": 2.1943, "step": 4718 }, { "epoch": 1.4793103448275862, "grad_norm": 23.227548599243164, "learning_rate": 2.3595e-06, "loss": 3.5659, "step": 4719 }, { "epoch": 1.4796238244514106, "grad_norm": 12.566490173339844, "learning_rate": 2.3600000000000003e-06, "loss": 2.4076, "step": 4720 }, { "epoch": 1.479937304075235, "grad_norm": 30.77694320678711, "learning_rate": 2.3605e-06, "loss": 2.5922, "step": 4721 }, { "epoch": 1.4802507836990595, "grad_norm": 15.837915420532227, "learning_rate": 2.3610000000000003e-06, "loss": 2.3494, "step": 4722 }, { "epoch": 1.480564263322884, "grad_norm": 18.752384185791016, "learning_rate": 2.3615e-06, "loss": 2.6885, "step": 4723 }, { "epoch": 1.4808777429467086, "grad_norm": 21.009431838989258, "learning_rate": 2.362e-06, "loss": 2.3239, "step": 4724 }, { "epoch": 1.481191222570533, "grad_norm": 51.4163703918457, "learning_rate": 2.3625000000000003e-06, "loss": 2.4585, "step": 4725 }, { "epoch": 1.4815047021943575, "grad_norm": 10.593263626098633, "learning_rate": 2.3630000000000005e-06, "loss": 2.2508, "step": 4726 }, { "epoch": 1.481818181818182, "grad_norm": 27.31687355041504, "learning_rate": 2.3635000000000003e-06, "loss": 2.5703, "step": 4727 }, { "epoch": 1.4821316614420064, "grad_norm": 27.05815315246582, "learning_rate": 2.364e-06, "loss": 2.7473, "step": 4728 }, { "epoch": 1.4824451410658308, "grad_norm": 18.825220108032227, "learning_rate": 2.3645e-06, "loss": 2.3339, "step": 4729 }, { "epoch": 1.4827586206896552, "grad_norm": 45.73810958862305, "learning_rate": 2.3650000000000002e-06, "loss": 2.526, "step": 4730 }, { "epoch": 1.4830721003134797, "grad_norm": 17.639097213745117, "learning_rate": 2.3655000000000005e-06, "loss": 2.6365, "step": 4731 }, { "epoch": 1.4833855799373041, "grad_norm": 20.23157501220703, "learning_rate": 2.3660000000000003e-06, "loss": 2.3239, "step": 4732 }, { "epoch": 1.4836990595611286, "grad_norm": 13.595239639282227, "learning_rate": 2.3665e-06, "loss": 2.2712, "step": 4733 }, { "epoch": 1.484012539184953, "grad_norm": 90.09672546386719, "learning_rate": 2.367e-06, "loss": 3.2016, "step": 4734 }, { "epoch": 1.4843260188087775, "grad_norm": 96.69264221191406, "learning_rate": 2.3675e-06, "loss": 2.354, "step": 4735 }, { "epoch": 1.484639498432602, "grad_norm": 32.80999755859375, "learning_rate": 2.3680000000000005e-06, "loss": 3.1623, "step": 4736 }, { "epoch": 1.4849529780564263, "grad_norm": 29.02104377746582, "learning_rate": 2.3685000000000003e-06, "loss": 2.4772, "step": 4737 }, { "epoch": 1.4852664576802508, "grad_norm": 19.033187866210938, "learning_rate": 2.369e-06, "loss": 2.3753, "step": 4738 }, { "epoch": 1.4855799373040752, "grad_norm": 20.396085739135742, "learning_rate": 2.3695e-06, "loss": 2.3814, "step": 4739 }, { "epoch": 1.4858934169278997, "grad_norm": 43.776119232177734, "learning_rate": 2.37e-06, "loss": 2.4351, "step": 4740 }, { "epoch": 1.486206896551724, "grad_norm": 16.356319427490234, "learning_rate": 2.3705000000000004e-06, "loss": 2.385, "step": 4741 }, { "epoch": 1.4865203761755486, "grad_norm": 27.0319766998291, "learning_rate": 2.3710000000000003e-06, "loss": 2.6299, "step": 4742 }, { "epoch": 1.486833855799373, "grad_norm": 28.409624099731445, "learning_rate": 2.3715e-06, "loss": 2.5679, "step": 4743 }, { "epoch": 1.4871473354231974, "grad_norm": 15.29839038848877, "learning_rate": 2.3720000000000003e-06, "loss": 2.2037, "step": 4744 }, { "epoch": 1.4874608150470219, "grad_norm": 44.7656364440918, "learning_rate": 2.3725e-06, "loss": 2.4191, "step": 4745 }, { "epoch": 1.4877742946708463, "grad_norm": 33.07748031616211, "learning_rate": 2.3730000000000004e-06, "loss": 2.9531, "step": 4746 }, { "epoch": 1.4880877742946708, "grad_norm": 15.354388236999512, "learning_rate": 2.3735000000000002e-06, "loss": 2.264, "step": 4747 }, { "epoch": 1.4884012539184952, "grad_norm": 52.4686164855957, "learning_rate": 2.374e-06, "loss": 2.2487, "step": 4748 }, { "epoch": 1.4887147335423196, "grad_norm": 13.158205032348633, "learning_rate": 2.3745000000000003e-06, "loss": 2.4897, "step": 4749 }, { "epoch": 1.489028213166144, "grad_norm": 32.53001403808594, "learning_rate": 2.375e-06, "loss": 2.7913, "step": 4750 }, { "epoch": 1.4893416927899685, "grad_norm": 15.134367942810059, "learning_rate": 2.3755000000000004e-06, "loss": 2.4354, "step": 4751 }, { "epoch": 1.489655172413793, "grad_norm": 14.022037506103516, "learning_rate": 2.376e-06, "loss": 2.8385, "step": 4752 }, { "epoch": 1.4899686520376176, "grad_norm": 12.180167198181152, "learning_rate": 2.3765e-06, "loss": 2.355, "step": 4753 }, { "epoch": 1.490282131661442, "grad_norm": 24.225900650024414, "learning_rate": 2.3770000000000003e-06, "loss": 2.7134, "step": 4754 }, { "epoch": 1.4905956112852665, "grad_norm": 14.048494338989258, "learning_rate": 2.3775e-06, "loss": 2.3146, "step": 4755 }, { "epoch": 1.490909090909091, "grad_norm": 52.83781433105469, "learning_rate": 2.3780000000000004e-06, "loss": 2.6985, "step": 4756 }, { "epoch": 1.4912225705329154, "grad_norm": 16.474990844726562, "learning_rate": 2.3785e-06, "loss": 2.3649, "step": 4757 }, { "epoch": 1.4915360501567398, "grad_norm": 71.45556640625, "learning_rate": 2.379e-06, "loss": 2.3475, "step": 4758 }, { "epoch": 1.4918495297805643, "grad_norm": 115.68441009521484, "learning_rate": 2.3795000000000003e-06, "loss": 3.3305, "step": 4759 }, { "epoch": 1.4921630094043887, "grad_norm": 33.583370208740234, "learning_rate": 2.38e-06, "loss": 2.6491, "step": 4760 }, { "epoch": 1.4924764890282132, "grad_norm": 67.39979553222656, "learning_rate": 2.3805000000000003e-06, "loss": 3.0342, "step": 4761 }, { "epoch": 1.4927899686520376, "grad_norm": 18.05873680114746, "learning_rate": 2.381e-06, "loss": 2.2855, "step": 4762 }, { "epoch": 1.493103448275862, "grad_norm": 19.0296573638916, "learning_rate": 2.3815000000000004e-06, "loss": 2.3671, "step": 4763 }, { "epoch": 1.4934169278996865, "grad_norm": 20.803741455078125, "learning_rate": 2.3820000000000002e-06, "loss": 2.1159, "step": 4764 }, { "epoch": 1.493730407523511, "grad_norm": 21.71002197265625, "learning_rate": 2.3825e-06, "loss": 2.1941, "step": 4765 }, { "epoch": 1.4940438871473354, "grad_norm": 27.901458740234375, "learning_rate": 2.3830000000000003e-06, "loss": 2.3726, "step": 4766 }, { "epoch": 1.4943573667711598, "grad_norm": 22.505205154418945, "learning_rate": 2.3835e-06, "loss": 2.2415, "step": 4767 }, { "epoch": 1.4946708463949843, "grad_norm": 59.68888473510742, "learning_rate": 2.3840000000000004e-06, "loss": 2.935, "step": 4768 }, { "epoch": 1.4949843260188087, "grad_norm": 24.86325454711914, "learning_rate": 2.3845e-06, "loss": 2.3544, "step": 4769 }, { "epoch": 1.4952978056426331, "grad_norm": 19.455236434936523, "learning_rate": 2.385e-06, "loss": 2.2056, "step": 4770 }, { "epoch": 1.4956112852664578, "grad_norm": 12.818082809448242, "learning_rate": 2.3855000000000003e-06, "loss": 2.2469, "step": 4771 }, { "epoch": 1.4959247648902823, "grad_norm": 41.96872329711914, "learning_rate": 2.386e-06, "loss": 2.7509, "step": 4772 }, { "epoch": 1.4962382445141067, "grad_norm": 19.305591583251953, "learning_rate": 2.3865000000000004e-06, "loss": 2.315, "step": 4773 }, { "epoch": 1.4965517241379311, "grad_norm": 37.54866027832031, "learning_rate": 2.387e-06, "loss": 2.1162, "step": 4774 }, { "epoch": 1.4968652037617556, "grad_norm": 21.865880966186523, "learning_rate": 2.3875e-06, "loss": 2.6231, "step": 4775 }, { "epoch": 1.49717868338558, "grad_norm": 16.861860275268555, "learning_rate": 2.3880000000000003e-06, "loss": 2.2176, "step": 4776 }, { "epoch": 1.4974921630094045, "grad_norm": 31.959991455078125, "learning_rate": 2.3885e-06, "loss": 2.1846, "step": 4777 }, { "epoch": 1.497805642633229, "grad_norm": 32.17866897583008, "learning_rate": 2.3890000000000003e-06, "loss": 2.3376, "step": 4778 }, { "epoch": 1.4981191222570533, "grad_norm": 22.17281723022461, "learning_rate": 2.3895e-06, "loss": 2.3265, "step": 4779 }, { "epoch": 1.4984326018808778, "grad_norm": 19.636741638183594, "learning_rate": 2.39e-06, "loss": 2.3045, "step": 4780 }, { "epoch": 1.4987460815047022, "grad_norm": 13.541577339172363, "learning_rate": 2.3905000000000002e-06, "loss": 2.8675, "step": 4781 }, { "epoch": 1.4990595611285267, "grad_norm": 14.708398818969727, "learning_rate": 2.391e-06, "loss": 2.4103, "step": 4782 }, { "epoch": 1.4993730407523511, "grad_norm": 20.451107025146484, "learning_rate": 2.3915000000000003e-06, "loss": 2.6472, "step": 4783 }, { "epoch": 1.4996865203761756, "grad_norm": 46.59673309326172, "learning_rate": 2.392e-06, "loss": 2.3958, "step": 4784 }, { "epoch": 1.5, "grad_norm": 22.707223892211914, "learning_rate": 2.3925e-06, "loss": 2.0873, "step": 4785 }, { "epoch": 1.5003134796238244, "grad_norm": 51.4575080871582, "learning_rate": 2.3930000000000002e-06, "loss": 2.4381, "step": 4786 }, { "epoch": 1.5006269592476489, "grad_norm": 17.58340072631836, "learning_rate": 2.3935000000000005e-06, "loss": 2.2001, "step": 4787 }, { "epoch": 1.5009404388714733, "grad_norm": 94.2140884399414, "learning_rate": 2.3940000000000003e-06, "loss": 2.8093, "step": 4788 }, { "epoch": 1.5009404388714733, "eval_loss": 3.0960214138031006, "eval_runtime": 21.033, "eval_samples_per_second": 127.752, "eval_steps_per_second": 7.987, "step": 4788 }, { "epoch": 1.5012539184952978, "grad_norm": 27.263410568237305, "learning_rate": 2.3945e-06, "loss": 2.5208, "step": 4789 }, { "epoch": 1.5015673981191222, "grad_norm": 46.74039077758789, "learning_rate": 2.395e-06, "loss": 2.3394, "step": 4790 }, { "epoch": 1.5018808777429467, "grad_norm": 29.4154109954834, "learning_rate": 2.3955e-06, "loss": 2.3911, "step": 4791 }, { "epoch": 1.502194357366771, "grad_norm": 21.53428840637207, "learning_rate": 2.3960000000000004e-06, "loss": 2.2403, "step": 4792 }, { "epoch": 1.5025078369905955, "grad_norm": 12.628145217895508, "learning_rate": 2.3965000000000003e-06, "loss": 2.8572, "step": 4793 }, { "epoch": 1.50282131661442, "grad_norm": 44.54032516479492, "learning_rate": 2.397e-06, "loss": 2.4828, "step": 4794 }, { "epoch": 1.5031347962382444, "grad_norm": 16.90373420715332, "learning_rate": 2.3975e-06, "loss": 2.3013, "step": 4795 }, { "epoch": 1.5034482758620689, "grad_norm": 17.77326774597168, "learning_rate": 2.398e-06, "loss": 2.414, "step": 4796 }, { "epoch": 1.5037617554858933, "grad_norm": 19.9771728515625, "learning_rate": 2.3985000000000004e-06, "loss": 2.236, "step": 4797 }, { "epoch": 1.5040752351097177, "grad_norm": 16.979183197021484, "learning_rate": 2.3990000000000002e-06, "loss": 2.2729, "step": 4798 }, { "epoch": 1.5043887147335422, "grad_norm": 28.2900333404541, "learning_rate": 2.3995e-06, "loss": 2.4509, "step": 4799 }, { "epoch": 1.5047021943573666, "grad_norm": 16.051536560058594, "learning_rate": 2.4000000000000003e-06, "loss": 2.4083, "step": 4800 }, { "epoch": 1.505015673981191, "grad_norm": 43.82846450805664, "learning_rate": 2.4005e-06, "loss": 2.3903, "step": 4801 }, { "epoch": 1.5053291536050155, "grad_norm": 24.0675106048584, "learning_rate": 2.4010000000000004e-06, "loss": 3.2002, "step": 4802 }, { "epoch": 1.5056426332288402, "grad_norm": 16.599925994873047, "learning_rate": 2.4015000000000002e-06, "loss": 2.2859, "step": 4803 }, { "epoch": 1.5059561128526646, "grad_norm": 15.7088623046875, "learning_rate": 2.402e-06, "loss": 2.1144, "step": 4804 }, { "epoch": 1.506269592476489, "grad_norm": 17.363964080810547, "learning_rate": 2.4025000000000003e-06, "loss": 2.4393, "step": 4805 }, { "epoch": 1.5065830721003135, "grad_norm": 40.69404602050781, "learning_rate": 2.4030000000000005e-06, "loss": 2.4038, "step": 4806 }, { "epoch": 1.506896551724138, "grad_norm": 11.931968688964844, "learning_rate": 2.4035000000000004e-06, "loss": 2.2415, "step": 4807 }, { "epoch": 1.5072100313479624, "grad_norm": 15.336716651916504, "learning_rate": 2.404e-06, "loss": 2.3751, "step": 4808 }, { "epoch": 1.5075235109717868, "grad_norm": 26.773588180541992, "learning_rate": 2.4045e-06, "loss": 2.4447, "step": 4809 }, { "epoch": 1.5078369905956113, "grad_norm": 10.982498168945312, "learning_rate": 2.4050000000000003e-06, "loss": 2.2266, "step": 4810 }, { "epoch": 1.5081504702194357, "grad_norm": 18.58058738708496, "learning_rate": 2.4055000000000005e-06, "loss": 2.4837, "step": 4811 }, { "epoch": 1.5084639498432602, "grad_norm": 25.92458152770996, "learning_rate": 2.4060000000000003e-06, "loss": 2.3429, "step": 4812 }, { "epoch": 1.5087774294670846, "grad_norm": 22.95157241821289, "learning_rate": 2.4065e-06, "loss": 2.3551, "step": 4813 }, { "epoch": 1.509090909090909, "grad_norm": 12.792777061462402, "learning_rate": 2.407e-06, "loss": 2.1755, "step": 4814 }, { "epoch": 1.5094043887147337, "grad_norm": 20.182924270629883, "learning_rate": 2.4075000000000002e-06, "loss": 2.0571, "step": 4815 }, { "epoch": 1.5097178683385581, "grad_norm": 16.23738670349121, "learning_rate": 2.408e-06, "loss": 2.681, "step": 4816 }, { "epoch": 1.5100313479623826, "grad_norm": 12.84005355834961, "learning_rate": 2.4085000000000003e-06, "loss": 2.9673, "step": 4817 }, { "epoch": 1.510344827586207, "grad_norm": 19.747169494628906, "learning_rate": 2.409e-06, "loss": 2.3013, "step": 4818 }, { "epoch": 1.5106583072100315, "grad_norm": 27.003398895263672, "learning_rate": 2.4095e-06, "loss": 3.092, "step": 4819 }, { "epoch": 1.510971786833856, "grad_norm": 26.648151397705078, "learning_rate": 2.4100000000000002e-06, "loss": 2.2164, "step": 4820 }, { "epoch": 1.5112852664576804, "grad_norm": 12.34875202178955, "learning_rate": 2.4105e-06, "loss": 2.6219, "step": 4821 }, { "epoch": 1.5115987460815048, "grad_norm": 27.272109985351562, "learning_rate": 2.4110000000000003e-06, "loss": 2.3687, "step": 4822 }, { "epoch": 1.5119122257053292, "grad_norm": 33.198692321777344, "learning_rate": 2.4115e-06, "loss": 2.7016, "step": 4823 }, { "epoch": 1.5122257053291537, "grad_norm": 15.936403274536133, "learning_rate": 2.4120000000000004e-06, "loss": 2.6341, "step": 4824 }, { "epoch": 1.5125391849529781, "grad_norm": 37.71348190307617, "learning_rate": 2.4125e-06, "loss": 2.3681, "step": 4825 }, { "epoch": 1.5128526645768026, "grad_norm": 13.023741722106934, "learning_rate": 2.413e-06, "loss": 2.473, "step": 4826 }, { "epoch": 1.513166144200627, "grad_norm": 13.259977340698242, "learning_rate": 2.4135000000000003e-06, "loss": 2.703, "step": 4827 }, { "epoch": 1.5134796238244514, "grad_norm": 59.40862274169922, "learning_rate": 2.414e-06, "loss": 3.2762, "step": 4828 }, { "epoch": 1.513793103448276, "grad_norm": 24.48963737487793, "learning_rate": 2.4145000000000003e-06, "loss": 2.906, "step": 4829 }, { "epoch": 1.5141065830721003, "grad_norm": 13.027275085449219, "learning_rate": 2.415e-06, "loss": 2.5957, "step": 4830 }, { "epoch": 1.5144200626959248, "grad_norm": 13.824039459228516, "learning_rate": 2.4155e-06, "loss": 2.2586, "step": 4831 }, { "epoch": 1.5147335423197492, "grad_norm": 22.69841194152832, "learning_rate": 2.4160000000000002e-06, "loss": 2.7997, "step": 4832 }, { "epoch": 1.5150470219435737, "grad_norm": 28.543188095092773, "learning_rate": 2.4165e-06, "loss": 2.6812, "step": 4833 }, { "epoch": 1.515360501567398, "grad_norm": 66.29344940185547, "learning_rate": 2.4170000000000003e-06, "loss": 3.7326, "step": 4834 }, { "epoch": 1.5156739811912225, "grad_norm": 35.89635467529297, "learning_rate": 2.4175e-06, "loss": 2.3554, "step": 4835 }, { "epoch": 1.515987460815047, "grad_norm": 21.20792579650879, "learning_rate": 2.418e-06, "loss": 2.4513, "step": 4836 }, { "epoch": 1.5163009404388714, "grad_norm": 24.268932342529297, "learning_rate": 2.4185000000000002e-06, "loss": 2.4245, "step": 4837 }, { "epoch": 1.5166144200626959, "grad_norm": 464.6249694824219, "learning_rate": 2.419e-06, "loss": 2.7334, "step": 4838 }, { "epoch": 1.5169278996865203, "grad_norm": 17.084293365478516, "learning_rate": 2.4195000000000003e-06, "loss": 2.1875, "step": 4839 }, { "epoch": 1.5172413793103448, "grad_norm": 47.49144744873047, "learning_rate": 2.42e-06, "loss": 2.5519, "step": 4840 }, { "epoch": 1.5175548589341692, "grad_norm": 19.089134216308594, "learning_rate": 2.4205e-06, "loss": 3.1702, "step": 4841 }, { "epoch": 1.5178683385579936, "grad_norm": 21.179597854614258, "learning_rate": 2.421e-06, "loss": 2.2937, "step": 4842 }, { "epoch": 1.518181818181818, "grad_norm": 12.67155933380127, "learning_rate": 2.4215000000000004e-06, "loss": 2.2309, "step": 4843 }, { "epoch": 1.5184952978056425, "grad_norm": 15.11013126373291, "learning_rate": 2.4220000000000003e-06, "loss": 2.2038, "step": 4844 }, { "epoch": 1.518808777429467, "grad_norm": 13.126248359680176, "learning_rate": 2.4225e-06, "loss": 2.4207, "step": 4845 }, { "epoch": 1.5191222570532914, "grad_norm": 55.65666198730469, "learning_rate": 2.423e-06, "loss": 2.9468, "step": 4846 }, { "epoch": 1.5194357366771158, "grad_norm": 12.58440113067627, "learning_rate": 2.4235e-06, "loss": 2.3262, "step": 4847 }, { "epoch": 1.5197492163009403, "grad_norm": 37.81396484375, "learning_rate": 2.4240000000000004e-06, "loss": 2.6327, "step": 4848 }, { "epoch": 1.5200626959247647, "grad_norm": 16.613920211791992, "learning_rate": 2.4245000000000002e-06, "loss": 2.4177, "step": 4849 }, { "epoch": 1.5203761755485894, "grad_norm": 25.54189682006836, "learning_rate": 2.425e-06, "loss": 2.4132, "step": 4850 }, { "epoch": 1.5206896551724138, "grad_norm": 11.360184669494629, "learning_rate": 2.4255e-06, "loss": 2.3275, "step": 4851 }, { "epoch": 1.5210031347962383, "grad_norm": 22.406675338745117, "learning_rate": 2.426e-06, "loss": 2.4512, "step": 4852 }, { "epoch": 1.5213166144200627, "grad_norm": 15.627935409545898, "learning_rate": 2.4265000000000004e-06, "loss": 2.3716, "step": 4853 }, { "epoch": 1.5216300940438872, "grad_norm": 42.328453063964844, "learning_rate": 2.4270000000000002e-06, "loss": 2.3246, "step": 4854 }, { "epoch": 1.5219435736677116, "grad_norm": 16.802934646606445, "learning_rate": 2.4275e-06, "loss": 2.1961, "step": 4855 }, { "epoch": 1.522257053291536, "grad_norm": 17.857563018798828, "learning_rate": 2.428e-06, "loss": 2.2421, "step": 4856 }, { "epoch": 1.5225705329153605, "grad_norm": 12.803881645202637, "learning_rate": 2.4285e-06, "loss": 2.4404, "step": 4857 }, { "epoch": 1.522884012539185, "grad_norm": 18.709604263305664, "learning_rate": 2.4290000000000004e-06, "loss": 2.6333, "step": 4858 }, { "epoch": 1.5231974921630094, "grad_norm": 25.00420379638672, "learning_rate": 2.4295e-06, "loss": 2.4124, "step": 4859 }, { "epoch": 1.5235109717868338, "grad_norm": 13.070844650268555, "learning_rate": 2.43e-06, "loss": 2.4096, "step": 4860 }, { "epoch": 1.5238244514106583, "grad_norm": 20.735050201416016, "learning_rate": 2.4305000000000003e-06, "loss": 2.0841, "step": 4861 }, { "epoch": 1.524137931034483, "grad_norm": 30.139467239379883, "learning_rate": 2.431e-06, "loss": 2.527, "step": 4862 }, { "epoch": 1.5244514106583074, "grad_norm": 31.38504409790039, "learning_rate": 2.4315000000000004e-06, "loss": 2.6405, "step": 4863 }, { "epoch": 1.5247648902821318, "grad_norm": 14.734469413757324, "learning_rate": 2.432e-06, "loss": 2.147, "step": 4864 }, { "epoch": 1.5250783699059562, "grad_norm": 21.783294677734375, "learning_rate": 2.4325e-06, "loss": 2.2365, "step": 4865 }, { "epoch": 1.5253918495297807, "grad_norm": 20.588727951049805, "learning_rate": 2.4330000000000003e-06, "loss": 2.3363, "step": 4866 }, { "epoch": 1.5257053291536051, "grad_norm": 17.136247634887695, "learning_rate": 2.4335000000000005e-06, "loss": 2.4411, "step": 4867 }, { "epoch": 1.5260188087774296, "grad_norm": 11.98731517791748, "learning_rate": 2.4340000000000003e-06, "loss": 2.3687, "step": 4868 }, { "epoch": 1.526332288401254, "grad_norm": 18.351879119873047, "learning_rate": 2.4345e-06, "loss": 2.4216, "step": 4869 }, { "epoch": 1.5266457680250785, "grad_norm": 15.112627029418945, "learning_rate": 2.435e-06, "loss": 2.2907, "step": 4870 }, { "epoch": 1.526959247648903, "grad_norm": 33.07174301147461, "learning_rate": 2.4355000000000002e-06, "loss": 2.5456, "step": 4871 }, { "epoch": 1.5272727272727273, "grad_norm": 13.481283187866211, "learning_rate": 2.4360000000000005e-06, "loss": 2.4042, "step": 4872 }, { "epoch": 1.5275862068965518, "grad_norm": 16.563289642333984, "learning_rate": 2.4365000000000003e-06, "loss": 2.2126, "step": 4873 }, { "epoch": 1.5278996865203762, "grad_norm": 15.1494779586792, "learning_rate": 2.437e-06, "loss": 2.4135, "step": 4874 }, { "epoch": 1.5282131661442007, "grad_norm": 27.80925941467285, "learning_rate": 2.4375e-06, "loss": 2.4168, "step": 4875 }, { "epoch": 1.528526645768025, "grad_norm": 13.605854034423828, "learning_rate": 2.438e-06, "loss": 2.4283, "step": 4876 }, { "epoch": 1.5288401253918495, "grad_norm": 24.888622283935547, "learning_rate": 2.4385000000000005e-06, "loss": 2.242, "step": 4877 }, { "epoch": 1.529153605015674, "grad_norm": 23.643573760986328, "learning_rate": 2.4390000000000003e-06, "loss": 2.5237, "step": 4878 }, { "epoch": 1.5294670846394984, "grad_norm": 14.883064270019531, "learning_rate": 2.4395e-06, "loss": 2.3608, "step": 4879 }, { "epoch": 1.5297805642633229, "grad_norm": 38.991031646728516, "learning_rate": 2.4400000000000004e-06, "loss": 2.3225, "step": 4880 }, { "epoch": 1.5300940438871473, "grad_norm": 112.9505386352539, "learning_rate": 2.4405e-06, "loss": 2.8802, "step": 4881 }, { "epoch": 1.5304075235109718, "grad_norm": 8.81878662109375, "learning_rate": 2.4410000000000004e-06, "loss": 2.2658, "step": 4882 }, { "epoch": 1.5307210031347962, "grad_norm": 27.886585235595703, "learning_rate": 2.4415000000000003e-06, "loss": 2.8646, "step": 4883 }, { "epoch": 1.5310344827586206, "grad_norm": 17.828750610351562, "learning_rate": 2.442e-06, "loss": 2.484, "step": 4884 }, { "epoch": 1.531347962382445, "grad_norm": 16.510709762573242, "learning_rate": 2.4425000000000003e-06, "loss": 2.273, "step": 4885 }, { "epoch": 1.5316614420062695, "grad_norm": 17.356098175048828, "learning_rate": 2.443e-06, "loss": 2.3664, "step": 4886 }, { "epoch": 1.531974921630094, "grad_norm": 22.031627655029297, "learning_rate": 2.4435000000000004e-06, "loss": 2.6871, "step": 4887 }, { "epoch": 1.5322884012539184, "grad_norm": 15.652859687805176, "learning_rate": 2.4440000000000002e-06, "loss": 2.469, "step": 4888 }, { "epoch": 1.5326018808777429, "grad_norm": 18.956735610961914, "learning_rate": 2.4445e-06, "loss": 2.2246, "step": 4889 }, { "epoch": 1.5329153605015673, "grad_norm": 16.19228744506836, "learning_rate": 2.4450000000000003e-06, "loss": 2.7894, "step": 4890 }, { "epoch": 1.5332288401253917, "grad_norm": 36.512962341308594, "learning_rate": 2.4455e-06, "loss": 3.2316, "step": 4891 }, { "epoch": 1.5335423197492162, "grad_norm": 194.37013244628906, "learning_rate": 2.4460000000000004e-06, "loss": 3.3238, "step": 4892 }, { "epoch": 1.5338557993730406, "grad_norm": 27.753740310668945, "learning_rate": 2.4465e-06, "loss": 2.5502, "step": 4893 }, { "epoch": 1.534169278996865, "grad_norm": 18.22922706604004, "learning_rate": 2.447e-06, "loss": 2.3534, "step": 4894 }, { "epoch": 1.5344827586206895, "grad_norm": 14.241109848022461, "learning_rate": 2.4475000000000003e-06, "loss": 2.2446, "step": 4895 }, { "epoch": 1.534796238244514, "grad_norm": 38.08067321777344, "learning_rate": 2.448e-06, "loss": 2.9727, "step": 4896 }, { "epoch": 1.5351097178683386, "grad_norm": 52.77937698364258, "learning_rate": 2.4485000000000004e-06, "loss": 2.7533, "step": 4897 }, { "epoch": 1.535423197492163, "grad_norm": 19.61460304260254, "learning_rate": 2.449e-06, "loss": 2.3417, "step": 4898 }, { "epoch": 1.5357366771159875, "grad_norm": 38.986412048339844, "learning_rate": 2.4495e-06, "loss": 2.6992, "step": 4899 }, { "epoch": 1.536050156739812, "grad_norm": 33.46477508544922, "learning_rate": 2.4500000000000003e-06, "loss": 2.2636, "step": 4900 }, { "epoch": 1.5363636363636364, "grad_norm": 59.14734649658203, "learning_rate": 2.4505e-06, "loss": 2.4793, "step": 4901 }, { "epoch": 1.5366771159874608, "grad_norm": 16.077693939208984, "learning_rate": 2.4510000000000003e-06, "loss": 2.2937, "step": 4902 }, { "epoch": 1.5369905956112853, "grad_norm": 25.643287658691406, "learning_rate": 2.4515e-06, "loss": 2.6153, "step": 4903 }, { "epoch": 1.5373040752351097, "grad_norm": 25.702070236206055, "learning_rate": 2.4520000000000004e-06, "loss": 2.3666, "step": 4904 }, { "epoch": 1.5376175548589341, "grad_norm": 44.60346603393555, "learning_rate": 2.4525000000000002e-06, "loss": 2.481, "step": 4905 }, { "epoch": 1.5379310344827586, "grad_norm": 28.352134704589844, "learning_rate": 2.453e-06, "loss": 2.6935, "step": 4906 }, { "epoch": 1.538244514106583, "grad_norm": 15.22026252746582, "learning_rate": 2.4535000000000003e-06, "loss": 2.3636, "step": 4907 }, { "epoch": 1.5385579937304075, "grad_norm": 29.389680862426758, "learning_rate": 2.454e-06, "loss": 2.2899, "step": 4908 }, { "epoch": 1.5388714733542321, "grad_norm": 16.111085891723633, "learning_rate": 2.4545000000000004e-06, "loss": 2.5567, "step": 4909 }, { "epoch": 1.5391849529780566, "grad_norm": 17.4854679107666, "learning_rate": 2.4550000000000002e-06, "loss": 2.4468, "step": 4910 }, { "epoch": 1.539498432601881, "grad_norm": 26.088729858398438, "learning_rate": 2.4555e-06, "loss": 2.4288, "step": 4911 }, { "epoch": 1.5398119122257055, "grad_norm": 18.86445426940918, "learning_rate": 2.4560000000000003e-06, "loss": 2.2143, "step": 4912 }, { "epoch": 1.54012539184953, "grad_norm": 11.737680435180664, "learning_rate": 2.4565e-06, "loss": 2.4803, "step": 4913 }, { "epoch": 1.5404388714733543, "grad_norm": 13.126084327697754, "learning_rate": 2.4570000000000004e-06, "loss": 2.6635, "step": 4914 }, { "epoch": 1.5407523510971788, "grad_norm": 14.669913291931152, "learning_rate": 2.4575e-06, "loss": 2.1352, "step": 4915 }, { "epoch": 1.5410658307210032, "grad_norm": 8.558469772338867, "learning_rate": 2.458e-06, "loss": 2.4981, "step": 4916 }, { "epoch": 1.5413793103448277, "grad_norm": 37.0605583190918, "learning_rate": 2.4585000000000003e-06, "loss": 2.7061, "step": 4917 }, { "epoch": 1.541692789968652, "grad_norm": 25.720998764038086, "learning_rate": 2.459e-06, "loss": 2.3515, "step": 4918 }, { "epoch": 1.5420062695924766, "grad_norm": 17.545047760009766, "learning_rate": 2.4595000000000003e-06, "loss": 2.2531, "step": 4919 }, { "epoch": 1.542319749216301, "grad_norm": 14.890344619750977, "learning_rate": 2.46e-06, "loss": 2.5966, "step": 4920 }, { "epoch": 1.5426332288401254, "grad_norm": 15.457768440246582, "learning_rate": 2.4605e-06, "loss": 2.1787, "step": 4921 }, { "epoch": 1.5429467084639499, "grad_norm": 10.848100662231445, "learning_rate": 2.4610000000000002e-06, "loss": 2.1762, "step": 4922 }, { "epoch": 1.5432601880877743, "grad_norm": 18.135581970214844, "learning_rate": 2.4615000000000005e-06, "loss": 2.0205, "step": 4923 }, { "epoch": 1.5435736677115988, "grad_norm": 16.26011848449707, "learning_rate": 2.4620000000000003e-06, "loss": 2.4316, "step": 4924 }, { "epoch": 1.5438871473354232, "grad_norm": 19.314348220825195, "learning_rate": 2.4625e-06, "loss": 2.4738, "step": 4925 }, { "epoch": 1.5442006269592476, "grad_norm": 21.843786239624023, "learning_rate": 2.463e-06, "loss": 2.2117, "step": 4926 }, { "epoch": 1.544514106583072, "grad_norm": 13.87489128112793, "learning_rate": 2.4635000000000002e-06, "loss": 2.2631, "step": 4927 }, { "epoch": 1.5448275862068965, "grad_norm": 19.082761764526367, "learning_rate": 2.4640000000000005e-06, "loss": 2.4595, "step": 4928 }, { "epoch": 1.545141065830721, "grad_norm": 49.517860412597656, "learning_rate": 2.4645000000000003e-06, "loss": 2.4809, "step": 4929 }, { "epoch": 1.5454545454545454, "grad_norm": 12.239781379699707, "learning_rate": 2.465e-06, "loss": 2.3829, "step": 4930 }, { "epoch": 1.5457680250783699, "grad_norm": 17.61081886291504, "learning_rate": 2.4655e-06, "loss": 2.2738, "step": 4931 }, { "epoch": 1.5460815047021943, "grad_norm": 9.764055252075195, "learning_rate": 2.466e-06, "loss": 2.2263, "step": 4932 }, { "epoch": 1.5463949843260187, "grad_norm": 16.245092391967773, "learning_rate": 2.4665000000000004e-06, "loss": 2.2703, "step": 4933 }, { "epoch": 1.5467084639498432, "grad_norm": 47.172786712646484, "learning_rate": 2.4670000000000003e-06, "loss": 2.4811, "step": 4934 }, { "epoch": 1.5470219435736676, "grad_norm": 36.98639678955078, "learning_rate": 2.4675e-06, "loss": 2.4616, "step": 4935 }, { "epoch": 1.547335423197492, "grad_norm": 88.0708236694336, "learning_rate": 2.468e-06, "loss": 3.2185, "step": 4936 }, { "epoch": 1.5476489028213165, "grad_norm": 20.56644058227539, "learning_rate": 2.4685e-06, "loss": 2.3528, "step": 4937 }, { "epoch": 1.547962382445141, "grad_norm": 22.679750442504883, "learning_rate": 2.4690000000000004e-06, "loss": 2.2352, "step": 4938 }, { "epoch": 1.5482758620689654, "grad_norm": 19.43711280822754, "learning_rate": 2.4695000000000002e-06, "loss": 2.2477, "step": 4939 }, { "epoch": 1.5485893416927898, "grad_norm": 55.72245788574219, "learning_rate": 2.47e-06, "loss": 2.6602, "step": 4940 }, { "epoch": 1.5489028213166143, "grad_norm": 44.908382415771484, "learning_rate": 2.4705000000000003e-06, "loss": 2.7444, "step": 4941 }, { "epoch": 1.5492163009404387, "grad_norm": 13.042838096618652, "learning_rate": 2.471e-06, "loss": 2.6583, "step": 4942 }, { "epoch": 1.5495297805642632, "grad_norm": 20.27886962890625, "learning_rate": 2.4715000000000004e-06, "loss": 2.3545, "step": 4943 }, { "epoch": 1.5498432601880878, "grad_norm": 17.01507568359375, "learning_rate": 2.4720000000000002e-06, "loss": 2.388, "step": 4944 }, { "epoch": 1.5501567398119123, "grad_norm": 12.890861511230469, "learning_rate": 2.4725e-06, "loss": 2.4163, "step": 4945 }, { "epoch": 1.5504702194357367, "grad_norm": 70.1278305053711, "learning_rate": 2.4730000000000003e-06, "loss": 2.9749, "step": 4946 }, { "epoch": 1.5507836990595611, "grad_norm": 21.87778091430664, "learning_rate": 2.4735e-06, "loss": 2.2962, "step": 4947 }, { "epoch": 1.5510971786833856, "grad_norm": 29.6918888092041, "learning_rate": 2.4740000000000004e-06, "loss": 2.5881, "step": 4948 }, { "epoch": 1.55141065830721, "grad_norm": 30.08953285217285, "learning_rate": 2.4745e-06, "loss": 2.466, "step": 4949 }, { "epoch": 1.5517241379310345, "grad_norm": 122.53662872314453, "learning_rate": 2.475e-06, "loss": 2.8568, "step": 4950 }, { "epoch": 1.552037617554859, "grad_norm": 11.565924644470215, "learning_rate": 2.4755000000000003e-06, "loss": 2.0998, "step": 4951 }, { "epoch": 1.5523510971786834, "grad_norm": 14.999190330505371, "learning_rate": 2.476e-06, "loss": 2.2514, "step": 4952 }, { "epoch": 1.5526645768025078, "grad_norm": 49.61758041381836, "learning_rate": 2.4765000000000003e-06, "loss": 2.2886, "step": 4953 }, { "epoch": 1.5529780564263322, "grad_norm": 38.12818908691406, "learning_rate": 2.477e-06, "loss": 2.3246, "step": 4954 }, { "epoch": 1.5532915360501567, "grad_norm": 19.00833511352539, "learning_rate": 2.4775e-06, "loss": 2.3158, "step": 4955 }, { "epoch": 1.5536050156739813, "grad_norm": 10.164694786071777, "learning_rate": 2.4780000000000002e-06, "loss": 2.3208, "step": 4956 }, { "epoch": 1.5539184952978058, "grad_norm": 19.940147399902344, "learning_rate": 2.4785e-06, "loss": 2.2967, "step": 4957 }, { "epoch": 1.5542319749216302, "grad_norm": 46.21580123901367, "learning_rate": 2.4790000000000003e-06, "loss": 2.3477, "step": 4958 }, { "epoch": 1.5545454545454547, "grad_norm": 13.118972778320312, "learning_rate": 2.4795e-06, "loss": 1.9755, "step": 4959 }, { "epoch": 1.5548589341692791, "grad_norm": 42.64442825317383, "learning_rate": 2.4800000000000004e-06, "loss": 2.5682, "step": 4960 }, { "epoch": 1.5551724137931036, "grad_norm": 13.624639511108398, "learning_rate": 2.4805000000000002e-06, "loss": 2.4254, "step": 4961 }, { "epoch": 1.555485893416928, "grad_norm": 21.184711456298828, "learning_rate": 2.481e-06, "loss": 2.2923, "step": 4962 }, { "epoch": 1.5557993730407524, "grad_norm": 15.5918607711792, "learning_rate": 2.4815000000000003e-06, "loss": 2.0308, "step": 4963 }, { "epoch": 1.5561128526645769, "grad_norm": 19.084026336669922, "learning_rate": 2.482e-06, "loss": 2.5449, "step": 4964 }, { "epoch": 1.5564263322884013, "grad_norm": 14.544201850891113, "learning_rate": 2.4825000000000004e-06, "loss": 2.2216, "step": 4965 }, { "epoch": 1.5567398119122258, "grad_norm": 51.081851959228516, "learning_rate": 2.483e-06, "loss": 2.478, "step": 4966 }, { "epoch": 1.5570532915360502, "grad_norm": 45.603790283203125, "learning_rate": 2.4835e-06, "loss": 2.3317, "step": 4967 }, { "epoch": 1.5573667711598747, "grad_norm": 23.09310531616211, "learning_rate": 2.4840000000000003e-06, "loss": 2.3886, "step": 4968 }, { "epoch": 1.557680250783699, "grad_norm": 36.93960952758789, "learning_rate": 2.4845e-06, "loss": 2.4891, "step": 4969 }, { "epoch": 1.5579937304075235, "grad_norm": 34.09645080566406, "learning_rate": 2.4850000000000003e-06, "loss": 3.6792, "step": 4970 }, { "epoch": 1.558307210031348, "grad_norm": 21.661108016967773, "learning_rate": 2.4855e-06, "loss": 2.4421, "step": 4971 }, { "epoch": 1.5586206896551724, "grad_norm": 31.483070373535156, "learning_rate": 2.486e-06, "loss": 3.076, "step": 4972 }, { "epoch": 1.5589341692789969, "grad_norm": 12.45557689666748, "learning_rate": 2.4865000000000002e-06, "loss": 2.2351, "step": 4973 }, { "epoch": 1.5592476489028213, "grad_norm": 19.75053596496582, "learning_rate": 2.487e-06, "loss": 2.4616, "step": 4974 }, { "epoch": 1.5595611285266457, "grad_norm": 26.347949981689453, "learning_rate": 2.4875000000000003e-06, "loss": 2.1774, "step": 4975 }, { "epoch": 1.5598746081504702, "grad_norm": 55.684993743896484, "learning_rate": 2.488e-06, "loss": 2.4876, "step": 4976 }, { "epoch": 1.5601880877742946, "grad_norm": 46.71628952026367, "learning_rate": 2.4885e-06, "loss": 2.73, "step": 4977 }, { "epoch": 1.560501567398119, "grad_norm": 18.326326370239258, "learning_rate": 2.4890000000000002e-06, "loss": 2.3658, "step": 4978 }, { "epoch": 1.5608150470219435, "grad_norm": 14.326557159423828, "learning_rate": 2.4895e-06, "loss": 2.0581, "step": 4979 }, { "epoch": 1.561128526645768, "grad_norm": 16.678058624267578, "learning_rate": 2.4900000000000003e-06, "loss": 2.356, "step": 4980 }, { "epoch": 1.5614420062695924, "grad_norm": 60.66374588012695, "learning_rate": 2.4905e-06, "loss": 2.6371, "step": 4981 }, { "epoch": 1.5617554858934168, "grad_norm": 19.981624603271484, "learning_rate": 2.491e-06, "loss": 2.2885, "step": 4982 }, { "epoch": 1.5620689655172413, "grad_norm": 161.30465698242188, "learning_rate": 2.4915e-06, "loss": 2.5611, "step": 4983 }, { "epoch": 1.5623824451410657, "grad_norm": 21.220386505126953, "learning_rate": 2.4920000000000005e-06, "loss": 2.3193, "step": 4984 }, { "epoch": 1.5626959247648902, "grad_norm": 114.13520050048828, "learning_rate": 2.4925000000000003e-06, "loss": 2.5983, "step": 4985 }, { "epoch": 1.5630094043887146, "grad_norm": 23.770221710205078, "learning_rate": 2.493e-06, "loss": 2.325, "step": 4986 }, { "epoch": 1.563322884012539, "grad_norm": 18.08496856689453, "learning_rate": 2.4935e-06, "loss": 2.4323, "step": 4987 }, { "epoch": 1.5636363636363635, "grad_norm": 35.3812370300293, "learning_rate": 2.494e-06, "loss": 2.4329, "step": 4988 }, { "epoch": 1.563949843260188, "grad_norm": 21.316783905029297, "learning_rate": 2.4945000000000004e-06, "loss": 2.3539, "step": 4989 }, { "epoch": 1.5642633228840124, "grad_norm": 17.182456970214844, "learning_rate": 2.4950000000000003e-06, "loss": 2.3882, "step": 4990 }, { "epoch": 1.564576802507837, "grad_norm": 18.844646453857422, "learning_rate": 2.4955e-06, "loss": 2.2841, "step": 4991 }, { "epoch": 1.5648902821316615, "grad_norm": 38.493377685546875, "learning_rate": 2.496e-06, "loss": 2.9194, "step": 4992 }, { "epoch": 1.565203761755486, "grad_norm": 14.904982566833496, "learning_rate": 2.4965e-06, "loss": 2.2876, "step": 4993 }, { "epoch": 1.5655172413793104, "grad_norm": 16.828540802001953, "learning_rate": 2.4970000000000004e-06, "loss": 2.4224, "step": 4994 }, { "epoch": 1.5658307210031348, "grad_norm": 25.141571044921875, "learning_rate": 2.4975000000000002e-06, "loss": 2.3427, "step": 4995 }, { "epoch": 1.5661442006269592, "grad_norm": 42.96630859375, "learning_rate": 2.498e-06, "loss": 2.5007, "step": 4996 }, { "epoch": 1.5664576802507837, "grad_norm": 39.685176849365234, "learning_rate": 2.4985000000000003e-06, "loss": 2.5697, "step": 4997 }, { "epoch": 1.5667711598746081, "grad_norm": 34.34861755371094, "learning_rate": 2.499e-06, "loss": 2.1571, "step": 4998 }, { "epoch": 1.5670846394984326, "grad_norm": 33.744346618652344, "learning_rate": 2.4995000000000004e-06, "loss": 2.2946, "step": 4999 }, { "epoch": 1.567398119122257, "grad_norm": 22.578693389892578, "learning_rate": 2.5e-06, "loss": 2.2518, "step": 5000 }, { "epoch": 1.5677115987460815, "grad_norm": 26.208200454711914, "learning_rate": 2.5005e-06, "loss": 2.383, "step": 5001 }, { "epoch": 1.5680250783699061, "grad_norm": 27.605117797851562, "learning_rate": 2.5010000000000003e-06, "loss": 2.265, "step": 5002 }, { "epoch": 1.5683385579937306, "grad_norm": 56.79507064819336, "learning_rate": 2.5015e-06, "loss": 2.8529, "step": 5003 }, { "epoch": 1.568652037617555, "grad_norm": 16.00729751586914, "learning_rate": 2.502e-06, "loss": 2.2934, "step": 5004 }, { "epoch": 1.5689655172413794, "grad_norm": 15.617908477783203, "learning_rate": 2.5024999999999998e-06, "loss": 2.713, "step": 5005 }, { "epoch": 1.569278996865204, "grad_norm": 25.584407806396484, "learning_rate": 2.5030000000000004e-06, "loss": 2.6503, "step": 5006 }, { "epoch": 1.5695924764890283, "grad_norm": 34.239315032958984, "learning_rate": 2.5035000000000003e-06, "loss": 2.1223, "step": 5007 }, { "epoch": 1.5699059561128528, "grad_norm": 42.1166877746582, "learning_rate": 2.5040000000000005e-06, "loss": 2.6302, "step": 5008 }, { "epoch": 1.5702194357366772, "grad_norm": 24.463972091674805, "learning_rate": 2.5045000000000003e-06, "loss": 2.5372, "step": 5009 }, { "epoch": 1.5705329153605017, "grad_norm": 50.24067687988281, "learning_rate": 2.505e-06, "loss": 2.6691, "step": 5010 }, { "epoch": 1.570846394984326, "grad_norm": 33.488521575927734, "learning_rate": 2.5055e-06, "loss": 2.6461, "step": 5011 }, { "epoch": 1.5711598746081505, "grad_norm": 13.341475486755371, "learning_rate": 2.5060000000000002e-06, "loss": 2.2295, "step": 5012 }, { "epoch": 1.571473354231975, "grad_norm": 20.728574752807617, "learning_rate": 2.5065e-06, "loss": 2.5787, "step": 5013 }, { "epoch": 1.5717868338557994, "grad_norm": 21.363683700561523, "learning_rate": 2.507e-06, "loss": 2.4587, "step": 5014 }, { "epoch": 1.5721003134796239, "grad_norm": 23.327157974243164, "learning_rate": 2.5075e-06, "loss": 2.6306, "step": 5015 }, { "epoch": 1.5724137931034483, "grad_norm": 12.774349212646484, "learning_rate": 2.5080000000000004e-06, "loss": 2.285, "step": 5016 }, { "epoch": 1.5727272727272728, "grad_norm": 15.191110610961914, "learning_rate": 2.5085e-06, "loss": 2.406, "step": 5017 }, { "epoch": 1.5730407523510972, "grad_norm": 36.179500579833984, "learning_rate": 2.5090000000000005e-06, "loss": 2.7932, "step": 5018 }, { "epoch": 1.5733542319749216, "grad_norm": 16.943391799926758, "learning_rate": 2.5095000000000003e-06, "loss": 2.5186, "step": 5019 }, { "epoch": 1.573667711598746, "grad_norm": 51.045005798339844, "learning_rate": 2.51e-06, "loss": 2.5958, "step": 5020 }, { "epoch": 1.5739811912225705, "grad_norm": 38.750431060791016, "learning_rate": 2.5105000000000004e-06, "loss": 3.3926, "step": 5021 }, { "epoch": 1.574294670846395, "grad_norm": 12.69948959350586, "learning_rate": 2.511e-06, "loss": 2.1353, "step": 5022 }, { "epoch": 1.5746081504702194, "grad_norm": 24.718772888183594, "learning_rate": 2.5115e-06, "loss": 2.467, "step": 5023 }, { "epoch": 1.5749216300940438, "grad_norm": 17.55669593811035, "learning_rate": 2.512e-06, "loss": 2.4588, "step": 5024 }, { "epoch": 1.5752351097178683, "grad_norm": 7.0970587730407715, "learning_rate": 2.5125e-06, "loss": 2.311, "step": 5025 }, { "epoch": 1.5755485893416927, "grad_norm": 25.72663688659668, "learning_rate": 2.5130000000000003e-06, "loss": 2.9244, "step": 5026 }, { "epoch": 1.5758620689655172, "grad_norm": 29.8355655670166, "learning_rate": 2.5135000000000006e-06, "loss": 2.3327, "step": 5027 }, { "epoch": 1.5761755485893416, "grad_norm": 14.82983684539795, "learning_rate": 2.5140000000000004e-06, "loss": 2.4196, "step": 5028 }, { "epoch": 1.576489028213166, "grad_norm": 23.94029998779297, "learning_rate": 2.5145000000000002e-06, "loss": 2.4763, "step": 5029 }, { "epoch": 1.5768025078369905, "grad_norm": 18.115259170532227, "learning_rate": 2.515e-06, "loss": 2.4923, "step": 5030 }, { "epoch": 1.577115987460815, "grad_norm": 12.026752471923828, "learning_rate": 2.5155000000000003e-06, "loss": 2.3874, "step": 5031 }, { "epoch": 1.5774294670846394, "grad_norm": 29.68210792541504, "learning_rate": 2.516e-06, "loss": 2.8552, "step": 5032 }, { "epoch": 1.5777429467084638, "grad_norm": 15.086202621459961, "learning_rate": 2.5165e-06, "loss": 2.1304, "step": 5033 }, { "epoch": 1.5780564263322883, "grad_norm": 24.435060501098633, "learning_rate": 2.5169999999999998e-06, "loss": 2.339, "step": 5034 }, { "epoch": 1.5783699059561127, "grad_norm": 18.48145866394043, "learning_rate": 2.5175e-06, "loss": 2.3635, "step": 5035 }, { "epoch": 1.5786833855799371, "grad_norm": 21.17127227783203, "learning_rate": 2.5180000000000003e-06, "loss": 2.18, "step": 5036 }, { "epoch": 1.5789968652037616, "grad_norm": 26.255979537963867, "learning_rate": 2.5185000000000005e-06, "loss": 2.5204, "step": 5037 }, { "epoch": 1.5793103448275863, "grad_norm": 260.4898986816406, "learning_rate": 2.5190000000000004e-06, "loss": 2.4303, "step": 5038 }, { "epoch": 1.5796238244514107, "grad_norm": 53.10416793823242, "learning_rate": 2.5195e-06, "loss": 2.4221, "step": 5039 }, { "epoch": 1.5799373040752351, "grad_norm": 18.819931030273438, "learning_rate": 2.52e-06, "loss": 2.5317, "step": 5040 }, { "epoch": 1.5802507836990596, "grad_norm": 26.026790618896484, "learning_rate": 2.5205000000000003e-06, "loss": 1.9606, "step": 5041 }, { "epoch": 1.580564263322884, "grad_norm": 23.0932559967041, "learning_rate": 2.521e-06, "loss": 2.3511, "step": 5042 }, { "epoch": 1.5808777429467085, "grad_norm": 20.482589721679688, "learning_rate": 2.5215e-06, "loss": 2.6349, "step": 5043 }, { "epoch": 1.581191222570533, "grad_norm": 22.782913208007812, "learning_rate": 2.522e-06, "loss": 2.5406, "step": 5044 }, { "epoch": 1.5815047021943573, "grad_norm": 17.503292083740234, "learning_rate": 2.5225e-06, "loss": 2.0932, "step": 5045 }, { "epoch": 1.5818181818181818, "grad_norm": 11.492012023925781, "learning_rate": 2.5230000000000002e-06, "loss": 2.4, "step": 5046 }, { "epoch": 1.5821316614420062, "grad_norm": 13.962905883789062, "learning_rate": 2.5235000000000005e-06, "loss": 2.2931, "step": 5047 }, { "epoch": 1.5824451410658307, "grad_norm": 13.378860473632812, "learning_rate": 2.5240000000000003e-06, "loss": 2.0573, "step": 5048 }, { "epoch": 1.5827586206896553, "grad_norm": 35.7122802734375, "learning_rate": 2.5245e-06, "loss": 2.5505, "step": 5049 }, { "epoch": 1.5830721003134798, "grad_norm": 19.624452590942383, "learning_rate": 2.5250000000000004e-06, "loss": 2.2378, "step": 5050 }, { "epoch": 1.5833855799373042, "grad_norm": 41.233428955078125, "learning_rate": 2.5255000000000002e-06, "loss": 2.5732, "step": 5051 }, { "epoch": 1.5836990595611287, "grad_norm": 15.03004264831543, "learning_rate": 2.526e-06, "loss": 2.6814, "step": 5052 }, { "epoch": 1.584012539184953, "grad_norm": 358.0602722167969, "learning_rate": 2.5265e-06, "loss": 2.7063, "step": 5053 }, { "epoch": 1.5843260188087775, "grad_norm": 33.94559860229492, "learning_rate": 2.527e-06, "loss": 2.775, "step": 5054 }, { "epoch": 1.584639498432602, "grad_norm": 14.883088111877441, "learning_rate": 2.5275e-06, "loss": 2.8067, "step": 5055 }, { "epoch": 1.5849529780564264, "grad_norm": 51.98908615112305, "learning_rate": 2.5280000000000006e-06, "loss": 2.34, "step": 5056 }, { "epoch": 1.5852664576802509, "grad_norm": 21.30936050415039, "learning_rate": 2.5285000000000004e-06, "loss": 2.4628, "step": 5057 }, { "epoch": 1.5855799373040753, "grad_norm": 18.566699981689453, "learning_rate": 2.5290000000000003e-06, "loss": 2.4734, "step": 5058 }, { "epoch": 1.5858934169278998, "grad_norm": 34.335670471191406, "learning_rate": 2.5295e-06, "loss": 2.5809, "step": 5059 }, { "epoch": 1.5862068965517242, "grad_norm": 27.54088592529297, "learning_rate": 2.5300000000000003e-06, "loss": 2.2176, "step": 5060 }, { "epoch": 1.5865203761755486, "grad_norm": 103.77539825439453, "learning_rate": 2.5305e-06, "loss": 2.7314, "step": 5061 }, { "epoch": 1.586833855799373, "grad_norm": 25.50882339477539, "learning_rate": 2.531e-06, "loss": 2.2157, "step": 5062 }, { "epoch": 1.5871473354231975, "grad_norm": 90.0643539428711, "learning_rate": 2.5315000000000002e-06, "loss": 2.1978, "step": 5063 }, { "epoch": 1.587460815047022, "grad_norm": 69.36904907226562, "learning_rate": 2.532e-06, "loss": 2.9566, "step": 5064 }, { "epoch": 1.5877742946708464, "grad_norm": 38.16106414794922, "learning_rate": 2.5325e-06, "loss": 2.3429, "step": 5065 }, { "epoch": 1.5880877742946709, "grad_norm": 29.278573989868164, "learning_rate": 2.5330000000000006e-06, "loss": 2.2119, "step": 5066 }, { "epoch": 1.5884012539184953, "grad_norm": 89.26683044433594, "learning_rate": 2.5335000000000004e-06, "loss": 2.4715, "step": 5067 }, { "epoch": 1.5887147335423197, "grad_norm": 21.57378578186035, "learning_rate": 2.5340000000000002e-06, "loss": 2.3059, "step": 5068 }, { "epoch": 1.5890282131661442, "grad_norm": 35.4509162902832, "learning_rate": 2.5345000000000005e-06, "loss": 2.3637, "step": 5069 }, { "epoch": 1.5893416927899686, "grad_norm": 21.950578689575195, "learning_rate": 2.5350000000000003e-06, "loss": 2.2167, "step": 5070 }, { "epoch": 1.589655172413793, "grad_norm": 52.572540283203125, "learning_rate": 2.5355e-06, "loss": 2.7111, "step": 5071 }, { "epoch": 1.5899686520376175, "grad_norm": 77.87381744384766, "learning_rate": 2.536e-06, "loss": 2.7671, "step": 5072 }, { "epoch": 1.590282131661442, "grad_norm": 15.252903938293457, "learning_rate": 2.5365e-06, "loss": 2.1193, "step": 5073 }, { "epoch": 1.5905956112852664, "grad_norm": 25.91535758972168, "learning_rate": 2.537e-06, "loss": 2.7424, "step": 5074 }, { "epoch": 1.5909090909090908, "grad_norm": 25.11017608642578, "learning_rate": 2.5375e-06, "loss": 2.8389, "step": 5075 }, { "epoch": 1.5912225705329153, "grad_norm": 15.80587100982666, "learning_rate": 2.5380000000000005e-06, "loss": 2.4494, "step": 5076 }, { "epoch": 1.5915360501567397, "grad_norm": 97.02642822265625, "learning_rate": 2.5385000000000003e-06, "loss": 2.6632, "step": 5077 }, { "epoch": 1.5918495297805642, "grad_norm": 16.840496063232422, "learning_rate": 2.539e-06, "loss": 2.4333, "step": 5078 }, { "epoch": 1.5921630094043886, "grad_norm": 27.470895767211914, "learning_rate": 2.5395000000000004e-06, "loss": 2.7408, "step": 5079 }, { "epoch": 1.592476489028213, "grad_norm": 11.823321342468262, "learning_rate": 2.5400000000000002e-06, "loss": 2.129, "step": 5080 }, { "epoch": 1.5927899686520375, "grad_norm": 43.7274169921875, "learning_rate": 2.5405e-06, "loss": 2.368, "step": 5081 }, { "epoch": 1.593103448275862, "grad_norm": 12.855867385864258, "learning_rate": 2.5410000000000003e-06, "loss": 2.1715, "step": 5082 }, { "epoch": 1.5934169278996864, "grad_norm": 12.842061996459961, "learning_rate": 2.5415e-06, "loss": 2.0901, "step": 5083 }, { "epoch": 1.5937304075235108, "grad_norm": 33.203914642333984, "learning_rate": 2.542e-06, "loss": 2.51, "step": 5084 }, { "epoch": 1.5940438871473355, "grad_norm": 15.070207595825195, "learning_rate": 2.5425e-06, "loss": 2.3985, "step": 5085 }, { "epoch": 1.59435736677116, "grad_norm": 24.810152053833008, "learning_rate": 2.5430000000000005e-06, "loss": 2.2131, "step": 5086 }, { "epoch": 1.5946708463949844, "grad_norm": 22.131668090820312, "learning_rate": 2.5435000000000003e-06, "loss": 2.276, "step": 5087 }, { "epoch": 1.5949843260188088, "grad_norm": 19.9857234954834, "learning_rate": 2.5440000000000005e-06, "loss": 2.5681, "step": 5088 }, { "epoch": 1.5952978056426332, "grad_norm": 33.42539978027344, "learning_rate": 2.5445000000000004e-06, "loss": 3.0566, "step": 5089 }, { "epoch": 1.5956112852664577, "grad_norm": 41.5034065246582, "learning_rate": 2.545e-06, "loss": 2.5322, "step": 5090 }, { "epoch": 1.5959247648902821, "grad_norm": 9.396846771240234, "learning_rate": 2.5455e-06, "loss": 2.5493, "step": 5091 }, { "epoch": 1.5962382445141066, "grad_norm": 17.213747024536133, "learning_rate": 2.5460000000000003e-06, "loss": 2.3311, "step": 5092 }, { "epoch": 1.596551724137931, "grad_norm": 25.562578201293945, "learning_rate": 2.5465e-06, "loss": 2.5074, "step": 5093 }, { "epoch": 1.5968652037617554, "grad_norm": 36.171207427978516, "learning_rate": 2.547e-06, "loss": 2.9993, "step": 5094 }, { "epoch": 1.59717868338558, "grad_norm": 19.676252365112305, "learning_rate": 2.5475e-06, "loss": 2.343, "step": 5095 }, { "epoch": 1.5974921630094046, "grad_norm": 14.119227409362793, "learning_rate": 2.5480000000000004e-06, "loss": 2.7787, "step": 5096 }, { "epoch": 1.597805642633229, "grad_norm": 23.950241088867188, "learning_rate": 2.5485000000000002e-06, "loss": 2.3251, "step": 5097 }, { "epoch": 1.5981191222570534, "grad_norm": 23.108810424804688, "learning_rate": 2.5490000000000005e-06, "loss": 2.6356, "step": 5098 }, { "epoch": 1.5984326018808779, "grad_norm": 12.062274932861328, "learning_rate": 2.5495000000000003e-06, "loss": 2.2522, "step": 5099 }, { "epoch": 1.5987460815047023, "grad_norm": 17.99264907836914, "learning_rate": 2.55e-06, "loss": 2.1981, "step": 5100 }, { "epoch": 1.5990595611285268, "grad_norm": 75.7223129272461, "learning_rate": 2.5505000000000004e-06, "loss": 2.5135, "step": 5101 }, { "epoch": 1.5993730407523512, "grad_norm": 16.194011688232422, "learning_rate": 2.5510000000000002e-06, "loss": 2.4676, "step": 5102 }, { "epoch": 1.5996865203761756, "grad_norm": 12.641945838928223, "learning_rate": 2.5515e-06, "loss": 2.1026, "step": 5103 }, { "epoch": 1.6, "grad_norm": 13.797097206115723, "learning_rate": 2.552e-06, "loss": 2.1526, "step": 5104 }, { "epoch": 1.6003134796238245, "grad_norm": 26.577638626098633, "learning_rate": 2.5525e-06, "loss": 2.4505, "step": 5105 }, { "epoch": 1.600626959247649, "grad_norm": 46.32077407836914, "learning_rate": 2.5530000000000004e-06, "loss": 2.4153, "step": 5106 }, { "epoch": 1.6009404388714734, "grad_norm": 33.867061614990234, "learning_rate": 2.5535000000000006e-06, "loss": 2.4845, "step": 5107 }, { "epoch": 1.6012539184952979, "grad_norm": 29.492074966430664, "learning_rate": 2.5540000000000004e-06, "loss": 2.2374, "step": 5108 }, { "epoch": 1.6015673981191223, "grad_norm": 15.982016563415527, "learning_rate": 2.5545000000000003e-06, "loss": 2.1256, "step": 5109 }, { "epoch": 1.6018808777429467, "grad_norm": 23.65928840637207, "learning_rate": 2.555e-06, "loss": 2.5441, "step": 5110 }, { "epoch": 1.6021943573667712, "grad_norm": 31.558923721313477, "learning_rate": 2.5555000000000004e-06, "loss": 2.4304, "step": 5111 }, { "epoch": 1.6025078369905956, "grad_norm": 22.344242095947266, "learning_rate": 2.556e-06, "loss": 2.3139, "step": 5112 }, { "epoch": 1.60282131661442, "grad_norm": 70.99754333496094, "learning_rate": 2.5565e-06, "loss": 3.1882, "step": 5113 }, { "epoch": 1.6031347962382445, "grad_norm": 11.966924667358398, "learning_rate": 2.557e-06, "loss": 2.2407, "step": 5114 }, { "epoch": 1.603448275862069, "grad_norm": 14.561931610107422, "learning_rate": 2.5575e-06, "loss": 2.5646, "step": 5115 }, { "epoch": 1.6037617554858934, "grad_norm": 18.54060173034668, "learning_rate": 2.5580000000000003e-06, "loss": 2.6713, "step": 5116 }, { "epoch": 1.6040752351097178, "grad_norm": 18.346193313598633, "learning_rate": 2.5585000000000006e-06, "loss": 2.2139, "step": 5117 }, { "epoch": 1.6043887147335423, "grad_norm": 16.431509017944336, "learning_rate": 2.5590000000000004e-06, "loss": 2.22, "step": 5118 }, { "epoch": 1.6047021943573667, "grad_norm": 17.47113037109375, "learning_rate": 2.5595000000000002e-06, "loss": 2.3653, "step": 5119 }, { "epoch": 1.6050156739811912, "grad_norm": 28.77997398376465, "learning_rate": 2.56e-06, "loss": 2.9677, "step": 5120 }, { "epoch": 1.6053291536050156, "grad_norm": 14.653223991394043, "learning_rate": 2.5605000000000003e-06, "loss": 2.2911, "step": 5121 }, { "epoch": 1.60564263322884, "grad_norm": 20.274412155151367, "learning_rate": 2.561e-06, "loss": 2.3815, "step": 5122 }, { "epoch": 1.6059561128526645, "grad_norm": 23.261613845825195, "learning_rate": 2.5615e-06, "loss": 2.4761, "step": 5123 }, { "epoch": 1.606269592476489, "grad_norm": 26.230159759521484, "learning_rate": 2.562e-06, "loss": 2.4104, "step": 5124 }, { "epoch": 1.6065830721003134, "grad_norm": 12.218854904174805, "learning_rate": 2.5625e-06, "loss": 2.2321, "step": 5125 }, { "epoch": 1.6068965517241378, "grad_norm": 30.996965408325195, "learning_rate": 2.563e-06, "loss": 2.6327, "step": 5126 }, { "epoch": 1.6072100313479623, "grad_norm": 21.883953094482422, "learning_rate": 2.5635000000000005e-06, "loss": 2.3285, "step": 5127 }, { "epoch": 1.6075235109717867, "grad_norm": 22.208730697631836, "learning_rate": 2.5640000000000004e-06, "loss": 2.2301, "step": 5128 }, { "epoch": 1.6078369905956111, "grad_norm": 41.55567169189453, "learning_rate": 2.5645e-06, "loss": 2.5755, "step": 5129 }, { "epoch": 1.6081504702194356, "grad_norm": 36.985572814941406, "learning_rate": 2.5650000000000004e-06, "loss": 2.9383, "step": 5130 }, { "epoch": 1.6084639498432602, "grad_norm": 10.896848678588867, "learning_rate": 2.5655000000000003e-06, "loss": 2.278, "step": 5131 }, { "epoch": 1.6087774294670847, "grad_norm": 26.881319046020508, "learning_rate": 2.566e-06, "loss": 2.5914, "step": 5132 }, { "epoch": 1.6090909090909091, "grad_norm": 43.129730224609375, "learning_rate": 2.5665e-06, "loss": 2.4239, "step": 5133 }, { "epoch": 1.6094043887147336, "grad_norm": 35.279144287109375, "learning_rate": 2.567e-06, "loss": 2.339, "step": 5134 }, { "epoch": 1.609717868338558, "grad_norm": 20.479421615600586, "learning_rate": 2.5675e-06, "loss": 2.5922, "step": 5135 }, { "epoch": 1.6100313479623825, "grad_norm": 47.79603576660156, "learning_rate": 2.568e-06, "loss": 2.3022, "step": 5136 }, { "epoch": 1.610344827586207, "grad_norm": 21.276691436767578, "learning_rate": 2.5685000000000005e-06, "loss": 2.4997, "step": 5137 }, { "epoch": 1.6106583072100313, "grad_norm": 32.518619537353516, "learning_rate": 2.5690000000000003e-06, "loss": 2.3542, "step": 5138 }, { "epoch": 1.6109717868338558, "grad_norm": 11.362214088439941, "learning_rate": 2.5695e-06, "loss": 2.4764, "step": 5139 }, { "epoch": 1.6112852664576802, "grad_norm": 18.297727584838867, "learning_rate": 2.5700000000000004e-06, "loss": 2.359, "step": 5140 }, { "epoch": 1.6115987460815047, "grad_norm": 35.51220703125, "learning_rate": 2.5705e-06, "loss": 2.5266, "step": 5141 }, { "epoch": 1.611912225705329, "grad_norm": 37.287601470947266, "learning_rate": 2.571e-06, "loss": 2.4847, "step": 5142 }, { "epoch": 1.6122257053291538, "grad_norm": 28.55080223083496, "learning_rate": 2.5715000000000003e-06, "loss": 2.6724, "step": 5143 }, { "epoch": 1.6125391849529782, "grad_norm": 24.460296630859375, "learning_rate": 2.572e-06, "loss": 2.4058, "step": 5144 }, { "epoch": 1.6128526645768027, "grad_norm": 13.860005378723145, "learning_rate": 2.5725e-06, "loss": 2.5218, "step": 5145 }, { "epoch": 1.613166144200627, "grad_norm": 20.911226272583008, "learning_rate": 2.5729999999999998e-06, "loss": 2.738, "step": 5146 }, { "epoch": 1.6134796238244515, "grad_norm": 20.03626251220703, "learning_rate": 2.5735000000000004e-06, "loss": 2.4337, "step": 5147 }, { "epoch": 1.613793103448276, "grad_norm": 15.790233612060547, "learning_rate": 2.5740000000000003e-06, "loss": 2.2734, "step": 5148 }, { "epoch": 1.6141065830721004, "grad_norm": 15.679661750793457, "learning_rate": 2.5745000000000005e-06, "loss": 2.7092, "step": 5149 }, { "epoch": 1.6144200626959249, "grad_norm": 16.354408264160156, "learning_rate": 2.5750000000000003e-06, "loss": 2.1285, "step": 5150 }, { "epoch": 1.6147335423197493, "grad_norm": 26.114641189575195, "learning_rate": 2.5755e-06, "loss": 2.459, "step": 5151 }, { "epoch": 1.6150470219435737, "grad_norm": 36.885169982910156, "learning_rate": 2.576e-06, "loss": 2.3374, "step": 5152 }, { "epoch": 1.6153605015673982, "grad_norm": 18.303848266601562, "learning_rate": 2.5765000000000002e-06, "loss": 2.8493, "step": 5153 }, { "epoch": 1.6156739811912226, "grad_norm": 29.80558204650879, "learning_rate": 2.577e-06, "loss": 2.535, "step": 5154 }, { "epoch": 1.615987460815047, "grad_norm": 12.516839981079102, "learning_rate": 2.5775e-06, "loss": 2.1776, "step": 5155 }, { "epoch": 1.6163009404388715, "grad_norm": 15.919252395629883, "learning_rate": 2.578e-06, "loss": 2.1872, "step": 5156 }, { "epoch": 1.616614420062696, "grad_norm": 11.775092124938965, "learning_rate": 2.5785000000000004e-06, "loss": 2.1423, "step": 5157 }, { "epoch": 1.6169278996865204, "grad_norm": 31.506715774536133, "learning_rate": 2.579e-06, "loss": 2.2092, "step": 5158 }, { "epoch": 1.6172413793103448, "grad_norm": 43.23335266113281, "learning_rate": 2.5795000000000005e-06, "loss": 2.4301, "step": 5159 }, { "epoch": 1.6175548589341693, "grad_norm": 17.751665115356445, "learning_rate": 2.5800000000000003e-06, "loss": 2.2749, "step": 5160 }, { "epoch": 1.6178683385579937, "grad_norm": 34.12142562866211, "learning_rate": 2.5805e-06, "loss": 2.522, "step": 5161 }, { "epoch": 1.6181818181818182, "grad_norm": 25.955385208129883, "learning_rate": 2.5810000000000004e-06, "loss": 2.3837, "step": 5162 }, { "epoch": 1.6184952978056426, "grad_norm": 18.65509605407715, "learning_rate": 2.5815e-06, "loss": 2.3719, "step": 5163 }, { "epoch": 1.618808777429467, "grad_norm": 13.978825569152832, "learning_rate": 2.582e-06, "loss": 2.1543, "step": 5164 }, { "epoch": 1.6191222570532915, "grad_norm": 32.66872024536133, "learning_rate": 2.5825e-06, "loss": 2.4058, "step": 5165 }, { "epoch": 1.619435736677116, "grad_norm": 13.46345329284668, "learning_rate": 2.583e-06, "loss": 2.4266, "step": 5166 }, { "epoch": 1.6197492163009404, "grad_norm": 21.65625, "learning_rate": 2.5835000000000003e-06, "loss": 2.2746, "step": 5167 }, { "epoch": 1.6200626959247648, "grad_norm": 47.85426712036133, "learning_rate": 2.5840000000000006e-06, "loss": 2.7069, "step": 5168 }, { "epoch": 1.6203761755485893, "grad_norm": 12.973807334899902, "learning_rate": 2.5845000000000004e-06, "loss": 2.5359, "step": 5169 }, { "epoch": 1.6206896551724137, "grad_norm": 59.425838470458984, "learning_rate": 2.5850000000000002e-06, "loss": 2.5992, "step": 5170 }, { "epoch": 1.6210031347962381, "grad_norm": 42.35601043701172, "learning_rate": 2.5855e-06, "loss": 3.1424, "step": 5171 }, { "epoch": 1.6213166144200626, "grad_norm": 18.894989013671875, "learning_rate": 2.5860000000000003e-06, "loss": 2.4605, "step": 5172 }, { "epoch": 1.621630094043887, "grad_norm": 24.31694221496582, "learning_rate": 2.5865e-06, "loss": 2.3184, "step": 5173 }, { "epoch": 1.6219435736677115, "grad_norm": 44.74848175048828, "learning_rate": 2.587e-06, "loss": 2.7195, "step": 5174 }, { "epoch": 1.622257053291536, "grad_norm": 49.74072265625, "learning_rate": 2.5875000000000002e-06, "loss": 2.1722, "step": 5175 }, { "epoch": 1.6225705329153604, "grad_norm": 48.17268753051758, "learning_rate": 2.588e-06, "loss": 2.392, "step": 5176 }, { "epoch": 1.6228840125391848, "grad_norm": 10.797576904296875, "learning_rate": 2.5885000000000003e-06, "loss": 2.3357, "step": 5177 }, { "epoch": 1.6231974921630095, "grad_norm": 18.476648330688477, "learning_rate": 2.5890000000000005e-06, "loss": 2.2672, "step": 5178 }, { "epoch": 1.623510971786834, "grad_norm": 18.288114547729492, "learning_rate": 2.5895000000000004e-06, "loss": 2.0863, "step": 5179 }, { "epoch": 1.6238244514106583, "grad_norm": 19.165102005004883, "learning_rate": 2.59e-06, "loss": 2.6391, "step": 5180 }, { "epoch": 1.6241379310344828, "grad_norm": 14.676997184753418, "learning_rate": 2.5905000000000004e-06, "loss": 2.2653, "step": 5181 }, { "epoch": 1.6244514106583072, "grad_norm": 20.614850997924805, "learning_rate": 2.5910000000000003e-06, "loss": 2.2439, "step": 5182 }, { "epoch": 1.6247648902821317, "grad_norm": 97.47027587890625, "learning_rate": 2.5915e-06, "loss": 2.2794, "step": 5183 }, { "epoch": 1.625078369905956, "grad_norm": 24.965999603271484, "learning_rate": 2.592e-06, "loss": 2.1568, "step": 5184 }, { "epoch": 1.6253918495297806, "grad_norm": 29.646013259887695, "learning_rate": 2.5925e-06, "loss": 2.3762, "step": 5185 }, { "epoch": 1.625705329153605, "grad_norm": 12.286757469177246, "learning_rate": 2.593e-06, "loss": 2.0198, "step": 5186 }, { "epoch": 1.6260188087774294, "grad_norm": 13.670647621154785, "learning_rate": 2.5935000000000007e-06, "loss": 2.2541, "step": 5187 }, { "epoch": 1.6263322884012539, "grad_norm": 14.53826904296875, "learning_rate": 2.5940000000000005e-06, "loss": 2.3772, "step": 5188 }, { "epoch": 1.6266457680250783, "grad_norm": 14.540467262268066, "learning_rate": 2.5945000000000003e-06, "loss": 2.2134, "step": 5189 }, { "epoch": 1.626959247648903, "grad_norm": 11.156545639038086, "learning_rate": 2.595e-06, "loss": 2.4183, "step": 5190 }, { "epoch": 1.6272727272727274, "grad_norm": 39.29147720336914, "learning_rate": 2.5955000000000004e-06, "loss": 2.5317, "step": 5191 }, { "epoch": 1.6275862068965519, "grad_norm": 23.9459171295166, "learning_rate": 2.5960000000000002e-06, "loss": 2.36, "step": 5192 }, { "epoch": 1.6278996865203763, "grad_norm": 17.24390411376953, "learning_rate": 2.5965e-06, "loss": 2.7135, "step": 5193 }, { "epoch": 1.6282131661442008, "grad_norm": 18.405895233154297, "learning_rate": 2.597e-06, "loss": 2.3673, "step": 5194 }, { "epoch": 1.6285266457680252, "grad_norm": 13.795333862304688, "learning_rate": 2.5975e-06, "loss": 2.3724, "step": 5195 }, { "epoch": 1.6288401253918496, "grad_norm": 20.32740592956543, "learning_rate": 2.598e-06, "loss": 2.2114, "step": 5196 }, { "epoch": 1.629153605015674, "grad_norm": 11.09119987487793, "learning_rate": 2.5985000000000006e-06, "loss": 2.2799, "step": 5197 }, { "epoch": 1.6294670846394985, "grad_norm": 37.84120559692383, "learning_rate": 2.5990000000000004e-06, "loss": 2.7828, "step": 5198 }, { "epoch": 1.629780564263323, "grad_norm": 16.424787521362305, "learning_rate": 2.5995000000000003e-06, "loss": 2.335, "step": 5199 }, { "epoch": 1.6300940438871474, "grad_norm": 19.634714126586914, "learning_rate": 2.6e-06, "loss": 2.3331, "step": 5200 }, { "epoch": 1.6304075235109718, "grad_norm": 16.779815673828125, "learning_rate": 2.6005000000000003e-06, "loss": 2.168, "step": 5201 }, { "epoch": 1.6307210031347963, "grad_norm": 27.600574493408203, "learning_rate": 2.601e-06, "loss": 2.5557, "step": 5202 }, { "epoch": 1.6310344827586207, "grad_norm": 38.47439193725586, "learning_rate": 2.6015e-06, "loss": 2.7327, "step": 5203 }, { "epoch": 1.6313479623824452, "grad_norm": 19.682342529296875, "learning_rate": 2.6020000000000002e-06, "loss": 2.5308, "step": 5204 }, { "epoch": 1.6316614420062696, "grad_norm": 53.40803909301758, "learning_rate": 2.6025e-06, "loss": 2.278, "step": 5205 }, { "epoch": 1.631974921630094, "grad_norm": 16.326505661010742, "learning_rate": 2.603e-06, "loss": 2.6542, "step": 5206 }, { "epoch": 1.6322884012539185, "grad_norm": 13.8661527633667, "learning_rate": 2.6035000000000006e-06, "loss": 2.1928, "step": 5207 }, { "epoch": 1.632601880877743, "grad_norm": 16.33017349243164, "learning_rate": 2.6040000000000004e-06, "loss": 2.3582, "step": 5208 }, { "epoch": 1.6329153605015674, "grad_norm": 51.326820373535156, "learning_rate": 2.6045000000000002e-06, "loss": 2.0451, "step": 5209 }, { "epoch": 1.6332288401253918, "grad_norm": 17.90423011779785, "learning_rate": 2.6050000000000005e-06, "loss": 2.5162, "step": 5210 }, { "epoch": 1.6335423197492163, "grad_norm": 40.12839126586914, "learning_rate": 2.6055000000000003e-06, "loss": 2.2756, "step": 5211 }, { "epoch": 1.6338557993730407, "grad_norm": 17.745290756225586, "learning_rate": 2.606e-06, "loss": 2.1294, "step": 5212 }, { "epoch": 1.6341692789968651, "grad_norm": 13.970556259155273, "learning_rate": 2.6065e-06, "loss": 2.1501, "step": 5213 }, { "epoch": 1.6344827586206896, "grad_norm": 22.84992027282715, "learning_rate": 2.607e-06, "loss": 2.8207, "step": 5214 }, { "epoch": 1.634796238244514, "grad_norm": 43.31232452392578, "learning_rate": 2.6075e-06, "loss": 3.122, "step": 5215 }, { "epoch": 1.6351097178683385, "grad_norm": 14.463607788085938, "learning_rate": 2.608e-06, "loss": 2.2365, "step": 5216 }, { "epoch": 1.635423197492163, "grad_norm": 23.331586837768555, "learning_rate": 2.6085000000000005e-06, "loss": 2.984, "step": 5217 }, { "epoch": 1.6357366771159874, "grad_norm": 10.62083625793457, "learning_rate": 2.6090000000000003e-06, "loss": 2.1679, "step": 5218 }, { "epoch": 1.6360501567398118, "grad_norm": 40.45970153808594, "learning_rate": 2.6095e-06, "loss": 2.3978, "step": 5219 }, { "epoch": 1.6363636363636362, "grad_norm": 12.194794654846191, "learning_rate": 2.6100000000000004e-06, "loss": 2.4125, "step": 5220 }, { "epoch": 1.6366771159874607, "grad_norm": 12.354119300842285, "learning_rate": 2.6105000000000002e-06, "loss": 2.2636, "step": 5221 }, { "epoch": 1.6369905956112851, "grad_norm": 15.73239517211914, "learning_rate": 2.611e-06, "loss": 2.2861, "step": 5222 }, { "epoch": 1.6373040752351096, "grad_norm": 13.624801635742188, "learning_rate": 2.6115000000000003e-06, "loss": 2.3806, "step": 5223 }, { "epoch": 1.637617554858934, "grad_norm": 38.89716720581055, "learning_rate": 2.612e-06, "loss": 2.7064, "step": 5224 }, { "epoch": 1.6379310344827587, "grad_norm": 104.40101623535156, "learning_rate": 2.6125e-06, "loss": 3.6707, "step": 5225 }, { "epoch": 1.6382445141065831, "grad_norm": 15.195369720458984, "learning_rate": 2.613e-06, "loss": 2.2893, "step": 5226 }, { "epoch": 1.6385579937304076, "grad_norm": 63.90957260131836, "learning_rate": 2.6135000000000005e-06, "loss": 2.4651, "step": 5227 }, { "epoch": 1.638871473354232, "grad_norm": 13.292878150939941, "learning_rate": 2.6140000000000003e-06, "loss": 2.3637, "step": 5228 }, { "epoch": 1.6391849529780564, "grad_norm": 25.40592384338379, "learning_rate": 2.6145000000000005e-06, "loss": 2.0852, "step": 5229 }, { "epoch": 1.6394984326018809, "grad_norm": 13.220362663269043, "learning_rate": 2.6150000000000004e-06, "loss": 2.1517, "step": 5230 }, { "epoch": 1.6398119122257053, "grad_norm": 11.700094223022461, "learning_rate": 2.6155e-06, "loss": 2.3451, "step": 5231 }, { "epoch": 1.6401253918495298, "grad_norm": 15.420618057250977, "learning_rate": 2.616e-06, "loss": 2.141, "step": 5232 }, { "epoch": 1.6404388714733542, "grad_norm": 17.454343795776367, "learning_rate": 2.6165000000000003e-06, "loss": 2.2411, "step": 5233 }, { "epoch": 1.6407523510971787, "grad_norm": 13.41075325012207, "learning_rate": 2.617e-06, "loss": 2.3167, "step": 5234 }, { "epoch": 1.641065830721003, "grad_norm": 24.065536499023438, "learning_rate": 2.6175e-06, "loss": 2.5559, "step": 5235 }, { "epoch": 1.6413793103448275, "grad_norm": 33.53857421875, "learning_rate": 2.618e-06, "loss": 2.0863, "step": 5236 }, { "epoch": 1.6416927899686522, "grad_norm": 69.07901763916016, "learning_rate": 2.6185000000000004e-06, "loss": 2.4614, "step": 5237 }, { "epoch": 1.6420062695924766, "grad_norm": 29.503938674926758, "learning_rate": 2.6190000000000003e-06, "loss": 2.262, "step": 5238 }, { "epoch": 1.642319749216301, "grad_norm": 13.055183410644531, "learning_rate": 2.6195000000000005e-06, "loss": 2.0189, "step": 5239 }, { "epoch": 1.6426332288401255, "grad_norm": 19.09827995300293, "learning_rate": 2.6200000000000003e-06, "loss": 2.3149, "step": 5240 }, { "epoch": 1.64294670846395, "grad_norm": 16.848682403564453, "learning_rate": 2.6205e-06, "loss": 2.3306, "step": 5241 }, { "epoch": 1.6432601880877744, "grad_norm": 37.780418395996094, "learning_rate": 2.6210000000000004e-06, "loss": 2.4372, "step": 5242 }, { "epoch": 1.6435736677115989, "grad_norm": 14.445202827453613, "learning_rate": 2.6215000000000002e-06, "loss": 2.3163, "step": 5243 }, { "epoch": 1.6438871473354233, "grad_norm": 32.84733200073242, "learning_rate": 2.622e-06, "loss": 2.4044, "step": 5244 }, { "epoch": 1.6442006269592477, "grad_norm": 11.114374160766602, "learning_rate": 2.6225e-06, "loss": 2.394, "step": 5245 }, { "epoch": 1.6445141065830722, "grad_norm": 11.89560317993164, "learning_rate": 2.623e-06, "loss": 2.3219, "step": 5246 }, { "epoch": 1.6448275862068966, "grad_norm": 39.383079528808594, "learning_rate": 2.6235000000000004e-06, "loss": 2.4139, "step": 5247 }, { "epoch": 1.645141065830721, "grad_norm": 59.19121551513672, "learning_rate": 2.6240000000000006e-06, "loss": 2.2169, "step": 5248 }, { "epoch": 1.6454545454545455, "grad_norm": 34.60609436035156, "learning_rate": 2.6245000000000005e-06, "loss": 2.5338, "step": 5249 }, { "epoch": 1.64576802507837, "grad_norm": 21.994110107421875, "learning_rate": 2.6250000000000003e-06, "loss": 2.6902, "step": 5250 }, { "epoch": 1.6460815047021944, "grad_norm": 17.888416290283203, "learning_rate": 2.6255e-06, "loss": 2.2012, "step": 5251 }, { "epoch": 1.6463949843260188, "grad_norm": 126.86553192138672, "learning_rate": 2.6260000000000004e-06, "loss": 2.5988, "step": 5252 }, { "epoch": 1.6467084639498433, "grad_norm": 20.460132598876953, "learning_rate": 2.6265e-06, "loss": 2.1957, "step": 5253 }, { "epoch": 1.6470219435736677, "grad_norm": 16.447261810302734, "learning_rate": 2.627e-06, "loss": 2.5429, "step": 5254 }, { "epoch": 1.6473354231974922, "grad_norm": 23.05217170715332, "learning_rate": 2.6275000000000003e-06, "loss": 2.5669, "step": 5255 }, { "epoch": 1.6476489028213166, "grad_norm": 27.731290817260742, "learning_rate": 2.628e-06, "loss": 2.3274, "step": 5256 }, { "epoch": 1.647962382445141, "grad_norm": 18.66347312927246, "learning_rate": 2.6285e-06, "loss": 2.2186, "step": 5257 }, { "epoch": 1.6482758620689655, "grad_norm": 12.416943550109863, "learning_rate": 2.6290000000000006e-06, "loss": 2.0702, "step": 5258 }, { "epoch": 1.64858934169279, "grad_norm": 58.2420539855957, "learning_rate": 2.6295000000000004e-06, "loss": 2.832, "step": 5259 }, { "epoch": 1.6489028213166144, "grad_norm": 18.872852325439453, "learning_rate": 2.6300000000000002e-06, "loss": 2.4728, "step": 5260 }, { "epoch": 1.6492163009404388, "grad_norm": 27.66944122314453, "learning_rate": 2.6305000000000005e-06, "loss": 2.3424, "step": 5261 }, { "epoch": 1.6495297805642632, "grad_norm": 21.679662704467773, "learning_rate": 2.6310000000000003e-06, "loss": 1.9519, "step": 5262 }, { "epoch": 1.6498432601880877, "grad_norm": 31.611692428588867, "learning_rate": 2.6315e-06, "loss": 2.9403, "step": 5263 }, { "epoch": 1.6501567398119121, "grad_norm": 31.275747299194336, "learning_rate": 2.632e-06, "loss": 2.1951, "step": 5264 }, { "epoch": 1.6504702194357366, "grad_norm": 15.326471328735352, "learning_rate": 2.6325e-06, "loss": 2.2404, "step": 5265 }, { "epoch": 1.650783699059561, "grad_norm": 16.914216995239258, "learning_rate": 2.633e-06, "loss": 2.2977, "step": 5266 }, { "epoch": 1.6510971786833855, "grad_norm": 14.189790725708008, "learning_rate": 2.6335e-06, "loss": 2.2385, "step": 5267 }, { "epoch": 1.65141065830721, "grad_norm": 74.67144775390625, "learning_rate": 2.6340000000000005e-06, "loss": 2.6448, "step": 5268 }, { "epoch": 1.6517241379310343, "grad_norm": 7.973588466644287, "learning_rate": 2.6345000000000004e-06, "loss": 2.3253, "step": 5269 }, { "epoch": 1.6520376175548588, "grad_norm": 17.322906494140625, "learning_rate": 2.635e-06, "loss": 2.3823, "step": 5270 }, { "epoch": 1.6523510971786832, "grad_norm": 20.03339385986328, "learning_rate": 2.6355000000000004e-06, "loss": 2.4951, "step": 5271 }, { "epoch": 1.652664576802508, "grad_norm": 35.52084732055664, "learning_rate": 2.6360000000000003e-06, "loss": 2.3325, "step": 5272 }, { "epoch": 1.6529780564263323, "grad_norm": 16.430076599121094, "learning_rate": 2.6365e-06, "loss": 2.3975, "step": 5273 }, { "epoch": 1.6532915360501568, "grad_norm": 14.450660705566406, "learning_rate": 2.637e-06, "loss": 2.2319, "step": 5274 }, { "epoch": 1.6536050156739812, "grad_norm": 16.825803756713867, "learning_rate": 2.6375e-06, "loss": 2.1143, "step": 5275 }, { "epoch": 1.6539184952978057, "grad_norm": 38.360015869140625, "learning_rate": 2.638e-06, "loss": 2.2986, "step": 5276 }, { "epoch": 1.65423197492163, "grad_norm": 19.999114990234375, "learning_rate": 2.6385e-06, "loss": 2.7058, "step": 5277 }, { "epoch": 1.6545454545454545, "grad_norm": 22.12860107421875, "learning_rate": 2.6390000000000005e-06, "loss": 2.3911, "step": 5278 }, { "epoch": 1.654858934169279, "grad_norm": 11.342117309570312, "learning_rate": 2.6395000000000003e-06, "loss": 2.0603, "step": 5279 }, { "epoch": 1.6551724137931034, "grad_norm": 25.63351821899414, "learning_rate": 2.64e-06, "loss": 2.3815, "step": 5280 }, { "epoch": 1.6554858934169279, "grad_norm": 10.051604270935059, "learning_rate": 2.6405000000000004e-06, "loss": 2.2003, "step": 5281 }, { "epoch": 1.6557993730407523, "grad_norm": 14.631726264953613, "learning_rate": 2.641e-06, "loss": 2.3557, "step": 5282 }, { "epoch": 1.6561128526645768, "grad_norm": 21.072256088256836, "learning_rate": 2.6415e-06, "loss": 2.607, "step": 5283 }, { "epoch": 1.6564263322884014, "grad_norm": 21.111793518066406, "learning_rate": 2.6420000000000003e-06, "loss": 2.3936, "step": 5284 }, { "epoch": 1.6567398119122259, "grad_norm": 31.564790725708008, "learning_rate": 2.6425e-06, "loss": 2.3416, "step": 5285 }, { "epoch": 1.6570532915360503, "grad_norm": 15.514556884765625, "learning_rate": 2.643e-06, "loss": 2.462, "step": 5286 }, { "epoch": 1.6573667711598747, "grad_norm": 49.09897994995117, "learning_rate": 2.6434999999999998e-06, "loss": 2.4764, "step": 5287 }, { "epoch": 1.6576802507836992, "grad_norm": 22.08146095275879, "learning_rate": 2.6440000000000004e-06, "loss": 2.3708, "step": 5288 }, { "epoch": 1.6579937304075236, "grad_norm": 37.27674102783203, "learning_rate": 2.6445000000000003e-06, "loss": 2.382, "step": 5289 }, { "epoch": 1.658307210031348, "grad_norm": 102.43901062011719, "learning_rate": 2.6450000000000005e-06, "loss": 2.4155, "step": 5290 }, { "epoch": 1.6586206896551725, "grad_norm": 22.695934295654297, "learning_rate": 2.6455000000000003e-06, "loss": 2.4774, "step": 5291 }, { "epoch": 1.658934169278997, "grad_norm": 57.112098693847656, "learning_rate": 2.646e-06, "loss": 4.0764, "step": 5292 }, { "epoch": 1.6592476489028214, "grad_norm": 18.36469268798828, "learning_rate": 2.6465e-06, "loss": 2.3275, "step": 5293 }, { "epoch": 1.6595611285266458, "grad_norm": 32.93498992919922, "learning_rate": 2.6470000000000002e-06, "loss": 2.4024, "step": 5294 }, { "epoch": 1.6598746081504703, "grad_norm": 15.336421966552734, "learning_rate": 2.6475e-06, "loss": 2.4848, "step": 5295 }, { "epoch": 1.6601880877742947, "grad_norm": 14.629399299621582, "learning_rate": 2.648e-06, "loss": 2.3644, "step": 5296 }, { "epoch": 1.6605015673981192, "grad_norm": 12.701495170593262, "learning_rate": 2.6485e-06, "loss": 2.1344, "step": 5297 }, { "epoch": 1.6608150470219436, "grad_norm": 13.902689933776855, "learning_rate": 2.6490000000000004e-06, "loss": 2.3507, "step": 5298 }, { "epoch": 1.661128526645768, "grad_norm": 21.512542724609375, "learning_rate": 2.6495000000000002e-06, "loss": 2.6414, "step": 5299 }, { "epoch": 1.6614420062695925, "grad_norm": 25.147584915161133, "learning_rate": 2.6500000000000005e-06, "loss": 2.5541, "step": 5300 }, { "epoch": 1.661755485893417, "grad_norm": 13.754753112792969, "learning_rate": 2.6505000000000003e-06, "loss": 2.1547, "step": 5301 }, { "epoch": 1.6620689655172414, "grad_norm": 12.000065803527832, "learning_rate": 2.651e-06, "loss": 2.3954, "step": 5302 }, { "epoch": 1.6623824451410658, "grad_norm": 26.096010208129883, "learning_rate": 2.6515000000000004e-06, "loss": 2.3964, "step": 5303 }, { "epoch": 1.6626959247648903, "grad_norm": 12.003847122192383, "learning_rate": 2.652e-06, "loss": 2.0897, "step": 5304 }, { "epoch": 1.6630094043887147, "grad_norm": 9.624919891357422, "learning_rate": 2.6525e-06, "loss": 2.3602, "step": 5305 }, { "epoch": 1.6633228840125391, "grad_norm": 17.856704711914062, "learning_rate": 2.653e-06, "loss": 2.5554, "step": 5306 }, { "epoch": 1.6636363636363636, "grad_norm": 13.328954696655273, "learning_rate": 2.6535e-06, "loss": 2.316, "step": 5307 }, { "epoch": 1.663949843260188, "grad_norm": 18.4957218170166, "learning_rate": 2.6540000000000003e-06, "loss": 2.3456, "step": 5308 }, { "epoch": 1.6642633228840125, "grad_norm": 27.82377815246582, "learning_rate": 2.6545000000000006e-06, "loss": 2.9891, "step": 5309 }, { "epoch": 1.664576802507837, "grad_norm": 23.645793914794922, "learning_rate": 2.6550000000000004e-06, "loss": 2.3558, "step": 5310 }, { "epoch": 1.6648902821316613, "grad_norm": 12.790231704711914, "learning_rate": 2.6555000000000002e-06, "loss": 2.1346, "step": 5311 }, { "epoch": 1.6652037617554858, "grad_norm": 33.11914825439453, "learning_rate": 2.656e-06, "loss": 2.3526, "step": 5312 }, { "epoch": 1.6655172413793102, "grad_norm": 17.447622299194336, "learning_rate": 2.6565000000000003e-06, "loss": 2.4022, "step": 5313 }, { "epoch": 1.6658307210031347, "grad_norm": 25.620925903320312, "learning_rate": 2.657e-06, "loss": 2.3914, "step": 5314 }, { "epoch": 1.6661442006269591, "grad_norm": 17.788557052612305, "learning_rate": 2.6575e-06, "loss": 2.2504, "step": 5315 }, { "epoch": 1.6664576802507836, "grad_norm": 34.506771087646484, "learning_rate": 2.6580000000000002e-06, "loss": 2.6078, "step": 5316 }, { "epoch": 1.666771159874608, "grad_norm": 15.220390319824219, "learning_rate": 2.6585e-06, "loss": 2.3253, "step": 5317 }, { "epoch": 1.6670846394984324, "grad_norm": 60.35187911987305, "learning_rate": 2.6590000000000003e-06, "loss": 2.5263, "step": 5318 }, { "epoch": 1.667398119122257, "grad_norm": 11.838387489318848, "learning_rate": 2.6595000000000005e-06, "loss": 2.2337, "step": 5319 }, { "epoch": 1.6677115987460815, "grad_norm": 13.644763946533203, "learning_rate": 2.6600000000000004e-06, "loss": 2.4781, "step": 5320 }, { "epoch": 1.668025078369906, "grad_norm": 28.51342010498047, "learning_rate": 2.6605e-06, "loss": 2.7712, "step": 5321 }, { "epoch": 1.6683385579937304, "grad_norm": 14.968554496765137, "learning_rate": 2.6610000000000004e-06, "loss": 2.4118, "step": 5322 }, { "epoch": 1.6686520376175549, "grad_norm": 16.145463943481445, "learning_rate": 2.6615000000000003e-06, "loss": 2.3086, "step": 5323 }, { "epoch": 1.6689655172413793, "grad_norm": 24.526046752929688, "learning_rate": 2.662e-06, "loss": 2.1476, "step": 5324 }, { "epoch": 1.6692789968652038, "grad_norm": 25.00034523010254, "learning_rate": 2.6625e-06, "loss": 2.2905, "step": 5325 }, { "epoch": 1.6695924764890282, "grad_norm": 39.29100036621094, "learning_rate": 2.663e-06, "loss": 2.4597, "step": 5326 }, { "epoch": 1.6699059561128526, "grad_norm": 31.492595672607422, "learning_rate": 2.6635e-06, "loss": 2.418, "step": 5327 }, { "epoch": 1.670219435736677, "grad_norm": 56.26333999633789, "learning_rate": 2.6640000000000007e-06, "loss": 2.2798, "step": 5328 }, { "epoch": 1.6705329153605015, "grad_norm": 19.510744094848633, "learning_rate": 2.6645000000000005e-06, "loss": 2.5782, "step": 5329 }, { "epoch": 1.670846394984326, "grad_norm": 21.897645950317383, "learning_rate": 2.6650000000000003e-06, "loss": 2.3679, "step": 5330 }, { "epoch": 1.6711598746081506, "grad_norm": 16.739225387573242, "learning_rate": 2.6655e-06, "loss": 2.4401, "step": 5331 }, { "epoch": 1.671473354231975, "grad_norm": 17.856504440307617, "learning_rate": 2.6660000000000004e-06, "loss": 2.3782, "step": 5332 }, { "epoch": 1.6717868338557995, "grad_norm": 16.302274703979492, "learning_rate": 2.6665000000000002e-06, "loss": 2.2191, "step": 5333 }, { "epoch": 1.672100313479624, "grad_norm": 12.230741500854492, "learning_rate": 2.667e-06, "loss": 2.4313, "step": 5334 }, { "epoch": 1.6724137931034484, "grad_norm": 37.70825958251953, "learning_rate": 2.6675000000000003e-06, "loss": 2.8394, "step": 5335 }, { "epoch": 1.6727272727272728, "grad_norm": 19.169376373291016, "learning_rate": 2.668e-06, "loss": 2.4162, "step": 5336 }, { "epoch": 1.6730407523510973, "grad_norm": 143.8944549560547, "learning_rate": 2.6685e-06, "loss": 2.3096, "step": 5337 }, { "epoch": 1.6733542319749217, "grad_norm": 25.016977310180664, "learning_rate": 2.6690000000000006e-06, "loss": 2.3672, "step": 5338 }, { "epoch": 1.6736677115987462, "grad_norm": 11.875692367553711, "learning_rate": 2.6695000000000004e-06, "loss": 2.2422, "step": 5339 }, { "epoch": 1.6739811912225706, "grad_norm": 14.99013614654541, "learning_rate": 2.6700000000000003e-06, "loss": 2.1051, "step": 5340 }, { "epoch": 1.674294670846395, "grad_norm": 7.908255100250244, "learning_rate": 2.6705000000000005e-06, "loss": 2.2055, "step": 5341 }, { "epoch": 1.6746081504702195, "grad_norm": 29.556995391845703, "learning_rate": 2.6710000000000003e-06, "loss": 2.1735, "step": 5342 }, { "epoch": 1.674921630094044, "grad_norm": 14.469071388244629, "learning_rate": 2.6715e-06, "loss": 2.2748, "step": 5343 }, { "epoch": 1.6752351097178684, "grad_norm": 12.449281692504883, "learning_rate": 2.672e-06, "loss": 2.3401, "step": 5344 }, { "epoch": 1.6755485893416928, "grad_norm": 12.840116500854492, "learning_rate": 2.6725000000000002e-06, "loss": 2.5153, "step": 5345 }, { "epoch": 1.6758620689655173, "grad_norm": 13.369140625, "learning_rate": 2.673e-06, "loss": 2.445, "step": 5346 }, { "epoch": 1.6761755485893417, "grad_norm": 26.555614471435547, "learning_rate": 2.6735e-06, "loss": 2.3638, "step": 5347 }, { "epoch": 1.6764890282131661, "grad_norm": 13.960328102111816, "learning_rate": 2.6740000000000006e-06, "loss": 2.1247, "step": 5348 }, { "epoch": 1.6768025078369906, "grad_norm": 17.971302032470703, "learning_rate": 2.6745000000000004e-06, "loss": 2.352, "step": 5349 }, { "epoch": 1.677115987460815, "grad_norm": 15.672755241394043, "learning_rate": 2.6750000000000002e-06, "loss": 2.0947, "step": 5350 }, { "epoch": 1.6774294670846395, "grad_norm": 27.532419204711914, "learning_rate": 2.6755000000000005e-06, "loss": 3.1038, "step": 5351 }, { "epoch": 1.677742946708464, "grad_norm": 32.5295295715332, "learning_rate": 2.6760000000000003e-06, "loss": 2.0789, "step": 5352 }, { "epoch": 1.6780564263322884, "grad_norm": 10.138176918029785, "learning_rate": 2.6765e-06, "loss": 2.6218, "step": 5353 }, { "epoch": 1.6783699059561128, "grad_norm": 22.55307960510254, "learning_rate": 2.677e-06, "loss": 1.9356, "step": 5354 }, { "epoch": 1.6786833855799372, "grad_norm": 47.21834182739258, "learning_rate": 2.6775e-06, "loss": 2.6484, "step": 5355 }, { "epoch": 1.6789968652037617, "grad_norm": 52.77558517456055, "learning_rate": 2.678e-06, "loss": 2.319, "step": 5356 }, { "epoch": 1.6793103448275861, "grad_norm": 13.631694793701172, "learning_rate": 2.6785e-06, "loss": 2.0864, "step": 5357 }, { "epoch": 1.6796238244514106, "grad_norm": 17.103946685791016, "learning_rate": 2.6790000000000005e-06, "loss": 2.2015, "step": 5358 }, { "epoch": 1.679937304075235, "grad_norm": 11.737060546875, "learning_rate": 2.6795000000000003e-06, "loss": 2.291, "step": 5359 }, { "epoch": 1.6802507836990594, "grad_norm": 34.89884948730469, "learning_rate": 2.68e-06, "loss": 2.3886, "step": 5360 }, { "epoch": 1.6805642633228839, "grad_norm": 25.996509552001953, "learning_rate": 2.6805000000000004e-06, "loss": 2.4901, "step": 5361 }, { "epoch": 1.6808777429467083, "grad_norm": 11.781593322753906, "learning_rate": 2.6810000000000003e-06, "loss": 2.2829, "step": 5362 }, { "epoch": 1.6811912225705328, "grad_norm": 21.603439331054688, "learning_rate": 2.6815e-06, "loss": 2.2104, "step": 5363 }, { "epoch": 1.6815047021943572, "grad_norm": 24.865238189697266, "learning_rate": 2.6820000000000003e-06, "loss": 2.3377, "step": 5364 }, { "epoch": 1.6818181818181817, "grad_norm": 16.560691833496094, "learning_rate": 2.6825e-06, "loss": 2.4748, "step": 5365 }, { "epoch": 1.6821316614420063, "grad_norm": 14.380731582641602, "learning_rate": 2.683e-06, "loss": 2.217, "step": 5366 }, { "epoch": 1.6824451410658308, "grad_norm": 13.259237289428711, "learning_rate": 2.6835e-06, "loss": 2.2073, "step": 5367 }, { "epoch": 1.6827586206896552, "grad_norm": 24.348876953125, "learning_rate": 2.6840000000000005e-06, "loss": 2.5637, "step": 5368 }, { "epoch": 1.6830721003134796, "grad_norm": 14.206762313842773, "learning_rate": 2.6845000000000003e-06, "loss": 2.3421, "step": 5369 }, { "epoch": 1.683385579937304, "grad_norm": 20.449251174926758, "learning_rate": 2.6850000000000006e-06, "loss": 2.2867, "step": 5370 }, { "epoch": 1.6836990595611285, "grad_norm": 29.787328720092773, "learning_rate": 2.6855000000000004e-06, "loss": 2.7544, "step": 5371 }, { "epoch": 1.684012539184953, "grad_norm": 107.39007568359375, "learning_rate": 2.686e-06, "loss": 2.6798, "step": 5372 }, { "epoch": 1.6843260188087774, "grad_norm": 10.828439712524414, "learning_rate": 2.6865e-06, "loss": 2.1149, "step": 5373 }, { "epoch": 1.6846394984326019, "grad_norm": 17.259565353393555, "learning_rate": 2.6870000000000003e-06, "loss": 2.4487, "step": 5374 }, { "epoch": 1.6849529780564263, "grad_norm": 9.159259796142578, "learning_rate": 2.6875e-06, "loss": 2.1542, "step": 5375 }, { "epoch": 1.6852664576802507, "grad_norm": 9.4286470413208, "learning_rate": 2.688e-06, "loss": 2.1916, "step": 5376 }, { "epoch": 1.6855799373040752, "grad_norm": 10.63680362701416, "learning_rate": 2.6885e-06, "loss": 2.1223, "step": 5377 }, { "epoch": 1.6858934169278998, "grad_norm": 15.974871635437012, "learning_rate": 2.689e-06, "loss": 2.473, "step": 5378 }, { "epoch": 1.6862068965517243, "grad_norm": 16.37948989868164, "learning_rate": 2.6895000000000003e-06, "loss": 2.6459, "step": 5379 }, { "epoch": 1.6865203761755487, "grad_norm": 17.967891693115234, "learning_rate": 2.6900000000000005e-06, "loss": 2.246, "step": 5380 }, { "epoch": 1.6868338557993732, "grad_norm": 15.976727485656738, "learning_rate": 2.6905000000000003e-06, "loss": 2.1751, "step": 5381 }, { "epoch": 1.6871473354231976, "grad_norm": 27.1754207611084, "learning_rate": 2.691e-06, "loss": 2.7939, "step": 5382 }, { "epoch": 1.687460815047022, "grad_norm": 14.75679874420166, "learning_rate": 2.6915000000000004e-06, "loss": 2.4285, "step": 5383 }, { "epoch": 1.6877742946708465, "grad_norm": 38.64543914794922, "learning_rate": 2.6920000000000002e-06, "loss": 2.2422, "step": 5384 }, { "epoch": 1.688087774294671, "grad_norm": 64.94852447509766, "learning_rate": 2.6925e-06, "loss": 2.3804, "step": 5385 }, { "epoch": 1.6884012539184954, "grad_norm": 17.144126892089844, "learning_rate": 2.693e-06, "loss": 2.2281, "step": 5386 }, { "epoch": 1.6887147335423198, "grad_norm": 29.98297882080078, "learning_rate": 2.6935e-06, "loss": 2.3383, "step": 5387 }, { "epoch": 1.6890282131661443, "grad_norm": 20.504491806030273, "learning_rate": 2.694e-06, "loss": 2.2015, "step": 5388 }, { "epoch": 1.6893416927899687, "grad_norm": 23.511560440063477, "learning_rate": 2.6945000000000006e-06, "loss": 2.2689, "step": 5389 }, { "epoch": 1.6896551724137931, "grad_norm": 13.887767791748047, "learning_rate": 2.6950000000000005e-06, "loss": 2.3898, "step": 5390 }, { "epoch": 1.6899686520376176, "grad_norm": 14.888626098632812, "learning_rate": 2.6955000000000003e-06, "loss": 2.3539, "step": 5391 }, { "epoch": 1.690282131661442, "grad_norm": 62.88370895385742, "learning_rate": 2.696e-06, "loss": 2.6187, "step": 5392 }, { "epoch": 1.6905956112852665, "grad_norm": 13.9352445602417, "learning_rate": 2.6965000000000004e-06, "loss": 2.3166, "step": 5393 }, { "epoch": 1.690909090909091, "grad_norm": 21.216232299804688, "learning_rate": 2.697e-06, "loss": 2.1817, "step": 5394 }, { "epoch": 1.6912225705329154, "grad_norm": 13.734606742858887, "learning_rate": 2.6975e-06, "loss": 2.5173, "step": 5395 }, { "epoch": 1.6915360501567398, "grad_norm": 68.65769958496094, "learning_rate": 2.6980000000000003e-06, "loss": 2.5343, "step": 5396 }, { "epoch": 1.6918495297805642, "grad_norm": 20.774269104003906, "learning_rate": 2.6985e-06, "loss": 2.9843, "step": 5397 }, { "epoch": 1.6921630094043887, "grad_norm": 37.50685119628906, "learning_rate": 2.699e-06, "loss": 2.4088, "step": 5398 }, { "epoch": 1.6924764890282131, "grad_norm": 137.178955078125, "learning_rate": 2.6995000000000006e-06, "loss": 3.5242, "step": 5399 }, { "epoch": 1.6927899686520376, "grad_norm": 47.885902404785156, "learning_rate": 2.7000000000000004e-06, "loss": 2.2528, "step": 5400 }, { "epoch": 1.693103448275862, "grad_norm": 18.669857025146484, "learning_rate": 2.7005000000000002e-06, "loss": 2.3694, "step": 5401 }, { "epoch": 1.6934169278996865, "grad_norm": 91.53276062011719, "learning_rate": 2.7010000000000005e-06, "loss": 2.7944, "step": 5402 }, { "epoch": 1.693730407523511, "grad_norm": 25.30732536315918, "learning_rate": 2.7015000000000003e-06, "loss": 2.3734, "step": 5403 }, { "epoch": 1.6940438871473353, "grad_norm": 22.524877548217773, "learning_rate": 2.702e-06, "loss": 2.5223, "step": 5404 }, { "epoch": 1.6943573667711598, "grad_norm": 14.78250503540039, "learning_rate": 2.7025e-06, "loss": 2.1247, "step": 5405 }, { "epoch": 1.6946708463949842, "grad_norm": 14.238978385925293, "learning_rate": 2.703e-06, "loss": 2.525, "step": 5406 }, { "epoch": 1.6949843260188087, "grad_norm": 27.516679763793945, "learning_rate": 2.7035e-06, "loss": 2.345, "step": 5407 }, { "epoch": 1.695297805642633, "grad_norm": 219.90899658203125, "learning_rate": 2.704e-06, "loss": 2.7112, "step": 5408 }, { "epoch": 1.6956112852664575, "grad_norm": 12.407668113708496, "learning_rate": 2.7045000000000005e-06, "loss": 2.3901, "step": 5409 }, { "epoch": 1.695924764890282, "grad_norm": 39.14141845703125, "learning_rate": 2.7050000000000004e-06, "loss": 2.2635, "step": 5410 }, { "epoch": 1.6962382445141064, "grad_norm": 16.694364547729492, "learning_rate": 2.7055e-06, "loss": 2.7107, "step": 5411 }, { "epoch": 1.6965517241379309, "grad_norm": 14.390851020812988, "learning_rate": 2.7060000000000004e-06, "loss": 2.1942, "step": 5412 }, { "epoch": 1.6968652037617555, "grad_norm": 18.806320190429688, "learning_rate": 2.7065000000000003e-06, "loss": 2.4038, "step": 5413 }, { "epoch": 1.69717868338558, "grad_norm": 13.088409423828125, "learning_rate": 2.707e-06, "loss": 2.1712, "step": 5414 }, { "epoch": 1.6974921630094044, "grad_norm": 19.5601749420166, "learning_rate": 2.7075000000000003e-06, "loss": 2.8417, "step": 5415 }, { "epoch": 1.6978056426332289, "grad_norm": 40.176918029785156, "learning_rate": 2.708e-06, "loss": 2.4874, "step": 5416 }, { "epoch": 1.6981191222570533, "grad_norm": 20.487564086914062, "learning_rate": 2.7085e-06, "loss": 2.2593, "step": 5417 }, { "epoch": 1.6984326018808777, "grad_norm": 13.035333633422852, "learning_rate": 2.709e-06, "loss": 2.2706, "step": 5418 }, { "epoch": 1.6987460815047022, "grad_norm": 10.95355224609375, "learning_rate": 2.7095000000000005e-06, "loss": 2.2284, "step": 5419 }, { "epoch": 1.6990595611285266, "grad_norm": 13.154052734375, "learning_rate": 2.7100000000000003e-06, "loss": 2.3685, "step": 5420 }, { "epoch": 1.699373040752351, "grad_norm": 26.198123931884766, "learning_rate": 2.7105000000000006e-06, "loss": 2.5497, "step": 5421 }, { "epoch": 1.6996865203761755, "grad_norm": 36.201995849609375, "learning_rate": 2.7110000000000004e-06, "loss": 3.0085, "step": 5422 }, { "epoch": 1.7, "grad_norm": 73.30572509765625, "learning_rate": 2.7115000000000002e-06, "loss": 1.8637, "step": 5423 }, { "epoch": 1.7003134796238244, "grad_norm": 59.0308723449707, "learning_rate": 2.712e-06, "loss": 2.7115, "step": 5424 }, { "epoch": 1.700626959247649, "grad_norm": 47.59678268432617, "learning_rate": 2.7125000000000003e-06, "loss": 2.8762, "step": 5425 }, { "epoch": 1.7009404388714735, "grad_norm": 11.48272705078125, "learning_rate": 2.713e-06, "loss": 2.2331, "step": 5426 }, { "epoch": 1.701253918495298, "grad_norm": 20.130481719970703, "learning_rate": 2.7135e-06, "loss": 2.4474, "step": 5427 }, { "epoch": 1.7015673981191224, "grad_norm": 15.507676124572754, "learning_rate": 2.7139999999999998e-06, "loss": 2.312, "step": 5428 }, { "epoch": 1.7018808777429468, "grad_norm": 15.507573127746582, "learning_rate": 2.7145000000000004e-06, "loss": 2.1403, "step": 5429 }, { "epoch": 1.7021943573667713, "grad_norm": 23.145816802978516, "learning_rate": 2.7150000000000003e-06, "loss": 2.5529, "step": 5430 }, { "epoch": 1.7025078369905957, "grad_norm": 38.59257888793945, "learning_rate": 2.7155000000000005e-06, "loss": 2.3534, "step": 5431 }, { "epoch": 1.7028213166144202, "grad_norm": 24.967159271240234, "learning_rate": 2.7160000000000003e-06, "loss": 2.9059, "step": 5432 }, { "epoch": 1.7031347962382446, "grad_norm": 29.68940544128418, "learning_rate": 2.7165e-06, "loss": 2.2956, "step": 5433 }, { "epoch": 1.703448275862069, "grad_norm": 23.613834381103516, "learning_rate": 2.717e-06, "loss": 2.2259, "step": 5434 }, { "epoch": 1.7037617554858935, "grad_norm": 18.64572525024414, "learning_rate": 2.7175000000000002e-06, "loss": 2.419, "step": 5435 }, { "epoch": 1.704075235109718, "grad_norm": 16.19615936279297, "learning_rate": 2.718e-06, "loss": 2.9591, "step": 5436 }, { "epoch": 1.7043887147335424, "grad_norm": 18.514894485473633, "learning_rate": 2.7185e-06, "loss": 2.215, "step": 5437 }, { "epoch": 1.7047021943573668, "grad_norm": 18.013778686523438, "learning_rate": 2.719e-06, "loss": 2.3559, "step": 5438 }, { "epoch": 1.7050156739811912, "grad_norm": 20.540124893188477, "learning_rate": 2.7195000000000004e-06, "loss": 2.5045, "step": 5439 }, { "epoch": 1.7053291536050157, "grad_norm": 13.501602172851562, "learning_rate": 2.7200000000000002e-06, "loss": 1.9096, "step": 5440 }, { "epoch": 1.7056426332288401, "grad_norm": 11.370809555053711, "learning_rate": 2.7205000000000005e-06, "loss": 2.1822, "step": 5441 }, { "epoch": 1.7059561128526646, "grad_norm": 21.86370086669922, "learning_rate": 2.7210000000000003e-06, "loss": 2.1602, "step": 5442 }, { "epoch": 1.706269592476489, "grad_norm": 30.073101043701172, "learning_rate": 2.7215e-06, "loss": 2.6491, "step": 5443 }, { "epoch": 1.7065830721003135, "grad_norm": 17.897363662719727, "learning_rate": 2.7220000000000004e-06, "loss": 2.2375, "step": 5444 }, { "epoch": 1.706896551724138, "grad_norm": 11.158616065979004, "learning_rate": 2.7225e-06, "loss": 2.0454, "step": 5445 }, { "epoch": 1.7072100313479623, "grad_norm": 8.605175018310547, "learning_rate": 2.723e-06, "loss": 2.2267, "step": 5446 }, { "epoch": 1.7075235109717868, "grad_norm": 27.836732864379883, "learning_rate": 2.7235e-06, "loss": 2.3088, "step": 5447 }, { "epoch": 1.7078369905956112, "grad_norm": 14.671130180358887, "learning_rate": 2.724e-06, "loss": 2.7475, "step": 5448 }, { "epoch": 1.7081504702194357, "grad_norm": 20.618295669555664, "learning_rate": 2.7245000000000003e-06, "loss": 2.8887, "step": 5449 }, { "epoch": 1.70846394984326, "grad_norm": 38.52854537963867, "learning_rate": 2.7250000000000006e-06, "loss": 2.4788, "step": 5450 }, { "epoch": 1.7087774294670846, "grad_norm": 11.359956741333008, "learning_rate": 2.7255000000000004e-06, "loss": 2.3325, "step": 5451 }, { "epoch": 1.709090909090909, "grad_norm": 36.49653625488281, "learning_rate": 2.7260000000000002e-06, "loss": 2.5183, "step": 5452 }, { "epoch": 1.7094043887147334, "grad_norm": 25.822084426879883, "learning_rate": 2.7265e-06, "loss": 2.372, "step": 5453 }, { "epoch": 1.7097178683385579, "grad_norm": 85.20195770263672, "learning_rate": 2.7270000000000003e-06, "loss": 2.3598, "step": 5454 }, { "epoch": 1.7100313479623823, "grad_norm": 12.452218055725098, "learning_rate": 2.7275e-06, "loss": 2.493, "step": 5455 }, { "epoch": 1.7103448275862068, "grad_norm": 10.54835319519043, "learning_rate": 2.728e-06, "loss": 2.2473, "step": 5456 }, { "epoch": 1.7106583072100312, "grad_norm": 12.409470558166504, "learning_rate": 2.7285000000000002e-06, "loss": 2.4374, "step": 5457 }, { "epoch": 1.7109717868338556, "grad_norm": 12.735477447509766, "learning_rate": 2.729e-06, "loss": 2.047, "step": 5458 }, { "epoch": 1.71128526645768, "grad_norm": 12.568780899047852, "learning_rate": 2.7295000000000003e-06, "loss": 2.2044, "step": 5459 }, { "epoch": 1.7115987460815048, "grad_norm": 11.628466606140137, "learning_rate": 2.7300000000000005e-06, "loss": 2.1427, "step": 5460 }, { "epoch": 1.7119122257053292, "grad_norm": 34.290645599365234, "learning_rate": 2.7305000000000004e-06, "loss": 2.2639, "step": 5461 }, { "epoch": 1.7122257053291536, "grad_norm": 16.477323532104492, "learning_rate": 2.731e-06, "loss": 2.4689, "step": 5462 }, { "epoch": 1.712539184952978, "grad_norm": 15.450041770935059, "learning_rate": 2.7315000000000004e-06, "loss": 2.4502, "step": 5463 }, { "epoch": 1.7128526645768025, "grad_norm": 45.1892204284668, "learning_rate": 2.7320000000000003e-06, "loss": 2.48, "step": 5464 }, { "epoch": 1.713166144200627, "grad_norm": 16.953018188476562, "learning_rate": 2.7325e-06, "loss": 1.9744, "step": 5465 }, { "epoch": 1.7134796238244514, "grad_norm": 14.956791877746582, "learning_rate": 2.733e-06, "loss": 2.1789, "step": 5466 }, { "epoch": 1.7137931034482758, "grad_norm": 8.133874893188477, "learning_rate": 2.7335e-06, "loss": 2.0267, "step": 5467 }, { "epoch": 1.7141065830721003, "grad_norm": 17.319612503051758, "learning_rate": 2.734e-06, "loss": 2.1611, "step": 5468 }, { "epoch": 1.7144200626959247, "grad_norm": 13.33533763885498, "learning_rate": 2.7345000000000007e-06, "loss": 2.2603, "step": 5469 }, { "epoch": 1.7147335423197492, "grad_norm": 18.363788604736328, "learning_rate": 2.7350000000000005e-06, "loss": 2.5405, "step": 5470 }, { "epoch": 1.7150470219435736, "grad_norm": 17.000051498413086, "learning_rate": 2.7355000000000003e-06, "loss": 2.9102, "step": 5471 }, { "epoch": 1.7153605015673983, "grad_norm": 10.053338050842285, "learning_rate": 2.736e-06, "loss": 2.263, "step": 5472 }, { "epoch": 1.7156739811912227, "grad_norm": 42.593536376953125, "learning_rate": 2.7365000000000004e-06, "loss": 2.4529, "step": 5473 }, { "epoch": 1.7159874608150472, "grad_norm": 12.775186538696289, "learning_rate": 2.7370000000000002e-06, "loss": 2.4782, "step": 5474 }, { "epoch": 1.7163009404388716, "grad_norm": 23.6019287109375, "learning_rate": 2.7375e-06, "loss": 2.2155, "step": 5475 }, { "epoch": 1.716614420062696, "grad_norm": 12.975761413574219, "learning_rate": 2.7380000000000003e-06, "loss": 2.4409, "step": 5476 }, { "epoch": 1.7169278996865205, "grad_norm": 14.86938762664795, "learning_rate": 2.7385e-06, "loss": 2.3281, "step": 5477 }, { "epoch": 1.717241379310345, "grad_norm": 13.768951416015625, "learning_rate": 2.739e-06, "loss": 2.267, "step": 5478 }, { "epoch": 1.7175548589341694, "grad_norm": 33.40892791748047, "learning_rate": 2.7395000000000006e-06, "loss": 2.019, "step": 5479 }, { "epoch": 1.7178683385579938, "grad_norm": 18.362037658691406, "learning_rate": 2.7400000000000004e-06, "loss": 2.4633, "step": 5480 }, { "epoch": 1.7181818181818183, "grad_norm": 20.025041580200195, "learning_rate": 2.7405000000000003e-06, "loss": 2.3384, "step": 5481 }, { "epoch": 1.7184952978056427, "grad_norm": 18.812082290649414, "learning_rate": 2.7410000000000005e-06, "loss": 2.2198, "step": 5482 }, { "epoch": 1.7188087774294671, "grad_norm": 20.752206802368164, "learning_rate": 2.7415000000000003e-06, "loss": 2.2079, "step": 5483 }, { "epoch": 1.7191222570532916, "grad_norm": 11.19144344329834, "learning_rate": 2.742e-06, "loss": 2.5804, "step": 5484 }, { "epoch": 1.719435736677116, "grad_norm": 20.350521087646484, "learning_rate": 2.7425e-06, "loss": 2.561, "step": 5485 }, { "epoch": 1.7197492163009405, "grad_norm": 40.37380599975586, "learning_rate": 2.7430000000000002e-06, "loss": 2.5229, "step": 5486 }, { "epoch": 1.720062695924765, "grad_norm": 51.770198822021484, "learning_rate": 2.7435e-06, "loss": 2.4845, "step": 5487 }, { "epoch": 1.7203761755485893, "grad_norm": 21.724157333374023, "learning_rate": 2.744e-06, "loss": 2.0513, "step": 5488 }, { "epoch": 1.7206896551724138, "grad_norm": 9.700921058654785, "learning_rate": 2.7445000000000006e-06, "loss": 2.2333, "step": 5489 }, { "epoch": 1.7210031347962382, "grad_norm": 33.1986198425293, "learning_rate": 2.7450000000000004e-06, "loss": 2.3553, "step": 5490 }, { "epoch": 1.7213166144200627, "grad_norm": 26.239349365234375, "learning_rate": 2.7455000000000002e-06, "loss": 2.4481, "step": 5491 }, { "epoch": 1.7216300940438871, "grad_norm": 25.467309951782227, "learning_rate": 2.7460000000000005e-06, "loss": 2.5357, "step": 5492 }, { "epoch": 1.7219435736677116, "grad_norm": 9.850618362426758, "learning_rate": 2.7465000000000003e-06, "loss": 2.0866, "step": 5493 }, { "epoch": 1.722257053291536, "grad_norm": 18.098697662353516, "learning_rate": 2.747e-06, "loss": 2.303, "step": 5494 }, { "epoch": 1.7225705329153604, "grad_norm": 27.904993057250977, "learning_rate": 2.7475000000000004e-06, "loss": 2.2298, "step": 5495 }, { "epoch": 1.7228840125391849, "grad_norm": 11.535282135009766, "learning_rate": 2.748e-06, "loss": 1.9315, "step": 5496 }, { "epoch": 1.7231974921630093, "grad_norm": 20.059200286865234, "learning_rate": 2.7485e-06, "loss": 2.3591, "step": 5497 }, { "epoch": 1.7235109717868338, "grad_norm": 12.67073917388916, "learning_rate": 2.749e-06, "loss": 2.4145, "step": 5498 }, { "epoch": 1.7238244514106582, "grad_norm": 21.71510124206543, "learning_rate": 2.7495000000000005e-06, "loss": 2.323, "step": 5499 }, { "epoch": 1.7241379310344827, "grad_norm": 65.52584075927734, "learning_rate": 2.7500000000000004e-06, "loss": 2.7043, "step": 5500 }, { "epoch": 1.724451410658307, "grad_norm": 11.966585159301758, "learning_rate": 2.7505000000000006e-06, "loss": 2.2303, "step": 5501 }, { "epoch": 1.7247648902821315, "grad_norm": 34.113651275634766, "learning_rate": 2.7510000000000004e-06, "loss": 2.1599, "step": 5502 }, { "epoch": 1.725078369905956, "grad_norm": 33.491355895996094, "learning_rate": 2.7515000000000003e-06, "loss": 2.5904, "step": 5503 }, { "epoch": 1.7253918495297804, "grad_norm": 16.249862670898438, "learning_rate": 2.752e-06, "loss": 2.6213, "step": 5504 }, { "epoch": 1.7257053291536049, "grad_norm": 20.483736038208008, "learning_rate": 2.7525000000000003e-06, "loss": 2.2492, "step": 5505 }, { "epoch": 1.7260188087774293, "grad_norm": 15.815305709838867, "learning_rate": 2.753e-06, "loss": 2.0444, "step": 5506 }, { "epoch": 1.726332288401254, "grad_norm": 8.834308624267578, "learning_rate": 2.7535e-06, "loss": 2.2711, "step": 5507 }, { "epoch": 1.7266457680250784, "grad_norm": 29.914960861206055, "learning_rate": 2.754e-06, "loss": 2.2056, "step": 5508 }, { "epoch": 1.7269592476489029, "grad_norm": 22.222673416137695, "learning_rate": 2.7545e-06, "loss": 2.6003, "step": 5509 }, { "epoch": 1.7272727272727273, "grad_norm": 59.530555725097656, "learning_rate": 2.7550000000000003e-06, "loss": 3.0075, "step": 5510 }, { "epoch": 1.7275862068965517, "grad_norm": 10.841995239257812, "learning_rate": 2.7555000000000006e-06, "loss": 2.108, "step": 5511 }, { "epoch": 1.7278996865203762, "grad_norm": 39.829383850097656, "learning_rate": 2.7560000000000004e-06, "loss": 2.0235, "step": 5512 }, { "epoch": 1.7282131661442006, "grad_norm": 9.398069381713867, "learning_rate": 2.7565e-06, "loss": 2.2911, "step": 5513 }, { "epoch": 1.728526645768025, "grad_norm": 17.04472541809082, "learning_rate": 2.757e-06, "loss": 2.0957, "step": 5514 }, { "epoch": 1.7288401253918495, "grad_norm": 21.816059112548828, "learning_rate": 2.7575000000000003e-06, "loss": 1.9723, "step": 5515 }, { "epoch": 1.729153605015674, "grad_norm": 16.998680114746094, "learning_rate": 2.758e-06, "loss": 2.2045, "step": 5516 }, { "epoch": 1.7294670846394984, "grad_norm": 16.111724853515625, "learning_rate": 2.7585e-06, "loss": 2.7597, "step": 5517 }, { "epoch": 1.729780564263323, "grad_norm": 20.748037338256836, "learning_rate": 2.759e-06, "loss": 2.6406, "step": 5518 }, { "epoch": 1.7300940438871475, "grad_norm": 47.98612976074219, "learning_rate": 2.7595e-06, "loss": 2.7046, "step": 5519 }, { "epoch": 1.730407523510972, "grad_norm": 15.667613983154297, "learning_rate": 2.7600000000000003e-06, "loss": 2.359, "step": 5520 }, { "epoch": 1.7307210031347964, "grad_norm": 10.290972709655762, "learning_rate": 2.7605000000000005e-06, "loss": 2.3374, "step": 5521 }, { "epoch": 1.7310344827586208, "grad_norm": 39.547481536865234, "learning_rate": 2.7610000000000003e-06, "loss": 2.2134, "step": 5522 }, { "epoch": 1.7313479623824453, "grad_norm": 9.93276309967041, "learning_rate": 2.7615e-06, "loss": 2.5338, "step": 5523 }, { "epoch": 1.7316614420062697, "grad_norm": 36.52473831176758, "learning_rate": 2.7620000000000004e-06, "loss": 2.8736, "step": 5524 }, { "epoch": 1.7319749216300941, "grad_norm": 16.293170928955078, "learning_rate": 2.7625000000000002e-06, "loss": 2.0691, "step": 5525 }, { "epoch": 1.7322884012539186, "grad_norm": 15.633368492126465, "learning_rate": 2.763e-06, "loss": 2.2657, "step": 5526 }, { "epoch": 1.732601880877743, "grad_norm": 11.91658878326416, "learning_rate": 2.7635e-06, "loss": 2.2018, "step": 5527 }, { "epoch": 1.7329153605015675, "grad_norm": 70.28834533691406, "learning_rate": 2.764e-06, "loss": 2.817, "step": 5528 }, { "epoch": 1.733228840125392, "grad_norm": 29.530969619750977, "learning_rate": 2.7645e-06, "loss": 2.3077, "step": 5529 }, { "epoch": 1.7335423197492164, "grad_norm": 13.929141998291016, "learning_rate": 2.7650000000000006e-06, "loss": 2.6901, "step": 5530 }, { "epoch": 1.7338557993730408, "grad_norm": 15.058401107788086, "learning_rate": 2.7655000000000005e-06, "loss": 2.2645, "step": 5531 }, { "epoch": 1.7341692789968652, "grad_norm": 17.164827346801758, "learning_rate": 2.7660000000000003e-06, "loss": 2.3818, "step": 5532 }, { "epoch": 1.7344827586206897, "grad_norm": 17.328838348388672, "learning_rate": 2.7665e-06, "loss": 2.4011, "step": 5533 }, { "epoch": 1.7347962382445141, "grad_norm": 11.72863483428955, "learning_rate": 2.7670000000000004e-06, "loss": 2.5156, "step": 5534 }, { "epoch": 1.7351097178683386, "grad_norm": 15.69046688079834, "learning_rate": 2.7675e-06, "loss": 2.2565, "step": 5535 }, { "epoch": 1.735423197492163, "grad_norm": 10.286733627319336, "learning_rate": 2.768e-06, "loss": 2.2191, "step": 5536 }, { "epoch": 1.7357366771159874, "grad_norm": 12.609308242797852, "learning_rate": 2.7685000000000003e-06, "loss": 2.3434, "step": 5537 }, { "epoch": 1.736050156739812, "grad_norm": 10.516945838928223, "learning_rate": 2.769e-06, "loss": 2.2088, "step": 5538 }, { "epoch": 1.7363636363636363, "grad_norm": 20.95050048828125, "learning_rate": 2.7695e-06, "loss": 2.4958, "step": 5539 }, { "epoch": 1.7366771159874608, "grad_norm": 10.20532512664795, "learning_rate": 2.7700000000000006e-06, "loss": 2.5765, "step": 5540 }, { "epoch": 1.7369905956112852, "grad_norm": 11.721442222595215, "learning_rate": 2.7705000000000004e-06, "loss": 2.336, "step": 5541 }, { "epoch": 1.7373040752351097, "grad_norm": 21.627376556396484, "learning_rate": 2.7710000000000002e-06, "loss": 2.3839, "step": 5542 }, { "epoch": 1.737617554858934, "grad_norm": 17.041444778442383, "learning_rate": 2.7715000000000005e-06, "loss": 2.4114, "step": 5543 }, { "epoch": 1.7379310344827585, "grad_norm": 37.384986877441406, "learning_rate": 2.7720000000000003e-06, "loss": 2.3266, "step": 5544 }, { "epoch": 1.738244514106583, "grad_norm": 34.521976470947266, "learning_rate": 2.7725e-06, "loss": 2.3097, "step": 5545 }, { "epoch": 1.7385579937304074, "grad_norm": 37.94413375854492, "learning_rate": 2.773e-06, "loss": 2.3119, "step": 5546 }, { "epoch": 1.7388714733542319, "grad_norm": 13.897578239440918, "learning_rate": 2.7735e-06, "loss": 2.3442, "step": 5547 }, { "epoch": 1.7391849529780563, "grad_norm": 26.50729751586914, "learning_rate": 2.774e-06, "loss": 2.1502, "step": 5548 }, { "epoch": 1.7394984326018808, "grad_norm": 12.668210983276367, "learning_rate": 2.7745e-06, "loss": 2.1628, "step": 5549 }, { "epoch": 1.7398119122257052, "grad_norm": 9.052404403686523, "learning_rate": 2.7750000000000005e-06, "loss": 2.1179, "step": 5550 }, { "epoch": 1.7401253918495296, "grad_norm": 16.18061637878418, "learning_rate": 2.7755000000000004e-06, "loss": 2.0622, "step": 5551 }, { "epoch": 1.740438871473354, "grad_norm": 45.02144241333008, "learning_rate": 2.776e-06, "loss": 2.3295, "step": 5552 }, { "epoch": 1.7407523510971785, "grad_norm": 13.236767768859863, "learning_rate": 2.7765000000000004e-06, "loss": 2.2401, "step": 5553 }, { "epoch": 1.7410658307210032, "grad_norm": 20.257795333862305, "learning_rate": 2.7770000000000003e-06, "loss": 2.545, "step": 5554 }, { "epoch": 1.7413793103448276, "grad_norm": 11.673392295837402, "learning_rate": 2.7775e-06, "loss": 2.0406, "step": 5555 }, { "epoch": 1.741692789968652, "grad_norm": 14.821595191955566, "learning_rate": 2.7780000000000003e-06, "loss": 2.215, "step": 5556 }, { "epoch": 1.7420062695924765, "grad_norm": 58.19940948486328, "learning_rate": 2.7785e-06, "loss": 2.8323, "step": 5557 }, { "epoch": 1.742319749216301, "grad_norm": 30.191211700439453, "learning_rate": 2.779e-06, "loss": 2.2626, "step": 5558 }, { "epoch": 1.7426332288401254, "grad_norm": 8.197135925292969, "learning_rate": 2.7795e-06, "loss": 2.214, "step": 5559 }, { "epoch": 1.7429467084639498, "grad_norm": 15.255378723144531, "learning_rate": 2.7800000000000005e-06, "loss": 2.1536, "step": 5560 }, { "epoch": 1.7432601880877743, "grad_norm": 86.08578491210938, "learning_rate": 2.7805000000000003e-06, "loss": 2.4542, "step": 5561 }, { "epoch": 1.7435736677115987, "grad_norm": 17.217504501342773, "learning_rate": 2.7810000000000006e-06, "loss": 2.9291, "step": 5562 }, { "epoch": 1.7438871473354232, "grad_norm": 65.4100112915039, "learning_rate": 2.7815000000000004e-06, "loss": 2.2653, "step": 5563 }, { "epoch": 1.7442006269592476, "grad_norm": 17.314762115478516, "learning_rate": 2.7820000000000002e-06, "loss": 2.2533, "step": 5564 }, { "epoch": 1.7445141065830723, "grad_norm": 26.037796020507812, "learning_rate": 2.7825e-06, "loss": 2.3206, "step": 5565 }, { "epoch": 1.7448275862068967, "grad_norm": 9.311919212341309, "learning_rate": 2.7830000000000003e-06, "loss": 2.3518, "step": 5566 }, { "epoch": 1.7451410658307211, "grad_norm": 20.615209579467773, "learning_rate": 2.7835e-06, "loss": 2.1305, "step": 5567 }, { "epoch": 1.7454545454545456, "grad_norm": 15.762939453125, "learning_rate": 2.784e-06, "loss": 2.4987, "step": 5568 }, { "epoch": 1.74576802507837, "grad_norm": 18.63046646118164, "learning_rate": 2.7845e-06, "loss": 2.0101, "step": 5569 }, { "epoch": 1.7460815047021945, "grad_norm": 11.48450756072998, "learning_rate": 2.7850000000000004e-06, "loss": 2.2357, "step": 5570 }, { "epoch": 1.746394984326019, "grad_norm": 96.20525360107422, "learning_rate": 2.7855000000000003e-06, "loss": 2.1663, "step": 5571 }, { "epoch": 1.7467084639498434, "grad_norm": 17.738521575927734, "learning_rate": 2.7860000000000005e-06, "loss": 2.1544, "step": 5572 }, { "epoch": 1.7470219435736678, "grad_norm": 15.63515853881836, "learning_rate": 2.7865000000000003e-06, "loss": 2.4514, "step": 5573 }, { "epoch": 1.7473354231974922, "grad_norm": 22.114669799804688, "learning_rate": 2.787e-06, "loss": 2.0831, "step": 5574 }, { "epoch": 1.7476489028213167, "grad_norm": 31.60894012451172, "learning_rate": 2.7875000000000004e-06, "loss": 2.2297, "step": 5575 }, { "epoch": 1.7479623824451411, "grad_norm": 14.62855339050293, "learning_rate": 2.7880000000000002e-06, "loss": 2.2535, "step": 5576 }, { "epoch": 1.7482758620689656, "grad_norm": 14.154539108276367, "learning_rate": 2.7885e-06, "loss": 2.3844, "step": 5577 }, { "epoch": 1.74858934169279, "grad_norm": 91.77095794677734, "learning_rate": 2.789e-06, "loss": 2.9025, "step": 5578 }, { "epoch": 1.7489028213166145, "grad_norm": 30.688533782958984, "learning_rate": 2.7895e-06, "loss": 2.7722, "step": 5579 }, { "epoch": 1.749216300940439, "grad_norm": 11.394512176513672, "learning_rate": 2.7900000000000004e-06, "loss": 2.3959, "step": 5580 }, { "epoch": 1.7495297805642633, "grad_norm": 43.727603912353516, "learning_rate": 2.7905000000000006e-06, "loss": 2.2847, "step": 5581 }, { "epoch": 1.7498432601880878, "grad_norm": 18.254802703857422, "learning_rate": 2.7910000000000005e-06, "loss": 2.4319, "step": 5582 }, { "epoch": 1.7501567398119122, "grad_norm": 12.743124961853027, "learning_rate": 2.7915000000000003e-06, "loss": 2.4469, "step": 5583 }, { "epoch": 1.7504702194357367, "grad_norm": 29.064430236816406, "learning_rate": 2.792e-06, "loss": 2.3628, "step": 5584 }, { "epoch": 1.750783699059561, "grad_norm": 16.731903076171875, "learning_rate": 2.7925000000000004e-06, "loss": 2.3773, "step": 5585 }, { "epoch": 1.7510971786833855, "grad_norm": 24.20734977722168, "learning_rate": 2.793e-06, "loss": 2.5311, "step": 5586 }, { "epoch": 1.7510971786833855, "eval_loss": 2.7825205326080322, "eval_runtime": 20.9022, "eval_samples_per_second": 128.551, "eval_steps_per_second": 8.037, "step": 5586 }, { "epoch": 1.75141065830721, "grad_norm": 36.4649658203125, "learning_rate": 2.7935e-06, "loss": 2.6414, "step": 5587 }, { "epoch": 1.7517241379310344, "grad_norm": 15.779841423034668, "learning_rate": 2.794e-06, "loss": 2.2283, "step": 5588 }, { "epoch": 1.7520376175548589, "grad_norm": 16.722549438476562, "learning_rate": 2.7945e-06, "loss": 1.9054, "step": 5589 }, { "epoch": 1.7523510971786833, "grad_norm": 21.327329635620117, "learning_rate": 2.7950000000000003e-06, "loss": 2.4145, "step": 5590 }, { "epoch": 1.7526645768025078, "grad_norm": 15.905557632446289, "learning_rate": 2.7955000000000006e-06, "loss": 2.3218, "step": 5591 }, { "epoch": 1.7529780564263322, "grad_norm": 38.292198181152344, "learning_rate": 2.7960000000000004e-06, "loss": 2.2133, "step": 5592 }, { "epoch": 1.7532915360501566, "grad_norm": 16.73972511291504, "learning_rate": 2.7965000000000002e-06, "loss": 2.2877, "step": 5593 }, { "epoch": 1.753605015673981, "grad_norm": 33.32880783081055, "learning_rate": 2.797e-06, "loss": 2.4803, "step": 5594 }, { "epoch": 1.7539184952978055, "grad_norm": 12.935139656066895, "learning_rate": 2.7975000000000003e-06, "loss": 2.2784, "step": 5595 }, { "epoch": 1.75423197492163, "grad_norm": 65.7669448852539, "learning_rate": 2.798e-06, "loss": 2.5523, "step": 5596 }, { "epoch": 1.7545454545454544, "grad_norm": 23.66747283935547, "learning_rate": 2.7985e-06, "loss": 2.8021, "step": 5597 }, { "epoch": 1.7548589341692789, "grad_norm": 19.805585861206055, "learning_rate": 2.7990000000000002e-06, "loss": 2.7297, "step": 5598 }, { "epoch": 1.7551724137931033, "grad_norm": 19.813541412353516, "learning_rate": 2.7995e-06, "loss": 2.6171, "step": 5599 }, { "epoch": 1.7554858934169277, "grad_norm": 10.666125297546387, "learning_rate": 2.8000000000000003e-06, "loss": 2.1845, "step": 5600 }, { "epoch": 1.7557993730407524, "grad_norm": 14.279791831970215, "learning_rate": 2.8005000000000005e-06, "loss": 2.2536, "step": 5601 }, { "epoch": 1.7561128526645768, "grad_norm": 11.552290916442871, "learning_rate": 2.8010000000000004e-06, "loss": 1.9127, "step": 5602 }, { "epoch": 1.7564263322884013, "grad_norm": 10.985512733459473, "learning_rate": 2.8015e-06, "loss": 2.3382, "step": 5603 }, { "epoch": 1.7567398119122257, "grad_norm": 35.681644439697266, "learning_rate": 2.8020000000000004e-06, "loss": 2.9624, "step": 5604 }, { "epoch": 1.7570532915360502, "grad_norm": 24.726045608520508, "learning_rate": 2.8025000000000003e-06, "loss": 2.3681, "step": 5605 }, { "epoch": 1.7573667711598746, "grad_norm": 38.52384567260742, "learning_rate": 2.803e-06, "loss": 2.2576, "step": 5606 }, { "epoch": 1.757680250783699, "grad_norm": 21.216524124145508, "learning_rate": 2.8035e-06, "loss": 3.0623, "step": 5607 }, { "epoch": 1.7579937304075235, "grad_norm": 13.406828880310059, "learning_rate": 2.804e-06, "loss": 2.317, "step": 5608 }, { "epoch": 1.758307210031348, "grad_norm": 40.402992248535156, "learning_rate": 2.8045e-06, "loss": 2.461, "step": 5609 }, { "epoch": 1.7586206896551724, "grad_norm": 14.416365623474121, "learning_rate": 2.8050000000000007e-06, "loss": 2.4303, "step": 5610 }, { "epoch": 1.7589341692789968, "grad_norm": 38.3383903503418, "learning_rate": 2.8055000000000005e-06, "loss": 2.8362, "step": 5611 }, { "epoch": 1.7592476489028215, "grad_norm": 14.47104549407959, "learning_rate": 2.8060000000000003e-06, "loss": 2.3346, "step": 5612 }, { "epoch": 1.759561128526646, "grad_norm": 30.86193084716797, "learning_rate": 2.8065e-06, "loss": 2.4582, "step": 5613 }, { "epoch": 1.7598746081504704, "grad_norm": 22.559818267822266, "learning_rate": 2.8070000000000004e-06, "loss": 2.6912, "step": 5614 }, { "epoch": 1.7601880877742948, "grad_norm": 16.12966537475586, "learning_rate": 2.8075000000000002e-06, "loss": 2.3455, "step": 5615 }, { "epoch": 1.7605015673981192, "grad_norm": 20.706560134887695, "learning_rate": 2.808e-06, "loss": 2.2846, "step": 5616 }, { "epoch": 1.7608150470219437, "grad_norm": 25.860445022583008, "learning_rate": 2.8085000000000003e-06, "loss": 2.3389, "step": 5617 }, { "epoch": 1.7611285266457681, "grad_norm": 20.793603897094727, "learning_rate": 2.809e-06, "loss": 2.3378, "step": 5618 }, { "epoch": 1.7614420062695926, "grad_norm": 14.1869535446167, "learning_rate": 2.8095e-06, "loss": 2.5403, "step": 5619 }, { "epoch": 1.761755485893417, "grad_norm": 17.897798538208008, "learning_rate": 2.8100000000000006e-06, "loss": 2.4763, "step": 5620 }, { "epoch": 1.7620689655172415, "grad_norm": 14.624987602233887, "learning_rate": 2.8105000000000005e-06, "loss": 2.2365, "step": 5621 }, { "epoch": 1.762382445141066, "grad_norm": 51.904640197753906, "learning_rate": 2.8110000000000003e-06, "loss": 2.3868, "step": 5622 }, { "epoch": 1.7626959247648903, "grad_norm": 14.12952995300293, "learning_rate": 2.8115000000000005e-06, "loss": 2.2413, "step": 5623 }, { "epoch": 1.7630094043887148, "grad_norm": 34.63715744018555, "learning_rate": 2.8120000000000004e-06, "loss": 2.5888, "step": 5624 }, { "epoch": 1.7633228840125392, "grad_norm": 53.19129943847656, "learning_rate": 2.8125e-06, "loss": 2.4327, "step": 5625 }, { "epoch": 1.7636363636363637, "grad_norm": 26.255016326904297, "learning_rate": 2.813e-06, "loss": 2.3355, "step": 5626 }, { "epoch": 1.763949843260188, "grad_norm": 21.07077407836914, "learning_rate": 2.8135000000000003e-06, "loss": 2.2259, "step": 5627 }, { "epoch": 1.7642633228840126, "grad_norm": 27.582504272460938, "learning_rate": 2.814e-06, "loss": 2.3778, "step": 5628 }, { "epoch": 1.764576802507837, "grad_norm": 15.880824089050293, "learning_rate": 2.8145e-06, "loss": 2.355, "step": 5629 }, { "epoch": 1.7648902821316614, "grad_norm": 10.928897857666016, "learning_rate": 2.815e-06, "loss": 2.1491, "step": 5630 }, { "epoch": 1.7652037617554859, "grad_norm": 26.47492027282715, "learning_rate": 2.8155000000000004e-06, "loss": 2.2021, "step": 5631 }, { "epoch": 1.7655172413793103, "grad_norm": 193.11053466796875, "learning_rate": 2.8160000000000002e-06, "loss": 2.6495, "step": 5632 }, { "epoch": 1.7658307210031348, "grad_norm": 26.360347747802734, "learning_rate": 2.8165000000000005e-06, "loss": 2.2046, "step": 5633 }, { "epoch": 1.7661442006269592, "grad_norm": 22.70421028137207, "learning_rate": 2.8170000000000003e-06, "loss": 2.2962, "step": 5634 }, { "epoch": 1.7664576802507836, "grad_norm": 14.537629127502441, "learning_rate": 2.8175e-06, "loss": 2.3563, "step": 5635 }, { "epoch": 1.766771159874608, "grad_norm": 31.513465881347656, "learning_rate": 2.8180000000000004e-06, "loss": 2.3216, "step": 5636 }, { "epoch": 1.7670846394984325, "grad_norm": 15.880422592163086, "learning_rate": 2.8185e-06, "loss": 2.4413, "step": 5637 }, { "epoch": 1.767398119122257, "grad_norm": 18.613235473632812, "learning_rate": 2.819e-06, "loss": 2.5158, "step": 5638 }, { "epoch": 1.7677115987460814, "grad_norm": 14.459510803222656, "learning_rate": 2.8195e-06, "loss": 2.3783, "step": 5639 }, { "epoch": 1.7680250783699059, "grad_norm": 11.37955093383789, "learning_rate": 2.82e-06, "loss": 2.141, "step": 5640 }, { "epoch": 1.7683385579937303, "grad_norm": 20.611160278320312, "learning_rate": 2.8205000000000004e-06, "loss": 2.4464, "step": 5641 }, { "epoch": 1.7686520376175547, "grad_norm": 82.4933090209961, "learning_rate": 2.8210000000000006e-06, "loss": 2.3389, "step": 5642 }, { "epoch": 1.7689655172413792, "grad_norm": 15.467883110046387, "learning_rate": 2.8215000000000004e-06, "loss": 2.4918, "step": 5643 }, { "epoch": 1.7692789968652036, "grad_norm": 9.933538436889648, "learning_rate": 2.8220000000000003e-06, "loss": 2.1153, "step": 5644 }, { "epoch": 1.769592476489028, "grad_norm": 40.362850189208984, "learning_rate": 2.8225e-06, "loss": 2.2969, "step": 5645 }, { "epoch": 1.7699059561128525, "grad_norm": 35.1439094543457, "learning_rate": 2.8230000000000003e-06, "loss": 2.2496, "step": 5646 }, { "epoch": 1.770219435736677, "grad_norm": 16.86162757873535, "learning_rate": 2.8235e-06, "loss": 2.4232, "step": 5647 }, { "epoch": 1.7705329153605016, "grad_norm": 22.56192970275879, "learning_rate": 2.824e-06, "loss": 2.3042, "step": 5648 }, { "epoch": 1.770846394984326, "grad_norm": 30.514720916748047, "learning_rate": 2.8245e-06, "loss": 2.6457, "step": 5649 }, { "epoch": 1.7711598746081505, "grad_norm": 10.613741874694824, "learning_rate": 2.825e-06, "loss": 2.3016, "step": 5650 }, { "epoch": 1.771473354231975, "grad_norm": 15.156309127807617, "learning_rate": 2.8255000000000003e-06, "loss": 2.231, "step": 5651 }, { "epoch": 1.7717868338557994, "grad_norm": 81.3517074584961, "learning_rate": 2.8260000000000006e-06, "loss": 2.0737, "step": 5652 }, { "epoch": 1.7721003134796238, "grad_norm": 14.351300239562988, "learning_rate": 2.8265000000000004e-06, "loss": 2.106, "step": 5653 }, { "epoch": 1.7724137931034483, "grad_norm": 14.630460739135742, "learning_rate": 2.827e-06, "loss": 2.2051, "step": 5654 }, { "epoch": 1.7727272727272727, "grad_norm": 14.13293170928955, "learning_rate": 2.8275e-06, "loss": 2.3056, "step": 5655 }, { "epoch": 1.7730407523510971, "grad_norm": 7.3408026695251465, "learning_rate": 2.8280000000000003e-06, "loss": 2.1562, "step": 5656 }, { "epoch": 1.7733542319749216, "grad_norm": 86.32215881347656, "learning_rate": 2.8285e-06, "loss": 2.5063, "step": 5657 }, { "epoch": 1.773667711598746, "grad_norm": 15.42455768585205, "learning_rate": 2.829e-06, "loss": 2.2239, "step": 5658 }, { "epoch": 1.7739811912225707, "grad_norm": 18.74974822998047, "learning_rate": 2.8295e-06, "loss": 2.3227, "step": 5659 }, { "epoch": 1.7742946708463951, "grad_norm": 16.02386474609375, "learning_rate": 2.83e-06, "loss": 2.3212, "step": 5660 }, { "epoch": 1.7746081504702196, "grad_norm": 12.063882827758789, "learning_rate": 2.8305000000000003e-06, "loss": 2.3197, "step": 5661 }, { "epoch": 1.774921630094044, "grad_norm": 11.073810577392578, "learning_rate": 2.8310000000000005e-06, "loss": 2.2106, "step": 5662 }, { "epoch": 1.7752351097178685, "grad_norm": 35.60942840576172, "learning_rate": 2.8315000000000003e-06, "loss": 2.3919, "step": 5663 }, { "epoch": 1.775548589341693, "grad_norm": 11.828412055969238, "learning_rate": 2.832e-06, "loss": 2.1175, "step": 5664 }, { "epoch": 1.7758620689655173, "grad_norm": 42.694332122802734, "learning_rate": 2.8325000000000004e-06, "loss": 2.7357, "step": 5665 }, { "epoch": 1.7761755485893418, "grad_norm": 12.503061294555664, "learning_rate": 2.8330000000000002e-06, "loss": 2.5552, "step": 5666 }, { "epoch": 1.7764890282131662, "grad_norm": 11.709190368652344, "learning_rate": 2.8335e-06, "loss": 2.1738, "step": 5667 }, { "epoch": 1.7768025078369907, "grad_norm": 25.45659637451172, "learning_rate": 2.834e-06, "loss": 2.4358, "step": 5668 }, { "epoch": 1.7771159874608151, "grad_norm": 38.15581512451172, "learning_rate": 2.8345e-06, "loss": 2.1994, "step": 5669 }, { "epoch": 1.7774294670846396, "grad_norm": 65.19048309326172, "learning_rate": 2.835e-06, "loss": 4.0908, "step": 5670 }, { "epoch": 1.777742946708464, "grad_norm": 11.405741691589355, "learning_rate": 2.8355000000000006e-06, "loss": 2.2353, "step": 5671 }, { "epoch": 1.7780564263322884, "grad_norm": 16.600887298583984, "learning_rate": 2.8360000000000005e-06, "loss": 1.9491, "step": 5672 }, { "epoch": 1.7783699059561129, "grad_norm": 19.279136657714844, "learning_rate": 2.8365000000000003e-06, "loss": 2.1248, "step": 5673 }, { "epoch": 1.7786833855799373, "grad_norm": 20.416065216064453, "learning_rate": 2.837e-06, "loss": 2.6189, "step": 5674 }, { "epoch": 1.7789968652037618, "grad_norm": 12.208967208862305, "learning_rate": 2.8375000000000004e-06, "loss": 2.1409, "step": 5675 }, { "epoch": 1.7793103448275862, "grad_norm": 92.15316009521484, "learning_rate": 2.838e-06, "loss": 3.3053, "step": 5676 }, { "epoch": 1.7796238244514107, "grad_norm": 62.69886016845703, "learning_rate": 2.8385e-06, "loss": 3.2674, "step": 5677 }, { "epoch": 1.779937304075235, "grad_norm": 36.62449264526367, "learning_rate": 2.8390000000000003e-06, "loss": 2.3739, "step": 5678 }, { "epoch": 1.7802507836990595, "grad_norm": 30.333797454833984, "learning_rate": 2.8395e-06, "loss": 2.167, "step": 5679 }, { "epoch": 1.780564263322884, "grad_norm": 18.453176498413086, "learning_rate": 2.84e-06, "loss": 2.1609, "step": 5680 }, { "epoch": 1.7808777429467084, "grad_norm": 13.545697212219238, "learning_rate": 2.8405000000000006e-06, "loss": 2.1917, "step": 5681 }, { "epoch": 1.7811912225705329, "grad_norm": 37.35604476928711, "learning_rate": 2.8410000000000004e-06, "loss": 2.1792, "step": 5682 }, { "epoch": 1.7815047021943573, "grad_norm": 39.16984558105469, "learning_rate": 2.8415000000000002e-06, "loss": 2.2692, "step": 5683 }, { "epoch": 1.7818181818181817, "grad_norm": 45.23867416381836, "learning_rate": 2.8420000000000005e-06, "loss": 2.1975, "step": 5684 }, { "epoch": 1.7821316614420062, "grad_norm": 16.841379165649414, "learning_rate": 2.8425000000000003e-06, "loss": 2.2613, "step": 5685 }, { "epoch": 1.7824451410658306, "grad_norm": 14.964749336242676, "learning_rate": 2.843e-06, "loss": 2.6653, "step": 5686 }, { "epoch": 1.782758620689655, "grad_norm": 45.70800018310547, "learning_rate": 2.8435e-06, "loss": 2.3517, "step": 5687 }, { "epoch": 1.7830721003134795, "grad_norm": 26.868568420410156, "learning_rate": 2.8440000000000002e-06, "loss": 2.0901, "step": 5688 }, { "epoch": 1.783385579937304, "grad_norm": 67.97185516357422, "learning_rate": 2.8445e-06, "loss": 2.3446, "step": 5689 }, { "epoch": 1.7836990595611284, "grad_norm": 16.462514877319336, "learning_rate": 2.845e-06, "loss": 2.2372, "step": 5690 }, { "epoch": 1.7840125391849528, "grad_norm": 20.244930267333984, "learning_rate": 2.8455000000000005e-06, "loss": 2.4724, "step": 5691 }, { "epoch": 1.7843260188087773, "grad_norm": 31.71061134338379, "learning_rate": 2.8460000000000004e-06, "loss": 2.4825, "step": 5692 }, { "epoch": 1.7846394984326017, "grad_norm": 15.576581954956055, "learning_rate": 2.8465e-06, "loss": 1.9315, "step": 5693 }, { "epoch": 1.7849529780564264, "grad_norm": 21.8217716217041, "learning_rate": 2.8470000000000004e-06, "loss": 2.428, "step": 5694 }, { "epoch": 1.7852664576802508, "grad_norm": 12.862072944641113, "learning_rate": 2.8475000000000003e-06, "loss": 2.0832, "step": 5695 }, { "epoch": 1.7855799373040753, "grad_norm": 26.172510147094727, "learning_rate": 2.848e-06, "loss": 2.2061, "step": 5696 }, { "epoch": 1.7858934169278997, "grad_norm": 9.237683296203613, "learning_rate": 2.8485000000000003e-06, "loss": 2.2316, "step": 5697 }, { "epoch": 1.7862068965517242, "grad_norm": 34.316192626953125, "learning_rate": 2.849e-06, "loss": 2.188, "step": 5698 }, { "epoch": 1.7865203761755486, "grad_norm": 41.66139221191406, "learning_rate": 2.8495e-06, "loss": 2.1467, "step": 5699 }, { "epoch": 1.786833855799373, "grad_norm": 24.990121841430664, "learning_rate": 2.85e-06, "loss": 2.2904, "step": 5700 }, { "epoch": 1.7871473354231975, "grad_norm": 15.360755920410156, "learning_rate": 2.8505000000000005e-06, "loss": 2.3998, "step": 5701 }, { "epoch": 1.787460815047022, "grad_norm": 13.665358543395996, "learning_rate": 2.8510000000000003e-06, "loss": 1.9507, "step": 5702 }, { "epoch": 1.7877742946708464, "grad_norm": 16.122602462768555, "learning_rate": 2.8515000000000006e-06, "loss": 2.5211, "step": 5703 }, { "epoch": 1.7880877742946708, "grad_norm": 18.225542068481445, "learning_rate": 2.8520000000000004e-06, "loss": 2.325, "step": 5704 }, { "epoch": 1.7884012539184952, "grad_norm": 10.050536155700684, "learning_rate": 2.8525000000000002e-06, "loss": 2.3043, "step": 5705 }, { "epoch": 1.78871473354232, "grad_norm": 15.066789627075195, "learning_rate": 2.853e-06, "loss": 2.0683, "step": 5706 }, { "epoch": 1.7890282131661444, "grad_norm": 22.185789108276367, "learning_rate": 2.8535000000000003e-06, "loss": 2.4388, "step": 5707 }, { "epoch": 1.7893416927899688, "grad_norm": 29.357311248779297, "learning_rate": 2.854e-06, "loss": 2.3013, "step": 5708 }, { "epoch": 1.7896551724137932, "grad_norm": 40.66057205200195, "learning_rate": 2.8545e-06, "loss": 2.5906, "step": 5709 }, { "epoch": 1.7899686520376177, "grad_norm": 26.080907821655273, "learning_rate": 2.855e-06, "loss": 2.2303, "step": 5710 }, { "epoch": 1.7902821316614421, "grad_norm": 21.70749282836914, "learning_rate": 2.8555000000000004e-06, "loss": 2.2158, "step": 5711 }, { "epoch": 1.7905956112852666, "grad_norm": 12.691455841064453, "learning_rate": 2.8560000000000003e-06, "loss": 2.4867, "step": 5712 }, { "epoch": 1.790909090909091, "grad_norm": 13.39187240600586, "learning_rate": 2.8565000000000005e-06, "loss": 2.1281, "step": 5713 }, { "epoch": 1.7912225705329154, "grad_norm": 75.04694366455078, "learning_rate": 2.8570000000000003e-06, "loss": 2.0658, "step": 5714 }, { "epoch": 1.79153605015674, "grad_norm": 27.390806198120117, "learning_rate": 2.8575e-06, "loss": 2.898, "step": 5715 }, { "epoch": 1.7918495297805643, "grad_norm": 22.34684181213379, "learning_rate": 2.8580000000000004e-06, "loss": 2.4863, "step": 5716 }, { "epoch": 1.7921630094043888, "grad_norm": 19.219614028930664, "learning_rate": 2.8585000000000002e-06, "loss": 2.5412, "step": 5717 }, { "epoch": 1.7924764890282132, "grad_norm": 16.021224975585938, "learning_rate": 2.859e-06, "loss": 2.3316, "step": 5718 }, { "epoch": 1.7927899686520377, "grad_norm": 8.465279579162598, "learning_rate": 2.8595e-06, "loss": 2.2175, "step": 5719 }, { "epoch": 1.793103448275862, "grad_norm": 15.827347755432129, "learning_rate": 2.86e-06, "loss": 2.2974, "step": 5720 }, { "epoch": 1.7934169278996865, "grad_norm": 15.751001358032227, "learning_rate": 2.8605000000000004e-06, "loss": 2.335, "step": 5721 }, { "epoch": 1.793730407523511, "grad_norm": 15.49514389038086, "learning_rate": 2.8610000000000006e-06, "loss": 2.2435, "step": 5722 }, { "epoch": 1.7940438871473354, "grad_norm": 53.53825759887695, "learning_rate": 2.8615000000000005e-06, "loss": 2.6453, "step": 5723 }, { "epoch": 1.7943573667711599, "grad_norm": 19.740135192871094, "learning_rate": 2.8620000000000003e-06, "loss": 2.3032, "step": 5724 }, { "epoch": 1.7946708463949843, "grad_norm": 40.47820281982422, "learning_rate": 2.8625e-06, "loss": 2.433, "step": 5725 }, { "epoch": 1.7949843260188088, "grad_norm": 13.667232513427734, "learning_rate": 2.8630000000000004e-06, "loss": 2.0633, "step": 5726 }, { "epoch": 1.7952978056426332, "grad_norm": 133.35665893554688, "learning_rate": 2.8635e-06, "loss": 2.2366, "step": 5727 }, { "epoch": 1.7956112852664576, "grad_norm": 19.358415603637695, "learning_rate": 2.864e-06, "loss": 2.5498, "step": 5728 }, { "epoch": 1.795924764890282, "grad_norm": 69.85310363769531, "learning_rate": 2.8645e-06, "loss": 2.8183, "step": 5729 }, { "epoch": 1.7962382445141065, "grad_norm": 21.108644485473633, "learning_rate": 2.865e-06, "loss": 2.4494, "step": 5730 }, { "epoch": 1.796551724137931, "grad_norm": 12.527911186218262, "learning_rate": 2.8655000000000003e-06, "loss": 2.2376, "step": 5731 }, { "epoch": 1.7968652037617554, "grad_norm": 18.049640655517578, "learning_rate": 2.8660000000000006e-06, "loss": 2.2159, "step": 5732 }, { "epoch": 1.7971786833855798, "grad_norm": 14.708891868591309, "learning_rate": 2.8665000000000004e-06, "loss": 2.0556, "step": 5733 }, { "epoch": 1.7974921630094043, "grad_norm": 20.81380271911621, "learning_rate": 2.8670000000000002e-06, "loss": 2.6582, "step": 5734 }, { "epoch": 1.7978056426332287, "grad_norm": 16.69767951965332, "learning_rate": 2.8675e-06, "loss": 2.5488, "step": 5735 }, { "epoch": 1.7981191222570532, "grad_norm": 25.40092658996582, "learning_rate": 2.8680000000000003e-06, "loss": 2.279, "step": 5736 }, { "epoch": 1.7984326018808776, "grad_norm": 8.9202241897583, "learning_rate": 2.8685e-06, "loss": 2.3573, "step": 5737 }, { "epoch": 1.798746081504702, "grad_norm": 26.168312072753906, "learning_rate": 2.869e-06, "loss": 2.3775, "step": 5738 }, { "epoch": 1.7990595611285265, "grad_norm": 14.53439712524414, "learning_rate": 2.8695000000000002e-06, "loss": 2.5065, "step": 5739 }, { "epoch": 1.799373040752351, "grad_norm": 16.31795883178711, "learning_rate": 2.87e-06, "loss": 2.4633, "step": 5740 }, { "epoch": 1.7996865203761756, "grad_norm": 28.0877742767334, "learning_rate": 2.8705000000000003e-06, "loss": 2.4924, "step": 5741 }, { "epoch": 1.8, "grad_norm": 18.114452362060547, "learning_rate": 2.8710000000000005e-06, "loss": 2.3432, "step": 5742 }, { "epoch": 1.8003134796238245, "grad_norm": 16.870342254638672, "learning_rate": 2.8715000000000004e-06, "loss": 2.624, "step": 5743 }, { "epoch": 1.800626959247649, "grad_norm": 11.62446403503418, "learning_rate": 2.872e-06, "loss": 2.2845, "step": 5744 }, { "epoch": 1.8009404388714734, "grad_norm": 13.54909896850586, "learning_rate": 2.8725000000000004e-06, "loss": 2.8479, "step": 5745 }, { "epoch": 1.8012539184952978, "grad_norm": 11.23021125793457, "learning_rate": 2.8730000000000003e-06, "loss": 2.3157, "step": 5746 }, { "epoch": 1.8015673981191223, "grad_norm": 24.773460388183594, "learning_rate": 2.8735e-06, "loss": 2.1324, "step": 5747 }, { "epoch": 1.8018808777429467, "grad_norm": 28.940399169921875, "learning_rate": 2.874e-06, "loss": 2.3426, "step": 5748 }, { "epoch": 1.8021943573667711, "grad_norm": 16.026710510253906, "learning_rate": 2.8745e-06, "loss": 2.1109, "step": 5749 }, { "epoch": 1.8025078369905956, "grad_norm": 24.745437622070312, "learning_rate": 2.875e-06, "loss": 2.5266, "step": 5750 }, { "epoch": 1.80282131661442, "grad_norm": 12.666834831237793, "learning_rate": 2.8755e-06, "loss": 2.153, "step": 5751 }, { "epoch": 1.8031347962382445, "grad_norm": 21.118148803710938, "learning_rate": 2.8760000000000005e-06, "loss": 2.0629, "step": 5752 }, { "epoch": 1.8034482758620691, "grad_norm": 40.68550491333008, "learning_rate": 2.8765000000000003e-06, "loss": 2.7818, "step": 5753 }, { "epoch": 1.8037617554858936, "grad_norm": 8.309558868408203, "learning_rate": 2.877e-06, "loss": 2.3525, "step": 5754 }, { "epoch": 1.804075235109718, "grad_norm": 18.94241714477539, "learning_rate": 2.8775000000000004e-06, "loss": 2.2492, "step": 5755 }, { "epoch": 1.8043887147335425, "grad_norm": 77.01664733886719, "learning_rate": 2.8780000000000002e-06, "loss": 2.0521, "step": 5756 }, { "epoch": 1.804702194357367, "grad_norm": 38.76837158203125, "learning_rate": 2.8785e-06, "loss": 3.3167, "step": 5757 }, { "epoch": 1.8050156739811913, "grad_norm": 24.91385841369629, "learning_rate": 2.8790000000000003e-06, "loss": 2.5723, "step": 5758 }, { "epoch": 1.8053291536050158, "grad_norm": 35.777259826660156, "learning_rate": 2.8795e-06, "loss": 2.3896, "step": 5759 }, { "epoch": 1.8056426332288402, "grad_norm": 39.020076751708984, "learning_rate": 2.88e-06, "loss": 2.5884, "step": 5760 }, { "epoch": 1.8059561128526647, "grad_norm": 31.04021453857422, "learning_rate": 2.8804999999999998e-06, "loss": 2.3261, "step": 5761 }, { "epoch": 1.806269592476489, "grad_norm": 17.670146942138672, "learning_rate": 2.8810000000000005e-06, "loss": 2.0315, "step": 5762 }, { "epoch": 1.8065830721003135, "grad_norm": 21.669349670410156, "learning_rate": 2.8815000000000003e-06, "loss": 2.3209, "step": 5763 }, { "epoch": 1.806896551724138, "grad_norm": 21.78011703491211, "learning_rate": 2.8820000000000005e-06, "loss": 2.4031, "step": 5764 }, { "epoch": 1.8072100313479624, "grad_norm": 16.515914916992188, "learning_rate": 2.8825000000000004e-06, "loss": 2.2467, "step": 5765 }, { "epoch": 1.8075235109717869, "grad_norm": 14.0270414352417, "learning_rate": 2.883e-06, "loss": 2.7854, "step": 5766 }, { "epoch": 1.8078369905956113, "grad_norm": 20.48775291442871, "learning_rate": 2.8835e-06, "loss": 2.5621, "step": 5767 }, { "epoch": 1.8081504702194358, "grad_norm": 11.038260459899902, "learning_rate": 2.8840000000000003e-06, "loss": 2.1779, "step": 5768 }, { "epoch": 1.8084639498432602, "grad_norm": 19.993820190429688, "learning_rate": 2.8845e-06, "loss": 2.2544, "step": 5769 }, { "epoch": 1.8087774294670846, "grad_norm": 18.76875114440918, "learning_rate": 2.885e-06, "loss": 2.74, "step": 5770 }, { "epoch": 1.809090909090909, "grad_norm": 29.65410804748535, "learning_rate": 2.8855e-06, "loss": 2.3125, "step": 5771 }, { "epoch": 1.8094043887147335, "grad_norm": 8.846988677978516, "learning_rate": 2.8860000000000004e-06, "loss": 2.4192, "step": 5772 }, { "epoch": 1.809717868338558, "grad_norm": 15.321855545043945, "learning_rate": 2.8865000000000002e-06, "loss": 2.3592, "step": 5773 }, { "epoch": 1.8100313479623824, "grad_norm": 68.62010955810547, "learning_rate": 2.8870000000000005e-06, "loss": 2.7464, "step": 5774 }, { "epoch": 1.8103448275862069, "grad_norm": 11.731450080871582, "learning_rate": 2.8875000000000003e-06, "loss": 2.1994, "step": 5775 }, { "epoch": 1.8106583072100313, "grad_norm": 22.625207901000977, "learning_rate": 2.888e-06, "loss": 2.3414, "step": 5776 }, { "epoch": 1.8109717868338557, "grad_norm": 12.00716781616211, "learning_rate": 2.8885000000000004e-06, "loss": 2.2781, "step": 5777 }, { "epoch": 1.8112852664576802, "grad_norm": 15.153823852539062, "learning_rate": 2.889e-06, "loss": 2.3895, "step": 5778 }, { "epoch": 1.8115987460815046, "grad_norm": 33.36376190185547, "learning_rate": 2.8895e-06, "loss": 2.9956, "step": 5779 }, { "epoch": 1.811912225705329, "grad_norm": 16.726226806640625, "learning_rate": 2.89e-06, "loss": 2.2121, "step": 5780 }, { "epoch": 1.8122257053291535, "grad_norm": 102.48206329345703, "learning_rate": 2.8905e-06, "loss": 2.3842, "step": 5781 }, { "epoch": 1.812539184952978, "grad_norm": 11.28581714630127, "learning_rate": 2.8910000000000004e-06, "loss": 2.3171, "step": 5782 }, { "epoch": 1.8128526645768024, "grad_norm": 13.72201156616211, "learning_rate": 2.8915000000000006e-06, "loss": 2.279, "step": 5783 }, { "epoch": 1.8131661442006268, "grad_norm": 23.668373107910156, "learning_rate": 2.8920000000000004e-06, "loss": 2.2777, "step": 5784 }, { "epoch": 1.8134796238244513, "grad_norm": 18.73801040649414, "learning_rate": 2.8925000000000003e-06, "loss": 2.475, "step": 5785 }, { "epoch": 1.8137931034482757, "grad_norm": 22.289701461791992, "learning_rate": 2.893e-06, "loss": 2.302, "step": 5786 }, { "epoch": 1.8141065830721002, "grad_norm": 12.893795013427734, "learning_rate": 2.8935000000000003e-06, "loss": 2.3009, "step": 5787 }, { "epoch": 1.8144200626959248, "grad_norm": 21.486906051635742, "learning_rate": 2.894e-06, "loss": 2.6981, "step": 5788 }, { "epoch": 1.8147335423197493, "grad_norm": 19.843420028686523, "learning_rate": 2.8945e-06, "loss": 1.9245, "step": 5789 }, { "epoch": 1.8150470219435737, "grad_norm": 16.991474151611328, "learning_rate": 2.8950000000000002e-06, "loss": 2.3261, "step": 5790 }, { "epoch": 1.8153605015673981, "grad_norm": 15.307354927062988, "learning_rate": 2.8955e-06, "loss": 2.215, "step": 5791 }, { "epoch": 1.8156739811912226, "grad_norm": 18.179475784301758, "learning_rate": 2.8960000000000003e-06, "loss": 2.4475, "step": 5792 }, { "epoch": 1.815987460815047, "grad_norm": 16.987171173095703, "learning_rate": 2.8965000000000006e-06, "loss": 2.2285, "step": 5793 }, { "epoch": 1.8163009404388715, "grad_norm": 11.264395713806152, "learning_rate": 2.8970000000000004e-06, "loss": 2.2343, "step": 5794 }, { "epoch": 1.816614420062696, "grad_norm": 28.56736946105957, "learning_rate": 2.8975e-06, "loss": 2.4624, "step": 5795 }, { "epoch": 1.8169278996865204, "grad_norm": 11.755648612976074, "learning_rate": 2.8980000000000005e-06, "loss": 2.2397, "step": 5796 }, { "epoch": 1.8172413793103448, "grad_norm": 9.479302406311035, "learning_rate": 2.8985000000000003e-06, "loss": 2.146, "step": 5797 }, { "epoch": 1.8175548589341692, "grad_norm": 82.61997985839844, "learning_rate": 2.899e-06, "loss": 2.3617, "step": 5798 }, { "epoch": 1.8178683385579937, "grad_norm": 10.157063484191895, "learning_rate": 2.8995e-06, "loss": 2.0932, "step": 5799 }, { "epoch": 1.8181818181818183, "grad_norm": 11.576388359069824, "learning_rate": 2.9e-06, "loss": 2.2464, "step": 5800 }, { "epoch": 1.8184952978056428, "grad_norm": 14.041361808776855, "learning_rate": 2.9005e-06, "loss": 2.2786, "step": 5801 }, { "epoch": 1.8188087774294672, "grad_norm": 44.641910552978516, "learning_rate": 2.9010000000000007e-06, "loss": 2.2961, "step": 5802 }, { "epoch": 1.8191222570532917, "grad_norm": 15.94238567352295, "learning_rate": 2.9015000000000005e-06, "loss": 1.8881, "step": 5803 }, { "epoch": 1.8194357366771161, "grad_norm": 46.97993087768555, "learning_rate": 2.9020000000000003e-06, "loss": 2.2066, "step": 5804 }, { "epoch": 1.8197492163009406, "grad_norm": 11.219070434570312, "learning_rate": 2.9025e-06, "loss": 2.0557, "step": 5805 }, { "epoch": 1.820062695924765, "grad_norm": 44.40027618408203, "learning_rate": 2.9030000000000004e-06, "loss": 2.4332, "step": 5806 }, { "epoch": 1.8203761755485894, "grad_norm": 23.074581146240234, "learning_rate": 2.9035000000000002e-06, "loss": 2.328, "step": 5807 }, { "epoch": 1.8206896551724139, "grad_norm": 23.892196655273438, "learning_rate": 2.904e-06, "loss": 2.2622, "step": 5808 }, { "epoch": 1.8210031347962383, "grad_norm": 68.89410400390625, "learning_rate": 2.9045e-06, "loss": 3.0924, "step": 5809 }, { "epoch": 1.8213166144200628, "grad_norm": 11.803862571716309, "learning_rate": 2.905e-06, "loss": 2.1911, "step": 5810 }, { "epoch": 1.8216300940438872, "grad_norm": 25.095535278320312, "learning_rate": 2.9055e-06, "loss": 2.3491, "step": 5811 }, { "epoch": 1.8219435736677116, "grad_norm": 13.671652793884277, "learning_rate": 2.9060000000000006e-06, "loss": 2.2188, "step": 5812 }, { "epoch": 1.822257053291536, "grad_norm": 25.322555541992188, "learning_rate": 2.9065000000000005e-06, "loss": 2.3493, "step": 5813 }, { "epoch": 1.8225705329153605, "grad_norm": 27.850494384765625, "learning_rate": 2.9070000000000003e-06, "loss": 2.192, "step": 5814 }, { "epoch": 1.822884012539185, "grad_norm": 23.03537368774414, "learning_rate": 2.9075e-06, "loss": 2.2677, "step": 5815 }, { "epoch": 1.8231974921630094, "grad_norm": 63.14573287963867, "learning_rate": 2.9080000000000004e-06, "loss": 2.7937, "step": 5816 }, { "epoch": 1.8235109717868339, "grad_norm": 93.47972869873047, "learning_rate": 2.9085e-06, "loss": 2.2098, "step": 5817 }, { "epoch": 1.8238244514106583, "grad_norm": 24.212627410888672, "learning_rate": 2.909e-06, "loss": 2.2872, "step": 5818 }, { "epoch": 1.8241379310344827, "grad_norm": 11.589085578918457, "learning_rate": 2.9095000000000003e-06, "loss": 2.227, "step": 5819 }, { "epoch": 1.8244514106583072, "grad_norm": 29.045310974121094, "learning_rate": 2.91e-06, "loss": 2.272, "step": 5820 }, { "epoch": 1.8247648902821316, "grad_norm": 112.3254623413086, "learning_rate": 2.9105e-06, "loss": 2.721, "step": 5821 }, { "epoch": 1.825078369905956, "grad_norm": 16.254894256591797, "learning_rate": 2.9110000000000006e-06, "loss": 2.5039, "step": 5822 }, { "epoch": 1.8253918495297805, "grad_norm": 18.320463180541992, "learning_rate": 2.9115000000000004e-06, "loss": 2.1283, "step": 5823 }, { "epoch": 1.825705329153605, "grad_norm": 16.942607879638672, "learning_rate": 2.9120000000000002e-06, "loss": 2.5775, "step": 5824 }, { "epoch": 1.8260188087774294, "grad_norm": 44.85321807861328, "learning_rate": 2.9125000000000005e-06, "loss": 2.0103, "step": 5825 }, { "epoch": 1.8263322884012538, "grad_norm": 16.66575050354004, "learning_rate": 2.9130000000000003e-06, "loss": 1.7175, "step": 5826 }, { "epoch": 1.8266457680250783, "grad_norm": 9.837501525878906, "learning_rate": 2.9135e-06, "loss": 2.1418, "step": 5827 }, { "epoch": 1.8269592476489027, "grad_norm": 75.40845489501953, "learning_rate": 2.914e-06, "loss": 2.7283, "step": 5828 }, { "epoch": 1.8272727272727272, "grad_norm": 23.27977752685547, "learning_rate": 2.9145000000000002e-06, "loss": 2.1375, "step": 5829 }, { "epoch": 1.8275862068965516, "grad_norm": 27.287527084350586, "learning_rate": 2.915e-06, "loss": 2.1722, "step": 5830 }, { "epoch": 1.827899686520376, "grad_norm": 23.660470962524414, "learning_rate": 2.9155e-06, "loss": 2.3881, "step": 5831 }, { "epoch": 1.8282131661442005, "grad_norm": 13.798928260803223, "learning_rate": 2.9160000000000005e-06, "loss": 2.1995, "step": 5832 }, { "epoch": 1.828526645768025, "grad_norm": 72.20987701416016, "learning_rate": 2.9165000000000004e-06, "loss": 2.6333, "step": 5833 }, { "epoch": 1.8288401253918494, "grad_norm": 19.457874298095703, "learning_rate": 2.917e-06, "loss": 2.2748, "step": 5834 }, { "epoch": 1.829153605015674, "grad_norm": 16.82941436767578, "learning_rate": 2.9175000000000004e-06, "loss": 2.2334, "step": 5835 }, { "epoch": 1.8294670846394985, "grad_norm": 46.999229431152344, "learning_rate": 2.9180000000000003e-06, "loss": 2.0499, "step": 5836 }, { "epoch": 1.829780564263323, "grad_norm": 11.347408294677734, "learning_rate": 2.9185e-06, "loss": 2.1089, "step": 5837 }, { "epoch": 1.8300940438871474, "grad_norm": 7.4026665687561035, "learning_rate": 2.9190000000000003e-06, "loss": 2.3338, "step": 5838 }, { "epoch": 1.8304075235109718, "grad_norm": 17.908647537231445, "learning_rate": 2.9195e-06, "loss": 2.3173, "step": 5839 }, { "epoch": 1.8307210031347962, "grad_norm": 36.890201568603516, "learning_rate": 2.92e-06, "loss": 2.7883, "step": 5840 }, { "epoch": 1.8310344827586207, "grad_norm": 19.727296829223633, "learning_rate": 2.9205e-06, "loss": 2.3895, "step": 5841 }, { "epoch": 1.8313479623824451, "grad_norm": 18.831193923950195, "learning_rate": 2.9210000000000005e-06, "loss": 2.6153, "step": 5842 }, { "epoch": 1.8316614420062696, "grad_norm": 79.4319839477539, "learning_rate": 2.9215000000000003e-06, "loss": 2.4548, "step": 5843 }, { "epoch": 1.831974921630094, "grad_norm": 14.499472618103027, "learning_rate": 2.9220000000000006e-06, "loss": 2.1656, "step": 5844 }, { "epoch": 1.8322884012539185, "grad_norm": 16.53804588317871, "learning_rate": 2.9225000000000004e-06, "loss": 2.2322, "step": 5845 }, { "epoch": 1.832601880877743, "grad_norm": 22.1581974029541, "learning_rate": 2.9230000000000002e-06, "loss": 2.0958, "step": 5846 }, { "epoch": 1.8329153605015676, "grad_norm": 31.330753326416016, "learning_rate": 2.9235e-06, "loss": 2.1026, "step": 5847 }, { "epoch": 1.833228840125392, "grad_norm": 19.017539978027344, "learning_rate": 2.9240000000000003e-06, "loss": 2.2064, "step": 5848 }, { "epoch": 1.8335423197492164, "grad_norm": 23.104690551757812, "learning_rate": 2.9245e-06, "loss": 2.5904, "step": 5849 }, { "epoch": 1.8338557993730409, "grad_norm": 24.975191116333008, "learning_rate": 2.925e-06, "loss": 1.9996, "step": 5850 }, { "epoch": 1.8341692789968653, "grad_norm": 15.302990913391113, "learning_rate": 2.9255e-06, "loss": 2.1248, "step": 5851 }, { "epoch": 1.8344827586206898, "grad_norm": 21.22588348388672, "learning_rate": 2.9260000000000004e-06, "loss": 2.5368, "step": 5852 }, { "epoch": 1.8347962382445142, "grad_norm": 184.5251007080078, "learning_rate": 2.9265000000000003e-06, "loss": 2.5012, "step": 5853 }, { "epoch": 1.8351097178683387, "grad_norm": 10.427804946899414, "learning_rate": 2.9270000000000005e-06, "loss": 2.2236, "step": 5854 }, { "epoch": 1.835423197492163, "grad_norm": 22.081449508666992, "learning_rate": 2.9275000000000003e-06, "loss": 2.3903, "step": 5855 }, { "epoch": 1.8357366771159875, "grad_norm": 11.947712898254395, "learning_rate": 2.928e-06, "loss": 2.1356, "step": 5856 }, { "epoch": 1.836050156739812, "grad_norm": 15.462569236755371, "learning_rate": 2.9285000000000004e-06, "loss": 2.1701, "step": 5857 }, { "epoch": 1.8363636363636364, "grad_norm": 13.894822120666504, "learning_rate": 2.9290000000000002e-06, "loss": 1.9988, "step": 5858 }, { "epoch": 1.8366771159874609, "grad_norm": 15.583471298217773, "learning_rate": 2.9295e-06, "loss": 2.6659, "step": 5859 }, { "epoch": 1.8369905956112853, "grad_norm": 15.877820014953613, "learning_rate": 2.93e-06, "loss": 2.5541, "step": 5860 }, { "epoch": 1.8373040752351097, "grad_norm": 13.1459379196167, "learning_rate": 2.9305e-06, "loss": 2.1646, "step": 5861 }, { "epoch": 1.8376175548589342, "grad_norm": 15.416597366333008, "learning_rate": 2.9310000000000004e-06, "loss": 2.2812, "step": 5862 }, { "epoch": 1.8379310344827586, "grad_norm": 32.908973693847656, "learning_rate": 2.9315000000000006e-06, "loss": 2.3487, "step": 5863 }, { "epoch": 1.838244514106583, "grad_norm": 9.497340202331543, "learning_rate": 2.9320000000000005e-06, "loss": 2.2336, "step": 5864 }, { "epoch": 1.8385579937304075, "grad_norm": 23.483781814575195, "learning_rate": 2.9325000000000003e-06, "loss": 2.5881, "step": 5865 }, { "epoch": 1.838871473354232, "grad_norm": 13.98792552947998, "learning_rate": 2.933e-06, "loss": 2.2051, "step": 5866 }, { "epoch": 1.8391849529780564, "grad_norm": 13.942326545715332, "learning_rate": 2.9335000000000004e-06, "loss": 2.2039, "step": 5867 }, { "epoch": 1.8394984326018808, "grad_norm": 10.772849082946777, "learning_rate": 2.934e-06, "loss": 2.1795, "step": 5868 }, { "epoch": 1.8398119122257053, "grad_norm": 22.885339736938477, "learning_rate": 2.9345e-06, "loss": 2.2539, "step": 5869 }, { "epoch": 1.8401253918495297, "grad_norm": 13.527241706848145, "learning_rate": 2.9350000000000003e-06, "loss": 2.2251, "step": 5870 }, { "epoch": 1.8404388714733542, "grad_norm": 32.79869079589844, "learning_rate": 2.9355e-06, "loss": 2.3778, "step": 5871 }, { "epoch": 1.8407523510971786, "grad_norm": 87.69104766845703, "learning_rate": 2.9360000000000003e-06, "loss": 2.5287, "step": 5872 }, { "epoch": 1.841065830721003, "grad_norm": 8.843754768371582, "learning_rate": 2.9365000000000006e-06, "loss": 2.1424, "step": 5873 }, { "epoch": 1.8413793103448275, "grad_norm": 20.374862670898438, "learning_rate": 2.9370000000000004e-06, "loss": 2.5718, "step": 5874 }, { "epoch": 1.841692789968652, "grad_norm": 13.062189102172852, "learning_rate": 2.9375000000000003e-06, "loss": 2.1401, "step": 5875 }, { "epoch": 1.8420062695924764, "grad_norm": 15.953989028930664, "learning_rate": 2.9380000000000005e-06, "loss": 2.4322, "step": 5876 }, { "epoch": 1.8423197492163008, "grad_norm": 11.689451217651367, "learning_rate": 2.9385000000000003e-06, "loss": 2.2732, "step": 5877 }, { "epoch": 1.8426332288401253, "grad_norm": 14.597411155700684, "learning_rate": 2.939e-06, "loss": 2.0253, "step": 5878 }, { "epoch": 1.8429467084639497, "grad_norm": 15.932793617248535, "learning_rate": 2.9395e-06, "loss": 2.3347, "step": 5879 }, { "epoch": 1.8432601880877741, "grad_norm": 16.13714599609375, "learning_rate": 2.9400000000000002e-06, "loss": 2.2366, "step": 5880 }, { "epoch": 1.8435736677115986, "grad_norm": 87.59251403808594, "learning_rate": 2.9405e-06, "loss": 2.5404, "step": 5881 }, { "epoch": 1.8438871473354232, "grad_norm": 13.624130249023438, "learning_rate": 2.941e-06, "loss": 2.4282, "step": 5882 }, { "epoch": 1.8442006269592477, "grad_norm": 15.835795402526855, "learning_rate": 2.9415000000000006e-06, "loss": 2.9211, "step": 5883 }, { "epoch": 1.8445141065830721, "grad_norm": 23.732357025146484, "learning_rate": 2.9420000000000004e-06, "loss": 1.9677, "step": 5884 }, { "epoch": 1.8448275862068966, "grad_norm": 12.846848487854004, "learning_rate": 2.9425e-06, "loss": 2.6445, "step": 5885 }, { "epoch": 1.845141065830721, "grad_norm": 96.78413391113281, "learning_rate": 2.9430000000000005e-06, "loss": 2.6282, "step": 5886 }, { "epoch": 1.8454545454545455, "grad_norm": 57.51097106933594, "learning_rate": 2.9435000000000003e-06, "loss": 2.7437, "step": 5887 }, { "epoch": 1.84576802507837, "grad_norm": 12.915363311767578, "learning_rate": 2.944e-06, "loss": 2.2672, "step": 5888 }, { "epoch": 1.8460815047021943, "grad_norm": 40.32135009765625, "learning_rate": 2.9445e-06, "loss": 2.1966, "step": 5889 }, { "epoch": 1.8463949843260188, "grad_norm": 11.176884651184082, "learning_rate": 2.945e-06, "loss": 2.4467, "step": 5890 }, { "epoch": 1.8467084639498432, "grad_norm": 12.287383079528809, "learning_rate": 2.9455e-06, "loss": 2.1618, "step": 5891 }, { "epoch": 1.8470219435736677, "grad_norm": 9.98103141784668, "learning_rate": 2.946e-06, "loss": 2.647, "step": 5892 }, { "epoch": 1.847335423197492, "grad_norm": 24.924226760864258, "learning_rate": 2.9465000000000005e-06, "loss": 2.3428, "step": 5893 }, { "epoch": 1.8476489028213168, "grad_norm": 18.21531105041504, "learning_rate": 2.9470000000000003e-06, "loss": 2.1155, "step": 5894 }, { "epoch": 1.8479623824451412, "grad_norm": 8.226020812988281, "learning_rate": 2.9475e-06, "loss": 2.15, "step": 5895 }, { "epoch": 1.8482758620689657, "grad_norm": 15.366581916809082, "learning_rate": 2.9480000000000004e-06, "loss": 2.0918, "step": 5896 }, { "epoch": 1.84858934169279, "grad_norm": 27.560998916625977, "learning_rate": 2.9485000000000002e-06, "loss": 2.3529, "step": 5897 }, { "epoch": 1.8489028213166145, "grad_norm": 11.260773658752441, "learning_rate": 2.949e-06, "loss": 2.1609, "step": 5898 }, { "epoch": 1.849216300940439, "grad_norm": 16.45672607421875, "learning_rate": 2.9495000000000003e-06, "loss": 2.0179, "step": 5899 }, { "epoch": 1.8495297805642634, "grad_norm": 11.358928680419922, "learning_rate": 2.95e-06, "loss": 2.5883, "step": 5900 }, { "epoch": 1.8498432601880879, "grad_norm": 13.168031692504883, "learning_rate": 2.9505e-06, "loss": 2.1458, "step": 5901 }, { "epoch": 1.8501567398119123, "grad_norm": 22.90192413330078, "learning_rate": 2.9509999999999998e-06, "loss": 2.3417, "step": 5902 }, { "epoch": 1.8504702194357368, "grad_norm": 11.460537910461426, "learning_rate": 2.9515000000000005e-06, "loss": 2.3871, "step": 5903 }, { "epoch": 1.8507836990595612, "grad_norm": 116.56676483154297, "learning_rate": 2.9520000000000003e-06, "loss": 2.7412, "step": 5904 }, { "epoch": 1.8510971786833856, "grad_norm": 17.3189640045166, "learning_rate": 2.9525000000000005e-06, "loss": 2.1462, "step": 5905 }, { "epoch": 1.85141065830721, "grad_norm": 25.294736862182617, "learning_rate": 2.9530000000000004e-06, "loss": 2.2008, "step": 5906 }, { "epoch": 1.8517241379310345, "grad_norm": 13.655937194824219, "learning_rate": 2.9535e-06, "loss": 2.2359, "step": 5907 }, { "epoch": 1.852037617554859, "grad_norm": 51.57292938232422, "learning_rate": 2.954e-06, "loss": 2.4588, "step": 5908 }, { "epoch": 1.8523510971786834, "grad_norm": 17.087007522583008, "learning_rate": 2.9545000000000003e-06, "loss": 2.1137, "step": 5909 }, { "epoch": 1.8526645768025078, "grad_norm": 20.294261932373047, "learning_rate": 2.955e-06, "loss": 2.1156, "step": 5910 }, { "epoch": 1.8529780564263323, "grad_norm": 12.528114318847656, "learning_rate": 2.9555e-06, "loss": 1.9428, "step": 5911 }, { "epoch": 1.8532915360501567, "grad_norm": 25.253681182861328, "learning_rate": 2.956e-06, "loss": 2.5651, "step": 5912 }, { "epoch": 1.8536050156739812, "grad_norm": 19.341289520263672, "learning_rate": 2.9565000000000004e-06, "loss": 2.2829, "step": 5913 }, { "epoch": 1.8539184952978056, "grad_norm": 24.48165512084961, "learning_rate": 2.9570000000000002e-06, "loss": 2.2775, "step": 5914 }, { "epoch": 1.85423197492163, "grad_norm": 124.07843780517578, "learning_rate": 2.9575000000000005e-06, "loss": 4.0833, "step": 5915 }, { "epoch": 1.8545454545454545, "grad_norm": 44.97792053222656, "learning_rate": 2.9580000000000003e-06, "loss": 2.6771, "step": 5916 }, { "epoch": 1.854858934169279, "grad_norm": 37.24640655517578, "learning_rate": 2.9585e-06, "loss": 3.0157, "step": 5917 }, { "epoch": 1.8551724137931034, "grad_norm": 19.38148307800293, "learning_rate": 2.9590000000000004e-06, "loss": 2.6838, "step": 5918 }, { "epoch": 1.8554858934169278, "grad_norm": 20.10344886779785, "learning_rate": 2.9595e-06, "loss": 2.3734, "step": 5919 }, { "epoch": 1.8557993730407523, "grad_norm": 28.982986450195312, "learning_rate": 2.96e-06, "loss": 2.5651, "step": 5920 }, { "epoch": 1.8561128526645767, "grad_norm": 12.973660469055176, "learning_rate": 2.9605e-06, "loss": 2.204, "step": 5921 }, { "epoch": 1.8564263322884011, "grad_norm": 10.518831253051758, "learning_rate": 2.961e-06, "loss": 2.1735, "step": 5922 }, { "epoch": 1.8567398119122256, "grad_norm": 13.175191879272461, "learning_rate": 2.9615000000000004e-06, "loss": 2.236, "step": 5923 }, { "epoch": 1.85705329153605, "grad_norm": 23.340044021606445, "learning_rate": 2.9620000000000006e-06, "loss": 2.7431, "step": 5924 }, { "epoch": 1.8573667711598745, "grad_norm": 19.4545955657959, "learning_rate": 2.9625000000000004e-06, "loss": 2.6503, "step": 5925 }, { "epoch": 1.857680250783699, "grad_norm": 12.167764663696289, "learning_rate": 2.9630000000000003e-06, "loss": 2.2549, "step": 5926 }, { "epoch": 1.8579937304075234, "grad_norm": 20.977935791015625, "learning_rate": 2.9635e-06, "loss": 2.432, "step": 5927 }, { "epoch": 1.8583072100313478, "grad_norm": 22.710613250732422, "learning_rate": 2.9640000000000003e-06, "loss": 2.5537, "step": 5928 }, { "epoch": 1.8586206896551725, "grad_norm": 21.324052810668945, "learning_rate": 2.9645e-06, "loss": 3.0388, "step": 5929 }, { "epoch": 1.858934169278997, "grad_norm": 12.704240798950195, "learning_rate": 2.965e-06, "loss": 2.2973, "step": 5930 }, { "epoch": 1.8592476489028213, "grad_norm": 23.687358856201172, "learning_rate": 2.9655000000000002e-06, "loss": 2.4357, "step": 5931 }, { "epoch": 1.8595611285266458, "grad_norm": 12.968914031982422, "learning_rate": 2.966e-06, "loss": 2.3901, "step": 5932 }, { "epoch": 1.8598746081504702, "grad_norm": 35.27592849731445, "learning_rate": 2.9665000000000003e-06, "loss": 2.6777, "step": 5933 }, { "epoch": 1.8601880877742947, "grad_norm": 12.709157943725586, "learning_rate": 2.9670000000000006e-06, "loss": 2.2181, "step": 5934 }, { "epoch": 1.8605015673981191, "grad_norm": 18.975936889648438, "learning_rate": 2.9675000000000004e-06, "loss": 2.5685, "step": 5935 }, { "epoch": 1.8608150470219436, "grad_norm": 17.399898529052734, "learning_rate": 2.9680000000000002e-06, "loss": 2.0833, "step": 5936 }, { "epoch": 1.861128526645768, "grad_norm": 13.010747909545898, "learning_rate": 2.9685000000000005e-06, "loss": 2.0245, "step": 5937 }, { "epoch": 1.8614420062695924, "grad_norm": 11.078125953674316, "learning_rate": 2.9690000000000003e-06, "loss": 2.1143, "step": 5938 }, { "epoch": 1.8617554858934169, "grad_norm": 27.35584831237793, "learning_rate": 2.9695e-06, "loss": 2.5801, "step": 5939 }, { "epoch": 1.8620689655172413, "grad_norm": 15.091031074523926, "learning_rate": 2.97e-06, "loss": 2.3546, "step": 5940 }, { "epoch": 1.862382445141066, "grad_norm": 9.473328590393066, "learning_rate": 2.9705e-06, "loss": 2.1142, "step": 5941 }, { "epoch": 1.8626959247648904, "grad_norm": 19.720460891723633, "learning_rate": 2.971e-06, "loss": 3.0768, "step": 5942 }, { "epoch": 1.8630094043887149, "grad_norm": 16.488967895507812, "learning_rate": 2.9715000000000007e-06, "loss": 2.4286, "step": 5943 }, { "epoch": 1.8633228840125393, "grad_norm": 22.054641723632812, "learning_rate": 2.9720000000000005e-06, "loss": 2.2125, "step": 5944 }, { "epoch": 1.8636363636363638, "grad_norm": 11.244823455810547, "learning_rate": 2.9725000000000003e-06, "loss": 2.3011, "step": 5945 }, { "epoch": 1.8639498432601882, "grad_norm": 13.354893684387207, "learning_rate": 2.973e-06, "loss": 2.172, "step": 5946 }, { "epoch": 1.8642633228840126, "grad_norm": 10.610267639160156, "learning_rate": 2.9735000000000004e-06, "loss": 2.3918, "step": 5947 }, { "epoch": 1.864576802507837, "grad_norm": 18.93093490600586, "learning_rate": 2.9740000000000002e-06, "loss": 2.3002, "step": 5948 }, { "epoch": 1.8648902821316615, "grad_norm": 14.058737754821777, "learning_rate": 2.9745e-06, "loss": 2.2429, "step": 5949 }, { "epoch": 1.865203761755486, "grad_norm": 11.233428955078125, "learning_rate": 2.9750000000000003e-06, "loss": 2.0587, "step": 5950 }, { "epoch": 1.8655172413793104, "grad_norm": 15.62847900390625, "learning_rate": 2.9755e-06, "loss": 2.2816, "step": 5951 }, { "epoch": 1.8658307210031349, "grad_norm": 11.445661544799805, "learning_rate": 2.976e-06, "loss": 2.0684, "step": 5952 }, { "epoch": 1.8661442006269593, "grad_norm": 43.25938415527344, "learning_rate": 2.9765000000000006e-06, "loss": 2.6958, "step": 5953 }, { "epoch": 1.8664576802507837, "grad_norm": 16.91835594177246, "learning_rate": 2.9770000000000005e-06, "loss": 2.2447, "step": 5954 }, { "epoch": 1.8667711598746082, "grad_norm": 81.19352722167969, "learning_rate": 2.9775000000000003e-06, "loss": 2.3908, "step": 5955 }, { "epoch": 1.8670846394984326, "grad_norm": 14.449911117553711, "learning_rate": 2.9780000000000005e-06, "loss": 2.2353, "step": 5956 }, { "epoch": 1.867398119122257, "grad_norm": 15.223895072937012, "learning_rate": 2.9785000000000004e-06, "loss": 2.3473, "step": 5957 }, { "epoch": 1.8677115987460815, "grad_norm": 10.657572746276855, "learning_rate": 2.979e-06, "loss": 2.4062, "step": 5958 }, { "epoch": 1.868025078369906, "grad_norm": 20.128049850463867, "learning_rate": 2.9795e-06, "loss": 2.1415, "step": 5959 }, { "epoch": 1.8683385579937304, "grad_norm": 19.867464065551758, "learning_rate": 2.9800000000000003e-06, "loss": 2.3222, "step": 5960 }, { "epoch": 1.8686520376175548, "grad_norm": 37.976341247558594, "learning_rate": 2.9805e-06, "loss": 3.0599, "step": 5961 }, { "epoch": 1.8689655172413793, "grad_norm": 13.279169082641602, "learning_rate": 2.981e-06, "loss": 2.72, "step": 5962 }, { "epoch": 1.8692789968652037, "grad_norm": 29.617717742919922, "learning_rate": 2.9815000000000006e-06, "loss": 3.9811, "step": 5963 }, { "epoch": 1.8695924764890282, "grad_norm": 13.578438758850098, "learning_rate": 2.9820000000000004e-06, "loss": 2.1697, "step": 5964 }, { "epoch": 1.8699059561128526, "grad_norm": 22.47218132019043, "learning_rate": 2.9825000000000002e-06, "loss": 2.1429, "step": 5965 }, { "epoch": 1.870219435736677, "grad_norm": 68.62042999267578, "learning_rate": 2.9830000000000005e-06, "loss": 2.5068, "step": 5966 }, { "epoch": 1.8705329153605015, "grad_norm": 14.457962036132812, "learning_rate": 2.9835000000000003e-06, "loss": 2.4458, "step": 5967 }, { "epoch": 1.870846394984326, "grad_norm": 46.994361877441406, "learning_rate": 2.984e-06, "loss": 2.5197, "step": 5968 }, { "epoch": 1.8711598746081504, "grad_norm": 8.362174034118652, "learning_rate": 2.9845e-06, "loss": 2.101, "step": 5969 }, { "epoch": 1.8714733542319748, "grad_norm": 31.05512046813965, "learning_rate": 2.9850000000000002e-06, "loss": 2.975, "step": 5970 }, { "epoch": 1.8717868338557992, "grad_norm": 14.238176345825195, "learning_rate": 2.9855e-06, "loss": 1.9055, "step": 5971 }, { "epoch": 1.8721003134796237, "grad_norm": 10.674015998840332, "learning_rate": 2.986e-06, "loss": 2.0822, "step": 5972 }, { "epoch": 1.8724137931034481, "grad_norm": 21.955867767333984, "learning_rate": 2.9865000000000005e-06, "loss": 2.4844, "step": 5973 }, { "epoch": 1.8727272727272726, "grad_norm": 51.78453063964844, "learning_rate": 2.9870000000000004e-06, "loss": 2.5129, "step": 5974 }, { "epoch": 1.873040752351097, "grad_norm": 11.626786231994629, "learning_rate": 2.9875e-06, "loss": 2.1576, "step": 5975 }, { "epoch": 1.8733542319749217, "grad_norm": 11.270774841308594, "learning_rate": 2.9880000000000004e-06, "loss": 2.5178, "step": 5976 }, { "epoch": 1.8736677115987461, "grad_norm": 12.440287590026855, "learning_rate": 2.9885000000000003e-06, "loss": 2.2678, "step": 5977 }, { "epoch": 1.8739811912225706, "grad_norm": 22.860023498535156, "learning_rate": 2.989e-06, "loss": 2.2348, "step": 5978 }, { "epoch": 1.874294670846395, "grad_norm": 45.02328109741211, "learning_rate": 2.9895000000000003e-06, "loss": 2.6181, "step": 5979 }, { "epoch": 1.8746081504702194, "grad_norm": 10.591276168823242, "learning_rate": 2.99e-06, "loss": 2.3716, "step": 5980 }, { "epoch": 1.874921630094044, "grad_norm": 27.081436157226562, "learning_rate": 2.9905e-06, "loss": 2.1792, "step": 5981 }, { "epoch": 1.8752351097178683, "grad_norm": 17.257184982299805, "learning_rate": 2.991e-06, "loss": 2.4155, "step": 5982 }, { "epoch": 1.8755485893416928, "grad_norm": 8.944378852844238, "learning_rate": 2.9915000000000005e-06, "loss": 2.0324, "step": 5983 }, { "epoch": 1.8758620689655172, "grad_norm": 13.029302597045898, "learning_rate": 2.9920000000000003e-06, "loss": 2.0769, "step": 5984 }, { "epoch": 1.8761755485893417, "grad_norm": 93.2248306274414, "learning_rate": 2.9925000000000006e-06, "loss": 2.7404, "step": 5985 }, { "epoch": 1.876489028213166, "grad_norm": 11.743170738220215, "learning_rate": 2.9930000000000004e-06, "loss": 2.0512, "step": 5986 }, { "epoch": 1.8768025078369905, "grad_norm": 13.233771324157715, "learning_rate": 2.9935000000000002e-06, "loss": 1.9644, "step": 5987 }, { "epoch": 1.8771159874608152, "grad_norm": 12.37627124786377, "learning_rate": 2.994e-06, "loss": 2.49, "step": 5988 }, { "epoch": 1.8774294670846396, "grad_norm": 19.133132934570312, "learning_rate": 2.9945000000000003e-06, "loss": 2.2471, "step": 5989 }, { "epoch": 1.877742946708464, "grad_norm": 14.256895065307617, "learning_rate": 2.995e-06, "loss": 2.1018, "step": 5990 }, { "epoch": 1.8780564263322885, "grad_norm": 17.488203048706055, "learning_rate": 2.9955e-06, "loss": 2.121, "step": 5991 }, { "epoch": 1.878369905956113, "grad_norm": 9.792834281921387, "learning_rate": 2.996e-06, "loss": 2.1461, "step": 5992 }, { "epoch": 1.8786833855799374, "grad_norm": 15.044869422912598, "learning_rate": 2.9965000000000004e-06, "loss": 2.1132, "step": 5993 }, { "epoch": 1.8789968652037619, "grad_norm": 12.447378158569336, "learning_rate": 2.9970000000000003e-06, "loss": 2.1201, "step": 5994 }, { "epoch": 1.8793103448275863, "grad_norm": 11.574535369873047, "learning_rate": 2.9975000000000005e-06, "loss": 2.2942, "step": 5995 }, { "epoch": 1.8796238244514107, "grad_norm": 18.180683135986328, "learning_rate": 2.9980000000000003e-06, "loss": 2.2917, "step": 5996 }, { "epoch": 1.8799373040752352, "grad_norm": 15.773523330688477, "learning_rate": 2.9985e-06, "loss": 2.1649, "step": 5997 }, { "epoch": 1.8802507836990596, "grad_norm": 12.05911922454834, "learning_rate": 2.9990000000000004e-06, "loss": 2.2594, "step": 5998 }, { "epoch": 1.880564263322884, "grad_norm": 13.821588516235352, "learning_rate": 2.9995000000000003e-06, "loss": 2.283, "step": 5999 }, { "epoch": 1.8808777429467085, "grad_norm": 9.15234661102295, "learning_rate": 3e-06, "loss": 2.194, "step": 6000 }, { "epoch": 1.881191222570533, "grad_norm": 18.25379753112793, "learning_rate": 3.0005e-06, "loss": 2.3789, "step": 6001 }, { "epoch": 1.8815047021943574, "grad_norm": 14.774020195007324, "learning_rate": 3.001e-06, "loss": 2.2303, "step": 6002 }, { "epoch": 1.8818181818181818, "grad_norm": 7.372211456298828, "learning_rate": 3.0015e-06, "loss": 2.0202, "step": 6003 }, { "epoch": 1.8821316614420063, "grad_norm": 29.366544723510742, "learning_rate": 3.0020000000000006e-06, "loss": 2.7209, "step": 6004 }, { "epoch": 1.8824451410658307, "grad_norm": 20.27481460571289, "learning_rate": 3.0025000000000005e-06, "loss": 2.2696, "step": 6005 }, { "epoch": 1.8827586206896552, "grad_norm": 17.354726791381836, "learning_rate": 3.0030000000000003e-06, "loss": 2.2796, "step": 6006 }, { "epoch": 1.8830721003134796, "grad_norm": 11.225373268127441, "learning_rate": 3.0035e-06, "loss": 2.4752, "step": 6007 }, { "epoch": 1.883385579937304, "grad_norm": 13.302656173706055, "learning_rate": 3.0040000000000004e-06, "loss": 2.2338, "step": 6008 }, { "epoch": 1.8836990595611285, "grad_norm": 36.54410934448242, "learning_rate": 3.0045e-06, "loss": 2.5358, "step": 6009 }, { "epoch": 1.884012539184953, "grad_norm": 12.383371353149414, "learning_rate": 3.005e-06, "loss": 2.2359, "step": 6010 }, { "epoch": 1.8843260188087774, "grad_norm": 10.54360294342041, "learning_rate": 3.0055000000000003e-06, "loss": 2.1926, "step": 6011 }, { "epoch": 1.8846394984326018, "grad_norm": 13.832954406738281, "learning_rate": 3.006e-06, "loss": 2.343, "step": 6012 }, { "epoch": 1.8849529780564263, "grad_norm": 18.631515502929688, "learning_rate": 3.0065e-06, "loss": 2.1568, "step": 6013 }, { "epoch": 1.8852664576802507, "grad_norm": 14.329492568969727, "learning_rate": 3.0070000000000006e-06, "loss": 2.2601, "step": 6014 }, { "epoch": 1.8855799373040751, "grad_norm": 64.08553314208984, "learning_rate": 3.0075000000000004e-06, "loss": 2.8621, "step": 6015 }, { "epoch": 1.8858934169278996, "grad_norm": 15.698051452636719, "learning_rate": 3.0080000000000003e-06, "loss": 2.3793, "step": 6016 }, { "epoch": 1.886206896551724, "grad_norm": 64.55384826660156, "learning_rate": 3.0085000000000005e-06, "loss": 2.6757, "step": 6017 }, { "epoch": 1.8865203761755485, "grad_norm": 96.30728149414062, "learning_rate": 3.0090000000000003e-06, "loss": 2.4954, "step": 6018 }, { "epoch": 1.886833855799373, "grad_norm": 46.265907287597656, "learning_rate": 3.0095e-06, "loss": 2.3806, "step": 6019 }, { "epoch": 1.8871473354231973, "grad_norm": 19.963144302368164, "learning_rate": 3.01e-06, "loss": 1.9748, "step": 6020 }, { "epoch": 1.8874608150470218, "grad_norm": 36.911582946777344, "learning_rate": 3.0105000000000002e-06, "loss": 2.2604, "step": 6021 }, { "epoch": 1.8877742946708462, "grad_norm": 14.679679870605469, "learning_rate": 3.011e-06, "loss": 2.3641, "step": 6022 }, { "epoch": 1.888087774294671, "grad_norm": 17.748821258544922, "learning_rate": 3.0115e-06, "loss": 2.8093, "step": 6023 }, { "epoch": 1.8884012539184953, "grad_norm": 59.87860870361328, "learning_rate": 3.0120000000000006e-06, "loss": 3.2332, "step": 6024 }, { "epoch": 1.8887147335423198, "grad_norm": 14.679272651672363, "learning_rate": 3.0125000000000004e-06, "loss": 2.2552, "step": 6025 }, { "epoch": 1.8890282131661442, "grad_norm": 12.415968894958496, "learning_rate": 3.013e-06, "loss": 2.2754, "step": 6026 }, { "epoch": 1.8893416927899687, "grad_norm": 25.21853256225586, "learning_rate": 3.0135000000000005e-06, "loss": 2.2498, "step": 6027 }, { "epoch": 1.889655172413793, "grad_norm": 66.48529815673828, "learning_rate": 3.0140000000000003e-06, "loss": 2.3706, "step": 6028 }, { "epoch": 1.8899686520376175, "grad_norm": 39.09945297241211, "learning_rate": 3.0145e-06, "loss": 2.4398, "step": 6029 }, { "epoch": 1.890282131661442, "grad_norm": 18.697851181030273, "learning_rate": 3.0150000000000004e-06, "loss": 2.4102, "step": 6030 }, { "epoch": 1.8905956112852664, "grad_norm": 19.043466567993164, "learning_rate": 3.0155e-06, "loss": 2.2626, "step": 6031 }, { "epoch": 1.8909090909090909, "grad_norm": 31.95065689086914, "learning_rate": 3.016e-06, "loss": 2.3616, "step": 6032 }, { "epoch": 1.8912225705329153, "grad_norm": 15.939237594604492, "learning_rate": 3.0165e-06, "loss": 2.207, "step": 6033 }, { "epoch": 1.8915360501567398, "grad_norm": 40.93842315673828, "learning_rate": 3.0170000000000005e-06, "loss": 2.2072, "step": 6034 }, { "epoch": 1.8918495297805644, "grad_norm": 20.635238647460938, "learning_rate": 3.0175000000000003e-06, "loss": 2.5603, "step": 6035 }, { "epoch": 1.8921630094043889, "grad_norm": 14.814455032348633, "learning_rate": 3.0180000000000006e-06, "loss": 2.4815, "step": 6036 }, { "epoch": 1.8924764890282133, "grad_norm": 11.78087043762207, "learning_rate": 3.0185000000000004e-06, "loss": 2.332, "step": 6037 }, { "epoch": 1.8927899686520377, "grad_norm": 9.055764198303223, "learning_rate": 3.0190000000000002e-06, "loss": 2.1643, "step": 6038 }, { "epoch": 1.8931034482758622, "grad_norm": 62.65747833251953, "learning_rate": 3.0195e-06, "loss": 2.3768, "step": 6039 }, { "epoch": 1.8934169278996866, "grad_norm": 10.645723342895508, "learning_rate": 3.0200000000000003e-06, "loss": 2.4308, "step": 6040 }, { "epoch": 1.893730407523511, "grad_norm": 7.89349889755249, "learning_rate": 3.0205e-06, "loss": 2.5163, "step": 6041 }, { "epoch": 1.8940438871473355, "grad_norm": 56.9163703918457, "learning_rate": 3.021e-06, "loss": 3.444, "step": 6042 }, { "epoch": 1.89435736677116, "grad_norm": 9.060624122619629, "learning_rate": 3.0214999999999998e-06, "loss": 2.0517, "step": 6043 }, { "epoch": 1.8946708463949844, "grad_norm": 18.83366584777832, "learning_rate": 3.0220000000000005e-06, "loss": 2.2112, "step": 6044 }, { "epoch": 1.8949843260188088, "grad_norm": 15.034554481506348, "learning_rate": 3.0225000000000003e-06, "loss": 2.1677, "step": 6045 }, { "epoch": 1.8952978056426333, "grad_norm": 19.451780319213867, "learning_rate": 3.0230000000000005e-06, "loss": 2.4123, "step": 6046 }, { "epoch": 1.8956112852664577, "grad_norm": 12.623496055603027, "learning_rate": 3.0235000000000004e-06, "loss": 2.629, "step": 6047 }, { "epoch": 1.8959247648902822, "grad_norm": 17.31379508972168, "learning_rate": 3.024e-06, "loss": 2.3774, "step": 6048 }, { "epoch": 1.8962382445141066, "grad_norm": 77.49359893798828, "learning_rate": 3.0245e-06, "loss": 3.0316, "step": 6049 }, { "epoch": 1.896551724137931, "grad_norm": 13.705031394958496, "learning_rate": 3.0250000000000003e-06, "loss": 2.1171, "step": 6050 }, { "epoch": 1.8968652037617555, "grad_norm": 17.037595748901367, "learning_rate": 3.0255e-06, "loss": 2.2576, "step": 6051 }, { "epoch": 1.89717868338558, "grad_norm": 7.457324981689453, "learning_rate": 3.026e-06, "loss": 2.1425, "step": 6052 }, { "epoch": 1.8974921630094044, "grad_norm": 60.89271545410156, "learning_rate": 3.0265e-06, "loss": 2.7768, "step": 6053 }, { "epoch": 1.8978056426332288, "grad_norm": 17.38759422302246, "learning_rate": 3.0270000000000004e-06, "loss": 1.9765, "step": 6054 }, { "epoch": 1.8981191222570533, "grad_norm": 13.200677871704102, "learning_rate": 3.0275000000000002e-06, "loss": 2.2556, "step": 6055 }, { "epoch": 1.8984326018808777, "grad_norm": 10.527555465698242, "learning_rate": 3.0280000000000005e-06, "loss": 2.0826, "step": 6056 }, { "epoch": 1.8987460815047021, "grad_norm": 81.97634887695312, "learning_rate": 3.0285000000000003e-06, "loss": 2.5342, "step": 6057 }, { "epoch": 1.8990595611285266, "grad_norm": 24.821269989013672, "learning_rate": 3.029e-06, "loss": 2.2356, "step": 6058 }, { "epoch": 1.899373040752351, "grad_norm": 20.83104133605957, "learning_rate": 3.0295000000000004e-06, "loss": 2.2116, "step": 6059 }, { "epoch": 1.8996865203761755, "grad_norm": 25.46314811706543, "learning_rate": 3.0300000000000002e-06, "loss": 2.5997, "step": 6060 }, { "epoch": 1.9, "grad_norm": 9.644169807434082, "learning_rate": 3.0305e-06, "loss": 2.0949, "step": 6061 }, { "epoch": 1.9003134796238244, "grad_norm": 35.5372200012207, "learning_rate": 3.031e-06, "loss": 2.5674, "step": 6062 }, { "epoch": 1.9006269592476488, "grad_norm": 16.273813247680664, "learning_rate": 3.0315e-06, "loss": 2.465, "step": 6063 }, { "epoch": 1.9009404388714732, "grad_norm": 31.08625030517578, "learning_rate": 3.0320000000000004e-06, "loss": 2.3942, "step": 6064 }, { "epoch": 1.9012539184952977, "grad_norm": 45.639808654785156, "learning_rate": 3.0325000000000006e-06, "loss": 2.5761, "step": 6065 }, { "epoch": 1.9015673981191221, "grad_norm": 34.586517333984375, "learning_rate": 3.0330000000000004e-06, "loss": 2.297, "step": 6066 }, { "epoch": 1.9018808777429466, "grad_norm": 14.389083862304688, "learning_rate": 3.0335000000000003e-06, "loss": 2.3685, "step": 6067 }, { "epoch": 1.902194357366771, "grad_norm": 54.00825881958008, "learning_rate": 3.034e-06, "loss": 2.7093, "step": 6068 }, { "epoch": 1.9025078369905954, "grad_norm": 8.659297943115234, "learning_rate": 3.0345000000000003e-06, "loss": 2.2077, "step": 6069 }, { "epoch": 1.90282131661442, "grad_norm": 40.629173278808594, "learning_rate": 3.035e-06, "loss": 2.1629, "step": 6070 }, { "epoch": 1.9031347962382446, "grad_norm": 13.120607376098633, "learning_rate": 3.0355e-06, "loss": 2.2908, "step": 6071 }, { "epoch": 1.903448275862069, "grad_norm": 15.2529878616333, "learning_rate": 3.0360000000000002e-06, "loss": 2.109, "step": 6072 }, { "epoch": 1.9037617554858934, "grad_norm": 19.432376861572266, "learning_rate": 3.0365e-06, "loss": 2.3507, "step": 6073 }, { "epoch": 1.9040752351097179, "grad_norm": 11.499388694763184, "learning_rate": 3.0370000000000003e-06, "loss": 2.3288, "step": 6074 }, { "epoch": 1.9043887147335423, "grad_norm": 13.910191535949707, "learning_rate": 3.0375000000000006e-06, "loss": 2.5768, "step": 6075 }, { "epoch": 1.9047021943573668, "grad_norm": 33.155765533447266, "learning_rate": 3.0380000000000004e-06, "loss": 2.279, "step": 6076 }, { "epoch": 1.9050156739811912, "grad_norm": 17.429643630981445, "learning_rate": 3.0385000000000002e-06, "loss": 2.4015, "step": 6077 }, { "epoch": 1.9053291536050156, "grad_norm": 9.968888282775879, "learning_rate": 3.0390000000000005e-06, "loss": 2.2738, "step": 6078 }, { "epoch": 1.90564263322884, "grad_norm": 19.311132431030273, "learning_rate": 3.0395000000000003e-06, "loss": 2.3267, "step": 6079 }, { "epoch": 1.9059561128526645, "grad_norm": 15.164823532104492, "learning_rate": 3.04e-06, "loss": 2.9306, "step": 6080 }, { "epoch": 1.9062695924764892, "grad_norm": 20.48821449279785, "learning_rate": 3.0405e-06, "loss": 2.3529, "step": 6081 }, { "epoch": 1.9065830721003136, "grad_norm": 42.66050338745117, "learning_rate": 3.041e-06, "loss": 2.5175, "step": 6082 }, { "epoch": 1.906896551724138, "grad_norm": 15.134031295776367, "learning_rate": 3.0415e-06, "loss": 2.2948, "step": 6083 }, { "epoch": 1.9072100313479625, "grad_norm": 12.295634269714355, "learning_rate": 3.0420000000000007e-06, "loss": 2.1349, "step": 6084 }, { "epoch": 1.907523510971787, "grad_norm": 11.567246437072754, "learning_rate": 3.0425000000000005e-06, "loss": 2.1183, "step": 6085 }, { "epoch": 1.9078369905956114, "grad_norm": 16.417787551879883, "learning_rate": 3.0430000000000003e-06, "loss": 2.1569, "step": 6086 }, { "epoch": 1.9081504702194358, "grad_norm": 21.135101318359375, "learning_rate": 3.0435e-06, "loss": 2.3614, "step": 6087 }, { "epoch": 1.9084639498432603, "grad_norm": 23.375978469848633, "learning_rate": 3.0440000000000004e-06, "loss": 2.5555, "step": 6088 }, { "epoch": 1.9087774294670847, "grad_norm": 11.53591537475586, "learning_rate": 3.0445000000000002e-06, "loss": 2.3772, "step": 6089 }, { "epoch": 1.9090909090909092, "grad_norm": 15.471573829650879, "learning_rate": 3.045e-06, "loss": 2.3107, "step": 6090 }, { "epoch": 1.9094043887147336, "grad_norm": 70.57958984375, "learning_rate": 3.0455000000000003e-06, "loss": 2.1541, "step": 6091 }, { "epoch": 1.909717868338558, "grad_norm": 24.529329299926758, "learning_rate": 3.046e-06, "loss": 2.9641, "step": 6092 }, { "epoch": 1.9100313479623825, "grad_norm": 9.5241117477417, "learning_rate": 3.0465e-06, "loss": 2.1275, "step": 6093 }, { "epoch": 1.910344827586207, "grad_norm": 14.434778213500977, "learning_rate": 3.0470000000000006e-06, "loss": 2.2802, "step": 6094 }, { "epoch": 1.9106583072100314, "grad_norm": 11.293213844299316, "learning_rate": 3.0475000000000005e-06, "loss": 2.3955, "step": 6095 }, { "epoch": 1.9109717868338558, "grad_norm": 53.23150634765625, "learning_rate": 3.0480000000000003e-06, "loss": 2.4416, "step": 6096 }, { "epoch": 1.9112852664576803, "grad_norm": 11.527998924255371, "learning_rate": 3.0485000000000005e-06, "loss": 2.2852, "step": 6097 }, { "epoch": 1.9115987460815047, "grad_norm": 15.593719482421875, "learning_rate": 3.0490000000000004e-06, "loss": 2.3357, "step": 6098 }, { "epoch": 1.9119122257053291, "grad_norm": 15.388204574584961, "learning_rate": 3.0495e-06, "loss": 2.2743, "step": 6099 }, { "epoch": 1.9122257053291536, "grad_norm": 9.700482368469238, "learning_rate": 3.05e-06, "loss": 2.3766, "step": 6100 }, { "epoch": 1.912539184952978, "grad_norm": 22.43489646911621, "learning_rate": 3.0505000000000003e-06, "loss": 2.169, "step": 6101 }, { "epoch": 1.9128526645768025, "grad_norm": 32.98069763183594, "learning_rate": 3.051e-06, "loss": 2.221, "step": 6102 }, { "epoch": 1.913166144200627, "grad_norm": 13.035593032836914, "learning_rate": 3.0515e-06, "loss": 2.0808, "step": 6103 }, { "epoch": 1.9134796238244514, "grad_norm": 14.784707069396973, "learning_rate": 3.0520000000000006e-06, "loss": 2.6694, "step": 6104 }, { "epoch": 1.9137931034482758, "grad_norm": 19.96974754333496, "learning_rate": 3.0525000000000004e-06, "loss": 2.2853, "step": 6105 }, { "epoch": 1.9141065830721002, "grad_norm": 21.09632682800293, "learning_rate": 3.0530000000000002e-06, "loss": 2.2353, "step": 6106 }, { "epoch": 1.9144200626959247, "grad_norm": 12.733104705810547, "learning_rate": 3.0535000000000005e-06, "loss": 2.1974, "step": 6107 }, { "epoch": 1.9147335423197491, "grad_norm": 21.158296585083008, "learning_rate": 3.0540000000000003e-06, "loss": 2.349, "step": 6108 }, { "epoch": 1.9150470219435736, "grad_norm": 12.683104515075684, "learning_rate": 3.0545e-06, "loss": 2.1925, "step": 6109 }, { "epoch": 1.915360501567398, "grad_norm": 21.3099422454834, "learning_rate": 3.0550000000000004e-06, "loss": 2.3814, "step": 6110 }, { "epoch": 1.9156739811912225, "grad_norm": 31.082923889160156, "learning_rate": 3.0555000000000002e-06, "loss": 2.4238, "step": 6111 }, { "epoch": 1.915987460815047, "grad_norm": 10.956926345825195, "learning_rate": 3.056e-06, "loss": 2.2932, "step": 6112 }, { "epoch": 1.9163009404388713, "grad_norm": 18.108543395996094, "learning_rate": 3.0565e-06, "loss": 2.3098, "step": 6113 }, { "epoch": 1.9166144200626958, "grad_norm": 8.127020835876465, "learning_rate": 3.0570000000000005e-06, "loss": 2.2654, "step": 6114 }, { "epoch": 1.9169278996865202, "grad_norm": 9.577781677246094, "learning_rate": 3.0575000000000004e-06, "loss": 2.1266, "step": 6115 }, { "epoch": 1.9172413793103447, "grad_norm": 34.873382568359375, "learning_rate": 3.0580000000000006e-06, "loss": 1.9902, "step": 6116 }, { "epoch": 1.9175548589341693, "grad_norm": 18.865312576293945, "learning_rate": 3.0585000000000004e-06, "loss": 2.2634, "step": 6117 }, { "epoch": 1.9178683385579938, "grad_norm": 59.097164154052734, "learning_rate": 3.0590000000000003e-06, "loss": 2.5326, "step": 6118 }, { "epoch": 1.9181818181818182, "grad_norm": 7.89798641204834, "learning_rate": 3.0595e-06, "loss": 2.201, "step": 6119 }, { "epoch": 1.9184952978056427, "grad_norm": 9.707578659057617, "learning_rate": 3.0600000000000003e-06, "loss": 2.2415, "step": 6120 }, { "epoch": 1.918808777429467, "grad_norm": 9.2268705368042, "learning_rate": 3.0605e-06, "loss": 2.0483, "step": 6121 }, { "epoch": 1.9191222570532915, "grad_norm": 30.342716217041016, "learning_rate": 3.061e-06, "loss": 2.4259, "step": 6122 }, { "epoch": 1.919435736677116, "grad_norm": 23.01906967163086, "learning_rate": 3.0615e-06, "loss": 2.2752, "step": 6123 }, { "epoch": 1.9197492163009404, "grad_norm": 9.22539234161377, "learning_rate": 3.0620000000000005e-06, "loss": 2.0471, "step": 6124 }, { "epoch": 1.9200626959247649, "grad_norm": 21.07687759399414, "learning_rate": 3.0625000000000003e-06, "loss": 2.4733, "step": 6125 }, { "epoch": 1.9203761755485893, "grad_norm": 19.99814224243164, "learning_rate": 3.0630000000000006e-06, "loss": 2.7217, "step": 6126 }, { "epoch": 1.9206896551724137, "grad_norm": 49.313392639160156, "learning_rate": 3.0635000000000004e-06, "loss": 3.131, "step": 6127 }, { "epoch": 1.9210031347962384, "grad_norm": 8.287384986877441, "learning_rate": 3.0640000000000002e-06, "loss": 2.1604, "step": 6128 }, { "epoch": 1.9213166144200629, "grad_norm": 16.254018783569336, "learning_rate": 3.0645e-06, "loss": 2.1469, "step": 6129 }, { "epoch": 1.9216300940438873, "grad_norm": 13.05605411529541, "learning_rate": 3.0650000000000003e-06, "loss": 2.2322, "step": 6130 }, { "epoch": 1.9219435736677117, "grad_norm": 13.967427253723145, "learning_rate": 3.0655e-06, "loss": 2.1692, "step": 6131 }, { "epoch": 1.9222570532915362, "grad_norm": 18.97654151916504, "learning_rate": 3.066e-06, "loss": 2.1141, "step": 6132 }, { "epoch": 1.9225705329153606, "grad_norm": 17.071155548095703, "learning_rate": 3.0665e-06, "loss": 2.0225, "step": 6133 }, { "epoch": 1.922884012539185, "grad_norm": 44.785980224609375, "learning_rate": 3.067e-06, "loss": 2.3243, "step": 6134 }, { "epoch": 1.9231974921630095, "grad_norm": 12.6852445602417, "learning_rate": 3.0675000000000003e-06, "loss": 2.3113, "step": 6135 }, { "epoch": 1.923510971786834, "grad_norm": 11.303502082824707, "learning_rate": 3.0680000000000005e-06, "loss": 1.9648, "step": 6136 }, { "epoch": 1.9238244514106584, "grad_norm": 9.669048309326172, "learning_rate": 3.0685000000000004e-06, "loss": 2.2378, "step": 6137 }, { "epoch": 1.9241379310344828, "grad_norm": 67.32134246826172, "learning_rate": 3.069e-06, "loss": 2.2429, "step": 6138 }, { "epoch": 1.9244514106583073, "grad_norm": 14.467636108398438, "learning_rate": 3.0695000000000004e-06, "loss": 2.3443, "step": 6139 }, { "epoch": 1.9247648902821317, "grad_norm": 26.40109634399414, "learning_rate": 3.0700000000000003e-06, "loss": 2.2199, "step": 6140 }, { "epoch": 1.9250783699059562, "grad_norm": 32.24627685546875, "learning_rate": 3.0705e-06, "loss": 2.0711, "step": 6141 }, { "epoch": 1.9253918495297806, "grad_norm": 43.16203689575195, "learning_rate": 3.071e-06, "loss": 2.0225, "step": 6142 }, { "epoch": 1.925705329153605, "grad_norm": 15.031079292297363, "learning_rate": 3.0715e-06, "loss": 2.7297, "step": 6143 }, { "epoch": 1.9260188087774295, "grad_norm": 11.947735786437988, "learning_rate": 3.072e-06, "loss": 2.2531, "step": 6144 }, { "epoch": 1.926332288401254, "grad_norm": 11.026966094970703, "learning_rate": 3.0725000000000007e-06, "loss": 2.1997, "step": 6145 }, { "epoch": 1.9266457680250784, "grad_norm": 12.493368148803711, "learning_rate": 3.0730000000000005e-06, "loss": 2.2338, "step": 6146 }, { "epoch": 1.9269592476489028, "grad_norm": 7.645673751831055, "learning_rate": 3.0735000000000003e-06, "loss": 2.2303, "step": 6147 }, { "epoch": 1.9272727272727272, "grad_norm": 42.979488372802734, "learning_rate": 3.074e-06, "loss": 2.0703, "step": 6148 }, { "epoch": 1.9275862068965517, "grad_norm": 9.205596923828125, "learning_rate": 3.0745000000000004e-06, "loss": 2.1779, "step": 6149 }, { "epoch": 1.9278996865203761, "grad_norm": 12.70852279663086, "learning_rate": 3.075e-06, "loss": 2.2296, "step": 6150 }, { "epoch": 1.9282131661442006, "grad_norm": 12.988320350646973, "learning_rate": 3.0755e-06, "loss": 2.1411, "step": 6151 }, { "epoch": 1.928526645768025, "grad_norm": 9.5177583694458, "learning_rate": 3.0760000000000003e-06, "loss": 2.3787, "step": 6152 }, { "epoch": 1.9288401253918495, "grad_norm": 16.0618839263916, "learning_rate": 3.0765e-06, "loss": 1.8497, "step": 6153 }, { "epoch": 1.929153605015674, "grad_norm": 51.27670669555664, "learning_rate": 3.077e-06, "loss": 2.3243, "step": 6154 }, { "epoch": 1.9294670846394983, "grad_norm": 19.865938186645508, "learning_rate": 3.0775000000000006e-06, "loss": 2.0024, "step": 6155 }, { "epoch": 1.9297805642633228, "grad_norm": 65.7314682006836, "learning_rate": 3.0780000000000004e-06, "loss": 2.3256, "step": 6156 }, { "epoch": 1.9300940438871472, "grad_norm": 55.74231719970703, "learning_rate": 3.0785000000000003e-06, "loss": 2.622, "step": 6157 }, { "epoch": 1.9304075235109717, "grad_norm": 15.176788330078125, "learning_rate": 3.0790000000000005e-06, "loss": 2.3372, "step": 6158 }, { "epoch": 1.930721003134796, "grad_norm": 12.638873100280762, "learning_rate": 3.0795000000000003e-06, "loss": 2.5779, "step": 6159 }, { "epoch": 1.9310344827586206, "grad_norm": 12.51217269897461, "learning_rate": 3.08e-06, "loss": 2.2006, "step": 6160 }, { "epoch": 1.931347962382445, "grad_norm": 16.80646514892578, "learning_rate": 3.0805e-06, "loss": 2.208, "step": 6161 }, { "epoch": 1.9316614420062694, "grad_norm": 19.71490478515625, "learning_rate": 3.0810000000000002e-06, "loss": 2.3028, "step": 6162 }, { "epoch": 1.9319749216300939, "grad_norm": 15.717697143554688, "learning_rate": 3.0815e-06, "loss": 2.3063, "step": 6163 }, { "epoch": 1.9322884012539185, "grad_norm": 31.971899032592773, "learning_rate": 3.082e-06, "loss": 1.9921, "step": 6164 }, { "epoch": 1.932601880877743, "grad_norm": 34.533424377441406, "learning_rate": 3.0825000000000006e-06, "loss": 2.0271, "step": 6165 }, { "epoch": 1.9329153605015674, "grad_norm": 81.34234619140625, "learning_rate": 3.0830000000000004e-06, "loss": 2.3326, "step": 6166 }, { "epoch": 1.9332288401253919, "grad_norm": 23.47624969482422, "learning_rate": 3.0835e-06, "loss": 2.111, "step": 6167 }, { "epoch": 1.9335423197492163, "grad_norm": 15.721684455871582, "learning_rate": 3.0840000000000005e-06, "loss": 2.1075, "step": 6168 }, { "epoch": 1.9338557993730408, "grad_norm": 41.16612243652344, "learning_rate": 3.0845000000000003e-06, "loss": 2.0034, "step": 6169 }, { "epoch": 1.9341692789968652, "grad_norm": 8.464729309082031, "learning_rate": 3.085e-06, "loss": 2.0479, "step": 6170 }, { "epoch": 1.9344827586206896, "grad_norm": 23.668128967285156, "learning_rate": 3.0855000000000004e-06, "loss": 2.283, "step": 6171 }, { "epoch": 1.934796238244514, "grad_norm": 18.2292537689209, "learning_rate": 3.086e-06, "loss": 2.5008, "step": 6172 }, { "epoch": 1.9351097178683385, "grad_norm": 15.80252456665039, "learning_rate": 3.0865e-06, "loss": 2.059, "step": 6173 }, { "epoch": 1.935423197492163, "grad_norm": 14.27904224395752, "learning_rate": 3.087e-06, "loss": 2.488, "step": 6174 }, { "epoch": 1.9357366771159876, "grad_norm": 21.141990661621094, "learning_rate": 3.0875000000000005e-06, "loss": 2.1918, "step": 6175 }, { "epoch": 1.936050156739812, "grad_norm": 13.165213584899902, "learning_rate": 3.0880000000000003e-06, "loss": 2.1132, "step": 6176 }, { "epoch": 1.9363636363636365, "grad_norm": 11.983834266662598, "learning_rate": 3.0885000000000006e-06, "loss": 2.1287, "step": 6177 }, { "epoch": 1.936677115987461, "grad_norm": 12.687564849853516, "learning_rate": 3.0890000000000004e-06, "loss": 2.433, "step": 6178 }, { "epoch": 1.9369905956112854, "grad_norm": 53.66118621826172, "learning_rate": 3.0895000000000002e-06, "loss": 2.3676, "step": 6179 }, { "epoch": 1.9373040752351098, "grad_norm": 28.613327026367188, "learning_rate": 3.09e-06, "loss": 2.297, "step": 6180 }, { "epoch": 1.9376175548589343, "grad_norm": 13.394174575805664, "learning_rate": 3.0905000000000003e-06, "loss": 2.0543, "step": 6181 }, { "epoch": 1.9379310344827587, "grad_norm": 15.933115005493164, "learning_rate": 3.091e-06, "loss": 2.2325, "step": 6182 }, { "epoch": 1.9382445141065832, "grad_norm": 26.931055068969727, "learning_rate": 3.0915e-06, "loss": 2.072, "step": 6183 }, { "epoch": 1.9385579937304076, "grad_norm": 21.357282638549805, "learning_rate": 3.092e-06, "loss": 2.1702, "step": 6184 }, { "epoch": 1.938871473354232, "grad_norm": 16.33864974975586, "learning_rate": 3.0925000000000005e-06, "loss": 2.5012, "step": 6185 }, { "epoch": 1.9391849529780565, "grad_norm": 17.6795711517334, "learning_rate": 3.0930000000000003e-06, "loss": 2.4467, "step": 6186 }, { "epoch": 1.939498432601881, "grad_norm": 83.75550079345703, "learning_rate": 3.0935000000000005e-06, "loss": 3.2387, "step": 6187 }, { "epoch": 1.9398119122257054, "grad_norm": 23.691688537597656, "learning_rate": 3.0940000000000004e-06, "loss": 2.1602, "step": 6188 }, { "epoch": 1.9401253918495298, "grad_norm": 15.03283977508545, "learning_rate": 3.0945e-06, "loss": 2.265, "step": 6189 }, { "epoch": 1.9404388714733543, "grad_norm": 19.019834518432617, "learning_rate": 3.0950000000000004e-06, "loss": 2.1407, "step": 6190 }, { "epoch": 1.9407523510971787, "grad_norm": 17.266340255737305, "learning_rate": 3.0955000000000003e-06, "loss": 2.2096, "step": 6191 }, { "epoch": 1.9410658307210031, "grad_norm": 19.639511108398438, "learning_rate": 3.096e-06, "loss": 2.2862, "step": 6192 }, { "epoch": 1.9413793103448276, "grad_norm": 32.09061813354492, "learning_rate": 3.0965e-06, "loss": 2.0854, "step": 6193 }, { "epoch": 1.941692789968652, "grad_norm": 12.486015319824219, "learning_rate": 3.097e-06, "loss": 2.2919, "step": 6194 }, { "epoch": 1.9420062695924765, "grad_norm": 15.492584228515625, "learning_rate": 3.0975000000000004e-06, "loss": 2.2623, "step": 6195 }, { "epoch": 1.942319749216301, "grad_norm": 9.039558410644531, "learning_rate": 3.0980000000000007e-06, "loss": 2.2022, "step": 6196 }, { "epoch": 1.9426332288401253, "grad_norm": 14.141021728515625, "learning_rate": 3.0985000000000005e-06, "loss": 2.7238, "step": 6197 }, { "epoch": 1.9429467084639498, "grad_norm": 7.6339945793151855, "learning_rate": 3.0990000000000003e-06, "loss": 2.6758, "step": 6198 }, { "epoch": 1.9432601880877742, "grad_norm": 10.205277442932129, "learning_rate": 3.0995e-06, "loss": 2.2798, "step": 6199 }, { "epoch": 1.9435736677115987, "grad_norm": 15.242185592651367, "learning_rate": 3.1000000000000004e-06, "loss": 2.6331, "step": 6200 }, { "epoch": 1.9438871473354231, "grad_norm": 25.186603546142578, "learning_rate": 3.1005000000000002e-06, "loss": 2.1404, "step": 6201 }, { "epoch": 1.9442006269592476, "grad_norm": 51.08753204345703, "learning_rate": 3.101e-06, "loss": 2.3513, "step": 6202 }, { "epoch": 1.944514106583072, "grad_norm": 9.865594863891602, "learning_rate": 3.1015e-06, "loss": 2.2377, "step": 6203 }, { "epoch": 1.9448275862068964, "grad_norm": 9.022784233093262, "learning_rate": 3.102e-06, "loss": 2.0802, "step": 6204 }, { "epoch": 1.9451410658307209, "grad_norm": 38.94915771484375, "learning_rate": 3.1025000000000004e-06, "loss": 2.7855, "step": 6205 }, { "epoch": 1.9454545454545453, "grad_norm": 16.95958137512207, "learning_rate": 3.1030000000000006e-06, "loss": 1.9462, "step": 6206 }, { "epoch": 1.9457680250783698, "grad_norm": 103.67754364013672, "learning_rate": 3.1035000000000004e-06, "loss": 2.4162, "step": 6207 }, { "epoch": 1.9460815047021942, "grad_norm": 20.15214729309082, "learning_rate": 3.1040000000000003e-06, "loss": 2.2354, "step": 6208 }, { "epoch": 1.9463949843260187, "grad_norm": 29.8538875579834, "learning_rate": 3.1045e-06, "loss": 2.1404, "step": 6209 }, { "epoch": 1.9467084639498433, "grad_norm": 18.50629997253418, "learning_rate": 3.1050000000000003e-06, "loss": 2.2808, "step": 6210 }, { "epoch": 1.9470219435736678, "grad_norm": 13.072752952575684, "learning_rate": 3.1055e-06, "loss": 2.1819, "step": 6211 }, { "epoch": 1.9473354231974922, "grad_norm": 12.113188743591309, "learning_rate": 3.106e-06, "loss": 2.2628, "step": 6212 }, { "epoch": 1.9476489028213166, "grad_norm": 19.383434295654297, "learning_rate": 3.1065000000000002e-06, "loss": 2.2511, "step": 6213 }, { "epoch": 1.947962382445141, "grad_norm": 13.1512451171875, "learning_rate": 3.107e-06, "loss": 1.8007, "step": 6214 }, { "epoch": 1.9482758620689655, "grad_norm": 29.93698501586914, "learning_rate": 3.1075000000000003e-06, "loss": 2.9913, "step": 6215 }, { "epoch": 1.94858934169279, "grad_norm": 51.03324890136719, "learning_rate": 3.1080000000000006e-06, "loss": 2.4474, "step": 6216 }, { "epoch": 1.9489028213166144, "grad_norm": 21.86216163635254, "learning_rate": 3.1085000000000004e-06, "loss": 2.3511, "step": 6217 }, { "epoch": 1.9492163009404389, "grad_norm": 44.02346420288086, "learning_rate": 3.1090000000000002e-06, "loss": 2.2321, "step": 6218 }, { "epoch": 1.9495297805642633, "grad_norm": 10.752887725830078, "learning_rate": 3.1095000000000005e-06, "loss": 2.1363, "step": 6219 }, { "epoch": 1.9498432601880877, "grad_norm": 9.247302055358887, "learning_rate": 3.1100000000000003e-06, "loss": 2.2588, "step": 6220 }, { "epoch": 1.9501567398119122, "grad_norm": 15.984376907348633, "learning_rate": 3.1105e-06, "loss": 2.384, "step": 6221 }, { "epoch": 1.9504702194357368, "grad_norm": 33.194580078125, "learning_rate": 3.111e-06, "loss": 2.8689, "step": 6222 }, { "epoch": 1.9507836990595613, "grad_norm": 10.841005325317383, "learning_rate": 3.1115e-06, "loss": 1.9113, "step": 6223 }, { "epoch": 1.9510971786833857, "grad_norm": 9.533475875854492, "learning_rate": 3.112e-06, "loss": 2.1351, "step": 6224 }, { "epoch": 1.9514106583072102, "grad_norm": 18.278017044067383, "learning_rate": 3.1125000000000007e-06, "loss": 2.4297, "step": 6225 }, { "epoch": 1.9517241379310346, "grad_norm": 14.732829093933105, "learning_rate": 3.1130000000000005e-06, "loss": 2.0193, "step": 6226 }, { "epoch": 1.952037617554859, "grad_norm": 14.808735847473145, "learning_rate": 3.1135000000000003e-06, "loss": 2.1229, "step": 6227 }, { "epoch": 1.9523510971786835, "grad_norm": 40.86724090576172, "learning_rate": 3.114e-06, "loss": 2.3927, "step": 6228 }, { "epoch": 1.952664576802508, "grad_norm": 16.639362335205078, "learning_rate": 3.1145000000000004e-06, "loss": 2.5477, "step": 6229 }, { "epoch": 1.9529780564263324, "grad_norm": 24.524995803833008, "learning_rate": 3.1150000000000002e-06, "loss": 2.2817, "step": 6230 }, { "epoch": 1.9532915360501568, "grad_norm": 32.46340560913086, "learning_rate": 3.1155e-06, "loss": 2.3405, "step": 6231 }, { "epoch": 1.9536050156739813, "grad_norm": 15.452544212341309, "learning_rate": 3.1160000000000003e-06, "loss": 2.4422, "step": 6232 }, { "epoch": 1.9539184952978057, "grad_norm": 15.222606658935547, "learning_rate": 3.1165e-06, "loss": 2.336, "step": 6233 }, { "epoch": 1.9542319749216301, "grad_norm": 25.828502655029297, "learning_rate": 3.117e-06, "loss": 2.2059, "step": 6234 }, { "epoch": 1.9545454545454546, "grad_norm": 15.925555229187012, "learning_rate": 3.1175000000000006e-06, "loss": 2.2686, "step": 6235 }, { "epoch": 1.954858934169279, "grad_norm": 13.170594215393066, "learning_rate": 3.1180000000000005e-06, "loss": 2.5014, "step": 6236 }, { "epoch": 1.9551724137931035, "grad_norm": 12.45872688293457, "learning_rate": 3.1185000000000003e-06, "loss": 2.0243, "step": 6237 }, { "epoch": 1.955485893416928, "grad_norm": 15.75046443939209, "learning_rate": 3.1190000000000005e-06, "loss": 2.0471, "step": 6238 }, { "epoch": 1.9557993730407524, "grad_norm": 92.76771545410156, "learning_rate": 3.1195000000000004e-06, "loss": 2.7066, "step": 6239 }, { "epoch": 1.9561128526645768, "grad_norm": 47.9924201965332, "learning_rate": 3.12e-06, "loss": 2.211, "step": 6240 }, { "epoch": 1.9564263322884012, "grad_norm": 13.54965591430664, "learning_rate": 3.1205e-06, "loss": 2.274, "step": 6241 }, { "epoch": 1.9567398119122257, "grad_norm": 11.754216194152832, "learning_rate": 3.1210000000000003e-06, "loss": 2.4049, "step": 6242 }, { "epoch": 1.9570532915360501, "grad_norm": 12.851913452148438, "learning_rate": 3.1215e-06, "loss": 2.3721, "step": 6243 }, { "epoch": 1.9573667711598746, "grad_norm": 11.860594749450684, "learning_rate": 3.122e-06, "loss": 2.1425, "step": 6244 }, { "epoch": 1.957680250783699, "grad_norm": 29.15382957458496, "learning_rate": 3.1225000000000006e-06, "loss": 2.2544, "step": 6245 }, { "epoch": 1.9579937304075234, "grad_norm": 32.79502487182617, "learning_rate": 3.1230000000000004e-06, "loss": 2.195, "step": 6246 }, { "epoch": 1.958307210031348, "grad_norm": 8.380058288574219, "learning_rate": 3.1235000000000002e-06, "loss": 2.2671, "step": 6247 }, { "epoch": 1.9586206896551723, "grad_norm": 41.89897918701172, "learning_rate": 3.1240000000000005e-06, "loss": 2.641, "step": 6248 }, { "epoch": 1.9589341692789968, "grad_norm": 10.770977020263672, "learning_rate": 3.1245000000000003e-06, "loss": 2.4666, "step": 6249 }, { "epoch": 1.9592476489028212, "grad_norm": 21.347427368164062, "learning_rate": 3.125e-06, "loss": 2.3943, "step": 6250 }, { "epoch": 1.9595611285266457, "grad_norm": 20.523212432861328, "learning_rate": 3.1255000000000004e-06, "loss": 2.3104, "step": 6251 }, { "epoch": 1.95987460815047, "grad_norm": 45.21733474731445, "learning_rate": 3.1260000000000002e-06, "loss": 2.0047, "step": 6252 }, { "epoch": 1.9601880877742945, "grad_norm": 19.431230545043945, "learning_rate": 3.1265e-06, "loss": 2.1989, "step": 6253 }, { "epoch": 1.960501567398119, "grad_norm": 43.393760681152344, "learning_rate": 3.127e-06, "loss": 2.1272, "step": 6254 }, { "epoch": 1.9608150470219434, "grad_norm": 19.50667381286621, "learning_rate": 3.1275e-06, "loss": 2.1979, "step": 6255 }, { "epoch": 1.9611285266457679, "grad_norm": 24.86729621887207, "learning_rate": 3.1280000000000004e-06, "loss": 2.296, "step": 6256 }, { "epoch": 1.9614420062695925, "grad_norm": 10.600141525268555, "learning_rate": 3.1285000000000006e-06, "loss": 2.0203, "step": 6257 }, { "epoch": 1.961755485893417, "grad_norm": 15.078958511352539, "learning_rate": 3.1290000000000005e-06, "loss": 2.7117, "step": 6258 }, { "epoch": 1.9620689655172414, "grad_norm": 31.340831756591797, "learning_rate": 3.1295000000000003e-06, "loss": 2.2201, "step": 6259 }, { "epoch": 1.9623824451410659, "grad_norm": 11.659078598022461, "learning_rate": 3.13e-06, "loss": 2.2339, "step": 6260 }, { "epoch": 1.9626959247648903, "grad_norm": 21.408418655395508, "learning_rate": 3.1305000000000004e-06, "loss": 2.2118, "step": 6261 }, { "epoch": 1.9630094043887147, "grad_norm": 32.824012756347656, "learning_rate": 3.131e-06, "loss": 2.5834, "step": 6262 }, { "epoch": 1.9633228840125392, "grad_norm": 22.480697631835938, "learning_rate": 3.1315e-06, "loss": 2.8845, "step": 6263 }, { "epoch": 1.9636363636363636, "grad_norm": 14.943920135498047, "learning_rate": 3.132e-06, "loss": 2.5693, "step": 6264 }, { "epoch": 1.963949843260188, "grad_norm": 21.227256774902344, "learning_rate": 3.1325e-06, "loss": 2.3999, "step": 6265 }, { "epoch": 1.9642633228840125, "grad_norm": 10.880897521972656, "learning_rate": 3.1330000000000003e-06, "loss": 2.2018, "step": 6266 }, { "epoch": 1.964576802507837, "grad_norm": 63.30581283569336, "learning_rate": 3.1335000000000006e-06, "loss": 1.9168, "step": 6267 }, { "epoch": 1.9648902821316614, "grad_norm": 10.682960510253906, "learning_rate": 3.1340000000000004e-06, "loss": 2.054, "step": 6268 }, { "epoch": 1.965203761755486, "grad_norm": 11.94248104095459, "learning_rate": 3.1345000000000002e-06, "loss": 2.3176, "step": 6269 }, { "epoch": 1.9655172413793105, "grad_norm": 16.106670379638672, "learning_rate": 3.135e-06, "loss": 2.0857, "step": 6270 }, { "epoch": 1.965830721003135, "grad_norm": 6.923719882965088, "learning_rate": 3.1355000000000003e-06, "loss": 2.1489, "step": 6271 }, { "epoch": 1.9661442006269594, "grad_norm": 12.050187110900879, "learning_rate": 3.136e-06, "loss": 2.2787, "step": 6272 }, { "epoch": 1.9664576802507838, "grad_norm": 33.00879669189453, "learning_rate": 3.1365e-06, "loss": 2.3985, "step": 6273 }, { "epoch": 1.9667711598746083, "grad_norm": 12.105003356933594, "learning_rate": 3.137e-06, "loss": 2.3161, "step": 6274 }, { "epoch": 1.9670846394984327, "grad_norm": 13.618671417236328, "learning_rate": 3.1375e-06, "loss": 2.2129, "step": 6275 }, { "epoch": 1.9673981191222571, "grad_norm": 12.591639518737793, "learning_rate": 3.1380000000000003e-06, "loss": 2.4553, "step": 6276 }, { "epoch": 1.9677115987460816, "grad_norm": 17.883012771606445, "learning_rate": 3.1385000000000005e-06, "loss": 2.1222, "step": 6277 }, { "epoch": 1.968025078369906, "grad_norm": 16.255971908569336, "learning_rate": 3.1390000000000004e-06, "loss": 2.2342, "step": 6278 }, { "epoch": 1.9683385579937305, "grad_norm": 16.84834861755371, "learning_rate": 3.1395e-06, "loss": 2.473, "step": 6279 }, { "epoch": 1.968652037617555, "grad_norm": 19.836959838867188, "learning_rate": 3.1400000000000004e-06, "loss": 2.2049, "step": 6280 }, { "epoch": 1.9689655172413794, "grad_norm": 46.876102447509766, "learning_rate": 3.1405000000000003e-06, "loss": 2.7783, "step": 6281 }, { "epoch": 1.9692789968652038, "grad_norm": 19.187061309814453, "learning_rate": 3.141e-06, "loss": 2.2508, "step": 6282 }, { "epoch": 1.9695924764890282, "grad_norm": 23.875375747680664, "learning_rate": 3.1415e-06, "loss": 2.7126, "step": 6283 }, { "epoch": 1.9699059561128527, "grad_norm": 15.11329460144043, "learning_rate": 3.142e-06, "loss": 2.1725, "step": 6284 }, { "epoch": 1.9702194357366771, "grad_norm": 64.44976806640625, "learning_rate": 3.1425e-06, "loss": 2.3958, "step": 6285 }, { "epoch": 1.9705329153605016, "grad_norm": 57.60660171508789, "learning_rate": 3.1430000000000007e-06, "loss": 2.4208, "step": 6286 }, { "epoch": 1.970846394984326, "grad_norm": 34.09165954589844, "learning_rate": 3.1435000000000005e-06, "loss": 2.2666, "step": 6287 }, { "epoch": 1.9711598746081505, "grad_norm": 32.35491180419922, "learning_rate": 3.1440000000000003e-06, "loss": 2.2841, "step": 6288 }, { "epoch": 1.971473354231975, "grad_norm": 11.348264694213867, "learning_rate": 3.1445e-06, "loss": 2.0545, "step": 6289 }, { "epoch": 1.9717868338557993, "grad_norm": 12.587067604064941, "learning_rate": 3.1450000000000004e-06, "loss": 2.0804, "step": 6290 }, { "epoch": 1.9721003134796238, "grad_norm": 19.97517967224121, "learning_rate": 3.1455e-06, "loss": 2.5052, "step": 6291 }, { "epoch": 1.9724137931034482, "grad_norm": 15.031810760498047, "learning_rate": 3.146e-06, "loss": 2.192, "step": 6292 }, { "epoch": 1.9727272727272727, "grad_norm": 10.090429306030273, "learning_rate": 3.1465000000000003e-06, "loss": 2.3494, "step": 6293 }, { "epoch": 1.973040752351097, "grad_norm": 16.087080001831055, "learning_rate": 3.147e-06, "loss": 2.1995, "step": 6294 }, { "epoch": 1.9733542319749215, "grad_norm": 50.80659866333008, "learning_rate": 3.1475e-06, "loss": 2.3741, "step": 6295 }, { "epoch": 1.973667711598746, "grad_norm": 8.694010734558105, "learning_rate": 3.1480000000000006e-06, "loss": 2.1153, "step": 6296 }, { "epoch": 1.9739811912225704, "grad_norm": 22.7263240814209, "learning_rate": 3.1485000000000004e-06, "loss": 2.2891, "step": 6297 }, { "epoch": 1.9742946708463949, "grad_norm": 6.88250207901001, "learning_rate": 3.1490000000000003e-06, "loss": 2.2056, "step": 6298 }, { "epoch": 1.9746081504702193, "grad_norm": 13.31670093536377, "learning_rate": 3.1495000000000005e-06, "loss": 2.0936, "step": 6299 }, { "epoch": 1.9749216300940438, "grad_norm": 11.524837493896484, "learning_rate": 3.1500000000000003e-06, "loss": 2.08, "step": 6300 }, { "epoch": 1.9752351097178682, "grad_norm": 8.468381881713867, "learning_rate": 3.1505e-06, "loss": 2.1082, "step": 6301 }, { "epoch": 1.9755485893416926, "grad_norm": 22.814212799072266, "learning_rate": 3.151e-06, "loss": 2.2338, "step": 6302 }, { "epoch": 1.975862068965517, "grad_norm": 15.636788368225098, "learning_rate": 3.1515000000000002e-06, "loss": 2.7083, "step": 6303 }, { "epoch": 1.9761755485893417, "grad_norm": 14.776278495788574, "learning_rate": 3.152e-06, "loss": 2.2016, "step": 6304 }, { "epoch": 1.9764890282131662, "grad_norm": 15.003888130187988, "learning_rate": 3.1525e-06, "loss": 2.2169, "step": 6305 }, { "epoch": 1.9768025078369906, "grad_norm": 17.022552490234375, "learning_rate": 3.1530000000000006e-06, "loss": 2.2239, "step": 6306 }, { "epoch": 1.977115987460815, "grad_norm": 12.406381607055664, "learning_rate": 3.1535000000000004e-06, "loss": 2.2029, "step": 6307 }, { "epoch": 1.9774294670846395, "grad_norm": 10.545319557189941, "learning_rate": 3.154e-06, "loss": 1.9441, "step": 6308 }, { "epoch": 1.977742946708464, "grad_norm": 11.90626049041748, "learning_rate": 3.1545000000000005e-06, "loss": 2.2987, "step": 6309 }, { "epoch": 1.9780564263322884, "grad_norm": 15.618658065795898, "learning_rate": 3.1550000000000003e-06, "loss": 1.8933, "step": 6310 }, { "epoch": 1.9783699059561128, "grad_norm": 9.833617210388184, "learning_rate": 3.1555e-06, "loss": 1.979, "step": 6311 }, { "epoch": 1.9786833855799373, "grad_norm": 13.40846061706543, "learning_rate": 3.1560000000000004e-06, "loss": 2.1384, "step": 6312 }, { "epoch": 1.9789968652037617, "grad_norm": 16.884504318237305, "learning_rate": 3.1565e-06, "loss": 2.3817, "step": 6313 }, { "epoch": 1.9793103448275862, "grad_norm": 30.242950439453125, "learning_rate": 3.157e-06, "loss": 2.4918, "step": 6314 }, { "epoch": 1.9796238244514106, "grad_norm": 12.3598051071167, "learning_rate": 3.1575e-06, "loss": 1.9898, "step": 6315 }, { "epoch": 1.9799373040752353, "grad_norm": 23.940189361572266, "learning_rate": 3.1580000000000005e-06, "loss": 2.2078, "step": 6316 }, { "epoch": 1.9802507836990597, "grad_norm": 6.898366928100586, "learning_rate": 3.1585000000000003e-06, "loss": 2.1173, "step": 6317 }, { "epoch": 1.9805642633228842, "grad_norm": 21.28299903869629, "learning_rate": 3.1590000000000006e-06, "loss": 2.7358, "step": 6318 }, { "epoch": 1.9808777429467086, "grad_norm": 14.182652473449707, "learning_rate": 3.1595000000000004e-06, "loss": 2.1299, "step": 6319 }, { "epoch": 1.981191222570533, "grad_norm": 9.222768783569336, "learning_rate": 3.1600000000000002e-06, "loss": 2.1796, "step": 6320 }, { "epoch": 1.9815047021943575, "grad_norm": 87.75921630859375, "learning_rate": 3.1605e-06, "loss": 2.3708, "step": 6321 }, { "epoch": 1.981818181818182, "grad_norm": 18.29620361328125, "learning_rate": 3.1610000000000003e-06, "loss": 2.4295, "step": 6322 }, { "epoch": 1.9821316614420064, "grad_norm": 14.226387977600098, "learning_rate": 3.1615e-06, "loss": 2.1621, "step": 6323 }, { "epoch": 1.9824451410658308, "grad_norm": 6.915779113769531, "learning_rate": 3.162e-06, "loss": 2.3139, "step": 6324 }, { "epoch": 1.9827586206896552, "grad_norm": 31.989805221557617, "learning_rate": 3.1625000000000002e-06, "loss": 2.4353, "step": 6325 }, { "epoch": 1.9830721003134797, "grad_norm": 10.334614753723145, "learning_rate": 3.1630000000000005e-06, "loss": 2.3891, "step": 6326 }, { "epoch": 1.9833855799373041, "grad_norm": 32.01905822753906, "learning_rate": 3.1635000000000003e-06, "loss": 2.1868, "step": 6327 }, { "epoch": 1.9836990595611286, "grad_norm": 19.62434959411621, "learning_rate": 3.1640000000000005e-06, "loss": 2.3079, "step": 6328 }, { "epoch": 1.984012539184953, "grad_norm": 12.64426040649414, "learning_rate": 3.1645000000000004e-06, "loss": 2.0242, "step": 6329 }, { "epoch": 1.9843260188087775, "grad_norm": 28.206012725830078, "learning_rate": 3.165e-06, "loss": 2.1971, "step": 6330 }, { "epoch": 1.984639498432602, "grad_norm": 17.226451873779297, "learning_rate": 3.1655000000000004e-06, "loss": 2.1305, "step": 6331 }, { "epoch": 1.9849529780564263, "grad_norm": 13.701321601867676, "learning_rate": 3.1660000000000003e-06, "loss": 2.0846, "step": 6332 }, { "epoch": 1.9852664576802508, "grad_norm": 53.87226104736328, "learning_rate": 3.1665e-06, "loss": 2.5442, "step": 6333 }, { "epoch": 1.9855799373040752, "grad_norm": 57.23225402832031, "learning_rate": 3.167e-06, "loss": 2.3749, "step": 6334 }, { "epoch": 1.9858934169278997, "grad_norm": 21.677019119262695, "learning_rate": 3.1675e-06, "loss": 2.0964, "step": 6335 }, { "epoch": 1.986206896551724, "grad_norm": 53.68826675415039, "learning_rate": 3.1680000000000004e-06, "loss": 3.0364, "step": 6336 }, { "epoch": 1.9865203761755486, "grad_norm": 33.148406982421875, "learning_rate": 3.1685000000000007e-06, "loss": 2.275, "step": 6337 }, { "epoch": 1.986833855799373, "grad_norm": 7.798027038574219, "learning_rate": 3.1690000000000005e-06, "loss": 1.9571, "step": 6338 }, { "epoch": 1.9871473354231974, "grad_norm": 9.131011962890625, "learning_rate": 3.1695000000000003e-06, "loss": 2.3012, "step": 6339 }, { "epoch": 1.9874608150470219, "grad_norm": 14.6558256149292, "learning_rate": 3.17e-06, "loss": 2.1853, "step": 6340 }, { "epoch": 1.9877742946708463, "grad_norm": 25.692052841186523, "learning_rate": 3.1705000000000004e-06, "loss": 2.2872, "step": 6341 }, { "epoch": 1.9880877742946708, "grad_norm": 43.00083541870117, "learning_rate": 3.1710000000000002e-06, "loss": 2.8004, "step": 6342 }, { "epoch": 1.9884012539184952, "grad_norm": 17.228382110595703, "learning_rate": 3.1715e-06, "loss": 2.2165, "step": 6343 }, { "epoch": 1.9887147335423196, "grad_norm": 19.439098358154297, "learning_rate": 3.172e-06, "loss": 2.3464, "step": 6344 }, { "epoch": 1.989028213166144, "grad_norm": 49.01555633544922, "learning_rate": 3.1725e-06, "loss": 2.4074, "step": 6345 }, { "epoch": 1.9893416927899685, "grad_norm": 14.276300430297852, "learning_rate": 3.1730000000000004e-06, "loss": 2.1991, "step": 6346 }, { "epoch": 1.989655172413793, "grad_norm": 23.116601943969727, "learning_rate": 3.1735000000000006e-06, "loss": 2.1692, "step": 6347 }, { "epoch": 1.9899686520376174, "grad_norm": 15.478215217590332, "learning_rate": 3.1740000000000004e-06, "loss": 2.0933, "step": 6348 }, { "epoch": 1.9902821316614419, "grad_norm": 17.975017547607422, "learning_rate": 3.1745000000000003e-06, "loss": 2.1837, "step": 6349 }, { "epoch": 1.9905956112852663, "grad_norm": 6.596172332763672, "learning_rate": 3.175e-06, "loss": 2.0905, "step": 6350 }, { "epoch": 1.990909090909091, "grad_norm": 13.028762817382812, "learning_rate": 3.1755000000000003e-06, "loss": 2.2581, "step": 6351 }, { "epoch": 1.9912225705329154, "grad_norm": 10.148426055908203, "learning_rate": 3.176e-06, "loss": 2.3209, "step": 6352 }, { "epoch": 1.9915360501567398, "grad_norm": 7.7970476150512695, "learning_rate": 3.1765e-06, "loss": 2.0534, "step": 6353 }, { "epoch": 1.9918495297805643, "grad_norm": 47.781227111816406, "learning_rate": 3.1770000000000002e-06, "loss": 2.2884, "step": 6354 }, { "epoch": 1.9921630094043887, "grad_norm": 13.419988632202148, "learning_rate": 3.1775e-06, "loss": 2.347, "step": 6355 }, { "epoch": 1.9924764890282132, "grad_norm": 14.526352882385254, "learning_rate": 3.1780000000000003e-06, "loss": 2.3624, "step": 6356 }, { "epoch": 1.9927899686520376, "grad_norm": 51.24506759643555, "learning_rate": 3.1785000000000006e-06, "loss": 2.3981, "step": 6357 }, { "epoch": 1.993103448275862, "grad_norm": 46.352943420410156, "learning_rate": 3.1790000000000004e-06, "loss": 2.729, "step": 6358 }, { "epoch": 1.9934169278996865, "grad_norm": 11.26982307434082, "learning_rate": 3.1795000000000002e-06, "loss": 2.1656, "step": 6359 }, { "epoch": 1.993730407523511, "grad_norm": 12.368547439575195, "learning_rate": 3.1800000000000005e-06, "loss": 2.3287, "step": 6360 }, { "epoch": 1.9940438871473354, "grad_norm": 12.606298446655273, "learning_rate": 3.1805000000000003e-06, "loss": 2.1878, "step": 6361 }, { "epoch": 1.9943573667711598, "grad_norm": 6.567707538604736, "learning_rate": 3.181e-06, "loss": 2.2835, "step": 6362 }, { "epoch": 1.9946708463949845, "grad_norm": 25.984466552734375, "learning_rate": 3.1815e-06, "loss": 2.4004, "step": 6363 }, { "epoch": 1.994984326018809, "grad_norm": 17.47072410583496, "learning_rate": 3.182e-06, "loss": 2.3559, "step": 6364 }, { "epoch": 1.9952978056426334, "grad_norm": 15.98811149597168, "learning_rate": 3.1825e-06, "loss": 2.1197, "step": 6365 }, { "epoch": 1.9956112852664578, "grad_norm": 6.573192119598389, "learning_rate": 3.1830000000000007e-06, "loss": 2.4837, "step": 6366 }, { "epoch": 1.9959247648902823, "grad_norm": 18.482707977294922, "learning_rate": 3.1835000000000005e-06, "loss": 2.2992, "step": 6367 }, { "epoch": 1.9962382445141067, "grad_norm": 19.308034896850586, "learning_rate": 3.1840000000000003e-06, "loss": 2.3165, "step": 6368 }, { "epoch": 1.9965517241379311, "grad_norm": 6.105698585510254, "learning_rate": 3.1845e-06, "loss": 1.9594, "step": 6369 }, { "epoch": 1.9968652037617556, "grad_norm": 32.73683547973633, "learning_rate": 3.1850000000000004e-06, "loss": 2.3862, "step": 6370 }, { "epoch": 1.99717868338558, "grad_norm": 13.382736206054688, "learning_rate": 3.1855000000000002e-06, "loss": 2.3056, "step": 6371 }, { "epoch": 1.9974921630094045, "grad_norm": 50.91844177246094, "learning_rate": 3.186e-06, "loss": 2.1562, "step": 6372 }, { "epoch": 1.997805642633229, "grad_norm": 16.6110897064209, "learning_rate": 3.1865000000000003e-06, "loss": 2.1552, "step": 6373 }, { "epoch": 1.9981191222570533, "grad_norm": 12.067429542541504, "learning_rate": 3.187e-06, "loss": 1.9695, "step": 6374 }, { "epoch": 1.9984326018808778, "grad_norm": 47.696414947509766, "learning_rate": 3.1875e-06, "loss": 2.3831, "step": 6375 }, { "epoch": 1.9987460815047022, "grad_norm": 22.687246322631836, "learning_rate": 3.188e-06, "loss": 2.4178, "step": 6376 }, { "epoch": 1.9990595611285267, "grad_norm": 16.281429290771484, "learning_rate": 3.1885000000000005e-06, "loss": 2.5671, "step": 6377 }, { "epoch": 1.9993730407523511, "grad_norm": 12.381708145141602, "learning_rate": 3.1890000000000003e-06, "loss": 2.4764, "step": 6378 }, { "epoch": 1.9996865203761756, "grad_norm": 12.664398193359375, "learning_rate": 3.1895000000000005e-06, "loss": 1.9393, "step": 6379 }, { "epoch": 2.0, "grad_norm": 11.222203254699707, "learning_rate": 3.1900000000000004e-06, "loss": 2.177, "step": 6380 }, { "epoch": 2.0003134796238244, "grad_norm": 12.29688549041748, "learning_rate": 3.1905e-06, "loss": 2.0168, "step": 6381 }, { "epoch": 2.000626959247649, "grad_norm": 27.42327880859375, "learning_rate": 3.191e-06, "loss": 1.999, "step": 6382 }, { "epoch": 2.0009404388714733, "grad_norm": 14.081686973571777, "learning_rate": 3.1915000000000003e-06, "loss": 2.2369, "step": 6383 }, { "epoch": 2.0012539184952978, "grad_norm": 12.910845756530762, "learning_rate": 3.192e-06, "loss": 1.9888, "step": 6384 }, { "epoch": 2.0012539184952978, "eval_loss": 2.602219820022583, "eval_runtime": 20.9002, "eval_samples_per_second": 128.563, "eval_steps_per_second": 8.038, "step": 6384 }, { "epoch": 2.001567398119122, "grad_norm": 25.162681579589844, "learning_rate": 3.1925e-06, "loss": 2.8597, "step": 6385 }, { "epoch": 2.0018808777429467, "grad_norm": 9.405834197998047, "learning_rate": 3.193e-06, "loss": 2.3257, "step": 6386 }, { "epoch": 2.002194357366771, "grad_norm": 11.864672660827637, "learning_rate": 3.1935000000000004e-06, "loss": 2.1454, "step": 6387 }, { "epoch": 2.0025078369905955, "grad_norm": 39.45462417602539, "learning_rate": 3.1940000000000003e-06, "loss": 2.7696, "step": 6388 }, { "epoch": 2.00282131661442, "grad_norm": 11.640250205993652, "learning_rate": 3.1945000000000005e-06, "loss": 2.2456, "step": 6389 }, { "epoch": 2.0031347962382444, "grad_norm": 17.460670471191406, "learning_rate": 3.1950000000000003e-06, "loss": 2.6799, "step": 6390 }, { "epoch": 2.003448275862069, "grad_norm": 52.15612030029297, "learning_rate": 3.1955e-06, "loss": 2.4082, "step": 6391 }, { "epoch": 2.0037617554858933, "grad_norm": 28.452922821044922, "learning_rate": 3.1960000000000004e-06, "loss": 2.2057, "step": 6392 }, { "epoch": 2.0040752351097177, "grad_norm": 12.585026741027832, "learning_rate": 3.1965000000000002e-06, "loss": 2.1995, "step": 6393 }, { "epoch": 2.004388714733542, "grad_norm": 34.237396240234375, "learning_rate": 3.197e-06, "loss": 2.265, "step": 6394 }, { "epoch": 2.0047021943573666, "grad_norm": 12.680131912231445, "learning_rate": 3.1975e-06, "loss": 2.0751, "step": 6395 }, { "epoch": 2.005015673981191, "grad_norm": 17.35845947265625, "learning_rate": 3.198e-06, "loss": 2.3537, "step": 6396 }, { "epoch": 2.0053291536050155, "grad_norm": 77.56371307373047, "learning_rate": 3.1985000000000004e-06, "loss": 2.4984, "step": 6397 }, { "epoch": 2.00564263322884, "grad_norm": 13.418671607971191, "learning_rate": 3.1990000000000006e-06, "loss": 2.3835, "step": 6398 }, { "epoch": 2.0059561128526644, "grad_norm": 39.91530227661133, "learning_rate": 3.1995000000000005e-06, "loss": 2.4927, "step": 6399 }, { "epoch": 2.006269592476489, "grad_norm": 9.506710052490234, "learning_rate": 3.2000000000000003e-06, "loss": 2.1323, "step": 6400 }, { "epoch": 2.0065830721003133, "grad_norm": 18.124570846557617, "learning_rate": 3.2005e-06, "loss": 2.2371, "step": 6401 }, { "epoch": 2.0068965517241377, "grad_norm": 20.50752067565918, "learning_rate": 3.2010000000000004e-06, "loss": 2.2125, "step": 6402 }, { "epoch": 2.007210031347962, "grad_norm": 10.371642112731934, "learning_rate": 3.2015e-06, "loss": 2.2344, "step": 6403 }, { "epoch": 2.007523510971787, "grad_norm": 13.89661693572998, "learning_rate": 3.202e-06, "loss": 2.2632, "step": 6404 }, { "epoch": 2.0078369905956115, "grad_norm": 71.66805267333984, "learning_rate": 3.2025000000000003e-06, "loss": 2.0427, "step": 6405 }, { "epoch": 2.008150470219436, "grad_norm": 11.179061889648438, "learning_rate": 3.203e-06, "loss": 2.218, "step": 6406 }, { "epoch": 2.0084639498432604, "grad_norm": 11.413198471069336, "learning_rate": 3.2035000000000003e-06, "loss": 2.303, "step": 6407 }, { "epoch": 2.008777429467085, "grad_norm": 8.672221183776855, "learning_rate": 3.2040000000000006e-06, "loss": 2.2851, "step": 6408 }, { "epoch": 2.0090909090909093, "grad_norm": 12.593329429626465, "learning_rate": 3.2045000000000004e-06, "loss": 2.0457, "step": 6409 }, { "epoch": 2.0094043887147337, "grad_norm": 29.556936264038086, "learning_rate": 3.2050000000000002e-06, "loss": 1.9656, "step": 6410 }, { "epoch": 2.009717868338558, "grad_norm": 156.7764129638672, "learning_rate": 3.2055000000000005e-06, "loss": 2.3355, "step": 6411 }, { "epoch": 2.0100313479623826, "grad_norm": 13.569225311279297, "learning_rate": 3.2060000000000003e-06, "loss": 2.469, "step": 6412 }, { "epoch": 2.010344827586207, "grad_norm": 102.1455078125, "learning_rate": 3.2065e-06, "loss": 2.127, "step": 6413 }, { "epoch": 2.0106583072100315, "grad_norm": 37.82844924926758, "learning_rate": 3.207e-06, "loss": 2.2039, "step": 6414 }, { "epoch": 2.010971786833856, "grad_norm": 7.729022026062012, "learning_rate": 3.2075e-06, "loss": 2.0836, "step": 6415 }, { "epoch": 2.0112852664576804, "grad_norm": 11.092761993408203, "learning_rate": 3.208e-06, "loss": 2.2979, "step": 6416 }, { "epoch": 2.011598746081505, "grad_norm": 12.816972732543945, "learning_rate": 3.2085000000000007e-06, "loss": 2.1071, "step": 6417 }, { "epoch": 2.0119122257053292, "grad_norm": 18.535078048706055, "learning_rate": 3.2090000000000005e-06, "loss": 2.1197, "step": 6418 }, { "epoch": 2.0122257053291537, "grad_norm": 11.44892406463623, "learning_rate": 3.2095000000000004e-06, "loss": 2.3013, "step": 6419 }, { "epoch": 2.012539184952978, "grad_norm": 9.000450134277344, "learning_rate": 3.21e-06, "loss": 2.3279, "step": 6420 }, { "epoch": 2.0128526645768026, "grad_norm": 11.124773025512695, "learning_rate": 3.2105000000000004e-06, "loss": 2.284, "step": 6421 }, { "epoch": 2.013166144200627, "grad_norm": 20.694377899169922, "learning_rate": 3.2110000000000003e-06, "loss": 2.2751, "step": 6422 }, { "epoch": 2.0134796238244514, "grad_norm": 15.966277122497559, "learning_rate": 3.2115e-06, "loss": 2.0844, "step": 6423 }, { "epoch": 2.013793103448276, "grad_norm": 25.242921829223633, "learning_rate": 3.212e-06, "loss": 2.2729, "step": 6424 }, { "epoch": 2.0141065830721003, "grad_norm": 15.954590797424316, "learning_rate": 3.2125e-06, "loss": 2.1691, "step": 6425 }, { "epoch": 2.0144200626959248, "grad_norm": 11.648881912231445, "learning_rate": 3.213e-06, "loss": 2.2821, "step": 6426 }, { "epoch": 2.014733542319749, "grad_norm": 8.528559684753418, "learning_rate": 3.2135000000000007e-06, "loss": 2.1041, "step": 6427 }, { "epoch": 2.0150470219435737, "grad_norm": 37.761295318603516, "learning_rate": 3.2140000000000005e-06, "loss": 2.4045, "step": 6428 }, { "epoch": 2.015360501567398, "grad_norm": 15.822317123413086, "learning_rate": 3.2145000000000003e-06, "loss": 2.5618, "step": 6429 }, { "epoch": 2.0156739811912225, "grad_norm": 12.833781242370605, "learning_rate": 3.215e-06, "loss": 2.1492, "step": 6430 }, { "epoch": 2.015987460815047, "grad_norm": 7.219548225402832, "learning_rate": 3.2155000000000004e-06, "loss": 2.3691, "step": 6431 }, { "epoch": 2.0163009404388714, "grad_norm": 8.92505931854248, "learning_rate": 3.216e-06, "loss": 2.1901, "step": 6432 }, { "epoch": 2.016614420062696, "grad_norm": 46.86988067626953, "learning_rate": 3.2165e-06, "loss": 2.6455, "step": 6433 }, { "epoch": 2.0169278996865203, "grad_norm": 9.709528923034668, "learning_rate": 3.2170000000000003e-06, "loss": 2.3117, "step": 6434 }, { "epoch": 2.0172413793103448, "grad_norm": 99.01873779296875, "learning_rate": 3.2175e-06, "loss": 3.4444, "step": 6435 }, { "epoch": 2.017554858934169, "grad_norm": 17.069381713867188, "learning_rate": 3.218e-06, "loss": 2.2857, "step": 6436 }, { "epoch": 2.0178683385579936, "grad_norm": 10.380099296569824, "learning_rate": 3.2185000000000006e-06, "loss": 2.2406, "step": 6437 }, { "epoch": 2.018181818181818, "grad_norm": 27.852876663208008, "learning_rate": 3.2190000000000004e-06, "loss": 2.0978, "step": 6438 }, { "epoch": 2.0184952978056425, "grad_norm": 26.04537582397461, "learning_rate": 3.2195000000000003e-06, "loss": 2.2703, "step": 6439 }, { "epoch": 2.018808777429467, "grad_norm": 11.134775161743164, "learning_rate": 3.2200000000000005e-06, "loss": 2.3914, "step": 6440 }, { "epoch": 2.0191222570532914, "grad_norm": 225.6446990966797, "learning_rate": 3.2205000000000003e-06, "loss": 2.6848, "step": 6441 }, { "epoch": 2.019435736677116, "grad_norm": 23.590717315673828, "learning_rate": 3.221e-06, "loss": 2.0387, "step": 6442 }, { "epoch": 2.0197492163009403, "grad_norm": 10.782753944396973, "learning_rate": 3.2215e-06, "loss": 2.0235, "step": 6443 }, { "epoch": 2.0200626959247647, "grad_norm": 15.721902847290039, "learning_rate": 3.2220000000000002e-06, "loss": 2.498, "step": 6444 }, { "epoch": 2.020376175548589, "grad_norm": 10.394645690917969, "learning_rate": 3.2225e-06, "loss": 2.3327, "step": 6445 }, { "epoch": 2.0206896551724136, "grad_norm": 13.19404411315918, "learning_rate": 3.223e-06, "loss": 2.486, "step": 6446 }, { "epoch": 2.021003134796238, "grad_norm": 11.165719985961914, "learning_rate": 3.2235000000000006e-06, "loss": 2.0418, "step": 6447 }, { "epoch": 2.0213166144200625, "grad_norm": 26.503110885620117, "learning_rate": 3.2240000000000004e-06, "loss": 2.379, "step": 6448 }, { "epoch": 2.021630094043887, "grad_norm": 20.466827392578125, "learning_rate": 3.2245000000000002e-06, "loss": 2.4795, "step": 6449 }, { "epoch": 2.0219435736677114, "grad_norm": 12.40024185180664, "learning_rate": 3.2250000000000005e-06, "loss": 2.1691, "step": 6450 }, { "epoch": 2.0222570532915363, "grad_norm": 12.075732231140137, "learning_rate": 3.2255000000000003e-06, "loss": 2.2837, "step": 6451 }, { "epoch": 2.0225705329153607, "grad_norm": 10.415963172912598, "learning_rate": 3.226e-06, "loss": 2.1337, "step": 6452 }, { "epoch": 2.022884012539185, "grad_norm": 18.45778465270996, "learning_rate": 3.2265000000000004e-06, "loss": 2.2322, "step": 6453 }, { "epoch": 2.0231974921630096, "grad_norm": 28.994972229003906, "learning_rate": 3.227e-06, "loss": 2.5254, "step": 6454 }, { "epoch": 2.023510971786834, "grad_norm": 15.62054443359375, "learning_rate": 3.2275e-06, "loss": 2.1627, "step": 6455 }, { "epoch": 2.0238244514106585, "grad_norm": 8.304019927978516, "learning_rate": 3.228e-06, "loss": 2.2992, "step": 6456 }, { "epoch": 2.024137931034483, "grad_norm": 26.025936126708984, "learning_rate": 3.2285000000000005e-06, "loss": 2.0311, "step": 6457 }, { "epoch": 2.0244514106583074, "grad_norm": 16.436079025268555, "learning_rate": 3.2290000000000003e-06, "loss": 2.5087, "step": 6458 }, { "epoch": 2.024764890282132, "grad_norm": 14.592082977294922, "learning_rate": 3.2295000000000006e-06, "loss": 2.1291, "step": 6459 }, { "epoch": 2.0250783699059562, "grad_norm": 51.50830841064453, "learning_rate": 3.2300000000000004e-06, "loss": 2.7696, "step": 6460 }, { "epoch": 2.0253918495297807, "grad_norm": 60.234981536865234, "learning_rate": 3.2305000000000002e-06, "loss": 2.505, "step": 6461 }, { "epoch": 2.025705329153605, "grad_norm": 13.99014663696289, "learning_rate": 3.231e-06, "loss": 2.1314, "step": 6462 }, { "epoch": 2.0260188087774296, "grad_norm": 40.045570373535156, "learning_rate": 3.2315000000000003e-06, "loss": 3.2125, "step": 6463 }, { "epoch": 2.026332288401254, "grad_norm": 22.33523941040039, "learning_rate": 3.232e-06, "loss": 2.3598, "step": 6464 }, { "epoch": 2.0266457680250785, "grad_norm": 23.795475006103516, "learning_rate": 3.2325e-06, "loss": 2.303, "step": 6465 }, { "epoch": 2.026959247648903, "grad_norm": 20.435598373413086, "learning_rate": 3.2330000000000002e-06, "loss": 2.3792, "step": 6466 }, { "epoch": 2.0272727272727273, "grad_norm": 45.86872100830078, "learning_rate": 3.2335000000000005e-06, "loss": 2.1971, "step": 6467 }, { "epoch": 2.027586206896552, "grad_norm": 13.451287269592285, "learning_rate": 3.2340000000000003e-06, "loss": 2.2069, "step": 6468 }, { "epoch": 2.027899686520376, "grad_norm": 41.58235549926758, "learning_rate": 3.2345000000000005e-06, "loss": 1.9855, "step": 6469 }, { "epoch": 2.0282131661442007, "grad_norm": 182.38726806640625, "learning_rate": 3.2350000000000004e-06, "loss": 2.9326, "step": 6470 }, { "epoch": 2.028526645768025, "grad_norm": 11.725469589233398, "learning_rate": 3.2355e-06, "loss": 2.1926, "step": 6471 }, { "epoch": 2.0288401253918495, "grad_norm": 9.941828727722168, "learning_rate": 3.2360000000000004e-06, "loss": 2.1713, "step": 6472 }, { "epoch": 2.029153605015674, "grad_norm": 11.165457725524902, "learning_rate": 3.2365000000000003e-06, "loss": 2.2665, "step": 6473 }, { "epoch": 2.0294670846394984, "grad_norm": 25.072948455810547, "learning_rate": 3.237e-06, "loss": 2.0548, "step": 6474 }, { "epoch": 2.029780564263323, "grad_norm": 20.833473205566406, "learning_rate": 3.2375e-06, "loss": 2.3578, "step": 6475 }, { "epoch": 2.0300940438871473, "grad_norm": 77.08531951904297, "learning_rate": 3.238e-06, "loss": 2.6897, "step": 6476 }, { "epoch": 2.0304075235109718, "grad_norm": 11.294124603271484, "learning_rate": 3.2385000000000004e-06, "loss": 2.21, "step": 6477 }, { "epoch": 2.030721003134796, "grad_norm": 10.167134284973145, "learning_rate": 3.2390000000000007e-06, "loss": 1.9652, "step": 6478 }, { "epoch": 2.0310344827586206, "grad_norm": 12.622770309448242, "learning_rate": 3.2395000000000005e-06, "loss": 2.1103, "step": 6479 }, { "epoch": 2.031347962382445, "grad_norm": 33.274600982666016, "learning_rate": 3.2400000000000003e-06, "loss": 2.1251, "step": 6480 }, { "epoch": 2.0316614420062695, "grad_norm": 16.832984924316406, "learning_rate": 3.2405e-06, "loss": 2.0308, "step": 6481 }, { "epoch": 2.031974921630094, "grad_norm": 19.317169189453125, "learning_rate": 3.2410000000000004e-06, "loss": 2.7644, "step": 6482 }, { "epoch": 2.0322884012539184, "grad_norm": 11.893217086791992, "learning_rate": 3.2415000000000002e-06, "loss": 2.3136, "step": 6483 }, { "epoch": 2.032601880877743, "grad_norm": 63.95478439331055, "learning_rate": 3.242e-06, "loss": 2.596, "step": 6484 }, { "epoch": 2.0329153605015673, "grad_norm": 21.883337020874023, "learning_rate": 3.2425000000000003e-06, "loss": 2.1207, "step": 6485 }, { "epoch": 2.0332288401253917, "grad_norm": 24.365686416625977, "learning_rate": 3.243e-06, "loss": 2.2292, "step": 6486 }, { "epoch": 2.033542319749216, "grad_norm": 15.450721740722656, "learning_rate": 3.2435000000000004e-06, "loss": 2.2825, "step": 6487 }, { "epoch": 2.0338557993730406, "grad_norm": 27.794113159179688, "learning_rate": 3.2440000000000006e-06, "loss": 2.1254, "step": 6488 }, { "epoch": 2.034169278996865, "grad_norm": 16.02458953857422, "learning_rate": 3.2445000000000004e-06, "loss": 2.2079, "step": 6489 }, { "epoch": 2.0344827586206895, "grad_norm": 11.653332710266113, "learning_rate": 3.2450000000000003e-06, "loss": 2.4478, "step": 6490 }, { "epoch": 2.034796238244514, "grad_norm": 11.9269380569458, "learning_rate": 3.2455000000000005e-06, "loss": 2.0945, "step": 6491 }, { "epoch": 2.0351097178683384, "grad_norm": 17.969438552856445, "learning_rate": 3.2460000000000003e-06, "loss": 2.0836, "step": 6492 }, { "epoch": 2.035423197492163, "grad_norm": 17.720714569091797, "learning_rate": 3.2465e-06, "loss": 2.2676, "step": 6493 }, { "epoch": 2.0357366771159873, "grad_norm": 22.731082916259766, "learning_rate": 3.247e-06, "loss": 2.3405, "step": 6494 }, { "epoch": 2.0360501567398117, "grad_norm": 11.327211380004883, "learning_rate": 3.2475000000000002e-06, "loss": 2.6774, "step": 6495 }, { "epoch": 2.036363636363636, "grad_norm": 8.33624267578125, "learning_rate": 3.248e-06, "loss": 2.1673, "step": 6496 }, { "epoch": 2.0366771159874606, "grad_norm": 10.953970909118652, "learning_rate": 3.2485000000000007e-06, "loss": 1.7377, "step": 6497 }, { "epoch": 2.0369905956112855, "grad_norm": 10.545639038085938, "learning_rate": 3.2490000000000006e-06, "loss": 2.1307, "step": 6498 }, { "epoch": 2.03730407523511, "grad_norm": 12.465002059936523, "learning_rate": 3.2495000000000004e-06, "loss": 2.2898, "step": 6499 }, { "epoch": 2.0376175548589344, "grad_norm": 11.045977592468262, "learning_rate": 3.2500000000000002e-06, "loss": 2.0794, "step": 6500 }, { "epoch": 2.037931034482759, "grad_norm": 12.869364738464355, "learning_rate": 3.2505000000000005e-06, "loss": 2.1772, "step": 6501 }, { "epoch": 2.0382445141065832, "grad_norm": 12.978128433227539, "learning_rate": 3.2510000000000003e-06, "loss": 1.9301, "step": 6502 }, { "epoch": 2.0385579937304077, "grad_norm": 10.306305885314941, "learning_rate": 3.2515e-06, "loss": 1.9183, "step": 6503 }, { "epoch": 2.038871473354232, "grad_norm": 16.00723648071289, "learning_rate": 3.252e-06, "loss": 2.0444, "step": 6504 }, { "epoch": 2.0391849529780566, "grad_norm": 11.812612533569336, "learning_rate": 3.2525e-06, "loss": 2.2402, "step": 6505 }, { "epoch": 2.039498432601881, "grad_norm": 71.21012878417969, "learning_rate": 3.253e-06, "loss": 2.2685, "step": 6506 }, { "epoch": 2.0398119122257055, "grad_norm": 77.43359375, "learning_rate": 3.2535e-06, "loss": 2.4774, "step": 6507 }, { "epoch": 2.04012539184953, "grad_norm": 8.491291046142578, "learning_rate": 3.2540000000000005e-06, "loss": 2.1702, "step": 6508 }, { "epoch": 2.0404388714733543, "grad_norm": 7.935294151306152, "learning_rate": 3.2545000000000003e-06, "loss": 2.233, "step": 6509 }, { "epoch": 2.040752351097179, "grad_norm": 20.703697204589844, "learning_rate": 3.255e-06, "loss": 2.514, "step": 6510 }, { "epoch": 2.0410658307210032, "grad_norm": 17.19257354736328, "learning_rate": 3.2555000000000004e-06, "loss": 2.1792, "step": 6511 }, { "epoch": 2.0413793103448277, "grad_norm": 15.406134605407715, "learning_rate": 3.2560000000000003e-06, "loss": 2.3231, "step": 6512 }, { "epoch": 2.041692789968652, "grad_norm": 16.585111618041992, "learning_rate": 3.2565e-06, "loss": 2.6569, "step": 6513 }, { "epoch": 2.0420062695924766, "grad_norm": 10.205801010131836, "learning_rate": 3.2570000000000003e-06, "loss": 2.3445, "step": 6514 }, { "epoch": 2.042319749216301, "grad_norm": 35.931095123291016, "learning_rate": 3.2575e-06, "loss": 2.2405, "step": 6515 }, { "epoch": 2.0426332288401254, "grad_norm": 17.4227352142334, "learning_rate": 3.258e-06, "loss": 2.4193, "step": 6516 }, { "epoch": 2.04294670846395, "grad_norm": 14.396782875061035, "learning_rate": 3.2585e-06, "loss": 2.0513, "step": 6517 }, { "epoch": 2.0432601880877743, "grad_norm": 11.218772888183594, "learning_rate": 3.2590000000000005e-06, "loss": 1.9101, "step": 6518 }, { "epoch": 2.0435736677115988, "grad_norm": 10.642355918884277, "learning_rate": 3.2595000000000003e-06, "loss": 2.0912, "step": 6519 }, { "epoch": 2.043887147335423, "grad_norm": 13.464892387390137, "learning_rate": 3.2600000000000006e-06, "loss": 2.5412, "step": 6520 }, { "epoch": 2.0442006269592476, "grad_norm": 23.115524291992188, "learning_rate": 3.2605000000000004e-06, "loss": 2.3612, "step": 6521 }, { "epoch": 2.044514106583072, "grad_norm": 9.855347633361816, "learning_rate": 3.261e-06, "loss": 2.1887, "step": 6522 }, { "epoch": 2.0448275862068965, "grad_norm": 19.134239196777344, "learning_rate": 3.2615e-06, "loss": 2.5038, "step": 6523 }, { "epoch": 2.045141065830721, "grad_norm": 10.420708656311035, "learning_rate": 3.2620000000000003e-06, "loss": 2.0677, "step": 6524 }, { "epoch": 2.0454545454545454, "grad_norm": 8.537455558776855, "learning_rate": 3.2625e-06, "loss": 2.2231, "step": 6525 }, { "epoch": 2.04576802507837, "grad_norm": 12.047176361083984, "learning_rate": 3.263e-06, "loss": 2.0603, "step": 6526 }, { "epoch": 2.0460815047021943, "grad_norm": 33.341251373291016, "learning_rate": 3.2635e-06, "loss": 2.5696, "step": 6527 }, { "epoch": 2.0463949843260187, "grad_norm": 13.5247802734375, "learning_rate": 3.2640000000000004e-06, "loss": 2.2928, "step": 6528 }, { "epoch": 2.046708463949843, "grad_norm": 11.721467018127441, "learning_rate": 3.2645000000000003e-06, "loss": 2.1597, "step": 6529 }, { "epoch": 2.0470219435736676, "grad_norm": 31.335716247558594, "learning_rate": 3.2650000000000005e-06, "loss": 2.9279, "step": 6530 }, { "epoch": 2.047335423197492, "grad_norm": 24.000282287597656, "learning_rate": 3.2655000000000003e-06, "loss": 3.0904, "step": 6531 }, { "epoch": 2.0476489028213165, "grad_norm": 33.47315216064453, "learning_rate": 3.266e-06, "loss": 2.1115, "step": 6532 }, { "epoch": 2.047962382445141, "grad_norm": 17.498619079589844, "learning_rate": 3.2665000000000004e-06, "loss": 1.9693, "step": 6533 }, { "epoch": 2.0482758620689654, "grad_norm": 10.471491813659668, "learning_rate": 3.2670000000000002e-06, "loss": 2.3144, "step": 6534 }, { "epoch": 2.04858934169279, "grad_norm": 12.451190948486328, "learning_rate": 3.2675e-06, "loss": 2.3217, "step": 6535 }, { "epoch": 2.0489028213166143, "grad_norm": 10.55508804321289, "learning_rate": 3.268e-06, "loss": 2.3057, "step": 6536 }, { "epoch": 2.0492163009404387, "grad_norm": 11.816363334655762, "learning_rate": 3.2685e-06, "loss": 2.2614, "step": 6537 }, { "epoch": 2.049529780564263, "grad_norm": 22.170557022094727, "learning_rate": 3.2690000000000004e-06, "loss": 2.1288, "step": 6538 }, { "epoch": 2.0498432601880876, "grad_norm": 22.232704162597656, "learning_rate": 3.2695000000000006e-06, "loss": 2.0275, "step": 6539 }, { "epoch": 2.050156739811912, "grad_norm": 12.408869743347168, "learning_rate": 3.2700000000000005e-06, "loss": 2.1547, "step": 6540 }, { "epoch": 2.0504702194357365, "grad_norm": 9.901449203491211, "learning_rate": 3.2705000000000003e-06, "loss": 2.2279, "step": 6541 }, { "epoch": 2.050783699059561, "grad_norm": 49.53593063354492, "learning_rate": 3.271e-06, "loss": 2.274, "step": 6542 }, { "epoch": 2.0510971786833854, "grad_norm": 13.901050567626953, "learning_rate": 3.2715000000000004e-06, "loss": 2.378, "step": 6543 }, { "epoch": 2.0514106583072103, "grad_norm": 15.067317962646484, "learning_rate": 3.272e-06, "loss": 2.2933, "step": 6544 }, { "epoch": 2.0517241379310347, "grad_norm": 14.043885231018066, "learning_rate": 3.2725e-06, "loss": 2.2197, "step": 6545 }, { "epoch": 2.052037617554859, "grad_norm": 47.624839782714844, "learning_rate": 3.2730000000000003e-06, "loss": 2.2144, "step": 6546 }, { "epoch": 2.0523510971786836, "grad_norm": 12.177602767944336, "learning_rate": 3.2735e-06, "loss": 2.2723, "step": 6547 }, { "epoch": 2.052664576802508, "grad_norm": 12.302322387695312, "learning_rate": 3.2740000000000003e-06, "loss": 2.2704, "step": 6548 }, { "epoch": 2.0529780564263325, "grad_norm": 12.53136920928955, "learning_rate": 3.2745000000000006e-06, "loss": 2.1673, "step": 6549 }, { "epoch": 2.053291536050157, "grad_norm": 23.230165481567383, "learning_rate": 3.2750000000000004e-06, "loss": 2.3295, "step": 6550 }, { "epoch": 2.0536050156739813, "grad_norm": 7.349822998046875, "learning_rate": 3.2755000000000002e-06, "loss": 2.1326, "step": 6551 }, { "epoch": 2.053918495297806, "grad_norm": 10.001233100891113, "learning_rate": 3.2760000000000005e-06, "loss": 2.4589, "step": 6552 }, { "epoch": 2.0542319749216302, "grad_norm": 11.291881561279297, "learning_rate": 3.2765000000000003e-06, "loss": 2.2089, "step": 6553 }, { "epoch": 2.0545454545454547, "grad_norm": 9.548392295837402, "learning_rate": 3.277e-06, "loss": 2.1072, "step": 6554 }, { "epoch": 2.054858934169279, "grad_norm": 68.1001968383789, "learning_rate": 3.2775e-06, "loss": 2.7502, "step": 6555 }, { "epoch": 2.0551724137931036, "grad_norm": 17.324678421020508, "learning_rate": 3.278e-06, "loss": 2.3864, "step": 6556 }, { "epoch": 2.055485893416928, "grad_norm": 13.538605690002441, "learning_rate": 3.2785e-06, "loss": 2.3208, "step": 6557 }, { "epoch": 2.0557993730407524, "grad_norm": 11.606280326843262, "learning_rate": 3.2790000000000007e-06, "loss": 2.18, "step": 6558 }, { "epoch": 2.056112852664577, "grad_norm": 14.955106735229492, "learning_rate": 3.2795000000000005e-06, "loss": 2.4096, "step": 6559 }, { "epoch": 2.0564263322884013, "grad_norm": 37.917640686035156, "learning_rate": 3.2800000000000004e-06, "loss": 2.3191, "step": 6560 }, { "epoch": 2.0567398119122258, "grad_norm": 22.407651901245117, "learning_rate": 3.2805e-06, "loss": 2.233, "step": 6561 }, { "epoch": 2.05705329153605, "grad_norm": 22.621692657470703, "learning_rate": 3.2810000000000004e-06, "loss": 2.0556, "step": 6562 }, { "epoch": 2.0573667711598747, "grad_norm": 12.181644439697266, "learning_rate": 3.2815000000000003e-06, "loss": 2.1268, "step": 6563 }, { "epoch": 2.057680250783699, "grad_norm": 15.706012725830078, "learning_rate": 3.282e-06, "loss": 2.1221, "step": 6564 }, { "epoch": 2.0579937304075235, "grad_norm": 52.52977752685547, "learning_rate": 3.2825000000000003e-06, "loss": 2.0932, "step": 6565 }, { "epoch": 2.058307210031348, "grad_norm": 17.487096786499023, "learning_rate": 3.283e-06, "loss": 2.4294, "step": 6566 }, { "epoch": 2.0586206896551724, "grad_norm": 17.99802589416504, "learning_rate": 3.2835e-06, "loss": 2.3314, "step": 6567 }, { "epoch": 2.058934169278997, "grad_norm": 18.688634872436523, "learning_rate": 3.2840000000000007e-06, "loss": 2.3654, "step": 6568 }, { "epoch": 2.0592476489028213, "grad_norm": 10.932000160217285, "learning_rate": 3.2845000000000005e-06, "loss": 1.8595, "step": 6569 }, { "epoch": 2.0595611285266457, "grad_norm": 118.30567932128906, "learning_rate": 3.2850000000000003e-06, "loss": 3.0121, "step": 6570 }, { "epoch": 2.05987460815047, "grad_norm": 26.23569107055664, "learning_rate": 3.2855000000000006e-06, "loss": 2.0933, "step": 6571 }, { "epoch": 2.0601880877742946, "grad_norm": 86.23782348632812, "learning_rate": 3.2860000000000004e-06, "loss": 2.0556, "step": 6572 }, { "epoch": 2.060501567398119, "grad_norm": 32.5789909362793, "learning_rate": 3.2865000000000002e-06, "loss": 2.2042, "step": 6573 }, { "epoch": 2.0608150470219435, "grad_norm": 17.33954429626465, "learning_rate": 3.287e-06, "loss": 2.2191, "step": 6574 }, { "epoch": 2.061128526645768, "grad_norm": 10.20433235168457, "learning_rate": 3.2875000000000003e-06, "loss": 2.0205, "step": 6575 }, { "epoch": 2.0614420062695924, "grad_norm": 10.898722648620605, "learning_rate": 3.288e-06, "loss": 2.1752, "step": 6576 }, { "epoch": 2.061755485893417, "grad_norm": 10.12350082397461, "learning_rate": 3.2885e-06, "loss": 2.2114, "step": 6577 }, { "epoch": 2.0620689655172413, "grad_norm": 26.5042724609375, "learning_rate": 3.2890000000000006e-06, "loss": 2.0912, "step": 6578 }, { "epoch": 2.0623824451410657, "grad_norm": 27.845170974731445, "learning_rate": 3.2895000000000004e-06, "loss": 2.5305, "step": 6579 }, { "epoch": 2.06269592476489, "grad_norm": 11.100144386291504, "learning_rate": 3.2900000000000003e-06, "loss": 2.1362, "step": 6580 }, { "epoch": 2.0630094043887146, "grad_norm": 16.818344116210938, "learning_rate": 3.2905000000000005e-06, "loss": 2.107, "step": 6581 }, { "epoch": 2.063322884012539, "grad_norm": 11.070076942443848, "learning_rate": 3.2910000000000003e-06, "loss": 1.9589, "step": 6582 }, { "epoch": 2.0636363636363635, "grad_norm": 20.741596221923828, "learning_rate": 3.2915e-06, "loss": 2.4316, "step": 6583 }, { "epoch": 2.063949843260188, "grad_norm": 8.86921501159668, "learning_rate": 3.292e-06, "loss": 2.3643, "step": 6584 }, { "epoch": 2.0642633228840124, "grad_norm": 9.465672492980957, "learning_rate": 3.2925000000000002e-06, "loss": 2.1525, "step": 6585 }, { "epoch": 2.064576802507837, "grad_norm": 11.008056640625, "learning_rate": 3.293e-06, "loss": 2.4556, "step": 6586 }, { "epoch": 2.0648902821316613, "grad_norm": 52.70744323730469, "learning_rate": 3.2935e-06, "loss": 2.629, "step": 6587 }, { "epoch": 2.0652037617554857, "grad_norm": 11.790030479431152, "learning_rate": 3.2940000000000006e-06, "loss": 2.2775, "step": 6588 }, { "epoch": 2.06551724137931, "grad_norm": 11.963207244873047, "learning_rate": 3.2945000000000004e-06, "loss": 2.2884, "step": 6589 }, { "epoch": 2.0658307210031346, "grad_norm": 11.591779708862305, "learning_rate": 3.2950000000000002e-06, "loss": 2.0195, "step": 6590 }, { "epoch": 2.066144200626959, "grad_norm": 80.11792755126953, "learning_rate": 3.2955000000000005e-06, "loss": 2.0462, "step": 6591 }, { "epoch": 2.066457680250784, "grad_norm": 7.185821056365967, "learning_rate": 3.2960000000000003e-06, "loss": 2.1338, "step": 6592 }, { "epoch": 2.0667711598746084, "grad_norm": 61.9261474609375, "learning_rate": 3.2965e-06, "loss": 2.3051, "step": 6593 }, { "epoch": 2.067084639498433, "grad_norm": 9.620954513549805, "learning_rate": 3.2970000000000004e-06, "loss": 2.1151, "step": 6594 }, { "epoch": 2.0673981191222572, "grad_norm": 72.38018035888672, "learning_rate": 3.2975e-06, "loss": 2.5258, "step": 6595 }, { "epoch": 2.0677115987460817, "grad_norm": 13.933908462524414, "learning_rate": 3.298e-06, "loss": 2.1151, "step": 6596 }, { "epoch": 2.068025078369906, "grad_norm": 13.310007095336914, "learning_rate": 3.2985e-06, "loss": 2.2821, "step": 6597 }, { "epoch": 2.0683385579937306, "grad_norm": 17.456371307373047, "learning_rate": 3.2990000000000005e-06, "loss": 2.2428, "step": 6598 }, { "epoch": 2.068652037617555, "grad_norm": 52.41962814331055, "learning_rate": 3.2995000000000003e-06, "loss": 2.8051, "step": 6599 }, { "epoch": 2.0689655172413794, "grad_norm": 17.86503791809082, "learning_rate": 3.3000000000000006e-06, "loss": 2.3175, "step": 6600 }, { "epoch": 2.069278996865204, "grad_norm": 10.019948959350586, "learning_rate": 3.3005000000000004e-06, "loss": 2.2459, "step": 6601 }, { "epoch": 2.0695924764890283, "grad_norm": 58.896514892578125, "learning_rate": 3.3010000000000002e-06, "loss": 2.5401, "step": 6602 }, { "epoch": 2.0699059561128528, "grad_norm": 23.040035247802734, "learning_rate": 3.3015e-06, "loss": 2.4952, "step": 6603 }, { "epoch": 2.070219435736677, "grad_norm": 16.154142379760742, "learning_rate": 3.3020000000000003e-06, "loss": 2.1321, "step": 6604 }, { "epoch": 2.0705329153605017, "grad_norm": 31.047504425048828, "learning_rate": 3.3025e-06, "loss": 2.9739, "step": 6605 }, { "epoch": 2.070846394984326, "grad_norm": 19.299121856689453, "learning_rate": 3.303e-06, "loss": 2.2749, "step": 6606 }, { "epoch": 2.0711598746081505, "grad_norm": 18.930706024169922, "learning_rate": 3.3035000000000002e-06, "loss": 2.337, "step": 6607 }, { "epoch": 2.071473354231975, "grad_norm": 13.805002212524414, "learning_rate": 3.3040000000000005e-06, "loss": 2.0813, "step": 6608 }, { "epoch": 2.0717868338557994, "grad_norm": 9.26939868927002, "learning_rate": 3.3045000000000003e-06, "loss": 2.2402, "step": 6609 }, { "epoch": 2.072100313479624, "grad_norm": 18.309066772460938, "learning_rate": 3.3050000000000005e-06, "loss": 2.6603, "step": 6610 }, { "epoch": 2.0724137931034483, "grad_norm": 98.1492919921875, "learning_rate": 3.3055000000000004e-06, "loss": 2.2323, "step": 6611 }, { "epoch": 2.0727272727272728, "grad_norm": 20.337392807006836, "learning_rate": 3.306e-06, "loss": 2.0218, "step": 6612 }, { "epoch": 2.073040752351097, "grad_norm": 14.744490623474121, "learning_rate": 3.3065000000000004e-06, "loss": 2.0046, "step": 6613 }, { "epoch": 2.0733542319749216, "grad_norm": 12.826772689819336, "learning_rate": 3.3070000000000003e-06, "loss": 2.7334, "step": 6614 }, { "epoch": 2.073667711598746, "grad_norm": 14.731560707092285, "learning_rate": 3.3075e-06, "loss": 2.179, "step": 6615 }, { "epoch": 2.0739811912225705, "grad_norm": 12.944982528686523, "learning_rate": 3.308e-06, "loss": 1.9857, "step": 6616 }, { "epoch": 2.074294670846395, "grad_norm": 8.561573028564453, "learning_rate": 3.3085e-06, "loss": 2.1118, "step": 6617 }, { "epoch": 2.0746081504702194, "grad_norm": 20.174453735351562, "learning_rate": 3.3090000000000004e-06, "loss": 2.2463, "step": 6618 }, { "epoch": 2.074921630094044, "grad_norm": 28.328346252441406, "learning_rate": 3.3095000000000007e-06, "loss": 2.0438, "step": 6619 }, { "epoch": 2.0752351097178683, "grad_norm": 14.029821395874023, "learning_rate": 3.3100000000000005e-06, "loss": 2.0556, "step": 6620 }, { "epoch": 2.0755485893416927, "grad_norm": 42.446876525878906, "learning_rate": 3.3105000000000003e-06, "loss": 2.3173, "step": 6621 }, { "epoch": 2.075862068965517, "grad_norm": 7.9354634284973145, "learning_rate": 3.311e-06, "loss": 2.4152, "step": 6622 }, { "epoch": 2.0761755485893416, "grad_norm": 13.0211181640625, "learning_rate": 3.3115000000000004e-06, "loss": 2.146, "step": 6623 }, { "epoch": 2.076489028213166, "grad_norm": 9.520026206970215, "learning_rate": 3.3120000000000002e-06, "loss": 2.0236, "step": 6624 }, { "epoch": 2.0768025078369905, "grad_norm": 18.808061599731445, "learning_rate": 3.3125e-06, "loss": 2.0592, "step": 6625 }, { "epoch": 2.077115987460815, "grad_norm": 14.794133186340332, "learning_rate": 3.3130000000000003e-06, "loss": 2.2309, "step": 6626 }, { "epoch": 2.0774294670846394, "grad_norm": 38.48018264770508, "learning_rate": 3.3135e-06, "loss": 2.1268, "step": 6627 }, { "epoch": 2.077742946708464, "grad_norm": 31.94816017150879, "learning_rate": 3.314e-06, "loss": 2.3691, "step": 6628 }, { "epoch": 2.0780564263322883, "grad_norm": 13.080465316772461, "learning_rate": 3.3145000000000006e-06, "loss": 2.2045, "step": 6629 }, { "epoch": 2.0783699059561127, "grad_norm": 13.168771743774414, "learning_rate": 3.3150000000000004e-06, "loss": 2.2219, "step": 6630 }, { "epoch": 2.078683385579937, "grad_norm": 18.66151237487793, "learning_rate": 3.3155000000000003e-06, "loss": 2.3957, "step": 6631 }, { "epoch": 2.0789968652037616, "grad_norm": 17.034591674804688, "learning_rate": 3.3160000000000005e-06, "loss": 2.4903, "step": 6632 }, { "epoch": 2.079310344827586, "grad_norm": 13.67409610748291, "learning_rate": 3.3165000000000003e-06, "loss": 2.267, "step": 6633 }, { "epoch": 2.0796238244514105, "grad_norm": 38.133419036865234, "learning_rate": 3.317e-06, "loss": 2.8496, "step": 6634 }, { "epoch": 2.079937304075235, "grad_norm": 16.589134216308594, "learning_rate": 3.3175e-06, "loss": 2.3411, "step": 6635 }, { "epoch": 2.0802507836990594, "grad_norm": 22.424087524414062, "learning_rate": 3.3180000000000003e-06, "loss": 2.1087, "step": 6636 }, { "epoch": 2.080564263322884, "grad_norm": 13.403560638427734, "learning_rate": 3.3185e-06, "loss": 2.202, "step": 6637 }, { "epoch": 2.0808777429467087, "grad_norm": 9.902822494506836, "learning_rate": 3.319e-06, "loss": 2.131, "step": 6638 }, { "epoch": 2.081191222570533, "grad_norm": 10.626757621765137, "learning_rate": 3.3195000000000006e-06, "loss": 2.5102, "step": 6639 }, { "epoch": 2.0815047021943576, "grad_norm": 9.363275527954102, "learning_rate": 3.3200000000000004e-06, "loss": 2.1632, "step": 6640 }, { "epoch": 2.081818181818182, "grad_norm": 6.962289333343506, "learning_rate": 3.3205000000000002e-06, "loss": 2.5366, "step": 6641 }, { "epoch": 2.0821316614420065, "grad_norm": 17.720054626464844, "learning_rate": 3.3210000000000005e-06, "loss": 2.1066, "step": 6642 }, { "epoch": 2.082445141065831, "grad_norm": 10.18647289276123, "learning_rate": 3.3215000000000003e-06, "loss": 2.1429, "step": 6643 }, { "epoch": 2.0827586206896553, "grad_norm": 18.66595458984375, "learning_rate": 3.322e-06, "loss": 2.3423, "step": 6644 }, { "epoch": 2.08307210031348, "grad_norm": 13.37343692779541, "learning_rate": 3.3225000000000004e-06, "loss": 2.2585, "step": 6645 }, { "epoch": 2.083385579937304, "grad_norm": 61.10220718383789, "learning_rate": 3.323e-06, "loss": 2.4428, "step": 6646 }, { "epoch": 2.0836990595611287, "grad_norm": 16.537601470947266, "learning_rate": 3.3235e-06, "loss": 2.2715, "step": 6647 }, { "epoch": 2.084012539184953, "grad_norm": 6.8676886558532715, "learning_rate": 3.324e-06, "loss": 2.0942, "step": 6648 }, { "epoch": 2.0843260188087775, "grad_norm": 9.526947021484375, "learning_rate": 3.3245000000000005e-06, "loss": 2.2249, "step": 6649 }, { "epoch": 2.084639498432602, "grad_norm": 13.596402168273926, "learning_rate": 3.3250000000000004e-06, "loss": 2.2283, "step": 6650 }, { "epoch": 2.0849529780564264, "grad_norm": 15.568804740905762, "learning_rate": 3.3255000000000006e-06, "loss": 2.3229, "step": 6651 }, { "epoch": 2.085266457680251, "grad_norm": 15.836637496948242, "learning_rate": 3.3260000000000004e-06, "loss": 2.9797, "step": 6652 }, { "epoch": 2.0855799373040753, "grad_norm": 10.93256664276123, "learning_rate": 3.3265000000000003e-06, "loss": 2.3805, "step": 6653 }, { "epoch": 2.0858934169278998, "grad_norm": 9.744269371032715, "learning_rate": 3.327e-06, "loss": 2.1817, "step": 6654 }, { "epoch": 2.086206896551724, "grad_norm": 17.95917510986328, "learning_rate": 3.3275000000000003e-06, "loss": 2.2258, "step": 6655 }, { "epoch": 2.0865203761755486, "grad_norm": 12.277441024780273, "learning_rate": 3.328e-06, "loss": 2.5802, "step": 6656 }, { "epoch": 2.086833855799373, "grad_norm": 20.651268005371094, "learning_rate": 3.3285e-06, "loss": 2.2499, "step": 6657 }, { "epoch": 2.0871473354231975, "grad_norm": 14.935371398925781, "learning_rate": 3.329e-06, "loss": 2.7205, "step": 6658 }, { "epoch": 2.087460815047022, "grad_norm": 15.045103073120117, "learning_rate": 3.3295000000000005e-06, "loss": 2.1623, "step": 6659 }, { "epoch": 2.0877742946708464, "grad_norm": 10.799389839172363, "learning_rate": 3.3300000000000003e-06, "loss": 2.3534, "step": 6660 }, { "epoch": 2.088087774294671, "grad_norm": 14.774282455444336, "learning_rate": 3.3305000000000006e-06, "loss": 2.1447, "step": 6661 }, { "epoch": 2.0884012539184953, "grad_norm": 17.469953536987305, "learning_rate": 3.3310000000000004e-06, "loss": 2.3119, "step": 6662 }, { "epoch": 2.0887147335423197, "grad_norm": 17.010204315185547, "learning_rate": 3.3315e-06, "loss": 2.2757, "step": 6663 }, { "epoch": 2.089028213166144, "grad_norm": 34.67298126220703, "learning_rate": 3.332e-06, "loss": 2.1586, "step": 6664 }, { "epoch": 2.0893416927899686, "grad_norm": 51.78559112548828, "learning_rate": 3.3325000000000003e-06, "loss": 2.4675, "step": 6665 }, { "epoch": 2.089655172413793, "grad_norm": 18.196996688842773, "learning_rate": 3.333e-06, "loss": 2.3452, "step": 6666 }, { "epoch": 2.0899686520376175, "grad_norm": 32.48048782348633, "learning_rate": 3.3335e-06, "loss": 1.957, "step": 6667 }, { "epoch": 2.090282131661442, "grad_norm": 17.36007308959961, "learning_rate": 3.334e-06, "loss": 2.5579, "step": 6668 }, { "epoch": 2.0905956112852664, "grad_norm": 36.04092025756836, "learning_rate": 3.3345000000000004e-06, "loss": 2.6253, "step": 6669 }, { "epoch": 2.090909090909091, "grad_norm": 24.759357452392578, "learning_rate": 3.3350000000000003e-06, "loss": 2.4658, "step": 6670 }, { "epoch": 2.0912225705329153, "grad_norm": 8.661863327026367, "learning_rate": 3.3355000000000005e-06, "loss": 2.1097, "step": 6671 }, { "epoch": 2.0915360501567397, "grad_norm": 43.173912048339844, "learning_rate": 3.3360000000000003e-06, "loss": 2.6263, "step": 6672 }, { "epoch": 2.091849529780564, "grad_norm": 12.342394828796387, "learning_rate": 3.3365e-06, "loss": 2.0878, "step": 6673 }, { "epoch": 2.0921630094043886, "grad_norm": 10.80349349975586, "learning_rate": 3.3370000000000004e-06, "loss": 2.2655, "step": 6674 }, { "epoch": 2.092476489028213, "grad_norm": 11.250377655029297, "learning_rate": 3.3375000000000002e-06, "loss": 2.332, "step": 6675 }, { "epoch": 2.0927899686520375, "grad_norm": 13.764281272888184, "learning_rate": 3.338e-06, "loss": 2.0459, "step": 6676 }, { "epoch": 2.093103448275862, "grad_norm": 15.49951457977295, "learning_rate": 3.3385e-06, "loss": 2.3769, "step": 6677 }, { "epoch": 2.0934169278996864, "grad_norm": 12.976837158203125, "learning_rate": 3.339e-06, "loss": 2.4025, "step": 6678 }, { "epoch": 2.093730407523511, "grad_norm": 16.374460220336914, "learning_rate": 3.3395000000000004e-06, "loss": 2.2005, "step": 6679 }, { "epoch": 2.0940438871473352, "grad_norm": 10.695979118347168, "learning_rate": 3.3400000000000006e-06, "loss": 2.1278, "step": 6680 }, { "epoch": 2.0943573667711597, "grad_norm": 9.359561920166016, "learning_rate": 3.3405000000000005e-06, "loss": 1.9865, "step": 6681 }, { "epoch": 2.094670846394984, "grad_norm": 18.71257972717285, "learning_rate": 3.3410000000000003e-06, "loss": 2.3642, "step": 6682 }, { "epoch": 2.0949843260188086, "grad_norm": 40.491127014160156, "learning_rate": 3.3415e-06, "loss": 2.4719, "step": 6683 }, { "epoch": 2.095297805642633, "grad_norm": 23.519136428833008, "learning_rate": 3.3420000000000004e-06, "loss": 2.1451, "step": 6684 }, { "epoch": 2.0956112852664575, "grad_norm": 7.2523298263549805, "learning_rate": 3.3425e-06, "loss": 2.4691, "step": 6685 }, { "epoch": 2.0959247648902823, "grad_norm": 18.274356842041016, "learning_rate": 3.343e-06, "loss": 2.4856, "step": 6686 }, { "epoch": 2.096238244514107, "grad_norm": 12.820414543151855, "learning_rate": 3.3435000000000003e-06, "loss": 2.0316, "step": 6687 }, { "epoch": 2.0965517241379312, "grad_norm": 14.49550724029541, "learning_rate": 3.344e-06, "loss": 2.4053, "step": 6688 }, { "epoch": 2.0968652037617557, "grad_norm": 34.82584762573242, "learning_rate": 3.3445000000000003e-06, "loss": 2.3544, "step": 6689 }, { "epoch": 2.09717868338558, "grad_norm": 14.53974437713623, "learning_rate": 3.3450000000000006e-06, "loss": 2.1333, "step": 6690 }, { "epoch": 2.0974921630094046, "grad_norm": 10.810935020446777, "learning_rate": 3.3455000000000004e-06, "loss": 2.1926, "step": 6691 }, { "epoch": 2.097805642633229, "grad_norm": 18.585914611816406, "learning_rate": 3.3460000000000002e-06, "loss": 2.3641, "step": 6692 }, { "epoch": 2.0981191222570534, "grad_norm": 60.79994583129883, "learning_rate": 3.3465000000000005e-06, "loss": 3.1442, "step": 6693 }, { "epoch": 2.098432601880878, "grad_norm": 8.86579704284668, "learning_rate": 3.3470000000000003e-06, "loss": 2.2946, "step": 6694 }, { "epoch": 2.0987460815047023, "grad_norm": 16.35152244567871, "learning_rate": 3.3475e-06, "loss": 2.2358, "step": 6695 }, { "epoch": 2.0990595611285268, "grad_norm": 22.779769897460938, "learning_rate": 3.348e-06, "loss": 2.4666, "step": 6696 }, { "epoch": 2.099373040752351, "grad_norm": 406.09893798828125, "learning_rate": 3.3485000000000002e-06, "loss": 2.2806, "step": 6697 }, { "epoch": 2.0996865203761756, "grad_norm": 14.422026634216309, "learning_rate": 3.349e-06, "loss": 2.2797, "step": 6698 }, { "epoch": 2.1, "grad_norm": 13.153593063354492, "learning_rate": 3.3495000000000007e-06, "loss": 2.4603, "step": 6699 }, { "epoch": 2.1003134796238245, "grad_norm": 8.306694030761719, "learning_rate": 3.3500000000000005e-06, "loss": 2.0856, "step": 6700 }, { "epoch": 2.100626959247649, "grad_norm": 9.133256912231445, "learning_rate": 3.3505000000000004e-06, "loss": 2.164, "step": 6701 }, { "epoch": 2.1009404388714734, "grad_norm": 37.87324523925781, "learning_rate": 3.351e-06, "loss": 2.5106, "step": 6702 }, { "epoch": 2.101253918495298, "grad_norm": 29.12238883972168, "learning_rate": 3.3515000000000004e-06, "loss": 2.1778, "step": 6703 }, { "epoch": 2.1015673981191223, "grad_norm": 12.899033546447754, "learning_rate": 3.3520000000000003e-06, "loss": 2.3426, "step": 6704 }, { "epoch": 2.1018808777429467, "grad_norm": 10.672435760498047, "learning_rate": 3.3525e-06, "loss": 2.5135, "step": 6705 }, { "epoch": 2.102194357366771, "grad_norm": 34.982059478759766, "learning_rate": 3.3530000000000003e-06, "loss": 2.8666, "step": 6706 }, { "epoch": 2.1025078369905956, "grad_norm": 12.343427658081055, "learning_rate": 3.3535e-06, "loss": 2.0481, "step": 6707 }, { "epoch": 2.10282131661442, "grad_norm": 12.975926399230957, "learning_rate": 3.354e-06, "loss": 1.8967, "step": 6708 }, { "epoch": 2.1031347962382445, "grad_norm": 11.27238941192627, "learning_rate": 3.3545000000000007e-06, "loss": 2.3304, "step": 6709 }, { "epoch": 2.103448275862069, "grad_norm": 57.91295623779297, "learning_rate": 3.3550000000000005e-06, "loss": 2.5935, "step": 6710 }, { "epoch": 2.1037617554858934, "grad_norm": 7.352316379547119, "learning_rate": 3.3555000000000003e-06, "loss": 2.2094, "step": 6711 }, { "epoch": 2.104075235109718, "grad_norm": 13.881854057312012, "learning_rate": 3.3560000000000006e-06, "loss": 2.11, "step": 6712 }, { "epoch": 2.1043887147335423, "grad_norm": 13.67516040802002, "learning_rate": 3.3565000000000004e-06, "loss": 2.3765, "step": 6713 }, { "epoch": 2.1047021943573667, "grad_norm": 22.974998474121094, "learning_rate": 3.3570000000000002e-06, "loss": 2.3384, "step": 6714 }, { "epoch": 2.105015673981191, "grad_norm": 55.85123825073242, "learning_rate": 3.3575e-06, "loss": 3.2571, "step": 6715 }, { "epoch": 2.1053291536050156, "grad_norm": 55.123199462890625, "learning_rate": 3.3580000000000003e-06, "loss": 2.9427, "step": 6716 }, { "epoch": 2.10564263322884, "grad_norm": 17.60451316833496, "learning_rate": 3.3585e-06, "loss": 2.2967, "step": 6717 }, { "epoch": 2.1059561128526645, "grad_norm": 33.52358627319336, "learning_rate": 3.359e-06, "loss": 2.5449, "step": 6718 }, { "epoch": 2.106269592476489, "grad_norm": 8.550735473632812, "learning_rate": 3.3595000000000006e-06, "loss": 2.4668, "step": 6719 }, { "epoch": 2.1065830721003134, "grad_norm": 12.040781021118164, "learning_rate": 3.3600000000000004e-06, "loss": 2.2412, "step": 6720 }, { "epoch": 2.106896551724138, "grad_norm": 16.591629028320312, "learning_rate": 3.3605000000000003e-06, "loss": 2.2766, "step": 6721 }, { "epoch": 2.1072100313479623, "grad_norm": 13.391143798828125, "learning_rate": 3.3610000000000005e-06, "loss": 2.3731, "step": 6722 }, { "epoch": 2.1075235109717867, "grad_norm": 15.255084991455078, "learning_rate": 3.3615000000000003e-06, "loss": 2.9471, "step": 6723 }, { "epoch": 2.107836990595611, "grad_norm": 15.728415489196777, "learning_rate": 3.362e-06, "loss": 2.3029, "step": 6724 }, { "epoch": 2.1081504702194356, "grad_norm": 23.847612380981445, "learning_rate": 3.3625000000000004e-06, "loss": 2.1601, "step": 6725 }, { "epoch": 2.10846394984326, "grad_norm": 17.340152740478516, "learning_rate": 3.3630000000000002e-06, "loss": 2.5864, "step": 6726 }, { "epoch": 2.1087774294670845, "grad_norm": 28.97724151611328, "learning_rate": 3.3635e-06, "loss": 2.4464, "step": 6727 }, { "epoch": 2.109090909090909, "grad_norm": 35.46091079711914, "learning_rate": 3.364e-06, "loss": 2.4796, "step": 6728 }, { "epoch": 2.1094043887147333, "grad_norm": 36.43714904785156, "learning_rate": 3.3645000000000006e-06, "loss": 2.2184, "step": 6729 }, { "epoch": 2.109717868338558, "grad_norm": 16.32804298400879, "learning_rate": 3.3650000000000004e-06, "loss": 2.5341, "step": 6730 }, { "epoch": 2.1100313479623822, "grad_norm": 10.126823425292969, "learning_rate": 3.3655000000000006e-06, "loss": 2.3222, "step": 6731 }, { "epoch": 2.110344827586207, "grad_norm": 19.259586334228516, "learning_rate": 3.3660000000000005e-06, "loss": 2.5427, "step": 6732 }, { "epoch": 2.1106583072100316, "grad_norm": 11.224645614624023, "learning_rate": 3.3665000000000003e-06, "loss": 2.2684, "step": 6733 }, { "epoch": 2.110971786833856, "grad_norm": 10.714729309082031, "learning_rate": 3.367e-06, "loss": 2.1432, "step": 6734 }, { "epoch": 2.1112852664576804, "grad_norm": 12.961616516113281, "learning_rate": 3.3675000000000004e-06, "loss": 2.2364, "step": 6735 }, { "epoch": 2.111598746081505, "grad_norm": 11.951560020446777, "learning_rate": 3.368e-06, "loss": 2.3489, "step": 6736 }, { "epoch": 2.1119122257053293, "grad_norm": 13.636985778808594, "learning_rate": 3.3685e-06, "loss": 2.1841, "step": 6737 }, { "epoch": 2.1122257053291538, "grad_norm": 14.269251823425293, "learning_rate": 3.369e-06, "loss": 2.074, "step": 6738 }, { "epoch": 2.112539184952978, "grad_norm": 103.82185363769531, "learning_rate": 3.3695000000000005e-06, "loss": 2.5077, "step": 6739 }, { "epoch": 2.1128526645768027, "grad_norm": 10.76889705657959, "learning_rate": 3.3700000000000003e-06, "loss": 2.1855, "step": 6740 }, { "epoch": 2.113166144200627, "grad_norm": 14.039129257202148, "learning_rate": 3.3705000000000006e-06, "loss": 2.2568, "step": 6741 }, { "epoch": 2.1134796238244515, "grad_norm": 11.721952438354492, "learning_rate": 3.3710000000000004e-06, "loss": 2.2501, "step": 6742 }, { "epoch": 2.113793103448276, "grad_norm": 15.525774955749512, "learning_rate": 3.3715000000000002e-06, "loss": 2.3719, "step": 6743 }, { "epoch": 2.1141065830721004, "grad_norm": 61.047393798828125, "learning_rate": 3.372e-06, "loss": 3.2326, "step": 6744 }, { "epoch": 2.114420062695925, "grad_norm": 10.93510627746582, "learning_rate": 3.3725000000000003e-06, "loss": 2.1868, "step": 6745 }, { "epoch": 2.1147335423197493, "grad_norm": 48.77324295043945, "learning_rate": 3.373e-06, "loss": 2.4688, "step": 6746 }, { "epoch": 2.1150470219435737, "grad_norm": 8.839903831481934, "learning_rate": 3.3735e-06, "loss": 2.3125, "step": 6747 }, { "epoch": 2.115360501567398, "grad_norm": 41.76405334472656, "learning_rate": 3.3740000000000002e-06, "loss": 2.2516, "step": 6748 }, { "epoch": 2.1156739811912226, "grad_norm": 11.765251159667969, "learning_rate": 3.3745000000000005e-06, "loss": 2.1609, "step": 6749 }, { "epoch": 2.115987460815047, "grad_norm": 17.074146270751953, "learning_rate": 3.3750000000000003e-06, "loss": 2.0306, "step": 6750 }, { "epoch": 2.1163009404388715, "grad_norm": 17.70394515991211, "learning_rate": 3.3755000000000005e-06, "loss": 2.1093, "step": 6751 }, { "epoch": 2.116614420062696, "grad_norm": 11.510361671447754, "learning_rate": 3.3760000000000004e-06, "loss": 2.0907, "step": 6752 }, { "epoch": 2.1169278996865204, "grad_norm": 21.996944427490234, "learning_rate": 3.3765e-06, "loss": 2.1179, "step": 6753 }, { "epoch": 2.117241379310345, "grad_norm": 12.12459659576416, "learning_rate": 3.3770000000000004e-06, "loss": 2.3918, "step": 6754 }, { "epoch": 2.1175548589341693, "grad_norm": 219.8904266357422, "learning_rate": 3.3775000000000003e-06, "loss": 2.4341, "step": 6755 }, { "epoch": 2.1178683385579937, "grad_norm": 42.828243255615234, "learning_rate": 3.378e-06, "loss": 2.4604, "step": 6756 }, { "epoch": 2.118181818181818, "grad_norm": 8.967055320739746, "learning_rate": 3.3785e-06, "loss": 2.2489, "step": 6757 }, { "epoch": 2.1184952978056426, "grad_norm": 11.2413969039917, "learning_rate": 3.379e-06, "loss": 2.2811, "step": 6758 }, { "epoch": 2.118808777429467, "grad_norm": 42.26744079589844, "learning_rate": 3.3795e-06, "loss": 2.6732, "step": 6759 }, { "epoch": 2.1191222570532915, "grad_norm": 34.583866119384766, "learning_rate": 3.3800000000000007e-06, "loss": 2.3893, "step": 6760 }, { "epoch": 2.119435736677116, "grad_norm": 88.51214599609375, "learning_rate": 3.3805000000000005e-06, "loss": 2.1316, "step": 6761 }, { "epoch": 2.1197492163009404, "grad_norm": 10.016351699829102, "learning_rate": 3.3810000000000003e-06, "loss": 2.1159, "step": 6762 }, { "epoch": 2.120062695924765, "grad_norm": 23.770122528076172, "learning_rate": 3.3815e-06, "loss": 2.1346, "step": 6763 }, { "epoch": 2.1203761755485893, "grad_norm": 12.543835639953613, "learning_rate": 3.3820000000000004e-06, "loss": 2.4389, "step": 6764 }, { "epoch": 2.1206896551724137, "grad_norm": 14.241379737854004, "learning_rate": 3.3825000000000002e-06, "loss": 2.3911, "step": 6765 }, { "epoch": 2.121003134796238, "grad_norm": 45.56835174560547, "learning_rate": 3.383e-06, "loss": 2.5627, "step": 6766 }, { "epoch": 2.1213166144200626, "grad_norm": 8.86864948272705, "learning_rate": 3.3835000000000003e-06, "loss": 2.1581, "step": 6767 }, { "epoch": 2.121630094043887, "grad_norm": 8.486713409423828, "learning_rate": 3.384e-06, "loss": 2.2056, "step": 6768 }, { "epoch": 2.1219435736677115, "grad_norm": 11.814230918884277, "learning_rate": 3.3845e-06, "loss": 2.2301, "step": 6769 }, { "epoch": 2.122257053291536, "grad_norm": 8.265667915344238, "learning_rate": 3.3850000000000006e-06, "loss": 2.1969, "step": 6770 }, { "epoch": 2.1225705329153604, "grad_norm": 9.525060653686523, "learning_rate": 3.3855000000000005e-06, "loss": 2.0402, "step": 6771 }, { "epoch": 2.122884012539185, "grad_norm": 28.402851104736328, "learning_rate": 3.3860000000000003e-06, "loss": 2.7888, "step": 6772 }, { "epoch": 2.1231974921630092, "grad_norm": 56.51530456542969, "learning_rate": 3.3865000000000005e-06, "loss": 2.7277, "step": 6773 }, { "epoch": 2.1235109717868337, "grad_norm": 15.09781551361084, "learning_rate": 3.3870000000000004e-06, "loss": 2.3442, "step": 6774 }, { "epoch": 2.123824451410658, "grad_norm": 20.094221115112305, "learning_rate": 3.3875e-06, "loss": 2.9136, "step": 6775 }, { "epoch": 2.1241379310344826, "grad_norm": 12.819809913635254, "learning_rate": 3.388e-06, "loss": 2.6969, "step": 6776 }, { "epoch": 2.124451410658307, "grad_norm": 47.905799865722656, "learning_rate": 3.3885000000000003e-06, "loss": 1.9744, "step": 6777 }, { "epoch": 2.1247648902821314, "grad_norm": 10.274794578552246, "learning_rate": 3.389e-06, "loss": 2.1343, "step": 6778 }, { "epoch": 2.125078369905956, "grad_norm": 17.00676918029785, "learning_rate": 3.3895e-06, "loss": 2.2527, "step": 6779 }, { "epoch": 2.1253918495297803, "grad_norm": 12.76203727722168, "learning_rate": 3.3900000000000006e-06, "loss": 2.0416, "step": 6780 }, { "epoch": 2.125705329153605, "grad_norm": 19.18592071533203, "learning_rate": 3.3905000000000004e-06, "loss": 1.9724, "step": 6781 }, { "epoch": 2.1260188087774297, "grad_norm": 8.213129043579102, "learning_rate": 3.3910000000000002e-06, "loss": 2.2069, "step": 6782 }, { "epoch": 2.126332288401254, "grad_norm": 14.306838989257812, "learning_rate": 3.3915000000000005e-06, "loss": 2.3, "step": 6783 }, { "epoch": 2.1266457680250785, "grad_norm": 20.52436637878418, "learning_rate": 3.3920000000000003e-06, "loss": 2.3876, "step": 6784 }, { "epoch": 2.126959247648903, "grad_norm": 15.298855781555176, "learning_rate": 3.3925e-06, "loss": 2.3102, "step": 6785 }, { "epoch": 2.1272727272727274, "grad_norm": 9.44373607635498, "learning_rate": 3.3930000000000004e-06, "loss": 2.0672, "step": 6786 }, { "epoch": 2.127586206896552, "grad_norm": 12.8731107711792, "learning_rate": 3.3935e-06, "loss": 2.7468, "step": 6787 }, { "epoch": 2.1278996865203763, "grad_norm": 9.178388595581055, "learning_rate": 3.394e-06, "loss": 2.2315, "step": 6788 }, { "epoch": 2.1282131661442008, "grad_norm": 14.290987014770508, "learning_rate": 3.3945e-06, "loss": 2.0909, "step": 6789 }, { "epoch": 2.128526645768025, "grad_norm": 17.735122680664062, "learning_rate": 3.3950000000000005e-06, "loss": 2.3217, "step": 6790 }, { "epoch": 2.1288401253918496, "grad_norm": 8.255292892456055, "learning_rate": 3.3955000000000004e-06, "loss": 2.1639, "step": 6791 }, { "epoch": 2.129153605015674, "grad_norm": 36.709346771240234, "learning_rate": 3.3960000000000006e-06, "loss": 2.4343, "step": 6792 }, { "epoch": 2.1294670846394985, "grad_norm": 11.496522903442383, "learning_rate": 3.3965000000000004e-06, "loss": 2.0047, "step": 6793 }, { "epoch": 2.129780564263323, "grad_norm": 28.03876304626465, "learning_rate": 3.3970000000000003e-06, "loss": 2.2701, "step": 6794 }, { "epoch": 2.1300940438871474, "grad_norm": 23.613258361816406, "learning_rate": 3.3975e-06, "loss": 2.1253, "step": 6795 }, { "epoch": 2.130407523510972, "grad_norm": 58.97169494628906, "learning_rate": 3.3980000000000003e-06, "loss": 3.3573, "step": 6796 }, { "epoch": 2.1307210031347963, "grad_norm": 14.830269813537598, "learning_rate": 3.3985e-06, "loss": 2.1736, "step": 6797 }, { "epoch": 2.1310344827586207, "grad_norm": 8.589932441711426, "learning_rate": 3.399e-06, "loss": 2.3833, "step": 6798 }, { "epoch": 2.131347962382445, "grad_norm": 21.1700382232666, "learning_rate": 3.3995000000000002e-06, "loss": 2.1513, "step": 6799 }, { "epoch": 2.1316614420062696, "grad_norm": 10.935470581054688, "learning_rate": 3.4000000000000005e-06, "loss": 2.3219, "step": 6800 }, { "epoch": 2.131974921630094, "grad_norm": 10.7236328125, "learning_rate": 3.4005000000000003e-06, "loss": 1.9308, "step": 6801 }, { "epoch": 2.1322884012539185, "grad_norm": 11.581204414367676, "learning_rate": 3.4010000000000006e-06, "loss": 1.8863, "step": 6802 }, { "epoch": 2.132601880877743, "grad_norm": 8.985054969787598, "learning_rate": 3.4015000000000004e-06, "loss": 2.3883, "step": 6803 }, { "epoch": 2.1329153605015674, "grad_norm": 69.11140441894531, "learning_rate": 3.402e-06, "loss": 2.5062, "step": 6804 }, { "epoch": 2.133228840125392, "grad_norm": 16.41836166381836, "learning_rate": 3.4025000000000005e-06, "loss": 2.1929, "step": 6805 }, { "epoch": 2.1335423197492163, "grad_norm": 12.03371524810791, "learning_rate": 3.4030000000000003e-06, "loss": 2.3241, "step": 6806 }, { "epoch": 2.1338557993730407, "grad_norm": 18.116392135620117, "learning_rate": 3.4035e-06, "loss": 2.7915, "step": 6807 }, { "epoch": 2.134169278996865, "grad_norm": 11.709860801696777, "learning_rate": 3.404e-06, "loss": 2.1731, "step": 6808 }, { "epoch": 2.1344827586206896, "grad_norm": 14.38310432434082, "learning_rate": 3.4045e-06, "loss": 2.0861, "step": 6809 }, { "epoch": 2.134796238244514, "grad_norm": 8.731871604919434, "learning_rate": 3.4050000000000004e-06, "loss": 2.1479, "step": 6810 }, { "epoch": 2.1351097178683385, "grad_norm": 11.81379508972168, "learning_rate": 3.4055000000000007e-06, "loss": 2.6132, "step": 6811 }, { "epoch": 2.135423197492163, "grad_norm": 31.87676429748535, "learning_rate": 3.4060000000000005e-06, "loss": 2.2791, "step": 6812 }, { "epoch": 2.1357366771159874, "grad_norm": 9.784947395324707, "learning_rate": 3.4065000000000003e-06, "loss": 2.0352, "step": 6813 }, { "epoch": 2.136050156739812, "grad_norm": 9.370558738708496, "learning_rate": 3.407e-06, "loss": 1.9851, "step": 6814 }, { "epoch": 2.1363636363636362, "grad_norm": 10.399176597595215, "learning_rate": 3.4075000000000004e-06, "loss": 2.2955, "step": 6815 }, { "epoch": 2.1366771159874607, "grad_norm": 26.81496238708496, "learning_rate": 3.4080000000000002e-06, "loss": 2.2285, "step": 6816 }, { "epoch": 2.136990595611285, "grad_norm": 8.112709999084473, "learning_rate": 3.4085e-06, "loss": 2.1929, "step": 6817 }, { "epoch": 2.1373040752351096, "grad_norm": 42.92530059814453, "learning_rate": 3.409e-06, "loss": 1.977, "step": 6818 }, { "epoch": 2.137617554858934, "grad_norm": 20.440536499023438, "learning_rate": 3.4095e-06, "loss": 2.1602, "step": 6819 }, { "epoch": 2.1379310344827585, "grad_norm": 15.455876350402832, "learning_rate": 3.4100000000000004e-06, "loss": 2.2529, "step": 6820 }, { "epoch": 2.138244514106583, "grad_norm": 12.394731521606445, "learning_rate": 3.4105000000000006e-06, "loss": 2.3339, "step": 6821 }, { "epoch": 2.1385579937304073, "grad_norm": 27.39763069152832, "learning_rate": 3.4110000000000005e-06, "loss": 2.4079, "step": 6822 }, { "epoch": 2.138871473354232, "grad_norm": 114.3840560913086, "learning_rate": 3.4115000000000003e-06, "loss": 3.0439, "step": 6823 }, { "epoch": 2.139184952978056, "grad_norm": 57.21638870239258, "learning_rate": 3.412e-06, "loss": 2.637, "step": 6824 }, { "epoch": 2.139498432601881, "grad_norm": 71.5346450805664, "learning_rate": 3.4125000000000004e-06, "loss": 2.305, "step": 6825 }, { "epoch": 2.1398119122257055, "grad_norm": 26.81305503845215, "learning_rate": 3.413e-06, "loss": 2.0555, "step": 6826 }, { "epoch": 2.14012539184953, "grad_norm": 10.337777137756348, "learning_rate": 3.4135e-06, "loss": 2.0905, "step": 6827 }, { "epoch": 2.1404388714733544, "grad_norm": 10.955779075622559, "learning_rate": 3.4140000000000003e-06, "loss": 2.2392, "step": 6828 }, { "epoch": 2.140752351097179, "grad_norm": 56.85873794555664, "learning_rate": 3.4145e-06, "loss": 2.3931, "step": 6829 }, { "epoch": 2.1410658307210033, "grad_norm": 16.437307357788086, "learning_rate": 3.4150000000000003e-06, "loss": 1.9202, "step": 6830 }, { "epoch": 2.1413793103448278, "grad_norm": 54.33417510986328, "learning_rate": 3.4155000000000006e-06, "loss": 2.3079, "step": 6831 }, { "epoch": 2.141692789968652, "grad_norm": 16.8037109375, "learning_rate": 3.4160000000000004e-06, "loss": 2.3048, "step": 6832 }, { "epoch": 2.1420062695924766, "grad_norm": 12.23515796661377, "learning_rate": 3.4165000000000002e-06, "loss": 2.1689, "step": 6833 }, { "epoch": 2.142319749216301, "grad_norm": 11.478507995605469, "learning_rate": 3.4170000000000005e-06, "loss": 2.2591, "step": 6834 }, { "epoch": 2.1426332288401255, "grad_norm": 10.895817756652832, "learning_rate": 3.4175000000000003e-06, "loss": 2.5474, "step": 6835 }, { "epoch": 2.14294670846395, "grad_norm": 35.13949966430664, "learning_rate": 3.418e-06, "loss": 2.7213, "step": 6836 }, { "epoch": 2.1432601880877744, "grad_norm": 18.97028350830078, "learning_rate": 3.4185e-06, "loss": 2.2668, "step": 6837 }, { "epoch": 2.143573667711599, "grad_norm": 10.949864387512207, "learning_rate": 3.4190000000000002e-06, "loss": 2.2087, "step": 6838 }, { "epoch": 2.1438871473354233, "grad_norm": 42.90413284301758, "learning_rate": 3.4195e-06, "loss": 2.2527, "step": 6839 }, { "epoch": 2.1442006269592477, "grad_norm": 23.613037109375, "learning_rate": 3.4200000000000007e-06, "loss": 2.1794, "step": 6840 }, { "epoch": 2.144514106583072, "grad_norm": 11.663117408752441, "learning_rate": 3.4205000000000005e-06, "loss": 2.1634, "step": 6841 }, { "epoch": 2.1448275862068966, "grad_norm": 23.462635040283203, "learning_rate": 3.4210000000000004e-06, "loss": 2.6146, "step": 6842 }, { "epoch": 2.145141065830721, "grad_norm": 23.625879287719727, "learning_rate": 3.4215e-06, "loss": 2.1591, "step": 6843 }, { "epoch": 2.1454545454545455, "grad_norm": 30.60991668701172, "learning_rate": 3.4220000000000004e-06, "loss": 2.7534, "step": 6844 }, { "epoch": 2.14576802507837, "grad_norm": 12.125800132751465, "learning_rate": 3.4225000000000003e-06, "loss": 2.0259, "step": 6845 }, { "epoch": 2.1460815047021944, "grad_norm": 22.787397384643555, "learning_rate": 3.423e-06, "loss": 2.2621, "step": 6846 }, { "epoch": 2.146394984326019, "grad_norm": 20.32537841796875, "learning_rate": 3.4235000000000003e-06, "loss": 2.0934, "step": 6847 }, { "epoch": 2.1467084639498433, "grad_norm": 9.994367599487305, "learning_rate": 3.424e-06, "loss": 2.2205, "step": 6848 }, { "epoch": 2.1470219435736677, "grad_norm": 11.022137641906738, "learning_rate": 3.4245e-06, "loss": 2.1032, "step": 6849 }, { "epoch": 2.147335423197492, "grad_norm": 24.02007484436035, "learning_rate": 3.4250000000000007e-06, "loss": 2.0436, "step": 6850 }, { "epoch": 2.1476489028213166, "grad_norm": 14.853202819824219, "learning_rate": 3.4255000000000005e-06, "loss": 2.2162, "step": 6851 }, { "epoch": 2.147962382445141, "grad_norm": 11.695042610168457, "learning_rate": 3.4260000000000003e-06, "loss": 2.1942, "step": 6852 }, { "epoch": 2.1482758620689655, "grad_norm": 13.188150405883789, "learning_rate": 3.4265000000000006e-06, "loss": 1.9549, "step": 6853 }, { "epoch": 2.14858934169279, "grad_norm": 28.15757179260254, "learning_rate": 3.4270000000000004e-06, "loss": 2.0309, "step": 6854 }, { "epoch": 2.1489028213166144, "grad_norm": 14.0144681930542, "learning_rate": 3.4275000000000002e-06, "loss": 2.2543, "step": 6855 }, { "epoch": 2.149216300940439, "grad_norm": 37.183753967285156, "learning_rate": 3.428e-06, "loss": 2.9988, "step": 6856 }, { "epoch": 2.1495297805642632, "grad_norm": 19.38912582397461, "learning_rate": 3.4285000000000003e-06, "loss": 2.1919, "step": 6857 }, { "epoch": 2.1498432601880877, "grad_norm": 25.731096267700195, "learning_rate": 3.429e-06, "loss": 2.2183, "step": 6858 }, { "epoch": 2.150156739811912, "grad_norm": 14.327131271362305, "learning_rate": 3.4295e-06, "loss": 2.0674, "step": 6859 }, { "epoch": 2.1504702194357366, "grad_norm": 15.840540885925293, "learning_rate": 3.4300000000000006e-06, "loss": 2.4225, "step": 6860 }, { "epoch": 2.150783699059561, "grad_norm": 11.5586576461792, "learning_rate": 3.4305000000000004e-06, "loss": 2.0981, "step": 6861 }, { "epoch": 2.1510971786833855, "grad_norm": 6.732176780700684, "learning_rate": 3.4310000000000003e-06, "loss": 2.1696, "step": 6862 }, { "epoch": 2.15141065830721, "grad_norm": 9.354598045349121, "learning_rate": 3.4315000000000005e-06, "loss": 2.2425, "step": 6863 }, { "epoch": 2.1517241379310343, "grad_norm": 27.88478660583496, "learning_rate": 3.4320000000000003e-06, "loss": 2.2616, "step": 6864 }, { "epoch": 2.152037617554859, "grad_norm": 10.490667343139648, "learning_rate": 3.4325e-06, "loss": 1.8864, "step": 6865 }, { "epoch": 2.1523510971786832, "grad_norm": 19.04070281982422, "learning_rate": 3.4330000000000004e-06, "loss": 2.1803, "step": 6866 }, { "epoch": 2.1526645768025077, "grad_norm": 14.13863754272461, "learning_rate": 3.4335000000000002e-06, "loss": 1.9035, "step": 6867 }, { "epoch": 2.152978056426332, "grad_norm": 16.341604232788086, "learning_rate": 3.434e-06, "loss": 2.1259, "step": 6868 }, { "epoch": 2.1532915360501566, "grad_norm": 10.861141204833984, "learning_rate": 3.4345e-06, "loss": 1.9963, "step": 6869 }, { "epoch": 2.153605015673981, "grad_norm": 15.914594650268555, "learning_rate": 3.4350000000000006e-06, "loss": 2.1514, "step": 6870 }, { "epoch": 2.1539184952978054, "grad_norm": 24.80222511291504, "learning_rate": 3.4355000000000004e-06, "loss": 2.0063, "step": 6871 }, { "epoch": 2.15423197492163, "grad_norm": 17.03633689880371, "learning_rate": 3.4360000000000006e-06, "loss": 1.87, "step": 6872 }, { "epoch": 2.1545454545454543, "grad_norm": 17.743072509765625, "learning_rate": 3.4365000000000005e-06, "loss": 2.2734, "step": 6873 }, { "epoch": 2.1548589341692788, "grad_norm": 13.45116901397705, "learning_rate": 3.4370000000000003e-06, "loss": 2.2056, "step": 6874 }, { "epoch": 2.1551724137931036, "grad_norm": 9.30422306060791, "learning_rate": 3.4375e-06, "loss": 2.1563, "step": 6875 }, { "epoch": 2.155485893416928, "grad_norm": 17.487060546875, "learning_rate": 3.4380000000000004e-06, "loss": 2.1787, "step": 6876 }, { "epoch": 2.1557993730407525, "grad_norm": 10.4694185256958, "learning_rate": 3.4385e-06, "loss": 2.1952, "step": 6877 }, { "epoch": 2.156112852664577, "grad_norm": 26.92039680480957, "learning_rate": 3.439e-06, "loss": 2.4277, "step": 6878 }, { "epoch": 2.1564263322884014, "grad_norm": 31.127328872680664, "learning_rate": 3.4395000000000003e-06, "loss": 2.3091, "step": 6879 }, { "epoch": 2.156739811912226, "grad_norm": 12.868185043334961, "learning_rate": 3.44e-06, "loss": 2.2072, "step": 6880 }, { "epoch": 2.1570532915360503, "grad_norm": 46.86790466308594, "learning_rate": 3.4405000000000003e-06, "loss": 2.6146, "step": 6881 }, { "epoch": 2.1573667711598747, "grad_norm": 9.953550338745117, "learning_rate": 3.4410000000000006e-06, "loss": 2.1342, "step": 6882 }, { "epoch": 2.157680250783699, "grad_norm": 8.743974685668945, "learning_rate": 3.4415000000000004e-06, "loss": 2.2264, "step": 6883 }, { "epoch": 2.1579937304075236, "grad_norm": 52.67411804199219, "learning_rate": 3.4420000000000002e-06, "loss": 2.4179, "step": 6884 }, { "epoch": 2.158307210031348, "grad_norm": 27.910085678100586, "learning_rate": 3.4425e-06, "loss": 2.015, "step": 6885 }, { "epoch": 2.1586206896551725, "grad_norm": 14.562686920166016, "learning_rate": 3.4430000000000003e-06, "loss": 2.4185, "step": 6886 }, { "epoch": 2.158934169278997, "grad_norm": 8.008809089660645, "learning_rate": 3.4435e-06, "loss": 2.2256, "step": 6887 }, { "epoch": 2.1592476489028214, "grad_norm": 11.626964569091797, "learning_rate": 3.444e-06, "loss": 2.1728, "step": 6888 }, { "epoch": 2.159561128526646, "grad_norm": 10.551536560058594, "learning_rate": 3.4445000000000002e-06, "loss": 2.0259, "step": 6889 }, { "epoch": 2.1598746081504703, "grad_norm": 80.46753692626953, "learning_rate": 3.445e-06, "loss": 2.4842, "step": 6890 }, { "epoch": 2.1601880877742947, "grad_norm": 12.3627290725708, "learning_rate": 3.4455000000000003e-06, "loss": 2.309, "step": 6891 }, { "epoch": 2.160501567398119, "grad_norm": 13.400481224060059, "learning_rate": 3.4460000000000005e-06, "loss": 2.2034, "step": 6892 }, { "epoch": 2.1608150470219436, "grad_norm": 10.977444648742676, "learning_rate": 3.4465000000000004e-06, "loss": 2.0528, "step": 6893 }, { "epoch": 2.161128526645768, "grad_norm": 30.91587257385254, "learning_rate": 3.447e-06, "loss": 2.1985, "step": 6894 }, { "epoch": 2.1614420062695925, "grad_norm": 12.722804069519043, "learning_rate": 3.4475000000000005e-06, "loss": 2.3122, "step": 6895 }, { "epoch": 2.161755485893417, "grad_norm": 27.230382919311523, "learning_rate": 3.4480000000000003e-06, "loss": 2.1554, "step": 6896 }, { "epoch": 2.1620689655172414, "grad_norm": 11.985507011413574, "learning_rate": 3.4485e-06, "loss": 2.3285, "step": 6897 }, { "epoch": 2.162382445141066, "grad_norm": 53.933021545410156, "learning_rate": 3.449e-06, "loss": 2.3472, "step": 6898 }, { "epoch": 2.1626959247648903, "grad_norm": 32.23548889160156, "learning_rate": 3.4495e-06, "loss": 2.1976, "step": 6899 }, { "epoch": 2.1630094043887147, "grad_norm": 9.508769035339355, "learning_rate": 3.45e-06, "loss": 2.1155, "step": 6900 }, { "epoch": 2.163322884012539, "grad_norm": 10.515083312988281, "learning_rate": 3.4505000000000007e-06, "loss": 2.1429, "step": 6901 }, { "epoch": 2.1636363636363636, "grad_norm": 30.78691864013672, "learning_rate": 3.4510000000000005e-06, "loss": 2.539, "step": 6902 }, { "epoch": 2.163949843260188, "grad_norm": 22.976768493652344, "learning_rate": 3.4515000000000003e-06, "loss": 2.7029, "step": 6903 }, { "epoch": 2.1642633228840125, "grad_norm": 10.03154468536377, "learning_rate": 3.452e-06, "loss": 2.3516, "step": 6904 }, { "epoch": 2.164576802507837, "grad_norm": 13.715048789978027, "learning_rate": 3.4525000000000004e-06, "loss": 2.1141, "step": 6905 }, { "epoch": 2.1648902821316613, "grad_norm": 9.703258514404297, "learning_rate": 3.4530000000000002e-06, "loss": 2.1848, "step": 6906 }, { "epoch": 2.165203761755486, "grad_norm": 9.807629585266113, "learning_rate": 3.4535e-06, "loss": 2.2151, "step": 6907 }, { "epoch": 2.1655172413793102, "grad_norm": 12.029234886169434, "learning_rate": 3.4540000000000003e-06, "loss": 2.0252, "step": 6908 }, { "epoch": 2.1658307210031347, "grad_norm": 18.455461502075195, "learning_rate": 3.4545e-06, "loss": 2.2914, "step": 6909 }, { "epoch": 2.166144200626959, "grad_norm": 11.100343704223633, "learning_rate": 3.455e-06, "loss": 2.2421, "step": 6910 }, { "epoch": 2.1664576802507836, "grad_norm": 7.689256191253662, "learning_rate": 3.4555000000000006e-06, "loss": 2.2564, "step": 6911 }, { "epoch": 2.166771159874608, "grad_norm": 15.277055740356445, "learning_rate": 3.4560000000000005e-06, "loss": 2.2835, "step": 6912 }, { "epoch": 2.1670846394984324, "grad_norm": 48.46110153198242, "learning_rate": 3.4565000000000003e-06, "loss": 2.1435, "step": 6913 }, { "epoch": 2.167398119122257, "grad_norm": 20.663454055786133, "learning_rate": 3.4570000000000005e-06, "loss": 2.5642, "step": 6914 }, { "epoch": 2.1677115987460813, "grad_norm": 13.22984504699707, "learning_rate": 3.4575000000000004e-06, "loss": 2.1452, "step": 6915 }, { "epoch": 2.1680250783699058, "grad_norm": 13.131516456604004, "learning_rate": 3.458e-06, "loss": 2.0397, "step": 6916 }, { "epoch": 2.16833855799373, "grad_norm": 28.660253524780273, "learning_rate": 3.4585e-06, "loss": 2.2928, "step": 6917 }, { "epoch": 2.1686520376175547, "grad_norm": 10.334028244018555, "learning_rate": 3.4590000000000003e-06, "loss": 2.1712, "step": 6918 }, { "epoch": 2.1689655172413795, "grad_norm": 10.435593605041504, "learning_rate": 3.4595e-06, "loss": 2.3303, "step": 6919 }, { "epoch": 2.169278996865204, "grad_norm": 14.95488452911377, "learning_rate": 3.46e-06, "loss": 2.0732, "step": 6920 }, { "epoch": 2.1695924764890284, "grad_norm": 72.69029235839844, "learning_rate": 3.4605000000000006e-06, "loss": 2.5956, "step": 6921 }, { "epoch": 2.169905956112853, "grad_norm": 15.435386657714844, "learning_rate": 3.4610000000000004e-06, "loss": 2.3033, "step": 6922 }, { "epoch": 2.1702194357366773, "grad_norm": 12.141288757324219, "learning_rate": 3.4615000000000002e-06, "loss": 2.1484, "step": 6923 }, { "epoch": 2.1705329153605017, "grad_norm": 9.972766876220703, "learning_rate": 3.4620000000000005e-06, "loss": 2.0766, "step": 6924 }, { "epoch": 2.170846394984326, "grad_norm": 11.411584854125977, "learning_rate": 3.4625000000000003e-06, "loss": 2.341, "step": 6925 }, { "epoch": 2.1711598746081506, "grad_norm": 16.873201370239258, "learning_rate": 3.463e-06, "loss": 2.4736, "step": 6926 }, { "epoch": 2.171473354231975, "grad_norm": 21.73773193359375, "learning_rate": 3.4635000000000004e-06, "loss": 2.1499, "step": 6927 }, { "epoch": 2.1717868338557995, "grad_norm": 22.63068199157715, "learning_rate": 3.464e-06, "loss": 2.6676, "step": 6928 }, { "epoch": 2.172100313479624, "grad_norm": 11.632658004760742, "learning_rate": 3.4645e-06, "loss": 2.1244, "step": 6929 }, { "epoch": 2.1724137931034484, "grad_norm": 9.442824363708496, "learning_rate": 3.465e-06, "loss": 1.9783, "step": 6930 }, { "epoch": 2.172727272727273, "grad_norm": 8.890788078308105, "learning_rate": 3.4655000000000005e-06, "loss": 2.1586, "step": 6931 }, { "epoch": 2.1730407523510973, "grad_norm": 32.60359191894531, "learning_rate": 3.4660000000000004e-06, "loss": 3.1686, "step": 6932 }, { "epoch": 2.1733542319749217, "grad_norm": 8.842353820800781, "learning_rate": 3.4665000000000006e-06, "loss": 2.1006, "step": 6933 }, { "epoch": 2.173667711598746, "grad_norm": 13.63758659362793, "learning_rate": 3.4670000000000004e-06, "loss": 2.5389, "step": 6934 }, { "epoch": 2.1739811912225706, "grad_norm": 10.475473403930664, "learning_rate": 3.4675000000000003e-06, "loss": 2.075, "step": 6935 }, { "epoch": 2.174294670846395, "grad_norm": 13.483911514282227, "learning_rate": 3.468e-06, "loss": 2.2035, "step": 6936 }, { "epoch": 2.1746081504702195, "grad_norm": 17.274625778198242, "learning_rate": 3.4685000000000003e-06, "loss": 2.2031, "step": 6937 }, { "epoch": 2.174921630094044, "grad_norm": 10.612244606018066, "learning_rate": 3.469e-06, "loss": 2.1475, "step": 6938 }, { "epoch": 2.1752351097178684, "grad_norm": 12.345919609069824, "learning_rate": 3.4695e-06, "loss": 2.81, "step": 6939 }, { "epoch": 2.175548589341693, "grad_norm": 21.124549865722656, "learning_rate": 3.4700000000000002e-06, "loss": 2.1106, "step": 6940 }, { "epoch": 2.1758620689655173, "grad_norm": 15.894444465637207, "learning_rate": 3.4705000000000005e-06, "loss": 2.1123, "step": 6941 }, { "epoch": 2.1761755485893417, "grad_norm": 18.159595489501953, "learning_rate": 3.4710000000000003e-06, "loss": 2.5442, "step": 6942 }, { "epoch": 2.176489028213166, "grad_norm": 15.514962196350098, "learning_rate": 3.4715000000000006e-06, "loss": 2.233, "step": 6943 }, { "epoch": 2.1768025078369906, "grad_norm": 15.110061645507812, "learning_rate": 3.4720000000000004e-06, "loss": 2.1864, "step": 6944 }, { "epoch": 2.177115987460815, "grad_norm": 55.93291091918945, "learning_rate": 3.4725e-06, "loss": 2.2317, "step": 6945 }, { "epoch": 2.1774294670846395, "grad_norm": 69.83547973632812, "learning_rate": 3.4730000000000005e-06, "loss": 2.2663, "step": 6946 }, { "epoch": 2.177742946708464, "grad_norm": 25.184389114379883, "learning_rate": 3.4735000000000003e-06, "loss": 2.0618, "step": 6947 }, { "epoch": 2.1780564263322884, "grad_norm": 11.246954917907715, "learning_rate": 3.474e-06, "loss": 2.0055, "step": 6948 }, { "epoch": 2.178369905956113, "grad_norm": 5.5993547439575195, "learning_rate": 3.4745e-06, "loss": 2.09, "step": 6949 }, { "epoch": 2.1786833855799372, "grad_norm": 39.72569274902344, "learning_rate": 3.475e-06, "loss": 2.1553, "step": 6950 }, { "epoch": 2.1789968652037617, "grad_norm": 17.756860733032227, "learning_rate": 3.4755000000000004e-06, "loss": 2.1759, "step": 6951 }, { "epoch": 2.179310344827586, "grad_norm": 23.598684310913086, "learning_rate": 3.4760000000000007e-06, "loss": 2.1523, "step": 6952 }, { "epoch": 2.1796238244514106, "grad_norm": 20.110124588012695, "learning_rate": 3.4765000000000005e-06, "loss": 2.162, "step": 6953 }, { "epoch": 2.179937304075235, "grad_norm": 9.745160102844238, "learning_rate": 3.4770000000000003e-06, "loss": 2.0619, "step": 6954 }, { "epoch": 2.1802507836990594, "grad_norm": 18.917219161987305, "learning_rate": 3.4775e-06, "loss": 2.2881, "step": 6955 }, { "epoch": 2.180564263322884, "grad_norm": 13.169961929321289, "learning_rate": 3.4780000000000004e-06, "loss": 2.4671, "step": 6956 }, { "epoch": 2.1808777429467083, "grad_norm": 14.872823715209961, "learning_rate": 3.4785000000000002e-06, "loss": 2.1184, "step": 6957 }, { "epoch": 2.1811912225705328, "grad_norm": 10.590782165527344, "learning_rate": 3.479e-06, "loss": 1.9362, "step": 6958 }, { "epoch": 2.181504702194357, "grad_norm": 20.34432029724121, "learning_rate": 3.4795e-06, "loss": 2.7275, "step": 6959 }, { "epoch": 2.1818181818181817, "grad_norm": 14.156587600708008, "learning_rate": 3.48e-06, "loss": 2.3055, "step": 6960 }, { "epoch": 2.182131661442006, "grad_norm": 27.308082580566406, "learning_rate": 3.4805000000000004e-06, "loss": 2.1747, "step": 6961 }, { "epoch": 2.1824451410658305, "grad_norm": 51.90277099609375, "learning_rate": 3.4810000000000006e-06, "loss": 2.3091, "step": 6962 }, { "epoch": 2.182758620689655, "grad_norm": 17.37398910522461, "learning_rate": 3.4815000000000005e-06, "loss": 1.9552, "step": 6963 }, { "epoch": 2.1830721003134794, "grad_norm": 51.738182067871094, "learning_rate": 3.4820000000000003e-06, "loss": 2.5037, "step": 6964 }, { "epoch": 2.183385579937304, "grad_norm": 28.63599967956543, "learning_rate": 3.4825e-06, "loss": 2.3401, "step": 6965 }, { "epoch": 2.1836990595611283, "grad_norm": 16.585054397583008, "learning_rate": 3.4830000000000004e-06, "loss": 2.0558, "step": 6966 }, { "epoch": 2.1840125391849528, "grad_norm": 20.072246551513672, "learning_rate": 3.4835e-06, "loss": 2.3002, "step": 6967 }, { "epoch": 2.1843260188087776, "grad_norm": 24.278078079223633, "learning_rate": 3.484e-06, "loss": 2.2806, "step": 6968 }, { "epoch": 2.184639498432602, "grad_norm": 46.744773864746094, "learning_rate": 3.4845000000000003e-06, "loss": 1.9996, "step": 6969 }, { "epoch": 2.1849529780564265, "grad_norm": 18.27585792541504, "learning_rate": 3.485e-06, "loss": 2.2772, "step": 6970 }, { "epoch": 2.185266457680251, "grad_norm": 22.770980834960938, "learning_rate": 3.4855000000000003e-06, "loss": 2.3376, "step": 6971 }, { "epoch": 2.1855799373040754, "grad_norm": 10.47475814819336, "learning_rate": 3.4860000000000006e-06, "loss": 2.2032, "step": 6972 }, { "epoch": 2.1858934169279, "grad_norm": 11.625728607177734, "learning_rate": 3.4865000000000004e-06, "loss": 2.469, "step": 6973 }, { "epoch": 2.1862068965517243, "grad_norm": 20.235750198364258, "learning_rate": 3.4870000000000002e-06, "loss": 2.4096, "step": 6974 }, { "epoch": 2.1865203761755487, "grad_norm": 8.474031448364258, "learning_rate": 3.4875000000000005e-06, "loss": 2.3613, "step": 6975 }, { "epoch": 2.186833855799373, "grad_norm": 12.229165077209473, "learning_rate": 3.4880000000000003e-06, "loss": 2.0266, "step": 6976 }, { "epoch": 2.1871473354231976, "grad_norm": 10.136529922485352, "learning_rate": 3.4885e-06, "loss": 2.0492, "step": 6977 }, { "epoch": 2.187460815047022, "grad_norm": 36.100502014160156, "learning_rate": 3.489e-06, "loss": 2.639, "step": 6978 }, { "epoch": 2.1877742946708465, "grad_norm": 9.78111743927002, "learning_rate": 3.4895000000000002e-06, "loss": 2.2953, "step": 6979 }, { "epoch": 2.188087774294671, "grad_norm": 23.95823097229004, "learning_rate": 3.49e-06, "loss": 2.8261, "step": 6980 }, { "epoch": 2.1884012539184954, "grad_norm": 31.69403076171875, "learning_rate": 3.4905000000000007e-06, "loss": 2.2804, "step": 6981 }, { "epoch": 2.18871473354232, "grad_norm": 13.984627723693848, "learning_rate": 3.4910000000000005e-06, "loss": 2.2058, "step": 6982 }, { "epoch": 2.1890282131661443, "grad_norm": 12.241868019104004, "learning_rate": 3.4915000000000004e-06, "loss": 2.1702, "step": 6983 }, { "epoch": 2.1893416927899687, "grad_norm": 11.750625610351562, "learning_rate": 3.492e-06, "loss": 2.16, "step": 6984 }, { "epoch": 2.189655172413793, "grad_norm": 31.47757911682129, "learning_rate": 3.4925000000000004e-06, "loss": 2.0741, "step": 6985 }, { "epoch": 2.1899686520376176, "grad_norm": 27.53549575805664, "learning_rate": 3.4930000000000003e-06, "loss": 2.7972, "step": 6986 }, { "epoch": 2.190282131661442, "grad_norm": 19.387630462646484, "learning_rate": 3.4935e-06, "loss": 2.3158, "step": 6987 }, { "epoch": 2.1905956112852665, "grad_norm": 28.865367889404297, "learning_rate": 3.4940000000000003e-06, "loss": 2.5238, "step": 6988 }, { "epoch": 2.190909090909091, "grad_norm": 19.169261932373047, "learning_rate": 3.4945e-06, "loss": 2.1497, "step": 6989 }, { "epoch": 2.1912225705329154, "grad_norm": 44.085880279541016, "learning_rate": 3.495e-06, "loss": 2.5714, "step": 6990 }, { "epoch": 2.19153605015674, "grad_norm": 12.665911674499512, "learning_rate": 3.4955000000000007e-06, "loss": 2.1209, "step": 6991 }, { "epoch": 2.1918495297805642, "grad_norm": 17.70903205871582, "learning_rate": 3.4960000000000005e-06, "loss": 2.0554, "step": 6992 }, { "epoch": 2.1921630094043887, "grad_norm": 128.4654083251953, "learning_rate": 3.4965000000000003e-06, "loss": 2.5496, "step": 6993 }, { "epoch": 2.192476489028213, "grad_norm": 7.527581691741943, "learning_rate": 3.4970000000000006e-06, "loss": 2.3946, "step": 6994 }, { "epoch": 2.1927899686520376, "grad_norm": 10.369160652160645, "learning_rate": 3.4975000000000004e-06, "loss": 2.2668, "step": 6995 }, { "epoch": 2.193103448275862, "grad_norm": 7.752161979675293, "learning_rate": 3.4980000000000002e-06, "loss": 2.1329, "step": 6996 }, { "epoch": 2.1934169278996865, "grad_norm": 14.886237144470215, "learning_rate": 3.4985e-06, "loss": 2.2909, "step": 6997 }, { "epoch": 2.193730407523511, "grad_norm": 10.20263671875, "learning_rate": 3.4990000000000003e-06, "loss": 1.9633, "step": 6998 }, { "epoch": 2.1940438871473353, "grad_norm": 9.909849166870117, "learning_rate": 3.4995e-06, "loss": 2.3337, "step": 6999 }, { "epoch": 2.19435736677116, "grad_norm": 12.828836441040039, "learning_rate": 3.5e-06, "loss": 2.1578, "step": 7000 }, { "epoch": 2.194670846394984, "grad_norm": 33.49946975708008, "learning_rate": 3.5005e-06, "loss": 2.4945, "step": 7001 }, { "epoch": 2.1949843260188087, "grad_norm": 11.534113883972168, "learning_rate": 3.5010000000000004e-06, "loss": 2.2808, "step": 7002 }, { "epoch": 2.195297805642633, "grad_norm": 12.272741317749023, "learning_rate": 3.5015000000000003e-06, "loss": 2.2166, "step": 7003 }, { "epoch": 2.1956112852664575, "grad_norm": 12.864968299865723, "learning_rate": 3.5020000000000005e-06, "loss": 1.7665, "step": 7004 }, { "epoch": 2.195924764890282, "grad_norm": 12.90298843383789, "learning_rate": 3.5025000000000003e-06, "loss": 2.1379, "step": 7005 }, { "epoch": 2.1962382445141064, "grad_norm": 7.430113315582275, "learning_rate": 3.503e-06, "loss": 2.0458, "step": 7006 }, { "epoch": 2.196551724137931, "grad_norm": 33.780372619628906, "learning_rate": 3.5035000000000004e-06, "loss": 2.7458, "step": 7007 }, { "epoch": 2.1968652037617553, "grad_norm": 31.221603393554688, "learning_rate": 3.5040000000000002e-06, "loss": 2.3979, "step": 7008 }, { "epoch": 2.1971786833855798, "grad_norm": 13.30338191986084, "learning_rate": 3.5045e-06, "loss": 1.8963, "step": 7009 }, { "epoch": 2.197492163009404, "grad_norm": 10.651296615600586, "learning_rate": 3.505e-06, "loss": 2.1804, "step": 7010 }, { "epoch": 2.1978056426332286, "grad_norm": 94.01361846923828, "learning_rate": 3.5055e-06, "loss": 2.8456, "step": 7011 }, { "epoch": 2.1981191222570535, "grad_norm": 9.76352310180664, "learning_rate": 3.5060000000000004e-06, "loss": 2.4617, "step": 7012 }, { "epoch": 2.198432601880878, "grad_norm": 14.140840530395508, "learning_rate": 3.5065000000000006e-06, "loss": 2.5596, "step": 7013 }, { "epoch": 2.1987460815047024, "grad_norm": 34.910648345947266, "learning_rate": 3.5070000000000005e-06, "loss": 2.2246, "step": 7014 }, { "epoch": 2.199059561128527, "grad_norm": 10.91124439239502, "learning_rate": 3.5075000000000003e-06, "loss": 1.8066, "step": 7015 }, { "epoch": 2.1993730407523513, "grad_norm": 12.093061447143555, "learning_rate": 3.508e-06, "loss": 2.2622, "step": 7016 }, { "epoch": 2.1996865203761757, "grad_norm": 50.25368118286133, "learning_rate": 3.5085000000000004e-06, "loss": 2.7275, "step": 7017 }, { "epoch": 2.2, "grad_norm": 8.938188552856445, "learning_rate": 3.509e-06, "loss": 2.2475, "step": 7018 }, { "epoch": 2.2003134796238246, "grad_norm": 8.015507698059082, "learning_rate": 3.5095e-06, "loss": 2.2311, "step": 7019 }, { "epoch": 2.200626959247649, "grad_norm": 23.359540939331055, "learning_rate": 3.5100000000000003e-06, "loss": 2.6065, "step": 7020 }, { "epoch": 2.2009404388714735, "grad_norm": 8.604899406433105, "learning_rate": 3.5105e-06, "loss": 2.3201, "step": 7021 }, { "epoch": 2.201253918495298, "grad_norm": 10.40499210357666, "learning_rate": 3.5110000000000004e-06, "loss": 2.1645, "step": 7022 }, { "epoch": 2.2015673981191224, "grad_norm": 9.35750961303711, "learning_rate": 3.5115000000000006e-06, "loss": 2.0373, "step": 7023 }, { "epoch": 2.201880877742947, "grad_norm": 8.183088302612305, "learning_rate": 3.5120000000000004e-06, "loss": 2.0476, "step": 7024 }, { "epoch": 2.2021943573667713, "grad_norm": 9.273404121398926, "learning_rate": 3.5125000000000003e-06, "loss": 2.1649, "step": 7025 }, { "epoch": 2.2025078369905957, "grad_norm": 17.128244400024414, "learning_rate": 3.5130000000000005e-06, "loss": 2.3145, "step": 7026 }, { "epoch": 2.20282131661442, "grad_norm": 54.616275787353516, "learning_rate": 3.5135000000000003e-06, "loss": 2.5012, "step": 7027 }, { "epoch": 2.2031347962382446, "grad_norm": 13.828841209411621, "learning_rate": 3.514e-06, "loss": 2.2098, "step": 7028 }, { "epoch": 2.203448275862069, "grad_norm": 20.072296142578125, "learning_rate": 3.5145e-06, "loss": 2.1528, "step": 7029 }, { "epoch": 2.2037617554858935, "grad_norm": 22.45638656616211, "learning_rate": 3.5150000000000002e-06, "loss": 2.2161, "step": 7030 }, { "epoch": 2.204075235109718, "grad_norm": 12.790304183959961, "learning_rate": 3.5155e-06, "loss": 2.2439, "step": 7031 }, { "epoch": 2.2043887147335424, "grad_norm": 16.862106323242188, "learning_rate": 3.5160000000000007e-06, "loss": 2.6222, "step": 7032 }, { "epoch": 2.204702194357367, "grad_norm": 46.84284210205078, "learning_rate": 3.5165000000000006e-06, "loss": 2.2979, "step": 7033 }, { "epoch": 2.2050156739811912, "grad_norm": 7.707181930541992, "learning_rate": 3.5170000000000004e-06, "loss": 2.1111, "step": 7034 }, { "epoch": 2.2053291536050157, "grad_norm": 61.13558578491211, "learning_rate": 3.5175e-06, "loss": 2.6039, "step": 7035 }, { "epoch": 2.20564263322884, "grad_norm": 14.961822509765625, "learning_rate": 3.5180000000000005e-06, "loss": 2.2589, "step": 7036 }, { "epoch": 2.2059561128526646, "grad_norm": 14.338484764099121, "learning_rate": 3.5185000000000003e-06, "loss": 2.2454, "step": 7037 }, { "epoch": 2.206269592476489, "grad_norm": 15.246488571166992, "learning_rate": 3.519e-06, "loss": 2.0901, "step": 7038 }, { "epoch": 2.2065830721003135, "grad_norm": 11.248987197875977, "learning_rate": 3.5195e-06, "loss": 2.1726, "step": 7039 }, { "epoch": 2.206896551724138, "grad_norm": 56.992340087890625, "learning_rate": 3.52e-06, "loss": 2.1786, "step": 7040 }, { "epoch": 2.2072100313479623, "grad_norm": 13.855323791503906, "learning_rate": 3.5205e-06, "loss": 2.0024, "step": 7041 }, { "epoch": 2.207523510971787, "grad_norm": 14.168869018554688, "learning_rate": 3.5210000000000007e-06, "loss": 2.0307, "step": 7042 }, { "epoch": 2.2078369905956112, "grad_norm": 20.478347778320312, "learning_rate": 3.5215000000000005e-06, "loss": 2.2844, "step": 7043 }, { "epoch": 2.2081504702194357, "grad_norm": 25.966365814208984, "learning_rate": 3.5220000000000003e-06, "loss": 2.3758, "step": 7044 }, { "epoch": 2.20846394984326, "grad_norm": 11.542539596557617, "learning_rate": 3.5225e-06, "loss": 2.2004, "step": 7045 }, { "epoch": 2.2087774294670846, "grad_norm": 8.398588180541992, "learning_rate": 3.5230000000000004e-06, "loss": 2.2516, "step": 7046 }, { "epoch": 2.209090909090909, "grad_norm": 16.954822540283203, "learning_rate": 3.5235000000000002e-06, "loss": 1.777, "step": 7047 }, { "epoch": 2.2094043887147334, "grad_norm": 33.871910095214844, "learning_rate": 3.524e-06, "loss": 2.4125, "step": 7048 }, { "epoch": 2.209717868338558, "grad_norm": 15.3339204788208, "learning_rate": 3.5245000000000003e-06, "loss": 2.2262, "step": 7049 }, { "epoch": 2.2100313479623823, "grad_norm": 12.728348731994629, "learning_rate": 3.525e-06, "loss": 2.158, "step": 7050 }, { "epoch": 2.2103448275862068, "grad_norm": 109.73506927490234, "learning_rate": 3.5255e-06, "loss": 2.3798, "step": 7051 }, { "epoch": 2.210658307210031, "grad_norm": 9.063850402832031, "learning_rate": 3.5260000000000006e-06, "loss": 2.0853, "step": 7052 }, { "epoch": 2.2109717868338556, "grad_norm": 9.89504337310791, "learning_rate": 3.5265000000000005e-06, "loss": 2.3841, "step": 7053 }, { "epoch": 2.21128526645768, "grad_norm": 9.473200798034668, "learning_rate": 3.5270000000000003e-06, "loss": 2.2786, "step": 7054 }, { "epoch": 2.2115987460815045, "grad_norm": 182.46815490722656, "learning_rate": 3.5275000000000005e-06, "loss": 2.3794, "step": 7055 }, { "epoch": 2.211912225705329, "grad_norm": 54.309852600097656, "learning_rate": 3.5280000000000004e-06, "loss": 2.2786, "step": 7056 }, { "epoch": 2.2122257053291534, "grad_norm": 106.93385314941406, "learning_rate": 3.5285e-06, "loss": 2.7345, "step": 7057 }, { "epoch": 2.212539184952978, "grad_norm": 9.693746566772461, "learning_rate": 3.529e-06, "loss": 2.1188, "step": 7058 }, { "epoch": 2.2128526645768023, "grad_norm": 12.784469604492188, "learning_rate": 3.5295000000000003e-06, "loss": 2.2525, "step": 7059 }, { "epoch": 2.2131661442006267, "grad_norm": 14.965514183044434, "learning_rate": 3.53e-06, "loss": 2.0136, "step": 7060 }, { "epoch": 2.213479623824451, "grad_norm": 11.913949966430664, "learning_rate": 3.5305e-06, "loss": 2.2113, "step": 7061 }, { "epoch": 2.213793103448276, "grad_norm": 26.847957611083984, "learning_rate": 3.5310000000000006e-06, "loss": 2.241, "step": 7062 }, { "epoch": 2.2141065830721005, "grad_norm": 11.093050956726074, "learning_rate": 3.5315000000000004e-06, "loss": 2.169, "step": 7063 }, { "epoch": 2.214420062695925, "grad_norm": 11.254499435424805, "learning_rate": 3.5320000000000002e-06, "loss": 2.0893, "step": 7064 }, { "epoch": 2.2147335423197494, "grad_norm": 12.10766887664795, "learning_rate": 3.5325000000000005e-06, "loss": 2.3673, "step": 7065 }, { "epoch": 2.215047021943574, "grad_norm": 20.228219985961914, "learning_rate": 3.5330000000000003e-06, "loss": 2.0338, "step": 7066 }, { "epoch": 2.2153605015673983, "grad_norm": 12.247249603271484, "learning_rate": 3.5335e-06, "loss": 2.3508, "step": 7067 }, { "epoch": 2.2156739811912227, "grad_norm": 9.960166931152344, "learning_rate": 3.5340000000000004e-06, "loss": 2.1375, "step": 7068 }, { "epoch": 2.215987460815047, "grad_norm": 16.839937210083008, "learning_rate": 3.5345e-06, "loss": 2.2619, "step": 7069 }, { "epoch": 2.2163009404388716, "grad_norm": 34.73561096191406, "learning_rate": 3.535e-06, "loss": 3.0058, "step": 7070 }, { "epoch": 2.216614420062696, "grad_norm": 21.49616050720215, "learning_rate": 3.5355e-06, "loss": 2.2393, "step": 7071 }, { "epoch": 2.2169278996865205, "grad_norm": 38.16383361816406, "learning_rate": 3.5360000000000005e-06, "loss": 2.0832, "step": 7072 }, { "epoch": 2.217241379310345, "grad_norm": 10.859467506408691, "learning_rate": 3.5365000000000004e-06, "loss": 2.2342, "step": 7073 }, { "epoch": 2.2175548589341694, "grad_norm": 19.308124542236328, "learning_rate": 3.5370000000000006e-06, "loss": 2.2912, "step": 7074 }, { "epoch": 2.217868338557994, "grad_norm": 8.347383499145508, "learning_rate": 3.5375000000000004e-06, "loss": 2.517, "step": 7075 }, { "epoch": 2.2181818181818183, "grad_norm": 10.600940704345703, "learning_rate": 3.5380000000000003e-06, "loss": 2.1054, "step": 7076 }, { "epoch": 2.2184952978056427, "grad_norm": 12.254573822021484, "learning_rate": 3.5385e-06, "loss": 2.1791, "step": 7077 }, { "epoch": 2.218808777429467, "grad_norm": 11.85391616821289, "learning_rate": 3.5390000000000003e-06, "loss": 2.1663, "step": 7078 }, { "epoch": 2.2191222570532916, "grad_norm": 9.994483947753906, "learning_rate": 3.5395e-06, "loss": 2.428, "step": 7079 }, { "epoch": 2.219435736677116, "grad_norm": 28.509777069091797, "learning_rate": 3.54e-06, "loss": 2.5909, "step": 7080 }, { "epoch": 2.2197492163009405, "grad_norm": 13.822752952575684, "learning_rate": 3.5405000000000002e-06, "loss": 2.3155, "step": 7081 }, { "epoch": 2.220062695924765, "grad_norm": 10.16942024230957, "learning_rate": 3.5410000000000005e-06, "loss": 2.1745, "step": 7082 }, { "epoch": 2.2203761755485893, "grad_norm": 10.339808464050293, "learning_rate": 3.5415000000000003e-06, "loss": 1.8173, "step": 7083 }, { "epoch": 2.220689655172414, "grad_norm": 58.65555953979492, "learning_rate": 3.5420000000000006e-06, "loss": 2.7708, "step": 7084 }, { "epoch": 2.2210031347962382, "grad_norm": 9.012490272521973, "learning_rate": 3.5425000000000004e-06, "loss": 2.1871, "step": 7085 }, { "epoch": 2.2213166144200627, "grad_norm": 9.79778003692627, "learning_rate": 3.5430000000000002e-06, "loss": 2.2072, "step": 7086 }, { "epoch": 2.221630094043887, "grad_norm": 10.99577808380127, "learning_rate": 3.5435000000000005e-06, "loss": 2.6023, "step": 7087 }, { "epoch": 2.2219435736677116, "grad_norm": 14.361542701721191, "learning_rate": 3.5440000000000003e-06, "loss": 2.4689, "step": 7088 }, { "epoch": 2.222257053291536, "grad_norm": 20.790489196777344, "learning_rate": 3.5445e-06, "loss": 2.2526, "step": 7089 }, { "epoch": 2.2225705329153604, "grad_norm": 11.628655433654785, "learning_rate": 3.545e-06, "loss": 2.2157, "step": 7090 }, { "epoch": 2.222884012539185, "grad_norm": 16.32400894165039, "learning_rate": 3.5455e-06, "loss": 2.3175, "step": 7091 }, { "epoch": 2.2231974921630093, "grad_norm": 13.606417655944824, "learning_rate": 3.5460000000000004e-06, "loss": 2.2204, "step": 7092 }, { "epoch": 2.2235109717868338, "grad_norm": 9.552896499633789, "learning_rate": 3.5465000000000007e-06, "loss": 2.0245, "step": 7093 }, { "epoch": 2.223824451410658, "grad_norm": 7.792943477630615, "learning_rate": 3.5470000000000005e-06, "loss": 2.1342, "step": 7094 }, { "epoch": 2.2241379310344827, "grad_norm": 8.231707572937012, "learning_rate": 3.5475000000000003e-06, "loss": 2.2138, "step": 7095 }, { "epoch": 2.224451410658307, "grad_norm": 13.393856048583984, "learning_rate": 3.548e-06, "loss": 2.0156, "step": 7096 }, { "epoch": 2.2247648902821315, "grad_norm": 14.117390632629395, "learning_rate": 3.5485000000000004e-06, "loss": 2.4265, "step": 7097 }, { "epoch": 2.225078369905956, "grad_norm": 22.97071647644043, "learning_rate": 3.5490000000000002e-06, "loss": 2.2683, "step": 7098 }, { "epoch": 2.2253918495297804, "grad_norm": 7.688055515289307, "learning_rate": 3.5495e-06, "loss": 2.3423, "step": 7099 }, { "epoch": 2.225705329153605, "grad_norm": 23.624292373657227, "learning_rate": 3.5500000000000003e-06, "loss": 2.3965, "step": 7100 }, { "epoch": 2.2260188087774293, "grad_norm": 32.03535842895508, "learning_rate": 3.5505e-06, "loss": 2.4757, "step": 7101 }, { "epoch": 2.2263322884012537, "grad_norm": 34.347042083740234, "learning_rate": 3.5510000000000004e-06, "loss": 2.3726, "step": 7102 }, { "epoch": 2.226645768025078, "grad_norm": 9.961214065551758, "learning_rate": 3.5515000000000006e-06, "loss": 2.296, "step": 7103 }, { "epoch": 2.2269592476489026, "grad_norm": 9.492974281311035, "learning_rate": 3.5520000000000005e-06, "loss": 2.0891, "step": 7104 }, { "epoch": 2.227272727272727, "grad_norm": 11.200065612792969, "learning_rate": 3.5525000000000003e-06, "loss": 2.303, "step": 7105 }, { "epoch": 2.227586206896552, "grad_norm": 13.679742813110352, "learning_rate": 3.5530000000000005e-06, "loss": 2.4089, "step": 7106 }, { "epoch": 2.2278996865203764, "grad_norm": 9.797566413879395, "learning_rate": 3.5535000000000004e-06, "loss": 2.2193, "step": 7107 }, { "epoch": 2.228213166144201, "grad_norm": 14.846070289611816, "learning_rate": 3.554e-06, "loss": 2.1373, "step": 7108 }, { "epoch": 2.2285266457680253, "grad_norm": 13.531272888183594, "learning_rate": 3.5545e-06, "loss": 2.2994, "step": 7109 }, { "epoch": 2.2288401253918497, "grad_norm": 42.00018310546875, "learning_rate": 3.5550000000000003e-06, "loss": 2.312, "step": 7110 }, { "epoch": 2.229153605015674, "grad_norm": 8.310403823852539, "learning_rate": 3.5555e-06, "loss": 2.1954, "step": 7111 }, { "epoch": 2.2294670846394986, "grad_norm": 9.747314453125, "learning_rate": 3.5560000000000008e-06, "loss": 2.0885, "step": 7112 }, { "epoch": 2.229780564263323, "grad_norm": 28.042261123657227, "learning_rate": 3.5565000000000006e-06, "loss": 2.3885, "step": 7113 }, { "epoch": 2.2300940438871475, "grad_norm": 55.079463958740234, "learning_rate": 3.5570000000000004e-06, "loss": 2.4243, "step": 7114 }, { "epoch": 2.230407523510972, "grad_norm": 74.71209716796875, "learning_rate": 3.5575000000000002e-06, "loss": 2.914, "step": 7115 }, { "epoch": 2.2307210031347964, "grad_norm": 16.395322799682617, "learning_rate": 3.5580000000000005e-06, "loss": 2.2497, "step": 7116 }, { "epoch": 2.231034482758621, "grad_norm": 10.165128707885742, "learning_rate": 3.5585000000000003e-06, "loss": 2.154, "step": 7117 }, { "epoch": 2.2313479623824453, "grad_norm": 9.329062461853027, "learning_rate": 3.559e-06, "loss": 1.9963, "step": 7118 }, { "epoch": 2.2316614420062697, "grad_norm": 37.05634307861328, "learning_rate": 3.5595e-06, "loss": 2.3787, "step": 7119 }, { "epoch": 2.231974921630094, "grad_norm": 13.289584159851074, "learning_rate": 3.5600000000000002e-06, "loss": 2.2813, "step": 7120 }, { "epoch": 2.2322884012539186, "grad_norm": 14.31875228881836, "learning_rate": 3.5605e-06, "loss": 2.2519, "step": 7121 }, { "epoch": 2.232601880877743, "grad_norm": 10.276305198669434, "learning_rate": 3.5610000000000007e-06, "loss": 2.0542, "step": 7122 }, { "epoch": 2.2329153605015675, "grad_norm": 11.904434204101562, "learning_rate": 3.5615000000000005e-06, "loss": 2.3037, "step": 7123 }, { "epoch": 2.233228840125392, "grad_norm": 11.562606811523438, "learning_rate": 3.5620000000000004e-06, "loss": 2.227, "step": 7124 }, { "epoch": 2.2335423197492164, "grad_norm": 15.81189250946045, "learning_rate": 3.5625e-06, "loss": 2.5752, "step": 7125 }, { "epoch": 2.233855799373041, "grad_norm": 51.10102844238281, "learning_rate": 3.5630000000000004e-06, "loss": 2.3682, "step": 7126 }, { "epoch": 2.2341692789968652, "grad_norm": 12.101628303527832, "learning_rate": 3.5635000000000003e-06, "loss": 2.1648, "step": 7127 }, { "epoch": 2.2344827586206897, "grad_norm": 21.33435821533203, "learning_rate": 3.564e-06, "loss": 2.3227, "step": 7128 }, { "epoch": 2.234796238244514, "grad_norm": 9.108460426330566, "learning_rate": 3.5645000000000003e-06, "loss": 2.14, "step": 7129 }, { "epoch": 2.2351097178683386, "grad_norm": 5.969293117523193, "learning_rate": 3.565e-06, "loss": 2.1229, "step": 7130 }, { "epoch": 2.235423197492163, "grad_norm": 12.906966209411621, "learning_rate": 3.5655e-06, "loss": 1.9585, "step": 7131 }, { "epoch": 2.2357366771159874, "grad_norm": 8.280203819274902, "learning_rate": 3.566e-06, "loss": 2.0501, "step": 7132 }, { "epoch": 2.236050156739812, "grad_norm": 11.277865409851074, "learning_rate": 3.5665000000000005e-06, "loss": 2.1179, "step": 7133 }, { "epoch": 2.2363636363636363, "grad_norm": 16.584272384643555, "learning_rate": 3.5670000000000003e-06, "loss": 2.4252, "step": 7134 }, { "epoch": 2.2366771159874608, "grad_norm": 31.076719284057617, "learning_rate": 3.5675000000000006e-06, "loss": 1.8485, "step": 7135 }, { "epoch": 2.236990595611285, "grad_norm": 21.546274185180664, "learning_rate": 3.5680000000000004e-06, "loss": 2.1926, "step": 7136 }, { "epoch": 2.2373040752351097, "grad_norm": 22.932079315185547, "learning_rate": 3.5685000000000002e-06, "loss": 2.2614, "step": 7137 }, { "epoch": 2.237617554858934, "grad_norm": 15.603130340576172, "learning_rate": 3.569e-06, "loss": 2.524, "step": 7138 }, { "epoch": 2.2379310344827585, "grad_norm": 12.728400230407715, "learning_rate": 3.5695000000000003e-06, "loss": 2.2551, "step": 7139 }, { "epoch": 2.238244514106583, "grad_norm": 12.261685371398926, "learning_rate": 3.57e-06, "loss": 2.3482, "step": 7140 }, { "epoch": 2.2385579937304074, "grad_norm": 37.186588287353516, "learning_rate": 3.5705e-06, "loss": 2.4016, "step": 7141 }, { "epoch": 2.238871473354232, "grad_norm": 9.2909517288208, "learning_rate": 3.571e-06, "loss": 2.6648, "step": 7142 }, { "epoch": 2.2391849529780563, "grad_norm": 10.105910301208496, "learning_rate": 3.5715000000000004e-06, "loss": 2.2289, "step": 7143 }, { "epoch": 2.2394984326018808, "grad_norm": 35.10897445678711, "learning_rate": 3.5720000000000003e-06, "loss": 2.3035, "step": 7144 }, { "epoch": 2.239811912225705, "grad_norm": 8.988752365112305, "learning_rate": 3.5725000000000005e-06, "loss": 2.1066, "step": 7145 }, { "epoch": 2.2401253918495296, "grad_norm": 15.336725234985352, "learning_rate": 3.5730000000000003e-06, "loss": 2.1121, "step": 7146 }, { "epoch": 2.240438871473354, "grad_norm": 31.607391357421875, "learning_rate": 3.5735e-06, "loss": 2.3258, "step": 7147 }, { "epoch": 2.2407523510971785, "grad_norm": 20.95800018310547, "learning_rate": 3.5740000000000004e-06, "loss": 2.1473, "step": 7148 }, { "epoch": 2.241065830721003, "grad_norm": 49.861175537109375, "learning_rate": 3.5745000000000003e-06, "loss": 2.3437, "step": 7149 }, { "epoch": 2.2413793103448274, "grad_norm": 12.217568397521973, "learning_rate": 3.575e-06, "loss": 2.1905, "step": 7150 }, { "epoch": 2.241692789968652, "grad_norm": 15.905807495117188, "learning_rate": 3.5755e-06, "loss": 2.1966, "step": 7151 }, { "epoch": 2.2420062695924763, "grad_norm": 9.921697616577148, "learning_rate": 3.576e-06, "loss": 1.9001, "step": 7152 }, { "epoch": 2.2423197492163007, "grad_norm": 30.135778427124023, "learning_rate": 3.5765000000000004e-06, "loss": 2.2015, "step": 7153 }, { "epoch": 2.242633228840125, "grad_norm": 7.372073650360107, "learning_rate": 3.5770000000000007e-06, "loss": 2.1252, "step": 7154 }, { "epoch": 2.2429467084639496, "grad_norm": 14.261728286743164, "learning_rate": 3.5775000000000005e-06, "loss": 2.4332, "step": 7155 }, { "epoch": 2.2432601880877745, "grad_norm": 21.214012145996094, "learning_rate": 3.5780000000000003e-06, "loss": 2.0557, "step": 7156 }, { "epoch": 2.243573667711599, "grad_norm": 9.037612915039062, "learning_rate": 3.5785e-06, "loss": 2.1991, "step": 7157 }, { "epoch": 2.2438871473354234, "grad_norm": 15.023518562316895, "learning_rate": 3.5790000000000004e-06, "loss": 2.2951, "step": 7158 }, { "epoch": 2.244200626959248, "grad_norm": 9.438035011291504, "learning_rate": 3.5795e-06, "loss": 2.2607, "step": 7159 }, { "epoch": 2.2445141065830723, "grad_norm": 15.13282585144043, "learning_rate": 3.58e-06, "loss": 2.1204, "step": 7160 }, { "epoch": 2.2448275862068967, "grad_norm": 85.73621368408203, "learning_rate": 3.5805000000000003e-06, "loss": 2.4441, "step": 7161 }, { "epoch": 2.245141065830721, "grad_norm": 14.550522804260254, "learning_rate": 3.581e-06, "loss": 2.1531, "step": 7162 }, { "epoch": 2.2454545454545456, "grad_norm": 18.766340255737305, "learning_rate": 3.5815000000000004e-06, "loss": 2.2147, "step": 7163 }, { "epoch": 2.24576802507837, "grad_norm": 22.684078216552734, "learning_rate": 3.5820000000000006e-06, "loss": 2.1919, "step": 7164 }, { "epoch": 2.2460815047021945, "grad_norm": 130.93141174316406, "learning_rate": 3.5825000000000004e-06, "loss": 2.658, "step": 7165 }, { "epoch": 2.246394984326019, "grad_norm": 16.606786727905273, "learning_rate": 3.5830000000000003e-06, "loss": 2.2346, "step": 7166 }, { "epoch": 2.2467084639498434, "grad_norm": 34.00636291503906, "learning_rate": 3.5835000000000005e-06, "loss": 2.2563, "step": 7167 }, { "epoch": 2.247021943573668, "grad_norm": 12.066088676452637, "learning_rate": 3.5840000000000003e-06, "loss": 1.7301, "step": 7168 }, { "epoch": 2.2473354231974922, "grad_norm": 10.510574340820312, "learning_rate": 3.5845e-06, "loss": 2.4187, "step": 7169 }, { "epoch": 2.2476489028213167, "grad_norm": 13.81126880645752, "learning_rate": 3.585e-06, "loss": 2.0827, "step": 7170 }, { "epoch": 2.247962382445141, "grad_norm": 45.93455123901367, "learning_rate": 3.5855000000000002e-06, "loss": 2.1119, "step": 7171 }, { "epoch": 2.2482758620689656, "grad_norm": 21.286945343017578, "learning_rate": 3.586e-06, "loss": 2.4371, "step": 7172 }, { "epoch": 2.24858934169279, "grad_norm": 65.36617279052734, "learning_rate": 3.5865000000000007e-06, "loss": 2.4069, "step": 7173 }, { "epoch": 2.2489028213166145, "grad_norm": 15.013283729553223, "learning_rate": 3.5870000000000006e-06, "loss": 2.3038, "step": 7174 }, { "epoch": 2.249216300940439, "grad_norm": 28.936763763427734, "learning_rate": 3.5875000000000004e-06, "loss": 2.2068, "step": 7175 }, { "epoch": 2.2495297805642633, "grad_norm": 16.521089553833008, "learning_rate": 3.588e-06, "loss": 1.9812, "step": 7176 }, { "epoch": 2.249843260188088, "grad_norm": 14.462285995483398, "learning_rate": 3.5885000000000005e-06, "loss": 2.2185, "step": 7177 }, { "epoch": 2.250156739811912, "grad_norm": 25.74633026123047, "learning_rate": 3.5890000000000003e-06, "loss": 2.1845, "step": 7178 }, { "epoch": 2.2504702194357367, "grad_norm": 50.065826416015625, "learning_rate": 3.5895e-06, "loss": 2.2768, "step": 7179 }, { "epoch": 2.250783699059561, "grad_norm": 6.686931133270264, "learning_rate": 3.5900000000000004e-06, "loss": 2.2578, "step": 7180 }, { "epoch": 2.2510971786833855, "grad_norm": 18.078153610229492, "learning_rate": 3.5905e-06, "loss": 2.1543, "step": 7181 }, { "epoch": 2.25141065830721, "grad_norm": 13.219353675842285, "learning_rate": 3.591e-06, "loss": 2.1802, "step": 7182 }, { "epoch": 2.25141065830721, "eval_loss": 2.4945011138916016, "eval_runtime": 21.0439, "eval_samples_per_second": 127.685, "eval_steps_per_second": 7.983, "step": 7182 }, { "epoch": 2.2517241379310344, "grad_norm": 14.665486335754395, "learning_rate": 3.5915000000000007e-06, "loss": 2.5557, "step": 7183 }, { "epoch": 2.252037617554859, "grad_norm": 17.79008674621582, "learning_rate": 3.5920000000000005e-06, "loss": 2.0346, "step": 7184 }, { "epoch": 2.2523510971786833, "grad_norm": 20.073755264282227, "learning_rate": 3.5925000000000003e-06, "loss": 2.1498, "step": 7185 }, { "epoch": 2.2526645768025078, "grad_norm": 12.581036567687988, "learning_rate": 3.5930000000000006e-06, "loss": 2.1402, "step": 7186 }, { "epoch": 2.252978056426332, "grad_norm": 19.35159683227539, "learning_rate": 3.5935000000000004e-06, "loss": 2.3435, "step": 7187 }, { "epoch": 2.2532915360501566, "grad_norm": 11.87389087677002, "learning_rate": 3.5940000000000002e-06, "loss": 2.109, "step": 7188 }, { "epoch": 2.253605015673981, "grad_norm": 11.6506929397583, "learning_rate": 3.5945e-06, "loss": 2.3019, "step": 7189 }, { "epoch": 2.2539184952978055, "grad_norm": 8.157673835754395, "learning_rate": 3.5950000000000003e-06, "loss": 1.7438, "step": 7190 }, { "epoch": 2.25423197492163, "grad_norm": 36.08637237548828, "learning_rate": 3.5955e-06, "loss": 2.7932, "step": 7191 }, { "epoch": 2.2545454545454544, "grad_norm": 12.120672225952148, "learning_rate": 3.596e-06, "loss": 1.872, "step": 7192 }, { "epoch": 2.254858934169279, "grad_norm": 25.82735252380371, "learning_rate": 3.5965000000000006e-06, "loss": 2.1613, "step": 7193 }, { "epoch": 2.2551724137931033, "grad_norm": 16.79880714416504, "learning_rate": 3.5970000000000005e-06, "loss": 2.3109, "step": 7194 }, { "epoch": 2.2554858934169277, "grad_norm": 56.27902603149414, "learning_rate": 3.5975000000000003e-06, "loss": 2.3316, "step": 7195 }, { "epoch": 2.255799373040752, "grad_norm": 8.648634910583496, "learning_rate": 3.5980000000000005e-06, "loss": 1.9602, "step": 7196 }, { "epoch": 2.2561128526645766, "grad_norm": 7.0171403884887695, "learning_rate": 3.5985000000000004e-06, "loss": 2.3148, "step": 7197 }, { "epoch": 2.256426332288401, "grad_norm": 40.01934051513672, "learning_rate": 3.599e-06, "loss": 2.8216, "step": 7198 }, { "epoch": 2.256739811912226, "grad_norm": 11.738056182861328, "learning_rate": 3.5995e-06, "loss": 2.0605, "step": 7199 }, { "epoch": 2.2570532915360504, "grad_norm": 11.64625358581543, "learning_rate": 3.6000000000000003e-06, "loss": 2.1984, "step": 7200 }, { "epoch": 2.257366771159875, "grad_norm": 10.108341217041016, "learning_rate": 3.6005e-06, "loss": 2.2411, "step": 7201 }, { "epoch": 2.2576802507836993, "grad_norm": 8.636689186096191, "learning_rate": 3.601e-06, "loss": 2.0594, "step": 7202 }, { "epoch": 2.2579937304075237, "grad_norm": 33.638492584228516, "learning_rate": 3.6015000000000006e-06, "loss": 2.8161, "step": 7203 }, { "epoch": 2.258307210031348, "grad_norm": 12.021764755249023, "learning_rate": 3.6020000000000004e-06, "loss": 2.3533, "step": 7204 }, { "epoch": 2.2586206896551726, "grad_norm": 10.805076599121094, "learning_rate": 3.6025000000000002e-06, "loss": 2.1351, "step": 7205 }, { "epoch": 2.258934169278997, "grad_norm": 26.758176803588867, "learning_rate": 3.6030000000000005e-06, "loss": 2.7063, "step": 7206 }, { "epoch": 2.2592476489028215, "grad_norm": 11.699298858642578, "learning_rate": 3.6035000000000003e-06, "loss": 2.1319, "step": 7207 }, { "epoch": 2.259561128526646, "grad_norm": 13.173660278320312, "learning_rate": 3.604e-06, "loss": 2.2169, "step": 7208 }, { "epoch": 2.2598746081504704, "grad_norm": 11.301003456115723, "learning_rate": 3.6045000000000004e-06, "loss": 2.2471, "step": 7209 }, { "epoch": 2.260188087774295, "grad_norm": 11.274521827697754, "learning_rate": 3.6050000000000002e-06, "loss": 1.8854, "step": 7210 }, { "epoch": 2.2605015673981192, "grad_norm": 8.971123695373535, "learning_rate": 3.6055e-06, "loss": 2.2822, "step": 7211 }, { "epoch": 2.2608150470219437, "grad_norm": 7.624792098999023, "learning_rate": 3.606e-06, "loss": 2.1283, "step": 7212 }, { "epoch": 2.261128526645768, "grad_norm": 6.846399784088135, "learning_rate": 3.6065000000000005e-06, "loss": 2.142, "step": 7213 }, { "epoch": 2.2614420062695926, "grad_norm": 9.79951286315918, "learning_rate": 3.6070000000000004e-06, "loss": 2.3404, "step": 7214 }, { "epoch": 2.261755485893417, "grad_norm": 43.51485824584961, "learning_rate": 3.6075000000000006e-06, "loss": 2.223, "step": 7215 }, { "epoch": 2.2620689655172415, "grad_norm": 13.380617141723633, "learning_rate": 3.6080000000000004e-06, "loss": 2.0451, "step": 7216 }, { "epoch": 2.262382445141066, "grad_norm": 10.865477561950684, "learning_rate": 3.6085000000000003e-06, "loss": 2.3691, "step": 7217 }, { "epoch": 2.2626959247648903, "grad_norm": 12.024353981018066, "learning_rate": 3.609e-06, "loss": 2.169, "step": 7218 }, { "epoch": 2.263009404388715, "grad_norm": 16.159467697143555, "learning_rate": 3.6095000000000003e-06, "loss": 2.014, "step": 7219 }, { "epoch": 2.2633228840125392, "grad_norm": 19.978511810302734, "learning_rate": 3.61e-06, "loss": 2.2969, "step": 7220 }, { "epoch": 2.2636363636363637, "grad_norm": 13.330805778503418, "learning_rate": 3.6105e-06, "loss": 2.0045, "step": 7221 }, { "epoch": 2.263949843260188, "grad_norm": 23.94683265686035, "learning_rate": 3.6110000000000002e-06, "loss": 2.1617, "step": 7222 }, { "epoch": 2.2642633228840126, "grad_norm": 14.688584327697754, "learning_rate": 3.6115000000000005e-06, "loss": 2.8749, "step": 7223 }, { "epoch": 2.264576802507837, "grad_norm": 37.662132263183594, "learning_rate": 3.6120000000000003e-06, "loss": 2.2973, "step": 7224 }, { "epoch": 2.2648902821316614, "grad_norm": 31.596712112426758, "learning_rate": 3.6125000000000006e-06, "loss": 2.2441, "step": 7225 }, { "epoch": 2.265203761755486, "grad_norm": 14.841413497924805, "learning_rate": 3.6130000000000004e-06, "loss": 1.9271, "step": 7226 }, { "epoch": 2.2655172413793103, "grad_norm": 24.38260841369629, "learning_rate": 3.6135000000000002e-06, "loss": 2.6305, "step": 7227 }, { "epoch": 2.2658307210031348, "grad_norm": 56.179420471191406, "learning_rate": 3.6140000000000005e-06, "loss": 2.6206, "step": 7228 }, { "epoch": 2.266144200626959, "grad_norm": 7.055506229400635, "learning_rate": 3.6145000000000003e-06, "loss": 2.0168, "step": 7229 }, { "epoch": 2.2664576802507836, "grad_norm": 36.492706298828125, "learning_rate": 3.615e-06, "loss": 2.1242, "step": 7230 }, { "epoch": 2.266771159874608, "grad_norm": 10.88786506652832, "learning_rate": 3.6155e-06, "loss": 2.4575, "step": 7231 }, { "epoch": 2.2670846394984325, "grad_norm": 7.990469455718994, "learning_rate": 3.616e-06, "loss": 2.1859, "step": 7232 }, { "epoch": 2.267398119122257, "grad_norm": 10.519044876098633, "learning_rate": 3.6165000000000004e-06, "loss": 2.0734, "step": 7233 }, { "epoch": 2.2677115987460814, "grad_norm": 11.63737678527832, "learning_rate": 3.6170000000000007e-06, "loss": 2.0344, "step": 7234 }, { "epoch": 2.268025078369906, "grad_norm": 24.95403480529785, "learning_rate": 3.6175000000000005e-06, "loss": 2.3296, "step": 7235 }, { "epoch": 2.2683385579937303, "grad_norm": 45.92222595214844, "learning_rate": 3.6180000000000003e-06, "loss": 2.2954, "step": 7236 }, { "epoch": 2.2686520376175547, "grad_norm": 15.379531860351562, "learning_rate": 3.6185e-06, "loss": 1.9429, "step": 7237 }, { "epoch": 2.268965517241379, "grad_norm": 13.637126922607422, "learning_rate": 3.6190000000000004e-06, "loss": 2.0125, "step": 7238 }, { "epoch": 2.2692789968652036, "grad_norm": 46.56778335571289, "learning_rate": 3.6195000000000002e-06, "loss": 2.2859, "step": 7239 }, { "epoch": 2.269592476489028, "grad_norm": 14.61168098449707, "learning_rate": 3.62e-06, "loss": 2.0966, "step": 7240 }, { "epoch": 2.2699059561128525, "grad_norm": 15.446404457092285, "learning_rate": 3.6205000000000003e-06, "loss": 2.3621, "step": 7241 }, { "epoch": 2.270219435736677, "grad_norm": 10.560173034667969, "learning_rate": 3.621e-06, "loss": 1.9773, "step": 7242 }, { "epoch": 2.2705329153605014, "grad_norm": 74.63044738769531, "learning_rate": 3.6215000000000004e-06, "loss": 2.4477, "step": 7243 }, { "epoch": 2.270846394984326, "grad_norm": 10.090686798095703, "learning_rate": 3.6220000000000006e-06, "loss": 2.1996, "step": 7244 }, { "epoch": 2.2711598746081503, "grad_norm": 11.0864839553833, "learning_rate": 3.6225000000000005e-06, "loss": 2.1407, "step": 7245 }, { "epoch": 2.2714733542319747, "grad_norm": 23.535518646240234, "learning_rate": 3.6230000000000003e-06, "loss": 1.9041, "step": 7246 }, { "epoch": 2.271786833855799, "grad_norm": 29.483312606811523, "learning_rate": 3.6235000000000005e-06, "loss": 2.4477, "step": 7247 }, { "epoch": 2.2721003134796236, "grad_norm": 10.477875709533691, "learning_rate": 3.6240000000000004e-06, "loss": 2.3691, "step": 7248 }, { "epoch": 2.272413793103448, "grad_norm": 26.94452667236328, "learning_rate": 3.6245e-06, "loss": 2.1515, "step": 7249 }, { "epoch": 2.2727272727272725, "grad_norm": 11.526517868041992, "learning_rate": 3.625e-06, "loss": 2.0815, "step": 7250 }, { "epoch": 2.2730407523510974, "grad_norm": 13.421350479125977, "learning_rate": 3.6255000000000003e-06, "loss": 2.8006, "step": 7251 }, { "epoch": 2.273354231974922, "grad_norm": 17.69158363342285, "learning_rate": 3.626e-06, "loss": 2.291, "step": 7252 }, { "epoch": 2.2736677115987463, "grad_norm": 13.186980247497559, "learning_rate": 3.6265e-06, "loss": 2.3254, "step": 7253 }, { "epoch": 2.2739811912225707, "grad_norm": 23.839916229248047, "learning_rate": 3.6270000000000006e-06, "loss": 2.3909, "step": 7254 }, { "epoch": 2.274294670846395, "grad_norm": 23.547060012817383, "learning_rate": 3.6275000000000004e-06, "loss": 2.1269, "step": 7255 }, { "epoch": 2.2746081504702196, "grad_norm": 47.0949821472168, "learning_rate": 3.6280000000000002e-06, "loss": 2.4669, "step": 7256 }, { "epoch": 2.274921630094044, "grad_norm": 36.64951705932617, "learning_rate": 3.6285000000000005e-06, "loss": 2.6099, "step": 7257 }, { "epoch": 2.2752351097178685, "grad_norm": 13.459712028503418, "learning_rate": 3.6290000000000003e-06, "loss": 2.091, "step": 7258 }, { "epoch": 2.275548589341693, "grad_norm": 70.17310333251953, "learning_rate": 3.6295e-06, "loss": 2.9814, "step": 7259 }, { "epoch": 2.2758620689655173, "grad_norm": 11.854323387145996, "learning_rate": 3.6300000000000004e-06, "loss": 2.0264, "step": 7260 }, { "epoch": 2.276175548589342, "grad_norm": 17.43655014038086, "learning_rate": 3.6305000000000002e-06, "loss": 2.1702, "step": 7261 }, { "epoch": 2.2764890282131662, "grad_norm": 9.961427688598633, "learning_rate": 3.631e-06, "loss": 2.3291, "step": 7262 }, { "epoch": 2.2768025078369907, "grad_norm": 6.855801105499268, "learning_rate": 3.6315e-06, "loss": 2.0679, "step": 7263 }, { "epoch": 2.277115987460815, "grad_norm": 9.185486793518066, "learning_rate": 3.6320000000000005e-06, "loss": 2.2961, "step": 7264 }, { "epoch": 2.2774294670846396, "grad_norm": 9.132794380187988, "learning_rate": 3.6325000000000004e-06, "loss": 2.1752, "step": 7265 }, { "epoch": 2.277742946708464, "grad_norm": 11.749226570129395, "learning_rate": 3.6330000000000006e-06, "loss": 2.3906, "step": 7266 }, { "epoch": 2.2780564263322884, "grad_norm": 57.306800842285156, "learning_rate": 3.6335000000000004e-06, "loss": 2.3411, "step": 7267 }, { "epoch": 2.278369905956113, "grad_norm": 15.80717658996582, "learning_rate": 3.6340000000000003e-06, "loss": 2.1394, "step": 7268 }, { "epoch": 2.2786833855799373, "grad_norm": 62.68533706665039, "learning_rate": 3.6345e-06, "loss": 2.4816, "step": 7269 }, { "epoch": 2.2789968652037618, "grad_norm": 83.24822235107422, "learning_rate": 3.6350000000000003e-06, "loss": 2.1257, "step": 7270 }, { "epoch": 2.279310344827586, "grad_norm": 10.266399383544922, "learning_rate": 3.6355e-06, "loss": 2.0159, "step": 7271 }, { "epoch": 2.2796238244514107, "grad_norm": 9.45529842376709, "learning_rate": 3.636e-06, "loss": 1.9788, "step": 7272 }, { "epoch": 2.279937304075235, "grad_norm": 11.23605728149414, "learning_rate": 3.6365e-06, "loss": 2.0858, "step": 7273 }, { "epoch": 2.2802507836990595, "grad_norm": 14.547463417053223, "learning_rate": 3.6370000000000005e-06, "loss": 2.3307, "step": 7274 }, { "epoch": 2.280564263322884, "grad_norm": 16.607769012451172, "learning_rate": 3.6375000000000003e-06, "loss": 2.2843, "step": 7275 }, { "epoch": 2.2808777429467084, "grad_norm": 8.159924507141113, "learning_rate": 3.6380000000000006e-06, "loss": 2.4309, "step": 7276 }, { "epoch": 2.281191222570533, "grad_norm": 139.78797912597656, "learning_rate": 3.6385000000000004e-06, "loss": 2.8586, "step": 7277 }, { "epoch": 2.2815047021943573, "grad_norm": 11.418096542358398, "learning_rate": 3.6390000000000002e-06, "loss": 2.0253, "step": 7278 }, { "epoch": 2.2818181818181817, "grad_norm": 12.863363265991211, "learning_rate": 3.6395e-06, "loss": 2.3567, "step": 7279 }, { "epoch": 2.282131661442006, "grad_norm": 8.575637817382812, "learning_rate": 3.6400000000000003e-06, "loss": 2.1688, "step": 7280 }, { "epoch": 2.2824451410658306, "grad_norm": 16.22159194946289, "learning_rate": 3.6405e-06, "loss": 2.1901, "step": 7281 }, { "epoch": 2.282758620689655, "grad_norm": 40.53306198120117, "learning_rate": 3.641e-06, "loss": 2.7036, "step": 7282 }, { "epoch": 2.2830721003134795, "grad_norm": 48.29307556152344, "learning_rate": 3.6415e-06, "loss": 2.1553, "step": 7283 }, { "epoch": 2.283385579937304, "grad_norm": 14.39148998260498, "learning_rate": 3.6420000000000005e-06, "loss": 2.7488, "step": 7284 }, { "epoch": 2.2836990595611284, "grad_norm": 11.778844833374023, "learning_rate": 3.6425000000000003e-06, "loss": 2.1678, "step": 7285 }, { "epoch": 2.284012539184953, "grad_norm": 11.584482192993164, "learning_rate": 3.6430000000000005e-06, "loss": 3.2497, "step": 7286 }, { "epoch": 2.2843260188087773, "grad_norm": 8.301648139953613, "learning_rate": 3.6435000000000004e-06, "loss": 2.0334, "step": 7287 }, { "epoch": 2.2846394984326017, "grad_norm": 10.004752159118652, "learning_rate": 3.644e-06, "loss": 1.9576, "step": 7288 }, { "epoch": 2.284952978056426, "grad_norm": 10.673555374145508, "learning_rate": 3.6445000000000004e-06, "loss": 1.9757, "step": 7289 }, { "epoch": 2.2852664576802506, "grad_norm": 28.051969528198242, "learning_rate": 3.6450000000000003e-06, "loss": 2.8219, "step": 7290 }, { "epoch": 2.285579937304075, "grad_norm": 12.56019115447998, "learning_rate": 3.6455e-06, "loss": 2.0277, "step": 7291 }, { "epoch": 2.2858934169279, "grad_norm": 13.937050819396973, "learning_rate": 3.646e-06, "loss": 2.6982, "step": 7292 }, { "epoch": 2.2862068965517244, "grad_norm": 11.295731544494629, "learning_rate": 3.6465e-06, "loss": 2.5053, "step": 7293 }, { "epoch": 2.286520376175549, "grad_norm": 11.401422500610352, "learning_rate": 3.6470000000000004e-06, "loss": 2.2891, "step": 7294 }, { "epoch": 2.2868338557993733, "grad_norm": 12.384406089782715, "learning_rate": 3.6475000000000007e-06, "loss": 1.9221, "step": 7295 }, { "epoch": 2.2871473354231977, "grad_norm": 21.355010986328125, "learning_rate": 3.6480000000000005e-06, "loss": 2.2827, "step": 7296 }, { "epoch": 2.287460815047022, "grad_norm": 9.771644592285156, "learning_rate": 3.6485000000000003e-06, "loss": 2.2188, "step": 7297 }, { "epoch": 2.2877742946708466, "grad_norm": 15.417769432067871, "learning_rate": 3.649e-06, "loss": 2.4738, "step": 7298 }, { "epoch": 2.288087774294671, "grad_norm": 8.464964866638184, "learning_rate": 3.6495000000000004e-06, "loss": 2.006, "step": 7299 }, { "epoch": 2.2884012539184955, "grad_norm": 15.349848747253418, "learning_rate": 3.65e-06, "loss": 2.1223, "step": 7300 }, { "epoch": 2.28871473354232, "grad_norm": 8.756331443786621, "learning_rate": 3.6505e-06, "loss": 2.1818, "step": 7301 }, { "epoch": 2.2890282131661444, "grad_norm": 10.621785163879395, "learning_rate": 3.6510000000000003e-06, "loss": 2.249, "step": 7302 }, { "epoch": 2.289341692789969, "grad_norm": 11.08302116394043, "learning_rate": 3.6515e-06, "loss": 2.0881, "step": 7303 }, { "epoch": 2.2896551724137932, "grad_norm": 35.1323127746582, "learning_rate": 3.6520000000000004e-06, "loss": 2.2882, "step": 7304 }, { "epoch": 2.2899686520376177, "grad_norm": 10.780580520629883, "learning_rate": 3.6525000000000006e-06, "loss": 1.9674, "step": 7305 }, { "epoch": 2.290282131661442, "grad_norm": 18.246137619018555, "learning_rate": 3.6530000000000004e-06, "loss": 2.507, "step": 7306 }, { "epoch": 2.2905956112852666, "grad_norm": 10.258757591247559, "learning_rate": 3.6535000000000003e-06, "loss": 2.1466, "step": 7307 }, { "epoch": 2.290909090909091, "grad_norm": 10.967842102050781, "learning_rate": 3.6540000000000005e-06, "loss": 2.2077, "step": 7308 }, { "epoch": 2.2912225705329154, "grad_norm": 9.059294700622559, "learning_rate": 3.6545000000000003e-06, "loss": 2.1443, "step": 7309 }, { "epoch": 2.29153605015674, "grad_norm": 14.1408052444458, "learning_rate": 3.655e-06, "loss": 2.2626, "step": 7310 }, { "epoch": 2.2918495297805643, "grad_norm": 408.2253112792969, "learning_rate": 3.6555e-06, "loss": 2.4134, "step": 7311 }, { "epoch": 2.2921630094043888, "grad_norm": 34.20561981201172, "learning_rate": 3.6560000000000002e-06, "loss": 2.2869, "step": 7312 }, { "epoch": 2.292476489028213, "grad_norm": 78.49437713623047, "learning_rate": 3.6565e-06, "loss": 2.1, "step": 7313 }, { "epoch": 2.2927899686520377, "grad_norm": 12.197046279907227, "learning_rate": 3.6570000000000007e-06, "loss": 3.1883, "step": 7314 }, { "epoch": 2.293103448275862, "grad_norm": 7.6918768882751465, "learning_rate": 3.6575000000000006e-06, "loss": 2.1498, "step": 7315 }, { "epoch": 2.2934169278996865, "grad_norm": 47.999473571777344, "learning_rate": 3.6580000000000004e-06, "loss": 2.3722, "step": 7316 }, { "epoch": 2.293730407523511, "grad_norm": 84.93740844726562, "learning_rate": 3.6585e-06, "loss": 2.3759, "step": 7317 }, { "epoch": 2.2940438871473354, "grad_norm": 13.031661033630371, "learning_rate": 3.6590000000000005e-06, "loss": 2.1309, "step": 7318 }, { "epoch": 2.29435736677116, "grad_norm": 14.5783052444458, "learning_rate": 3.6595000000000003e-06, "loss": 2.1356, "step": 7319 }, { "epoch": 2.2946708463949843, "grad_norm": 33.6772575378418, "learning_rate": 3.66e-06, "loss": 2.3004, "step": 7320 }, { "epoch": 2.2949843260188088, "grad_norm": 7.3739471435546875, "learning_rate": 3.6605000000000004e-06, "loss": 2.0734, "step": 7321 }, { "epoch": 2.295297805642633, "grad_norm": 8.614192962646484, "learning_rate": 3.661e-06, "loss": 1.9625, "step": 7322 }, { "epoch": 2.2956112852664576, "grad_norm": 19.887514114379883, "learning_rate": 3.6615e-06, "loss": 2.5214, "step": 7323 }, { "epoch": 2.295924764890282, "grad_norm": 13.122472763061523, "learning_rate": 3.6620000000000007e-06, "loss": 2.4287, "step": 7324 }, { "epoch": 2.2962382445141065, "grad_norm": 7.210230827331543, "learning_rate": 3.6625000000000005e-06, "loss": 2.1122, "step": 7325 }, { "epoch": 2.296551724137931, "grad_norm": 39.83118438720703, "learning_rate": 3.6630000000000003e-06, "loss": 2.1056, "step": 7326 }, { "epoch": 2.2968652037617554, "grad_norm": 11.875229835510254, "learning_rate": 3.6635000000000006e-06, "loss": 2.4499, "step": 7327 }, { "epoch": 2.29717868338558, "grad_norm": 40.137088775634766, "learning_rate": 3.6640000000000004e-06, "loss": 2.1489, "step": 7328 }, { "epoch": 2.2974921630094043, "grad_norm": 16.190235137939453, "learning_rate": 3.6645000000000002e-06, "loss": 2.3431, "step": 7329 }, { "epoch": 2.2978056426332287, "grad_norm": 14.192766189575195, "learning_rate": 3.665e-06, "loss": 2.2262, "step": 7330 }, { "epoch": 2.298119122257053, "grad_norm": 8.17026424407959, "learning_rate": 3.6655000000000003e-06, "loss": 2.0985, "step": 7331 }, { "epoch": 2.2984326018808776, "grad_norm": 19.303911209106445, "learning_rate": 3.666e-06, "loss": 2.2816, "step": 7332 }, { "epoch": 2.298746081504702, "grad_norm": 7.751400470733643, "learning_rate": 3.6665e-06, "loss": 2.1593, "step": 7333 }, { "epoch": 2.2990595611285265, "grad_norm": 17.76302146911621, "learning_rate": 3.6670000000000006e-06, "loss": 2.3251, "step": 7334 }, { "epoch": 2.299373040752351, "grad_norm": 11.700510025024414, "learning_rate": 3.6675000000000005e-06, "loss": 2.0066, "step": 7335 }, { "epoch": 2.2996865203761754, "grad_norm": 16.93220329284668, "learning_rate": 3.6680000000000003e-06, "loss": 2.3539, "step": 7336 }, { "epoch": 2.3, "grad_norm": 32.08696365356445, "learning_rate": 3.6685000000000005e-06, "loss": 2.6861, "step": 7337 }, { "epoch": 2.3003134796238243, "grad_norm": 51.448848724365234, "learning_rate": 3.6690000000000004e-06, "loss": 2.1885, "step": 7338 }, { "epoch": 2.3006269592476487, "grad_norm": 8.5531005859375, "learning_rate": 3.6695e-06, "loss": 2.0435, "step": 7339 }, { "epoch": 2.300940438871473, "grad_norm": 11.284357070922852, "learning_rate": 3.6700000000000004e-06, "loss": 2.8753, "step": 7340 }, { "epoch": 2.3012539184952976, "grad_norm": 17.045236587524414, "learning_rate": 3.6705000000000003e-06, "loss": 2.1916, "step": 7341 }, { "epoch": 2.301567398119122, "grad_norm": 12.128131866455078, "learning_rate": 3.671e-06, "loss": 2.2854, "step": 7342 }, { "epoch": 2.3018808777429465, "grad_norm": 10.835186004638672, "learning_rate": 3.6715e-06, "loss": 2.0896, "step": 7343 }, { "epoch": 2.302194357366771, "grad_norm": 6.936743259429932, "learning_rate": 3.6720000000000006e-06, "loss": 2.1959, "step": 7344 }, { "epoch": 2.302507836990596, "grad_norm": 17.3377742767334, "learning_rate": 3.6725000000000004e-06, "loss": 2.2892, "step": 7345 }, { "epoch": 2.3028213166144202, "grad_norm": 13.948187828063965, "learning_rate": 3.6730000000000007e-06, "loss": 2.167, "step": 7346 }, { "epoch": 2.3031347962382447, "grad_norm": 12.212050437927246, "learning_rate": 3.6735000000000005e-06, "loss": 2.2448, "step": 7347 }, { "epoch": 2.303448275862069, "grad_norm": 10.113065719604492, "learning_rate": 3.6740000000000003e-06, "loss": 2.0326, "step": 7348 }, { "epoch": 2.3037617554858936, "grad_norm": 11.750550270080566, "learning_rate": 3.6745e-06, "loss": 2.162, "step": 7349 }, { "epoch": 2.304075235109718, "grad_norm": 9.069467544555664, "learning_rate": 3.6750000000000004e-06, "loss": 2.0701, "step": 7350 }, { "epoch": 2.3043887147335425, "grad_norm": 7.077291965484619, "learning_rate": 3.6755000000000002e-06, "loss": 2.146, "step": 7351 }, { "epoch": 2.304702194357367, "grad_norm": 27.952428817749023, "learning_rate": 3.676e-06, "loss": 2.3666, "step": 7352 }, { "epoch": 2.3050156739811913, "grad_norm": 18.55834197998047, "learning_rate": 3.6765e-06, "loss": 2.164, "step": 7353 }, { "epoch": 2.305329153605016, "grad_norm": 8.361699104309082, "learning_rate": 3.6770000000000005e-06, "loss": 2.0193, "step": 7354 }, { "epoch": 2.30564263322884, "grad_norm": 7.437684535980225, "learning_rate": 3.6775000000000004e-06, "loss": 2.0273, "step": 7355 }, { "epoch": 2.3059561128526647, "grad_norm": 10.00115966796875, "learning_rate": 3.6780000000000006e-06, "loss": 1.9694, "step": 7356 }, { "epoch": 2.306269592476489, "grad_norm": 69.46129608154297, "learning_rate": 3.6785000000000004e-06, "loss": 2.3412, "step": 7357 }, { "epoch": 2.3065830721003135, "grad_norm": 23.599761962890625, "learning_rate": 3.6790000000000003e-06, "loss": 2.1828, "step": 7358 }, { "epoch": 2.306896551724138, "grad_norm": 7.547100067138672, "learning_rate": 3.6795e-06, "loss": 2.2643, "step": 7359 }, { "epoch": 2.3072100313479624, "grad_norm": 13.41345500946045, "learning_rate": 3.6800000000000003e-06, "loss": 2.3136, "step": 7360 }, { "epoch": 2.307523510971787, "grad_norm": 15.3207368850708, "learning_rate": 3.6805e-06, "loss": 2.2585, "step": 7361 }, { "epoch": 2.3078369905956113, "grad_norm": 8.973692893981934, "learning_rate": 3.681e-06, "loss": 2.1586, "step": 7362 }, { "epoch": 2.3081504702194358, "grad_norm": 11.297796249389648, "learning_rate": 3.6815000000000002e-06, "loss": 2.4528, "step": 7363 }, { "epoch": 2.30846394984326, "grad_norm": 8.479543685913086, "learning_rate": 3.6820000000000005e-06, "loss": 2.1428, "step": 7364 }, { "epoch": 2.3087774294670846, "grad_norm": 17.749982833862305, "learning_rate": 3.6825000000000003e-06, "loss": 2.1984, "step": 7365 }, { "epoch": 2.309090909090909, "grad_norm": 34.43549728393555, "learning_rate": 3.6830000000000006e-06, "loss": 1.9228, "step": 7366 }, { "epoch": 2.3094043887147335, "grad_norm": 92.1659164428711, "learning_rate": 3.6835000000000004e-06, "loss": 2.3925, "step": 7367 }, { "epoch": 2.309717868338558, "grad_norm": 89.55754852294922, "learning_rate": 3.6840000000000002e-06, "loss": 2.3736, "step": 7368 }, { "epoch": 2.3100313479623824, "grad_norm": 10.297791481018066, "learning_rate": 3.6845000000000005e-06, "loss": 2.2583, "step": 7369 }, { "epoch": 2.310344827586207, "grad_norm": 39.69812774658203, "learning_rate": 3.6850000000000003e-06, "loss": 2.8764, "step": 7370 }, { "epoch": 2.3106583072100313, "grad_norm": 11.99314022064209, "learning_rate": 3.6855e-06, "loss": 2.1349, "step": 7371 }, { "epoch": 2.3109717868338557, "grad_norm": 12.176197052001953, "learning_rate": 3.686e-06, "loss": 2.0121, "step": 7372 }, { "epoch": 2.31128526645768, "grad_norm": 8.143828392028809, "learning_rate": 3.6865e-06, "loss": 2.1111, "step": 7373 }, { "epoch": 2.3115987460815046, "grad_norm": 12.237077713012695, "learning_rate": 3.6870000000000004e-06, "loss": 2.1518, "step": 7374 }, { "epoch": 2.311912225705329, "grad_norm": 13.478866577148438, "learning_rate": 3.6875000000000007e-06, "loss": 2.1566, "step": 7375 }, { "epoch": 2.3122257053291535, "grad_norm": 13.143150329589844, "learning_rate": 3.6880000000000005e-06, "loss": 2.2333, "step": 7376 }, { "epoch": 2.312539184952978, "grad_norm": 13.364155769348145, "learning_rate": 3.6885000000000003e-06, "loss": 2.0975, "step": 7377 }, { "epoch": 2.3128526645768024, "grad_norm": 14.199429512023926, "learning_rate": 3.689e-06, "loss": 2.6997, "step": 7378 }, { "epoch": 2.313166144200627, "grad_norm": 16.23635482788086, "learning_rate": 3.6895000000000004e-06, "loss": 1.9973, "step": 7379 }, { "epoch": 2.3134796238244513, "grad_norm": 11.58362865447998, "learning_rate": 3.6900000000000002e-06, "loss": 2.2228, "step": 7380 }, { "epoch": 2.3137931034482757, "grad_norm": 15.334861755371094, "learning_rate": 3.6905e-06, "loss": 2.0141, "step": 7381 }, { "epoch": 2.3141065830721, "grad_norm": 8.142290115356445, "learning_rate": 3.6910000000000003e-06, "loss": 2.2552, "step": 7382 }, { "epoch": 2.3144200626959246, "grad_norm": 29.615467071533203, "learning_rate": 3.6915e-06, "loss": 2.4391, "step": 7383 }, { "epoch": 2.314733542319749, "grad_norm": 10.87438678741455, "learning_rate": 3.692e-06, "loss": 2.1896, "step": 7384 }, { "epoch": 2.3150470219435735, "grad_norm": 29.416385650634766, "learning_rate": 3.6925000000000006e-06, "loss": 2.1046, "step": 7385 }, { "epoch": 2.3153605015673984, "grad_norm": 9.442472457885742, "learning_rate": 3.6930000000000005e-06, "loss": 2.4763, "step": 7386 }, { "epoch": 2.315673981191223, "grad_norm": 9.818400382995605, "learning_rate": 3.6935000000000003e-06, "loss": 1.9637, "step": 7387 }, { "epoch": 2.3159874608150472, "grad_norm": 12.118255615234375, "learning_rate": 3.6940000000000005e-06, "loss": 2.328, "step": 7388 }, { "epoch": 2.3163009404388717, "grad_norm": 7.663362503051758, "learning_rate": 3.6945000000000004e-06, "loss": 2.2856, "step": 7389 }, { "epoch": 2.316614420062696, "grad_norm": 9.70523452758789, "learning_rate": 3.695e-06, "loss": 2.4039, "step": 7390 }, { "epoch": 2.3169278996865206, "grad_norm": 10.905033111572266, "learning_rate": 3.6955e-06, "loss": 2.1467, "step": 7391 }, { "epoch": 2.317241379310345, "grad_norm": 10.506361961364746, "learning_rate": 3.6960000000000003e-06, "loss": 2.222, "step": 7392 }, { "epoch": 2.3175548589341695, "grad_norm": 6.936699867248535, "learning_rate": 3.6965e-06, "loss": 2.0757, "step": 7393 }, { "epoch": 2.317868338557994, "grad_norm": 150.7834014892578, "learning_rate": 3.697e-06, "loss": 2.3572, "step": 7394 }, { "epoch": 2.3181818181818183, "grad_norm": 11.875110626220703, "learning_rate": 3.6975000000000006e-06, "loss": 1.9825, "step": 7395 }, { "epoch": 2.318495297805643, "grad_norm": 22.18080711364746, "learning_rate": 3.6980000000000004e-06, "loss": 2.7774, "step": 7396 }, { "epoch": 2.3188087774294672, "grad_norm": 17.41339683532715, "learning_rate": 3.6985000000000002e-06, "loss": 2.3306, "step": 7397 }, { "epoch": 2.3191222570532917, "grad_norm": 15.024280548095703, "learning_rate": 3.6990000000000005e-06, "loss": 2.2748, "step": 7398 }, { "epoch": 2.319435736677116, "grad_norm": 16.163711547851562, "learning_rate": 3.6995000000000003e-06, "loss": 2.1726, "step": 7399 }, { "epoch": 2.3197492163009406, "grad_norm": 33.926544189453125, "learning_rate": 3.7e-06, "loss": 2.5392, "step": 7400 }, { "epoch": 2.320062695924765, "grad_norm": 14.800670623779297, "learning_rate": 3.7005000000000004e-06, "loss": 2.1722, "step": 7401 }, { "epoch": 2.3203761755485894, "grad_norm": 7.829089164733887, "learning_rate": 3.7010000000000002e-06, "loss": 2.2216, "step": 7402 }, { "epoch": 2.320689655172414, "grad_norm": 33.6285285949707, "learning_rate": 3.7015e-06, "loss": 2.4368, "step": 7403 }, { "epoch": 2.3210031347962383, "grad_norm": 29.689800262451172, "learning_rate": 3.702e-06, "loss": 2.1404, "step": 7404 }, { "epoch": 2.3213166144200628, "grad_norm": 16.54550552368164, "learning_rate": 3.7025000000000005e-06, "loss": 1.9032, "step": 7405 }, { "epoch": 2.321630094043887, "grad_norm": 14.241104125976562, "learning_rate": 3.7030000000000004e-06, "loss": 2.2645, "step": 7406 }, { "epoch": 2.3219435736677116, "grad_norm": 11.232771873474121, "learning_rate": 3.7035000000000006e-06, "loss": 2.2511, "step": 7407 }, { "epoch": 2.322257053291536, "grad_norm": 10.890080451965332, "learning_rate": 3.7040000000000005e-06, "loss": 2.2161, "step": 7408 }, { "epoch": 2.3225705329153605, "grad_norm": 35.95681381225586, "learning_rate": 3.7045000000000003e-06, "loss": 2.7221, "step": 7409 }, { "epoch": 2.322884012539185, "grad_norm": 12.461686134338379, "learning_rate": 3.705e-06, "loss": 2.5535, "step": 7410 }, { "epoch": 2.3231974921630094, "grad_norm": 72.9301986694336, "learning_rate": 3.7055000000000004e-06, "loss": 2.4704, "step": 7411 }, { "epoch": 2.323510971786834, "grad_norm": 18.283411026000977, "learning_rate": 3.706e-06, "loss": 1.9793, "step": 7412 }, { "epoch": 2.3238244514106583, "grad_norm": 17.89227867126465, "learning_rate": 3.7065e-06, "loss": 2.309, "step": 7413 }, { "epoch": 2.3241379310344827, "grad_norm": 11.035760879516602, "learning_rate": 3.7070000000000003e-06, "loss": 2.0491, "step": 7414 }, { "epoch": 2.324451410658307, "grad_norm": 15.21668815612793, "learning_rate": 3.7075000000000005e-06, "loss": 2.611, "step": 7415 }, { "epoch": 2.3247648902821316, "grad_norm": 143.13308715820312, "learning_rate": 3.7080000000000003e-06, "loss": 2.7403, "step": 7416 }, { "epoch": 2.325078369905956, "grad_norm": 8.220003128051758, "learning_rate": 3.7085000000000006e-06, "loss": 2.0423, "step": 7417 }, { "epoch": 2.3253918495297805, "grad_norm": 21.011747360229492, "learning_rate": 3.7090000000000004e-06, "loss": 2.232, "step": 7418 }, { "epoch": 2.325705329153605, "grad_norm": 23.205533981323242, "learning_rate": 3.7095000000000002e-06, "loss": 2.2535, "step": 7419 }, { "epoch": 2.3260188087774294, "grad_norm": 27.258634567260742, "learning_rate": 3.7100000000000005e-06, "loss": 2.2473, "step": 7420 }, { "epoch": 2.326332288401254, "grad_norm": 12.285863876342773, "learning_rate": 3.7105000000000003e-06, "loss": 2.0902, "step": 7421 }, { "epoch": 2.3266457680250783, "grad_norm": 37.95127868652344, "learning_rate": 3.711e-06, "loss": 2.8228, "step": 7422 }, { "epoch": 2.3269592476489027, "grad_norm": 58.29522705078125, "learning_rate": 3.7115e-06, "loss": 2.219, "step": 7423 }, { "epoch": 2.327272727272727, "grad_norm": 11.065194129943848, "learning_rate": 3.712e-06, "loss": 2.2862, "step": 7424 }, { "epoch": 2.3275862068965516, "grad_norm": 9.926230430603027, "learning_rate": 3.7125000000000005e-06, "loss": 2.1678, "step": 7425 }, { "epoch": 2.327899686520376, "grad_norm": 11.107030868530273, "learning_rate": 3.7130000000000007e-06, "loss": 2.1579, "step": 7426 }, { "epoch": 2.3282131661442005, "grad_norm": 6.60050630569458, "learning_rate": 3.7135000000000005e-06, "loss": 2.3279, "step": 7427 }, { "epoch": 2.328526645768025, "grad_norm": 7.916518211364746, "learning_rate": 3.7140000000000004e-06, "loss": 2.1502, "step": 7428 }, { "epoch": 2.3288401253918494, "grad_norm": 9.64710807800293, "learning_rate": 3.7145e-06, "loss": 2.1599, "step": 7429 }, { "epoch": 2.329153605015674, "grad_norm": 11.696365356445312, "learning_rate": 3.7150000000000004e-06, "loss": 2.116, "step": 7430 }, { "epoch": 2.3294670846394983, "grad_norm": 6.389529705047607, "learning_rate": 3.7155000000000003e-06, "loss": 1.928, "step": 7431 }, { "epoch": 2.3297805642633227, "grad_norm": 12.57972240447998, "learning_rate": 3.716e-06, "loss": 2.0127, "step": 7432 }, { "epoch": 2.330094043887147, "grad_norm": 10.347940444946289, "learning_rate": 3.7165e-06, "loss": 2.241, "step": 7433 }, { "epoch": 2.3304075235109716, "grad_norm": 10.769411087036133, "learning_rate": 3.717e-06, "loss": 1.8849, "step": 7434 }, { "epoch": 2.330721003134796, "grad_norm": 9.45003890991211, "learning_rate": 3.7175000000000004e-06, "loss": 2.052, "step": 7435 }, { "epoch": 2.3310344827586205, "grad_norm": 12.043750762939453, "learning_rate": 3.7180000000000007e-06, "loss": 2.0595, "step": 7436 }, { "epoch": 2.331347962382445, "grad_norm": 12.208069801330566, "learning_rate": 3.7185000000000005e-06, "loss": 1.8874, "step": 7437 }, { "epoch": 2.3316614420062693, "grad_norm": 8.682051658630371, "learning_rate": 3.7190000000000003e-06, "loss": 1.8688, "step": 7438 }, { "epoch": 2.3319749216300942, "grad_norm": 13.93393325805664, "learning_rate": 3.7195e-06, "loss": 2.3054, "step": 7439 }, { "epoch": 2.3322884012539187, "grad_norm": 10.033586502075195, "learning_rate": 3.7200000000000004e-06, "loss": 2.8946, "step": 7440 }, { "epoch": 2.332601880877743, "grad_norm": 110.87235260009766, "learning_rate": 3.7205e-06, "loss": 3.1258, "step": 7441 }, { "epoch": 2.3329153605015676, "grad_norm": 12.784510612487793, "learning_rate": 3.721e-06, "loss": 2.1371, "step": 7442 }, { "epoch": 2.333228840125392, "grad_norm": 22.052978515625, "learning_rate": 3.7215000000000003e-06, "loss": 2.1272, "step": 7443 }, { "epoch": 2.3335423197492164, "grad_norm": 15.194536209106445, "learning_rate": 3.722e-06, "loss": 2.4269, "step": 7444 }, { "epoch": 2.333855799373041, "grad_norm": 13.074575424194336, "learning_rate": 3.7225000000000004e-06, "loss": 2.2919, "step": 7445 }, { "epoch": 2.3341692789968653, "grad_norm": 12.135882377624512, "learning_rate": 3.7230000000000006e-06, "loss": 2.053, "step": 7446 }, { "epoch": 2.3344827586206898, "grad_norm": 9.084222793579102, "learning_rate": 3.7235000000000004e-06, "loss": 2.2904, "step": 7447 }, { "epoch": 2.334796238244514, "grad_norm": 17.912174224853516, "learning_rate": 3.7240000000000003e-06, "loss": 2.5168, "step": 7448 }, { "epoch": 2.3351097178683387, "grad_norm": 14.255494117736816, "learning_rate": 3.7245000000000005e-06, "loss": 2.1658, "step": 7449 }, { "epoch": 2.335423197492163, "grad_norm": 33.94017028808594, "learning_rate": 3.7250000000000003e-06, "loss": 2.4092, "step": 7450 }, { "epoch": 2.3357366771159875, "grad_norm": 8.786423683166504, "learning_rate": 3.7255e-06, "loss": 2.3889, "step": 7451 }, { "epoch": 2.336050156739812, "grad_norm": 20.953718185424805, "learning_rate": 3.726e-06, "loss": 2.1842, "step": 7452 }, { "epoch": 2.3363636363636364, "grad_norm": 9.891927719116211, "learning_rate": 3.7265000000000002e-06, "loss": 2.2718, "step": 7453 }, { "epoch": 2.336677115987461, "grad_norm": 38.01201629638672, "learning_rate": 3.727e-06, "loss": 2.1813, "step": 7454 }, { "epoch": 2.3369905956112853, "grad_norm": 8.129267692565918, "learning_rate": 3.7275000000000007e-06, "loss": 2.0664, "step": 7455 }, { "epoch": 2.3373040752351097, "grad_norm": 7.101378440856934, "learning_rate": 3.7280000000000006e-06, "loss": 2.1878, "step": 7456 }, { "epoch": 2.337617554858934, "grad_norm": 10.147604942321777, "learning_rate": 3.7285000000000004e-06, "loss": 2.0015, "step": 7457 }, { "epoch": 2.3379310344827586, "grad_norm": 17.602445602416992, "learning_rate": 3.729e-06, "loss": 2.159, "step": 7458 }, { "epoch": 2.338244514106583, "grad_norm": 16.053327560424805, "learning_rate": 3.7295000000000005e-06, "loss": 2.0408, "step": 7459 }, { "epoch": 2.3385579937304075, "grad_norm": 8.142462730407715, "learning_rate": 3.7300000000000003e-06, "loss": 1.9878, "step": 7460 }, { "epoch": 2.338871473354232, "grad_norm": 45.563228607177734, "learning_rate": 3.7305e-06, "loss": 2.4716, "step": 7461 }, { "epoch": 2.3391849529780564, "grad_norm": 37.93796157836914, "learning_rate": 3.7310000000000004e-06, "loss": 2.5844, "step": 7462 }, { "epoch": 2.339498432601881, "grad_norm": 12.426397323608398, "learning_rate": 3.7315e-06, "loss": 2.2307, "step": 7463 }, { "epoch": 2.3398119122257053, "grad_norm": 8.1763334274292, "learning_rate": 3.732e-06, "loss": 1.9342, "step": 7464 }, { "epoch": 2.3401253918495297, "grad_norm": 17.687929153442383, "learning_rate": 3.7325000000000007e-06, "loss": 2.056, "step": 7465 }, { "epoch": 2.340438871473354, "grad_norm": 76.34940338134766, "learning_rate": 3.7330000000000005e-06, "loss": 2.3082, "step": 7466 }, { "epoch": 2.3407523510971786, "grad_norm": 7.995907783508301, "learning_rate": 3.7335000000000003e-06, "loss": 2.3062, "step": 7467 }, { "epoch": 2.341065830721003, "grad_norm": 36.199771881103516, "learning_rate": 3.7340000000000006e-06, "loss": 2.5566, "step": 7468 }, { "epoch": 2.3413793103448275, "grad_norm": 18.274425506591797, "learning_rate": 3.7345000000000004e-06, "loss": 2.2244, "step": 7469 }, { "epoch": 2.341692789968652, "grad_norm": 41.214717864990234, "learning_rate": 3.7350000000000002e-06, "loss": 2.237, "step": 7470 }, { "epoch": 2.3420062695924764, "grad_norm": 13.862223625183105, "learning_rate": 3.7355e-06, "loss": 2.1516, "step": 7471 }, { "epoch": 2.342319749216301, "grad_norm": 12.588922500610352, "learning_rate": 3.7360000000000003e-06, "loss": 1.9605, "step": 7472 }, { "epoch": 2.3426332288401253, "grad_norm": 10.872784614562988, "learning_rate": 3.7365e-06, "loss": 2.3185, "step": 7473 }, { "epoch": 2.3429467084639497, "grad_norm": 22.444738388061523, "learning_rate": 3.737e-06, "loss": 2.3358, "step": 7474 }, { "epoch": 2.343260188087774, "grad_norm": 24.353376388549805, "learning_rate": 3.7375000000000006e-06, "loss": 2.1146, "step": 7475 }, { "epoch": 2.3435736677115986, "grad_norm": 152.16514587402344, "learning_rate": 3.7380000000000005e-06, "loss": 2.4918, "step": 7476 }, { "epoch": 2.343887147335423, "grad_norm": 22.4973201751709, "learning_rate": 3.7385000000000003e-06, "loss": 2.554, "step": 7477 }, { "epoch": 2.3442006269592475, "grad_norm": 20.581514358520508, "learning_rate": 3.7390000000000005e-06, "loss": 2.1282, "step": 7478 }, { "epoch": 2.344514106583072, "grad_norm": 96.27397155761719, "learning_rate": 3.7395000000000004e-06, "loss": 2.3136, "step": 7479 }, { "epoch": 2.344827586206897, "grad_norm": 10.789430618286133, "learning_rate": 3.74e-06, "loss": 2.1371, "step": 7480 }, { "epoch": 2.3451410658307212, "grad_norm": 13.69425106048584, "learning_rate": 3.7405000000000004e-06, "loss": 2.3149, "step": 7481 }, { "epoch": 2.3454545454545457, "grad_norm": 8.174257278442383, "learning_rate": 3.7410000000000003e-06, "loss": 2.3051, "step": 7482 }, { "epoch": 2.34576802507837, "grad_norm": 55.180599212646484, "learning_rate": 3.7415e-06, "loss": 2.322, "step": 7483 }, { "epoch": 2.3460815047021946, "grad_norm": 25.564199447631836, "learning_rate": 3.742e-06, "loss": 2.2274, "step": 7484 }, { "epoch": 2.346394984326019, "grad_norm": 9.052996635437012, "learning_rate": 3.7425000000000006e-06, "loss": 2.034, "step": 7485 }, { "epoch": 2.3467084639498434, "grad_norm": 11.388501167297363, "learning_rate": 3.7430000000000004e-06, "loss": 2.1719, "step": 7486 }, { "epoch": 2.347021943573668, "grad_norm": 89.06910705566406, "learning_rate": 3.7435000000000007e-06, "loss": 2.4144, "step": 7487 }, { "epoch": 2.3473354231974923, "grad_norm": 12.16662883758545, "learning_rate": 3.7440000000000005e-06, "loss": 2.2557, "step": 7488 }, { "epoch": 2.3476489028213168, "grad_norm": 99.88227081298828, "learning_rate": 3.7445000000000003e-06, "loss": 2.332, "step": 7489 }, { "epoch": 2.347962382445141, "grad_norm": 23.34038734436035, "learning_rate": 3.745e-06, "loss": 2.0684, "step": 7490 }, { "epoch": 2.3482758620689657, "grad_norm": 9.094398498535156, "learning_rate": 3.7455000000000004e-06, "loss": 2.282, "step": 7491 }, { "epoch": 2.34858934169279, "grad_norm": 10.438407897949219, "learning_rate": 3.7460000000000002e-06, "loss": 2.2895, "step": 7492 }, { "epoch": 2.3489028213166145, "grad_norm": 15.971596717834473, "learning_rate": 3.7465e-06, "loss": 2.1715, "step": 7493 }, { "epoch": 2.349216300940439, "grad_norm": 15.316971778869629, "learning_rate": 3.7470000000000003e-06, "loss": 2.3835, "step": 7494 }, { "epoch": 2.3495297805642634, "grad_norm": 9.960250854492188, "learning_rate": 3.7475000000000005e-06, "loss": 2.2922, "step": 7495 }, { "epoch": 2.349843260188088, "grad_norm": 10.117279052734375, "learning_rate": 3.7480000000000004e-06, "loss": 2.5394, "step": 7496 }, { "epoch": 2.3501567398119123, "grad_norm": 9.93826675415039, "learning_rate": 3.7485000000000006e-06, "loss": 2.1615, "step": 7497 }, { "epoch": 2.3504702194357368, "grad_norm": 20.171283721923828, "learning_rate": 3.7490000000000004e-06, "loss": 2.2555, "step": 7498 }, { "epoch": 2.350783699059561, "grad_norm": 9.589591026306152, "learning_rate": 3.7495000000000003e-06, "loss": 2.2651, "step": 7499 }, { "epoch": 2.3510971786833856, "grad_norm": 13.686689376831055, "learning_rate": 3.7500000000000005e-06, "loss": 2.0974, "step": 7500 }, { "epoch": 2.35141065830721, "grad_norm": 11.247666358947754, "learning_rate": 3.7505000000000003e-06, "loss": 1.9652, "step": 7501 }, { "epoch": 2.3517241379310345, "grad_norm": 106.85468292236328, "learning_rate": 3.751e-06, "loss": 2.8487, "step": 7502 }, { "epoch": 2.352037617554859, "grad_norm": 6.946944236755371, "learning_rate": 3.7515e-06, "loss": 2.0811, "step": 7503 }, { "epoch": 2.3523510971786834, "grad_norm": 62.64019775390625, "learning_rate": 3.7520000000000002e-06, "loss": 2.3686, "step": 7504 }, { "epoch": 2.352664576802508, "grad_norm": 26.0736026763916, "learning_rate": 3.7525e-06, "loss": 2.4541, "step": 7505 }, { "epoch": 2.3529780564263323, "grad_norm": 11.253118515014648, "learning_rate": 3.7530000000000003e-06, "loss": 2.4254, "step": 7506 }, { "epoch": 2.3532915360501567, "grad_norm": 24.771671295166016, "learning_rate": 3.7535000000000006e-06, "loss": 2.1042, "step": 7507 }, { "epoch": 2.353605015673981, "grad_norm": 7.764426231384277, "learning_rate": 3.7540000000000004e-06, "loss": 2.26, "step": 7508 }, { "epoch": 2.3539184952978056, "grad_norm": 47.99844741821289, "learning_rate": 3.7545000000000002e-06, "loss": 2.4344, "step": 7509 }, { "epoch": 2.35423197492163, "grad_norm": 16.0629825592041, "learning_rate": 3.7550000000000005e-06, "loss": 2.3399, "step": 7510 }, { "epoch": 2.3545454545454545, "grad_norm": 70.71231079101562, "learning_rate": 3.7555000000000003e-06, "loss": 2.1898, "step": 7511 }, { "epoch": 2.354858934169279, "grad_norm": 10.193946838378906, "learning_rate": 3.756e-06, "loss": 2.6066, "step": 7512 }, { "epoch": 2.3551724137931034, "grad_norm": 15.010258674621582, "learning_rate": 3.7565e-06, "loss": 3.0704, "step": 7513 }, { "epoch": 2.355485893416928, "grad_norm": 44.823978424072266, "learning_rate": 3.757e-06, "loss": 2.3118, "step": 7514 }, { "epoch": 2.3557993730407523, "grad_norm": 14.34240436553955, "learning_rate": 3.7575e-06, "loss": 2.0778, "step": 7515 }, { "epoch": 2.3561128526645767, "grad_norm": 26.604570388793945, "learning_rate": 3.7580000000000007e-06, "loss": 2.2868, "step": 7516 }, { "epoch": 2.356426332288401, "grad_norm": 107.75370025634766, "learning_rate": 3.7585000000000005e-06, "loss": 2.8349, "step": 7517 }, { "epoch": 2.3567398119122256, "grad_norm": 6.907593727111816, "learning_rate": 3.7590000000000003e-06, "loss": 2.1112, "step": 7518 }, { "epoch": 2.35705329153605, "grad_norm": 12.936373710632324, "learning_rate": 3.7595e-06, "loss": 1.9877, "step": 7519 }, { "epoch": 2.3573667711598745, "grad_norm": 11.473402976989746, "learning_rate": 3.7600000000000004e-06, "loss": 2.0683, "step": 7520 }, { "epoch": 2.357680250783699, "grad_norm": 8.270030975341797, "learning_rate": 3.7605000000000002e-06, "loss": 2.0443, "step": 7521 }, { "epoch": 2.3579937304075234, "grad_norm": 11.990708351135254, "learning_rate": 3.761e-06, "loss": 2.2018, "step": 7522 }, { "epoch": 2.358307210031348, "grad_norm": 72.41227722167969, "learning_rate": 3.7615000000000003e-06, "loss": 2.2836, "step": 7523 }, { "epoch": 2.3586206896551722, "grad_norm": 29.735172271728516, "learning_rate": 3.762e-06, "loss": 2.5709, "step": 7524 }, { "epoch": 2.3589341692789967, "grad_norm": 12.274335861206055, "learning_rate": 3.7625e-06, "loss": 2.6152, "step": 7525 }, { "epoch": 2.359247648902821, "grad_norm": 10.732202529907227, "learning_rate": 3.7630000000000006e-06, "loss": 2.2102, "step": 7526 }, { "epoch": 2.3595611285266456, "grad_norm": 11.891633987426758, "learning_rate": 3.7635000000000005e-06, "loss": 2.0766, "step": 7527 }, { "epoch": 2.35987460815047, "grad_norm": 104.39857482910156, "learning_rate": 3.7640000000000003e-06, "loss": 2.1575, "step": 7528 }, { "epoch": 2.3601880877742945, "grad_norm": 97.5882568359375, "learning_rate": 3.7645000000000005e-06, "loss": 2.3062, "step": 7529 }, { "epoch": 2.360501567398119, "grad_norm": 93.4491958618164, "learning_rate": 3.7650000000000004e-06, "loss": 2.2928, "step": 7530 }, { "epoch": 2.3608150470219433, "grad_norm": 9.98805046081543, "learning_rate": 3.7655e-06, "loss": 2.2852, "step": 7531 }, { "epoch": 2.3611285266457678, "grad_norm": 8.228720664978027, "learning_rate": 3.766e-06, "loss": 2.0596, "step": 7532 }, { "epoch": 2.3614420062695927, "grad_norm": 15.195112228393555, "learning_rate": 3.7665000000000003e-06, "loss": 2.3139, "step": 7533 }, { "epoch": 2.361755485893417, "grad_norm": 10.088196754455566, "learning_rate": 3.767e-06, "loss": 2.2944, "step": 7534 }, { "epoch": 2.3620689655172415, "grad_norm": 14.441904067993164, "learning_rate": 3.7675e-06, "loss": 2.4763, "step": 7535 }, { "epoch": 2.362382445141066, "grad_norm": 20.123929977416992, "learning_rate": 3.7680000000000006e-06, "loss": 2.1489, "step": 7536 }, { "epoch": 2.3626959247648904, "grad_norm": 11.031922340393066, "learning_rate": 3.7685000000000004e-06, "loss": 2.4172, "step": 7537 }, { "epoch": 2.363009404388715, "grad_norm": 22.59804344177246, "learning_rate": 3.7690000000000003e-06, "loss": 2.313, "step": 7538 }, { "epoch": 2.3633228840125393, "grad_norm": 7.959091663360596, "learning_rate": 3.7695000000000005e-06, "loss": 2.2106, "step": 7539 }, { "epoch": 2.3636363636363638, "grad_norm": 21.194488525390625, "learning_rate": 3.7700000000000003e-06, "loss": 2.1134, "step": 7540 }, { "epoch": 2.363949843260188, "grad_norm": 12.03039264678955, "learning_rate": 3.7705e-06, "loss": 2.1883, "step": 7541 }, { "epoch": 2.3642633228840126, "grad_norm": 18.800586700439453, "learning_rate": 3.7710000000000004e-06, "loss": 2.6216, "step": 7542 }, { "epoch": 2.364576802507837, "grad_norm": 17.500911712646484, "learning_rate": 3.7715000000000002e-06, "loss": 1.8558, "step": 7543 }, { "epoch": 2.3648902821316615, "grad_norm": 10.920929908752441, "learning_rate": 3.772e-06, "loss": 2.1468, "step": 7544 }, { "epoch": 2.365203761755486, "grad_norm": 22.849397659301758, "learning_rate": 3.7725e-06, "loss": 2.3309, "step": 7545 }, { "epoch": 2.3655172413793104, "grad_norm": 11.280135154724121, "learning_rate": 3.7730000000000006e-06, "loss": 2.2391, "step": 7546 }, { "epoch": 2.365830721003135, "grad_norm": 10.751194953918457, "learning_rate": 3.7735000000000004e-06, "loss": 2.2596, "step": 7547 }, { "epoch": 2.3661442006269593, "grad_norm": 9.03420352935791, "learning_rate": 3.7740000000000006e-06, "loss": 2.1899, "step": 7548 }, { "epoch": 2.3664576802507837, "grad_norm": 8.728597640991211, "learning_rate": 3.7745000000000005e-06, "loss": 2.0032, "step": 7549 }, { "epoch": 2.366771159874608, "grad_norm": 28.473175048828125, "learning_rate": 3.7750000000000003e-06, "loss": 2.2924, "step": 7550 }, { "epoch": 2.3670846394984326, "grad_norm": 41.957313537597656, "learning_rate": 3.7755e-06, "loss": 2.8044, "step": 7551 }, { "epoch": 2.367398119122257, "grad_norm": 9.625332832336426, "learning_rate": 3.7760000000000004e-06, "loss": 2.2616, "step": 7552 }, { "epoch": 2.3677115987460815, "grad_norm": 12.028620719909668, "learning_rate": 3.7765e-06, "loss": 2.1488, "step": 7553 }, { "epoch": 2.368025078369906, "grad_norm": 11.937979698181152, "learning_rate": 3.777e-06, "loss": 2.2888, "step": 7554 }, { "epoch": 2.3683385579937304, "grad_norm": 8.078612327575684, "learning_rate": 3.7775000000000003e-06, "loss": 2.1348, "step": 7555 }, { "epoch": 2.368652037617555, "grad_norm": 15.714337348937988, "learning_rate": 3.7780000000000005e-06, "loss": 2.2854, "step": 7556 }, { "epoch": 2.3689655172413793, "grad_norm": 14.439678192138672, "learning_rate": 3.7785000000000003e-06, "loss": 2.4233, "step": 7557 }, { "epoch": 2.3692789968652037, "grad_norm": 18.254709243774414, "learning_rate": 3.7790000000000006e-06, "loss": 2.0725, "step": 7558 }, { "epoch": 2.369592476489028, "grad_norm": 39.31232452392578, "learning_rate": 3.7795000000000004e-06, "loss": 2.6343, "step": 7559 }, { "epoch": 2.3699059561128526, "grad_norm": 11.581707954406738, "learning_rate": 3.7800000000000002e-06, "loss": 2.1901, "step": 7560 }, { "epoch": 2.370219435736677, "grad_norm": 50.96880340576172, "learning_rate": 3.7805000000000005e-06, "loss": 2.9012, "step": 7561 }, { "epoch": 2.3705329153605015, "grad_norm": 10.946471214294434, "learning_rate": 3.7810000000000003e-06, "loss": 2.1995, "step": 7562 }, { "epoch": 2.370846394984326, "grad_norm": 14.892229080200195, "learning_rate": 3.7815e-06, "loss": 2.2715, "step": 7563 }, { "epoch": 2.3711598746081504, "grad_norm": 7.052079200744629, "learning_rate": 3.782e-06, "loss": 2.1672, "step": 7564 }, { "epoch": 2.371473354231975, "grad_norm": 18.24445343017578, "learning_rate": 3.7825e-06, "loss": 2.7559, "step": 7565 }, { "epoch": 2.3717868338557992, "grad_norm": 12.191713333129883, "learning_rate": 3.7830000000000005e-06, "loss": 2.0391, "step": 7566 }, { "epoch": 2.3721003134796237, "grad_norm": 11.00248908996582, "learning_rate": 3.7835000000000007e-06, "loss": 2.0831, "step": 7567 }, { "epoch": 2.372413793103448, "grad_norm": 29.58919334411621, "learning_rate": 3.7840000000000005e-06, "loss": 2.1157, "step": 7568 }, { "epoch": 2.3727272727272726, "grad_norm": 18.105939865112305, "learning_rate": 3.7845000000000004e-06, "loss": 2.2601, "step": 7569 }, { "epoch": 2.373040752351097, "grad_norm": 13.980188369750977, "learning_rate": 3.785e-06, "loss": 2.2157, "step": 7570 }, { "epoch": 2.3733542319749215, "grad_norm": 8.464113235473633, "learning_rate": 3.7855000000000004e-06, "loss": 2.3421, "step": 7571 }, { "epoch": 2.373667711598746, "grad_norm": 10.430408477783203, "learning_rate": 3.7860000000000003e-06, "loss": 2.1068, "step": 7572 }, { "epoch": 2.3739811912225703, "grad_norm": 8.265875816345215, "learning_rate": 3.7865e-06, "loss": 2.0397, "step": 7573 }, { "epoch": 2.3742946708463952, "grad_norm": 11.182906150817871, "learning_rate": 3.787e-06, "loss": 2.2343, "step": 7574 }, { "epoch": 2.3746081504702197, "grad_norm": 12.923723220825195, "learning_rate": 3.7875e-06, "loss": 2.3455, "step": 7575 }, { "epoch": 2.374921630094044, "grad_norm": 6.568053722381592, "learning_rate": 3.7880000000000004e-06, "loss": 2.3191, "step": 7576 }, { "epoch": 2.3752351097178686, "grad_norm": 19.378129959106445, "learning_rate": 3.7885000000000007e-06, "loss": 2.4547, "step": 7577 }, { "epoch": 2.375548589341693, "grad_norm": 10.531891822814941, "learning_rate": 3.7890000000000005e-06, "loss": 1.9784, "step": 7578 }, { "epoch": 2.3758620689655174, "grad_norm": 26.610008239746094, "learning_rate": 3.7895000000000003e-06, "loss": 2.1666, "step": 7579 }, { "epoch": 2.376175548589342, "grad_norm": 7.623846530914307, "learning_rate": 3.79e-06, "loss": 2.0211, "step": 7580 }, { "epoch": 2.3764890282131663, "grad_norm": 67.45047760009766, "learning_rate": 3.7905000000000004e-06, "loss": 2.2968, "step": 7581 }, { "epoch": 2.3768025078369908, "grad_norm": 7.856488227844238, "learning_rate": 3.791e-06, "loss": 2.265, "step": 7582 }, { "epoch": 2.377115987460815, "grad_norm": 103.0473861694336, "learning_rate": 3.7915e-06, "loss": 2.6797, "step": 7583 }, { "epoch": 2.3774294670846396, "grad_norm": 124.34622955322266, "learning_rate": 3.7920000000000003e-06, "loss": 2.0861, "step": 7584 }, { "epoch": 2.377742946708464, "grad_norm": 386.7806091308594, "learning_rate": 3.7925e-06, "loss": 3.0252, "step": 7585 }, { "epoch": 2.3780564263322885, "grad_norm": 11.590187072753906, "learning_rate": 3.7930000000000004e-06, "loss": 2.0093, "step": 7586 }, { "epoch": 2.378369905956113, "grad_norm": 7.444549560546875, "learning_rate": 3.7935000000000006e-06, "loss": 2.1162, "step": 7587 }, { "epoch": 2.3786833855799374, "grad_norm": 11.169967651367188, "learning_rate": 3.7940000000000004e-06, "loss": 2.1076, "step": 7588 }, { "epoch": 2.378996865203762, "grad_norm": 18.482463836669922, "learning_rate": 3.7945000000000003e-06, "loss": 2.2238, "step": 7589 }, { "epoch": 2.3793103448275863, "grad_norm": 11.902023315429688, "learning_rate": 3.7950000000000005e-06, "loss": 2.2008, "step": 7590 }, { "epoch": 2.3796238244514107, "grad_norm": 6.9084906578063965, "learning_rate": 3.7955000000000003e-06, "loss": 2.1053, "step": 7591 }, { "epoch": 2.379937304075235, "grad_norm": 10.605999946594238, "learning_rate": 3.796e-06, "loss": 2.2262, "step": 7592 }, { "epoch": 2.3802507836990596, "grad_norm": 9.843685150146484, "learning_rate": 3.7965e-06, "loss": 1.8289, "step": 7593 }, { "epoch": 2.380564263322884, "grad_norm": 52.68415069580078, "learning_rate": 3.7970000000000002e-06, "loss": 1.7944, "step": 7594 }, { "epoch": 2.3808777429467085, "grad_norm": 7.418448448181152, "learning_rate": 3.7975e-06, "loss": 1.9969, "step": 7595 }, { "epoch": 2.381191222570533, "grad_norm": 13.815657615661621, "learning_rate": 3.7980000000000007e-06, "loss": 2.0561, "step": 7596 }, { "epoch": 2.3815047021943574, "grad_norm": 9.356049537658691, "learning_rate": 3.7985000000000006e-06, "loss": 2.1875, "step": 7597 }, { "epoch": 2.381818181818182, "grad_norm": 9.415925979614258, "learning_rate": 3.7990000000000004e-06, "loss": 2.2282, "step": 7598 }, { "epoch": 2.3821316614420063, "grad_norm": 6.984994411468506, "learning_rate": 3.7995000000000002e-06, "loss": 2.1025, "step": 7599 }, { "epoch": 2.3824451410658307, "grad_norm": 11.518677711486816, "learning_rate": 3.8000000000000005e-06, "loss": 2.2879, "step": 7600 }, { "epoch": 2.382758620689655, "grad_norm": 9.700697898864746, "learning_rate": 3.8005000000000003e-06, "loss": 2.2406, "step": 7601 }, { "epoch": 2.3830721003134796, "grad_norm": 7.200320243835449, "learning_rate": 3.801e-06, "loss": 2.1947, "step": 7602 }, { "epoch": 2.383385579937304, "grad_norm": 10.102734565734863, "learning_rate": 3.8015000000000004e-06, "loss": 2.2103, "step": 7603 }, { "epoch": 2.3836990595611285, "grad_norm": 9.612333297729492, "learning_rate": 3.802e-06, "loss": 2.2205, "step": 7604 }, { "epoch": 2.384012539184953, "grad_norm": 10.362100601196289, "learning_rate": 3.8025e-06, "loss": 2.1581, "step": 7605 }, { "epoch": 2.3843260188087774, "grad_norm": 11.039507865905762, "learning_rate": 3.8030000000000007e-06, "loss": 1.7934, "step": 7606 }, { "epoch": 2.384639498432602, "grad_norm": 13.387303352355957, "learning_rate": 3.8035000000000005e-06, "loss": 2.08, "step": 7607 }, { "epoch": 2.3849529780564263, "grad_norm": 9.778520584106445, "learning_rate": 3.8040000000000003e-06, "loss": 1.9611, "step": 7608 }, { "epoch": 2.3852664576802507, "grad_norm": 42.11952209472656, "learning_rate": 3.8045000000000006e-06, "loss": 2.1163, "step": 7609 }, { "epoch": 2.385579937304075, "grad_norm": 9.9139986038208, "learning_rate": 3.8050000000000004e-06, "loss": 2.2484, "step": 7610 }, { "epoch": 2.3858934169278996, "grad_norm": 6.559987545013428, "learning_rate": 3.8055000000000002e-06, "loss": 2.2596, "step": 7611 }, { "epoch": 2.386206896551724, "grad_norm": 16.624814987182617, "learning_rate": 3.806e-06, "loss": 2.443, "step": 7612 }, { "epoch": 2.3865203761755485, "grad_norm": 8.37903118133545, "learning_rate": 3.8065000000000003e-06, "loss": 2.1529, "step": 7613 }, { "epoch": 2.386833855799373, "grad_norm": 14.952163696289062, "learning_rate": 3.807e-06, "loss": 2.1651, "step": 7614 }, { "epoch": 2.3871473354231973, "grad_norm": 17.216014862060547, "learning_rate": 3.8075e-06, "loss": 2.2995, "step": 7615 }, { "epoch": 2.387460815047022, "grad_norm": 47.170841217041016, "learning_rate": 3.8080000000000006e-06, "loss": 2.056, "step": 7616 }, { "epoch": 2.3877742946708462, "grad_norm": 37.809017181396484, "learning_rate": 3.8085000000000005e-06, "loss": 2.3327, "step": 7617 }, { "epoch": 2.3880877742946707, "grad_norm": 86.04744720458984, "learning_rate": 3.8090000000000003e-06, "loss": 2.3283, "step": 7618 }, { "epoch": 2.388401253918495, "grad_norm": 83.6880874633789, "learning_rate": 3.8095000000000005e-06, "loss": 2.3206, "step": 7619 }, { "epoch": 2.3887147335423196, "grad_norm": 16.333332061767578, "learning_rate": 3.8100000000000004e-06, "loss": 2.303, "step": 7620 }, { "epoch": 2.389028213166144, "grad_norm": 47.766605377197266, "learning_rate": 3.8105e-06, "loss": 2.18, "step": 7621 }, { "epoch": 2.3893416927899684, "grad_norm": 12.504522323608398, "learning_rate": 3.8110000000000004e-06, "loss": 2.0407, "step": 7622 }, { "epoch": 2.389655172413793, "grad_norm": 13.140955924987793, "learning_rate": 3.8115000000000003e-06, "loss": 2.2059, "step": 7623 }, { "epoch": 2.3899686520376173, "grad_norm": 37.38854217529297, "learning_rate": 3.812e-06, "loss": 2.2309, "step": 7624 }, { "epoch": 2.3902821316614418, "grad_norm": 16.521257400512695, "learning_rate": 3.8125e-06, "loss": 1.8793, "step": 7625 }, { "epoch": 2.390595611285266, "grad_norm": 24.827463150024414, "learning_rate": 3.813e-06, "loss": 2.1203, "step": 7626 }, { "epoch": 2.390909090909091, "grad_norm": 21.19013786315918, "learning_rate": 3.8135000000000004e-06, "loss": 2.1855, "step": 7627 }, { "epoch": 2.3912225705329155, "grad_norm": 44.844520568847656, "learning_rate": 3.8140000000000007e-06, "loss": 2.3199, "step": 7628 }, { "epoch": 2.39153605015674, "grad_norm": 14.009591102600098, "learning_rate": 3.8145000000000005e-06, "loss": 2.1491, "step": 7629 }, { "epoch": 2.3918495297805644, "grad_norm": 9.335655212402344, "learning_rate": 3.815000000000001e-06, "loss": 2.2375, "step": 7630 }, { "epoch": 2.392163009404389, "grad_norm": 10.362065315246582, "learning_rate": 3.8155e-06, "loss": 2.2177, "step": 7631 }, { "epoch": 2.3924764890282133, "grad_norm": 11.755572319030762, "learning_rate": 3.816e-06, "loss": 2.121, "step": 7632 }, { "epoch": 2.3927899686520377, "grad_norm": 10.188096046447754, "learning_rate": 3.8165e-06, "loss": 2.1035, "step": 7633 }, { "epoch": 2.393103448275862, "grad_norm": 29.14549446105957, "learning_rate": 3.817e-06, "loss": 2.1825, "step": 7634 }, { "epoch": 2.3934169278996866, "grad_norm": 8.220953941345215, "learning_rate": 3.8175e-06, "loss": 1.9614, "step": 7635 }, { "epoch": 2.393730407523511, "grad_norm": 11.84253978729248, "learning_rate": 3.818e-06, "loss": 2.052, "step": 7636 }, { "epoch": 2.3940438871473355, "grad_norm": 13.948365211486816, "learning_rate": 3.818500000000001e-06, "loss": 2.9696, "step": 7637 }, { "epoch": 2.39435736677116, "grad_norm": 68.55855560302734, "learning_rate": 3.819e-06, "loss": 2.4433, "step": 7638 }, { "epoch": 2.3946708463949844, "grad_norm": 10.881951332092285, "learning_rate": 3.8195000000000004e-06, "loss": 1.91, "step": 7639 }, { "epoch": 2.394984326018809, "grad_norm": 45.97584915161133, "learning_rate": 3.820000000000001e-06, "loss": 2.0456, "step": 7640 }, { "epoch": 2.3952978056426333, "grad_norm": 9.28108024597168, "learning_rate": 3.8205e-06, "loss": 2.0773, "step": 7641 }, { "epoch": 2.3956112852664577, "grad_norm": 36.159698486328125, "learning_rate": 3.821e-06, "loss": 2.3127, "step": 7642 }, { "epoch": 2.395924764890282, "grad_norm": 15.76865005493164, "learning_rate": 3.821500000000001e-06, "loss": 2.2897, "step": 7643 }, { "epoch": 2.3962382445141066, "grad_norm": 15.745697975158691, "learning_rate": 3.822e-06, "loss": 2.0437, "step": 7644 }, { "epoch": 2.396551724137931, "grad_norm": 14.115145683288574, "learning_rate": 3.8225e-06, "loss": 2.4627, "step": 7645 }, { "epoch": 2.3968652037617555, "grad_norm": 9.220922470092773, "learning_rate": 3.823e-06, "loss": 2.0782, "step": 7646 }, { "epoch": 2.39717868338558, "grad_norm": 10.528179168701172, "learning_rate": 3.823500000000001e-06, "loss": 2.2822, "step": 7647 }, { "epoch": 2.3974921630094044, "grad_norm": 17.768251419067383, "learning_rate": 3.824e-06, "loss": 2.8272, "step": 7648 }, { "epoch": 2.397805642633229, "grad_norm": 11.23782730102539, "learning_rate": 3.8245e-06, "loss": 2.1334, "step": 7649 }, { "epoch": 2.3981191222570533, "grad_norm": 15.89545726776123, "learning_rate": 3.825000000000001e-06, "loss": 2.0382, "step": 7650 }, { "epoch": 2.3984326018808777, "grad_norm": 10.835371017456055, "learning_rate": 3.8255e-06, "loss": 2.1747, "step": 7651 }, { "epoch": 2.398746081504702, "grad_norm": 8.47184944152832, "learning_rate": 3.826e-06, "loss": 2.1394, "step": 7652 }, { "epoch": 2.3990595611285266, "grad_norm": 11.712238311767578, "learning_rate": 3.8265000000000005e-06, "loss": 2.1518, "step": 7653 }, { "epoch": 2.399373040752351, "grad_norm": 18.89401626586914, "learning_rate": 3.827e-06, "loss": 2.1884, "step": 7654 }, { "epoch": 2.3996865203761755, "grad_norm": 54.23405075073242, "learning_rate": 3.8275e-06, "loss": 2.0512, "step": 7655 }, { "epoch": 2.4, "grad_norm": 30.92529296875, "learning_rate": 3.8280000000000004e-06, "loss": 2.2182, "step": 7656 }, { "epoch": 2.4003134796238244, "grad_norm": 14.092068672180176, "learning_rate": 3.828500000000001e-06, "loss": 1.8617, "step": 7657 }, { "epoch": 2.400626959247649, "grad_norm": 9.441667556762695, "learning_rate": 3.829e-06, "loss": 2.1969, "step": 7658 }, { "epoch": 2.4009404388714732, "grad_norm": 15.503475189208984, "learning_rate": 3.8295e-06, "loss": 2.0171, "step": 7659 }, { "epoch": 2.4012539184952977, "grad_norm": 11.037444114685059, "learning_rate": 3.830000000000001e-06, "loss": 2.2621, "step": 7660 }, { "epoch": 2.401567398119122, "grad_norm": 10.93976879119873, "learning_rate": 3.8305e-06, "loss": 2.137, "step": 7661 }, { "epoch": 2.4018808777429466, "grad_norm": 10.128861427307129, "learning_rate": 3.831e-06, "loss": 2.0624, "step": 7662 }, { "epoch": 2.402194357366771, "grad_norm": 11.597820281982422, "learning_rate": 3.8315000000000005e-06, "loss": 2.0284, "step": 7663 }, { "epoch": 2.4025078369905954, "grad_norm": 14.918848991394043, "learning_rate": 3.832e-06, "loss": 2.2572, "step": 7664 }, { "epoch": 2.40282131661442, "grad_norm": 9.468354225158691, "learning_rate": 3.8325e-06, "loss": 2.0405, "step": 7665 }, { "epoch": 2.4031347962382443, "grad_norm": 7.677601337432861, "learning_rate": 3.833e-06, "loss": 2.0677, "step": 7666 }, { "epoch": 2.4034482758620688, "grad_norm": 7.7597126960754395, "learning_rate": 3.833500000000001e-06, "loss": 2.2207, "step": 7667 }, { "epoch": 2.4037617554858937, "grad_norm": 9.098000526428223, "learning_rate": 3.834000000000001e-06, "loss": 2.1647, "step": 7668 }, { "epoch": 2.404075235109718, "grad_norm": 6.249232292175293, "learning_rate": 3.8345e-06, "loss": 2.2209, "step": 7669 }, { "epoch": 2.4043887147335425, "grad_norm": 28.894777297973633, "learning_rate": 3.8350000000000006e-06, "loss": 2.0591, "step": 7670 }, { "epoch": 2.404702194357367, "grad_norm": 10.601434707641602, "learning_rate": 3.8355e-06, "loss": 2.1746, "step": 7671 }, { "epoch": 2.4050156739811914, "grad_norm": 17.692670822143555, "learning_rate": 3.836e-06, "loss": 2.27, "step": 7672 }, { "epoch": 2.405329153605016, "grad_norm": 7.8094096183776855, "learning_rate": 3.8365000000000005e-06, "loss": 2.0961, "step": 7673 }, { "epoch": 2.4056426332288403, "grad_norm": 64.30963897705078, "learning_rate": 3.837e-06, "loss": 2.1397, "step": 7674 }, { "epoch": 2.4059561128526648, "grad_norm": 12.039095878601074, "learning_rate": 3.8375e-06, "loss": 2.102, "step": 7675 }, { "epoch": 2.406269592476489, "grad_norm": 9.88943099975586, "learning_rate": 3.838e-06, "loss": 2.3207, "step": 7676 }, { "epoch": 2.4065830721003136, "grad_norm": 221.2774200439453, "learning_rate": 3.838500000000001e-06, "loss": 2.4329, "step": 7677 }, { "epoch": 2.406896551724138, "grad_norm": 7.631561279296875, "learning_rate": 3.839000000000001e-06, "loss": 2.1533, "step": 7678 }, { "epoch": 2.4072100313479625, "grad_norm": 9.12175464630127, "learning_rate": 3.8395e-06, "loss": 2.376, "step": 7679 }, { "epoch": 2.407523510971787, "grad_norm": 11.242121696472168, "learning_rate": 3.8400000000000005e-06, "loss": 2.1876, "step": 7680 }, { "epoch": 2.4078369905956114, "grad_norm": 10.507530212402344, "learning_rate": 3.840500000000001e-06, "loss": 2.1519, "step": 7681 }, { "epoch": 2.408150470219436, "grad_norm": 33.44599533081055, "learning_rate": 3.841e-06, "loss": 2.3341, "step": 7682 }, { "epoch": 2.4084639498432603, "grad_norm": 9.884668350219727, "learning_rate": 3.8415e-06, "loss": 2.0574, "step": 7683 }, { "epoch": 2.4087774294670847, "grad_norm": 16.659543991088867, "learning_rate": 3.842e-06, "loss": 2.2833, "step": 7684 }, { "epoch": 2.409090909090909, "grad_norm": 9.122176170349121, "learning_rate": 3.8425e-06, "loss": 2.038, "step": 7685 }, { "epoch": 2.4094043887147336, "grad_norm": 14.15365982055664, "learning_rate": 3.843e-06, "loss": 2.3236, "step": 7686 }, { "epoch": 2.409717868338558, "grad_norm": 14.537882804870605, "learning_rate": 3.8435000000000006e-06, "loss": 1.9546, "step": 7687 }, { "epoch": 2.4100313479623825, "grad_norm": 36.1397590637207, "learning_rate": 3.844000000000001e-06, "loss": 2.0064, "step": 7688 }, { "epoch": 2.410344827586207, "grad_norm": 16.462356567382812, "learning_rate": 3.8445e-06, "loss": 2.1094, "step": 7689 }, { "epoch": 2.4106583072100314, "grad_norm": 11.6400728225708, "learning_rate": 3.8450000000000005e-06, "loss": 1.9072, "step": 7690 }, { "epoch": 2.410971786833856, "grad_norm": 13.747432708740234, "learning_rate": 3.845500000000001e-06, "loss": 2.1082, "step": 7691 }, { "epoch": 2.4112852664576803, "grad_norm": 15.540373802185059, "learning_rate": 3.846e-06, "loss": 2.6436, "step": 7692 }, { "epoch": 2.4115987460815047, "grad_norm": 6.160389423370361, "learning_rate": 3.8465e-06, "loss": 2.1155, "step": 7693 }, { "epoch": 2.411912225705329, "grad_norm": 8.378205299377441, "learning_rate": 3.847e-06, "loss": 2.1946, "step": 7694 }, { "epoch": 2.4122257053291536, "grad_norm": 11.588189125061035, "learning_rate": 3.8475e-06, "loss": 2.1241, "step": 7695 }, { "epoch": 2.412539184952978, "grad_norm": 13.196329116821289, "learning_rate": 3.848e-06, "loss": 2.0537, "step": 7696 }, { "epoch": 2.4128526645768025, "grad_norm": 8.22701358795166, "learning_rate": 3.8485000000000005e-06, "loss": 2.1139, "step": 7697 }, { "epoch": 2.413166144200627, "grad_norm": 11.089346885681152, "learning_rate": 3.849000000000001e-06, "loss": 2.2044, "step": 7698 }, { "epoch": 2.4134796238244514, "grad_norm": 19.3448429107666, "learning_rate": 3.8495e-06, "loss": 2.4141, "step": 7699 }, { "epoch": 2.413793103448276, "grad_norm": 11.318143844604492, "learning_rate": 3.85e-06, "loss": 2.1163, "step": 7700 }, { "epoch": 2.4141065830721002, "grad_norm": 14.12908935546875, "learning_rate": 3.850500000000001e-06, "loss": 2.2633, "step": 7701 }, { "epoch": 2.4144200626959247, "grad_norm": 13.211986541748047, "learning_rate": 3.851e-06, "loss": 2.2258, "step": 7702 }, { "epoch": 2.414733542319749, "grad_norm": 21.518695831298828, "learning_rate": 3.8515e-06, "loss": 2.2675, "step": 7703 }, { "epoch": 2.4150470219435736, "grad_norm": 12.789783477783203, "learning_rate": 3.8520000000000006e-06, "loss": 1.979, "step": 7704 }, { "epoch": 2.415360501567398, "grad_norm": 22.694408416748047, "learning_rate": 3.8525e-06, "loss": 2.7114, "step": 7705 }, { "epoch": 2.4156739811912225, "grad_norm": 9.9482421875, "learning_rate": 3.853e-06, "loss": 2.1214, "step": 7706 }, { "epoch": 2.415987460815047, "grad_norm": 33.53236389160156, "learning_rate": 3.8535000000000005e-06, "loss": 2.1029, "step": 7707 }, { "epoch": 2.4163009404388713, "grad_norm": 15.033732414245605, "learning_rate": 3.854000000000001e-06, "loss": 2.4906, "step": 7708 }, { "epoch": 2.416614420062696, "grad_norm": 9.195804595947266, "learning_rate": 3.8545e-06, "loss": 2.1426, "step": 7709 }, { "epoch": 2.41692789968652, "grad_norm": 49.578792572021484, "learning_rate": 3.855e-06, "loss": 2.2275, "step": 7710 }, { "epoch": 2.4172413793103447, "grad_norm": 7.01850700378418, "learning_rate": 3.855500000000001e-06, "loss": 1.9436, "step": 7711 }, { "epoch": 2.417554858934169, "grad_norm": 58.17573547363281, "learning_rate": 3.856e-06, "loss": 2.4019, "step": 7712 }, { "epoch": 2.4178683385579935, "grad_norm": 17.604841232299805, "learning_rate": 3.8565e-06, "loss": 2.1166, "step": 7713 }, { "epoch": 2.418181818181818, "grad_norm": 9.961703300476074, "learning_rate": 3.8570000000000005e-06, "loss": 2.0159, "step": 7714 }, { "epoch": 2.4184952978056424, "grad_norm": 17.007003784179688, "learning_rate": 3.8575e-06, "loss": 2.463, "step": 7715 }, { "epoch": 2.418808777429467, "grad_norm": 11.631433486938477, "learning_rate": 3.858e-06, "loss": 2.1919, "step": 7716 }, { "epoch": 2.4191222570532913, "grad_norm": 12.101078033447266, "learning_rate": 3.8585e-06, "loss": 2.0336, "step": 7717 }, { "epoch": 2.4194357366771158, "grad_norm": 60.939361572265625, "learning_rate": 3.859000000000001e-06, "loss": 2.3531, "step": 7718 }, { "epoch": 2.41974921630094, "grad_norm": 9.468650817871094, "learning_rate": 3.8595e-06, "loss": 2.1325, "step": 7719 }, { "epoch": 2.420062695924765, "grad_norm": 61.29365921020508, "learning_rate": 3.86e-06, "loss": 1.8754, "step": 7720 }, { "epoch": 2.4203761755485895, "grad_norm": 8.288795471191406, "learning_rate": 3.8605000000000006e-06, "loss": 2.1316, "step": 7721 }, { "epoch": 2.420689655172414, "grad_norm": 13.33533000946045, "learning_rate": 3.861e-06, "loss": 2.3631, "step": 7722 }, { "epoch": 2.4210031347962384, "grad_norm": 56.10675048828125, "learning_rate": 3.8615e-06, "loss": 2.5585, "step": 7723 }, { "epoch": 2.421316614420063, "grad_norm": 36.284515380859375, "learning_rate": 3.8620000000000005e-06, "loss": 2.0605, "step": 7724 }, { "epoch": 2.4216300940438873, "grad_norm": 80.10411834716797, "learning_rate": 3.8625e-06, "loss": 2.1775, "step": 7725 }, { "epoch": 2.4219435736677117, "grad_norm": 7.244259834289551, "learning_rate": 3.863e-06, "loss": 2.3176, "step": 7726 }, { "epoch": 2.422257053291536, "grad_norm": 15.228224754333496, "learning_rate": 3.8635e-06, "loss": 2.4194, "step": 7727 }, { "epoch": 2.4225705329153606, "grad_norm": 8.223821640014648, "learning_rate": 3.864000000000001e-06, "loss": 1.9635, "step": 7728 }, { "epoch": 2.422884012539185, "grad_norm": 9.506672859191895, "learning_rate": 3.864500000000001e-06, "loss": 2.126, "step": 7729 }, { "epoch": 2.4231974921630095, "grad_norm": 9.310214042663574, "learning_rate": 3.865e-06, "loss": 2.1113, "step": 7730 }, { "epoch": 2.423510971786834, "grad_norm": 10.558013916015625, "learning_rate": 3.8655000000000005e-06, "loss": 1.6246, "step": 7731 }, { "epoch": 2.4238244514106584, "grad_norm": 21.110506057739258, "learning_rate": 3.866e-06, "loss": 2.4633, "step": 7732 }, { "epoch": 2.424137931034483, "grad_norm": 9.753863334655762, "learning_rate": 3.8665e-06, "loss": 2.3482, "step": 7733 }, { "epoch": 2.4244514106583073, "grad_norm": 27.95020866394043, "learning_rate": 3.867e-06, "loss": 2.5685, "step": 7734 }, { "epoch": 2.4247648902821317, "grad_norm": 112.76493072509766, "learning_rate": 3.8675e-06, "loss": 2.3865, "step": 7735 }, { "epoch": 2.425078369905956, "grad_norm": 15.534668922424316, "learning_rate": 3.868e-06, "loss": 2.264, "step": 7736 }, { "epoch": 2.4253918495297806, "grad_norm": 10.583703994750977, "learning_rate": 3.8685e-06, "loss": 2.3579, "step": 7737 }, { "epoch": 2.425705329153605, "grad_norm": 14.666104316711426, "learning_rate": 3.8690000000000006e-06, "loss": 2.0503, "step": 7738 }, { "epoch": 2.4260188087774295, "grad_norm": 4.730828762054443, "learning_rate": 3.869500000000001e-06, "loss": 2.1273, "step": 7739 }, { "epoch": 2.426332288401254, "grad_norm": 16.89728546142578, "learning_rate": 3.87e-06, "loss": 2.328, "step": 7740 }, { "epoch": 2.4266457680250784, "grad_norm": 11.676743507385254, "learning_rate": 3.8705000000000005e-06, "loss": 2.2042, "step": 7741 }, { "epoch": 2.426959247648903, "grad_norm": 22.43178939819336, "learning_rate": 3.871000000000001e-06, "loss": 2.2966, "step": 7742 }, { "epoch": 2.4272727272727272, "grad_norm": 25.351423263549805, "learning_rate": 3.8715e-06, "loss": 2.392, "step": 7743 }, { "epoch": 2.4275862068965517, "grad_norm": 22.675395965576172, "learning_rate": 3.872e-06, "loss": 2.1451, "step": 7744 }, { "epoch": 2.427899686520376, "grad_norm": 7.179481506347656, "learning_rate": 3.8725e-06, "loss": 2.33, "step": 7745 }, { "epoch": 2.4282131661442006, "grad_norm": 15.644564628601074, "learning_rate": 3.873e-06, "loss": 2.0476, "step": 7746 }, { "epoch": 2.428526645768025, "grad_norm": 9.65941333770752, "learning_rate": 3.8735e-06, "loss": 2.2586, "step": 7747 }, { "epoch": 2.4288401253918495, "grad_norm": 8.368847846984863, "learning_rate": 3.8740000000000005e-06, "loss": 2.0072, "step": 7748 }, { "epoch": 2.429153605015674, "grad_norm": 20.4112548828125, "learning_rate": 3.874500000000001e-06, "loss": 2.5314, "step": 7749 }, { "epoch": 2.4294670846394983, "grad_norm": 8.056722640991211, "learning_rate": 3.875e-06, "loss": 1.9804, "step": 7750 }, { "epoch": 2.429780564263323, "grad_norm": 7.814377784729004, "learning_rate": 3.8755e-06, "loss": 2.0505, "step": 7751 }, { "epoch": 2.4300940438871472, "grad_norm": 8.06814956665039, "learning_rate": 3.876000000000001e-06, "loss": 2.1465, "step": 7752 }, { "epoch": 2.4304075235109717, "grad_norm": 10.24548625946045, "learning_rate": 3.8765e-06, "loss": 2.1174, "step": 7753 }, { "epoch": 2.430721003134796, "grad_norm": 14.822341918945312, "learning_rate": 3.877e-06, "loss": 2.1173, "step": 7754 }, { "epoch": 2.4310344827586206, "grad_norm": 19.235980987548828, "learning_rate": 3.8775000000000006e-06, "loss": 2.292, "step": 7755 }, { "epoch": 2.431347962382445, "grad_norm": 8.851950645446777, "learning_rate": 3.878e-06, "loss": 1.9077, "step": 7756 }, { "epoch": 2.4316614420062694, "grad_norm": 7.941638946533203, "learning_rate": 3.8785e-06, "loss": 2.2116, "step": 7757 }, { "epoch": 2.431974921630094, "grad_norm": 8.783881187438965, "learning_rate": 3.8790000000000005e-06, "loss": 2.0873, "step": 7758 }, { "epoch": 2.4322884012539183, "grad_norm": 12.741527557373047, "learning_rate": 3.879500000000001e-06, "loss": 2.3999, "step": 7759 }, { "epoch": 2.4326018808777428, "grad_norm": 9.406828880310059, "learning_rate": 3.88e-06, "loss": 2.2064, "step": 7760 }, { "epoch": 2.4329153605015676, "grad_norm": 40.19290542602539, "learning_rate": 3.8805e-06, "loss": 1.8874, "step": 7761 }, { "epoch": 2.433228840125392, "grad_norm": 18.26624298095703, "learning_rate": 3.881000000000001e-06, "loss": 2.0759, "step": 7762 }, { "epoch": 2.4335423197492165, "grad_norm": 48.6797981262207, "learning_rate": 3.8815e-06, "loss": 2.3124, "step": 7763 }, { "epoch": 2.433855799373041, "grad_norm": 34.276580810546875, "learning_rate": 3.882e-06, "loss": 2.2452, "step": 7764 }, { "epoch": 2.4341692789968654, "grad_norm": 12.125838279724121, "learning_rate": 3.8825000000000005e-06, "loss": 2.2757, "step": 7765 }, { "epoch": 2.43448275862069, "grad_norm": 5.842771530151367, "learning_rate": 3.883e-06, "loss": 2.664, "step": 7766 }, { "epoch": 2.4347962382445143, "grad_norm": 40.553768157958984, "learning_rate": 3.8835e-06, "loss": 2.3799, "step": 7767 }, { "epoch": 2.4351097178683387, "grad_norm": 119.73785400390625, "learning_rate": 3.884e-06, "loss": 2.4882, "step": 7768 }, { "epoch": 2.435423197492163, "grad_norm": 11.483431816101074, "learning_rate": 3.884500000000001e-06, "loss": 2.0818, "step": 7769 }, { "epoch": 2.4357366771159876, "grad_norm": 68.69570922851562, "learning_rate": 3.885e-06, "loss": 2.6068, "step": 7770 }, { "epoch": 2.436050156739812, "grad_norm": 7.407710552215576, "learning_rate": 3.8855e-06, "loss": 2.1665, "step": 7771 }, { "epoch": 2.4363636363636365, "grad_norm": 12.153562545776367, "learning_rate": 3.8860000000000006e-06, "loss": 2.1593, "step": 7772 }, { "epoch": 2.436677115987461, "grad_norm": 7.465826988220215, "learning_rate": 3.8865e-06, "loss": 2.0294, "step": 7773 }, { "epoch": 2.4369905956112854, "grad_norm": 15.646428108215332, "learning_rate": 3.887e-06, "loss": 2.1772, "step": 7774 }, { "epoch": 2.43730407523511, "grad_norm": 16.8114013671875, "learning_rate": 3.8875000000000005e-06, "loss": 2.2299, "step": 7775 }, { "epoch": 2.4376175548589343, "grad_norm": 11.693188667297363, "learning_rate": 3.888e-06, "loss": 2.1271, "step": 7776 }, { "epoch": 2.4379310344827587, "grad_norm": 7.9489312171936035, "learning_rate": 3.8885e-06, "loss": 2.446, "step": 7777 }, { "epoch": 2.438244514106583, "grad_norm": 10.517236709594727, "learning_rate": 3.889e-06, "loss": 2.1916, "step": 7778 }, { "epoch": 2.4385579937304076, "grad_norm": 34.65458679199219, "learning_rate": 3.889500000000001e-06, "loss": 2.3929, "step": 7779 }, { "epoch": 2.438871473354232, "grad_norm": 9.605621337890625, "learning_rate": 3.89e-06, "loss": 1.8686, "step": 7780 }, { "epoch": 2.4391849529780565, "grad_norm": 15.1489896774292, "learning_rate": 3.8905e-06, "loss": 2.2183, "step": 7781 }, { "epoch": 2.439498432601881, "grad_norm": 12.255752563476562, "learning_rate": 3.8910000000000005e-06, "loss": 2.0652, "step": 7782 }, { "epoch": 2.4398119122257054, "grad_norm": 17.215301513671875, "learning_rate": 3.8915e-06, "loss": 1.9074, "step": 7783 }, { "epoch": 2.44012539184953, "grad_norm": 6.352075576782227, "learning_rate": 3.892e-06, "loss": 2.1003, "step": 7784 }, { "epoch": 2.4404388714733543, "grad_norm": 80.23092651367188, "learning_rate": 3.8925000000000004e-06, "loss": 2.3653, "step": 7785 }, { "epoch": 2.4407523510971787, "grad_norm": 23.931217193603516, "learning_rate": 3.893e-06, "loss": 2.1019, "step": 7786 }, { "epoch": 2.441065830721003, "grad_norm": 22.55059814453125, "learning_rate": 3.8935e-06, "loss": 2.295, "step": 7787 }, { "epoch": 2.4413793103448276, "grad_norm": 13.864753723144531, "learning_rate": 3.894e-06, "loss": 2.2982, "step": 7788 }, { "epoch": 2.441692789968652, "grad_norm": 16.332931518554688, "learning_rate": 3.8945000000000006e-06, "loss": 2.3005, "step": 7789 }, { "epoch": 2.4420062695924765, "grad_norm": 17.263086318969727, "learning_rate": 3.895000000000001e-06, "loss": 2.8135, "step": 7790 }, { "epoch": 2.442319749216301, "grad_norm": 14.815372467041016, "learning_rate": 3.8955e-06, "loss": 1.9764, "step": 7791 }, { "epoch": 2.4426332288401253, "grad_norm": 35.921024322509766, "learning_rate": 3.8960000000000005e-06, "loss": 2.2631, "step": 7792 }, { "epoch": 2.44294670846395, "grad_norm": 33.36460494995117, "learning_rate": 3.8965e-06, "loss": 2.0601, "step": 7793 }, { "epoch": 2.4432601880877742, "grad_norm": 21.6844539642334, "learning_rate": 3.897e-06, "loss": 2.2292, "step": 7794 }, { "epoch": 2.4435736677115987, "grad_norm": 6.807336330413818, "learning_rate": 3.8975e-06, "loss": 2.0116, "step": 7795 }, { "epoch": 2.443887147335423, "grad_norm": 8.875222206115723, "learning_rate": 3.898e-06, "loss": 2.6671, "step": 7796 }, { "epoch": 2.4442006269592476, "grad_norm": 11.693392753601074, "learning_rate": 3.8985e-06, "loss": 2.0421, "step": 7797 }, { "epoch": 2.444514106583072, "grad_norm": 21.314062118530273, "learning_rate": 3.899e-06, "loss": 2.1601, "step": 7798 }, { "epoch": 2.4448275862068964, "grad_norm": 17.729360580444336, "learning_rate": 3.8995000000000005e-06, "loss": 2.1047, "step": 7799 }, { "epoch": 2.445141065830721, "grad_norm": 23.925479888916016, "learning_rate": 3.900000000000001e-06, "loss": 2.4176, "step": 7800 }, { "epoch": 2.4454545454545453, "grad_norm": 28.251436233520508, "learning_rate": 3.9005e-06, "loss": 2.081, "step": 7801 }, { "epoch": 2.4457680250783698, "grad_norm": 9.291318893432617, "learning_rate": 3.9010000000000004e-06, "loss": 2.1789, "step": 7802 }, { "epoch": 2.446081504702194, "grad_norm": 6.612745761871338, "learning_rate": 3.901500000000001e-06, "loss": 2.1279, "step": 7803 }, { "epoch": 2.4463949843260187, "grad_norm": 11.001721382141113, "learning_rate": 3.902e-06, "loss": 1.9823, "step": 7804 }, { "epoch": 2.446708463949843, "grad_norm": 31.40390968322754, "learning_rate": 3.9025e-06, "loss": 2.3552, "step": 7805 }, { "epoch": 2.4470219435736675, "grad_norm": 15.96174144744873, "learning_rate": 3.903e-06, "loss": 2.0833, "step": 7806 }, { "epoch": 2.447335423197492, "grad_norm": 17.228862762451172, "learning_rate": 3.9035e-06, "loss": 2.2744, "step": 7807 }, { "epoch": 2.4476489028213164, "grad_norm": 15.133646011352539, "learning_rate": 3.904e-06, "loss": 2.3296, "step": 7808 }, { "epoch": 2.447962382445141, "grad_norm": 8.409405708312988, "learning_rate": 3.9045000000000005e-06, "loss": 2.0565, "step": 7809 }, { "epoch": 2.4482758620689653, "grad_norm": 7.684636116027832, "learning_rate": 3.905000000000001e-06, "loss": 2.1734, "step": 7810 }, { "epoch": 2.4485893416927897, "grad_norm": 77.93966674804688, "learning_rate": 3.9055e-06, "loss": 2.0164, "step": 7811 }, { "epoch": 2.448902821316614, "grad_norm": 12.741106033325195, "learning_rate": 3.906e-06, "loss": 2.2782, "step": 7812 }, { "epoch": 2.4492163009404386, "grad_norm": 20.681453704833984, "learning_rate": 3.906500000000001e-06, "loss": 2.1242, "step": 7813 }, { "epoch": 2.4495297805642635, "grad_norm": 10.630790710449219, "learning_rate": 3.907e-06, "loss": 2.065, "step": 7814 }, { "epoch": 2.449843260188088, "grad_norm": 11.276067733764648, "learning_rate": 3.9075e-06, "loss": 2.6823, "step": 7815 }, { "epoch": 2.4501567398119124, "grad_norm": 89.85816192626953, "learning_rate": 3.9080000000000005e-06, "loss": 2.1715, "step": 7816 }, { "epoch": 2.450470219435737, "grad_norm": 8.77840805053711, "learning_rate": 3.9085e-06, "loss": 2.093, "step": 7817 }, { "epoch": 2.4507836990595613, "grad_norm": 13.168466567993164, "learning_rate": 3.909e-06, "loss": 2.2043, "step": 7818 }, { "epoch": 2.4510971786833857, "grad_norm": 17.20318031311035, "learning_rate": 3.9095000000000004e-06, "loss": 2.1502, "step": 7819 }, { "epoch": 2.45141065830721, "grad_norm": 18.02589225769043, "learning_rate": 3.910000000000001e-06, "loss": 2.0247, "step": 7820 }, { "epoch": 2.4517241379310346, "grad_norm": 7.482308387756348, "learning_rate": 3.9105e-06, "loss": 2.0707, "step": 7821 }, { "epoch": 2.452037617554859, "grad_norm": 32.53311538696289, "learning_rate": 3.911e-06, "loss": 2.1176, "step": 7822 }, { "epoch": 2.4523510971786835, "grad_norm": 13.442570686340332, "learning_rate": 3.911500000000001e-06, "loss": 2.2501, "step": 7823 }, { "epoch": 2.452664576802508, "grad_norm": 55.23815155029297, "learning_rate": 3.912e-06, "loss": 2.5541, "step": 7824 }, { "epoch": 2.4529780564263324, "grad_norm": 11.761247634887695, "learning_rate": 3.9125e-06, "loss": 2.2055, "step": 7825 }, { "epoch": 2.453291536050157, "grad_norm": 13.896100044250488, "learning_rate": 3.9130000000000005e-06, "loss": 2.1658, "step": 7826 }, { "epoch": 2.4536050156739813, "grad_norm": 104.66007995605469, "learning_rate": 3.9135e-06, "loss": 1.9465, "step": 7827 }, { "epoch": 2.4539184952978057, "grad_norm": 16.297744750976562, "learning_rate": 3.914000000000001e-06, "loss": 2.4097, "step": 7828 }, { "epoch": 2.45423197492163, "grad_norm": 10.359723091125488, "learning_rate": 3.9145e-06, "loss": 1.9537, "step": 7829 }, { "epoch": 2.4545454545454546, "grad_norm": 7.101844787597656, "learning_rate": 3.915000000000001e-06, "loss": 2.0768, "step": 7830 }, { "epoch": 2.454858934169279, "grad_norm": 36.98095703125, "learning_rate": 3.9155e-06, "loss": 2.4815, "step": 7831 }, { "epoch": 2.4551724137931035, "grad_norm": 6.188116073608398, "learning_rate": 3.916e-06, "loss": 2.1173, "step": 7832 }, { "epoch": 2.455485893416928, "grad_norm": 12.334293365478516, "learning_rate": 3.9165000000000005e-06, "loss": 2.1284, "step": 7833 }, { "epoch": 2.4557993730407524, "grad_norm": 58.01237487792969, "learning_rate": 3.917e-06, "loss": 2.1862, "step": 7834 }, { "epoch": 2.456112852664577, "grad_norm": 15.541260719299316, "learning_rate": 3.9175e-06, "loss": 2.083, "step": 7835 }, { "epoch": 2.4564263322884012, "grad_norm": 50.827598571777344, "learning_rate": 3.9180000000000004e-06, "loss": 2.352, "step": 7836 }, { "epoch": 2.4567398119122257, "grad_norm": 9.92031478881836, "learning_rate": 3.9185e-06, "loss": 2.3184, "step": 7837 }, { "epoch": 2.45705329153605, "grad_norm": 17.176191329956055, "learning_rate": 3.919000000000001e-06, "loss": 2.3656, "step": 7838 }, { "epoch": 2.4573667711598746, "grad_norm": 56.617942810058594, "learning_rate": 3.9195e-06, "loss": 2.2804, "step": 7839 }, { "epoch": 2.457680250783699, "grad_norm": 23.612031936645508, "learning_rate": 3.920000000000001e-06, "loss": 2.2752, "step": 7840 }, { "epoch": 2.4579937304075234, "grad_norm": 11.181166648864746, "learning_rate": 3.920500000000001e-06, "loss": 2.3052, "step": 7841 }, { "epoch": 2.458307210031348, "grad_norm": 14.611475944519043, "learning_rate": 3.921e-06, "loss": 2.0526, "step": 7842 }, { "epoch": 2.4586206896551723, "grad_norm": 25.896997451782227, "learning_rate": 3.9215000000000005e-06, "loss": 2.8072, "step": 7843 }, { "epoch": 2.4589341692789968, "grad_norm": 21.56471061706543, "learning_rate": 3.922e-06, "loss": 2.2713, "step": 7844 }, { "epoch": 2.459247648902821, "grad_norm": 17.851634979248047, "learning_rate": 3.9225e-06, "loss": 2.0162, "step": 7845 }, { "epoch": 2.4595611285266457, "grad_norm": 15.308432579040527, "learning_rate": 3.923e-06, "loss": 2.2073, "step": 7846 }, { "epoch": 2.45987460815047, "grad_norm": 46.11570358276367, "learning_rate": 3.9235e-06, "loss": 2.255, "step": 7847 }, { "epoch": 2.4601880877742945, "grad_norm": 11.713844299316406, "learning_rate": 3.924000000000001e-06, "loss": 2.1646, "step": 7848 }, { "epoch": 2.460501567398119, "grad_norm": 14.890658378601074, "learning_rate": 3.9245e-06, "loss": 2.1913, "step": 7849 }, { "epoch": 2.4608150470219434, "grad_norm": 20.108068466186523, "learning_rate": 3.9250000000000005e-06, "loss": 2.4157, "step": 7850 }, { "epoch": 2.461128526645768, "grad_norm": 6.695193290710449, "learning_rate": 3.925500000000001e-06, "loss": 2.1523, "step": 7851 }, { "epoch": 2.4614420062695923, "grad_norm": 19.464017868041992, "learning_rate": 3.926e-06, "loss": 1.8469, "step": 7852 }, { "epoch": 2.4617554858934168, "grad_norm": 10.35474967956543, "learning_rate": 3.9265000000000004e-06, "loss": 2.2338, "step": 7853 }, { "epoch": 2.462068965517241, "grad_norm": 8.11723804473877, "learning_rate": 3.927e-06, "loss": 2.0229, "step": 7854 }, { "epoch": 2.462382445141066, "grad_norm": 7.715476989746094, "learning_rate": 3.9275e-06, "loss": 2.0291, "step": 7855 }, { "epoch": 2.4626959247648905, "grad_norm": 21.6331787109375, "learning_rate": 3.928e-06, "loss": 2.4331, "step": 7856 }, { "epoch": 2.463009404388715, "grad_norm": 11.732404708862305, "learning_rate": 3.9285e-06, "loss": 2.1331, "step": 7857 }, { "epoch": 2.4633228840125394, "grad_norm": 9.480164527893066, "learning_rate": 3.929000000000001e-06, "loss": 2.1459, "step": 7858 }, { "epoch": 2.463636363636364, "grad_norm": 11.250545501708984, "learning_rate": 3.9295e-06, "loss": 2.0195, "step": 7859 }, { "epoch": 2.4639498432601883, "grad_norm": 12.897054672241211, "learning_rate": 3.9300000000000005e-06, "loss": 2.1599, "step": 7860 }, { "epoch": 2.4642633228840127, "grad_norm": 13.875755310058594, "learning_rate": 3.930500000000001e-06, "loss": 2.0265, "step": 7861 }, { "epoch": 2.464576802507837, "grad_norm": 12.537698745727539, "learning_rate": 3.931e-06, "loss": 2.1444, "step": 7862 }, { "epoch": 2.4648902821316616, "grad_norm": 8.874787330627441, "learning_rate": 3.9315e-06, "loss": 2.0368, "step": 7863 }, { "epoch": 2.465203761755486, "grad_norm": 107.7464599609375, "learning_rate": 3.932000000000001e-06, "loss": 2.0961, "step": 7864 }, { "epoch": 2.4655172413793105, "grad_norm": 13.913363456726074, "learning_rate": 3.9325e-06, "loss": 1.9006, "step": 7865 }, { "epoch": 2.465830721003135, "grad_norm": 7.523643970489502, "learning_rate": 3.933e-06, "loss": 2.0978, "step": 7866 }, { "epoch": 2.4661442006269594, "grad_norm": 11.282609939575195, "learning_rate": 3.9335e-06, "loss": 1.9047, "step": 7867 }, { "epoch": 2.466457680250784, "grad_norm": 10.75317096710205, "learning_rate": 3.934000000000001e-06, "loss": 2.226, "step": 7868 }, { "epoch": 2.4667711598746083, "grad_norm": 12.000752449035645, "learning_rate": 3.9345e-06, "loss": 2.1688, "step": 7869 }, { "epoch": 2.4670846394984327, "grad_norm": 9.888389587402344, "learning_rate": 3.9350000000000004e-06, "loss": 1.9675, "step": 7870 }, { "epoch": 2.467398119122257, "grad_norm": 11.848458290100098, "learning_rate": 3.935500000000001e-06, "loss": 2.4307, "step": 7871 }, { "epoch": 2.4677115987460816, "grad_norm": 14.710775375366211, "learning_rate": 3.936e-06, "loss": 2.4312, "step": 7872 }, { "epoch": 2.468025078369906, "grad_norm": 13.29966926574707, "learning_rate": 3.9365e-06, "loss": 2.1354, "step": 7873 }, { "epoch": 2.4683385579937305, "grad_norm": 9.703301429748535, "learning_rate": 3.937000000000001e-06, "loss": 2.6093, "step": 7874 }, { "epoch": 2.468652037617555, "grad_norm": 15.019454002380371, "learning_rate": 3.9375e-06, "loss": 2.1247, "step": 7875 }, { "epoch": 2.4689655172413794, "grad_norm": 14.30727767944336, "learning_rate": 3.938e-06, "loss": 2.0729, "step": 7876 }, { "epoch": 2.469278996865204, "grad_norm": 61.304931640625, "learning_rate": 3.9385000000000005e-06, "loss": 2.1572, "step": 7877 }, { "epoch": 2.4695924764890282, "grad_norm": 7.997250556945801, "learning_rate": 3.939e-06, "loss": 2.1451, "step": 7878 }, { "epoch": 2.4699059561128527, "grad_norm": 5.872831344604492, "learning_rate": 3.9395e-06, "loss": 2.0899, "step": 7879 }, { "epoch": 2.470219435736677, "grad_norm": 19.89234733581543, "learning_rate": 3.94e-06, "loss": 2.488, "step": 7880 }, { "epoch": 2.4705329153605016, "grad_norm": 7.986907482147217, "learning_rate": 3.940500000000001e-06, "loss": 2.326, "step": 7881 }, { "epoch": 2.470846394984326, "grad_norm": 16.151874542236328, "learning_rate": 3.941e-06, "loss": 2.2297, "step": 7882 }, { "epoch": 2.4711598746081505, "grad_norm": 17.96297836303711, "learning_rate": 3.9415e-06, "loss": 2.1452, "step": 7883 }, { "epoch": 2.471473354231975, "grad_norm": 40.4291877746582, "learning_rate": 3.9420000000000005e-06, "loss": 3.0487, "step": 7884 }, { "epoch": 2.4717868338557993, "grad_norm": 21.144515991210938, "learning_rate": 3.9425e-06, "loss": 2.0486, "step": 7885 }, { "epoch": 2.472100313479624, "grad_norm": 18.997873306274414, "learning_rate": 3.943e-06, "loss": 2.1212, "step": 7886 }, { "epoch": 2.472413793103448, "grad_norm": 16.178781509399414, "learning_rate": 3.9435000000000004e-06, "loss": 1.9373, "step": 7887 }, { "epoch": 2.4727272727272727, "grad_norm": 10.4570951461792, "learning_rate": 3.944e-06, "loss": 1.9494, "step": 7888 }, { "epoch": 2.473040752351097, "grad_norm": 7.207930088043213, "learning_rate": 3.944500000000001e-06, "loss": 1.9961, "step": 7889 }, { "epoch": 2.4733542319749215, "grad_norm": 27.152292251586914, "learning_rate": 3.945e-06, "loss": 2.6428, "step": 7890 }, { "epoch": 2.473667711598746, "grad_norm": 12.384553909301758, "learning_rate": 3.945500000000001e-06, "loss": 2.3265, "step": 7891 }, { "epoch": 2.4739811912225704, "grad_norm": 8.311858177185059, "learning_rate": 3.946e-06, "loss": 2.2302, "step": 7892 }, { "epoch": 2.474294670846395, "grad_norm": 11.149374008178711, "learning_rate": 3.9465e-06, "loss": 2.0841, "step": 7893 }, { "epoch": 2.4746081504702193, "grad_norm": 43.10945129394531, "learning_rate": 3.9470000000000005e-06, "loss": 1.9215, "step": 7894 }, { "epoch": 2.4749216300940438, "grad_norm": 9.949850082397461, "learning_rate": 3.9475e-06, "loss": 2.7061, "step": 7895 }, { "epoch": 2.475235109717868, "grad_norm": 23.711782455444336, "learning_rate": 3.948e-06, "loss": 2.1028, "step": 7896 }, { "epoch": 2.4755485893416926, "grad_norm": 42.231414794921875, "learning_rate": 3.9485e-06, "loss": 2.2975, "step": 7897 }, { "epoch": 2.475862068965517, "grad_norm": 23.38568687438965, "learning_rate": 3.949e-06, "loss": 2.1293, "step": 7898 }, { "epoch": 2.4761755485893415, "grad_norm": 27.997535705566406, "learning_rate": 3.949500000000001e-06, "loss": 1.957, "step": 7899 }, { "epoch": 2.476489028213166, "grad_norm": 10.403421401977539, "learning_rate": 3.95e-06, "loss": 1.9896, "step": 7900 }, { "epoch": 2.4768025078369904, "grad_norm": 5.756619453430176, "learning_rate": 3.9505000000000005e-06, "loss": 2.0056, "step": 7901 }, { "epoch": 2.477115987460815, "grad_norm": 15.1923189163208, "learning_rate": 3.951000000000001e-06, "loss": 2.249, "step": 7902 }, { "epoch": 2.4774294670846393, "grad_norm": 12.584705352783203, "learning_rate": 3.9515e-06, "loss": 2.195, "step": 7903 }, { "epoch": 2.4777429467084637, "grad_norm": 8.447481155395508, "learning_rate": 3.9520000000000004e-06, "loss": 2.0367, "step": 7904 }, { "epoch": 2.478056426332288, "grad_norm": 9.368947982788086, "learning_rate": 3.9525e-06, "loss": 2.0938, "step": 7905 }, { "epoch": 2.4783699059561126, "grad_norm": 64.52045440673828, "learning_rate": 3.953e-06, "loss": 2.3587, "step": 7906 }, { "epoch": 2.478683385579937, "grad_norm": 131.960205078125, "learning_rate": 3.9535e-06, "loss": 2.2805, "step": 7907 }, { "epoch": 2.478996865203762, "grad_norm": 10.197537422180176, "learning_rate": 3.954e-06, "loss": 2.4862, "step": 7908 }, { "epoch": 2.4793103448275864, "grad_norm": 8.383183479309082, "learning_rate": 3.954500000000001e-06, "loss": 2.0724, "step": 7909 }, { "epoch": 2.479623824451411, "grad_norm": 48.54277420043945, "learning_rate": 3.955e-06, "loss": 2.4348, "step": 7910 }, { "epoch": 2.4799373040752353, "grad_norm": 12.113147735595703, "learning_rate": 3.9555000000000005e-06, "loss": 2.4023, "step": 7911 }, { "epoch": 2.4802507836990597, "grad_norm": 10.172322273254395, "learning_rate": 3.956000000000001e-06, "loss": 2.0797, "step": 7912 }, { "epoch": 2.480564263322884, "grad_norm": 12.61081314086914, "learning_rate": 3.9565e-06, "loss": 2.0588, "step": 7913 }, { "epoch": 2.4808777429467086, "grad_norm": 11.703728675842285, "learning_rate": 3.957e-06, "loss": 2.1331, "step": 7914 }, { "epoch": 2.481191222570533, "grad_norm": 9.777523040771484, "learning_rate": 3.957500000000001e-06, "loss": 2.1439, "step": 7915 }, { "epoch": 2.4815047021943575, "grad_norm": 23.280132293701172, "learning_rate": 3.958e-06, "loss": 2.3701, "step": 7916 }, { "epoch": 2.481818181818182, "grad_norm": 11.687907218933105, "learning_rate": 3.9585e-06, "loss": 2.3128, "step": 7917 }, { "epoch": 2.4821316614420064, "grad_norm": 9.232254981994629, "learning_rate": 3.959e-06, "loss": 1.9846, "step": 7918 }, { "epoch": 2.482445141065831, "grad_norm": 11.229503631591797, "learning_rate": 3.959500000000001e-06, "loss": 2.2397, "step": 7919 }, { "epoch": 2.4827586206896552, "grad_norm": 14.284053802490234, "learning_rate": 3.96e-06, "loss": 2.2555, "step": 7920 }, { "epoch": 2.4830721003134797, "grad_norm": 11.733200073242188, "learning_rate": 3.9605000000000005e-06, "loss": 2.034, "step": 7921 }, { "epoch": 2.483385579937304, "grad_norm": 16.50055694580078, "learning_rate": 3.961000000000001e-06, "loss": 2.7439, "step": 7922 }, { "epoch": 2.4836990595611286, "grad_norm": 18.86034393310547, "learning_rate": 3.9615e-06, "loss": 1.9877, "step": 7923 }, { "epoch": 2.484012539184953, "grad_norm": 10.510337829589844, "learning_rate": 3.962e-06, "loss": 1.9931, "step": 7924 }, { "epoch": 2.4843260188087775, "grad_norm": 16.541913986206055, "learning_rate": 3.962500000000001e-06, "loss": 2.1255, "step": 7925 }, { "epoch": 2.484639498432602, "grad_norm": 9.874707221984863, "learning_rate": 3.963e-06, "loss": 2.0936, "step": 7926 }, { "epoch": 2.4849529780564263, "grad_norm": 12.12887191772461, "learning_rate": 3.9635e-06, "loss": 2.1847, "step": 7927 }, { "epoch": 2.485266457680251, "grad_norm": 6.921993255615234, "learning_rate": 3.964e-06, "loss": 2.3098, "step": 7928 }, { "epoch": 2.4855799373040752, "grad_norm": 10.193533897399902, "learning_rate": 3.964500000000001e-06, "loss": 2.0171, "step": 7929 }, { "epoch": 2.4858934169278997, "grad_norm": 22.761281967163086, "learning_rate": 3.965e-06, "loss": 2.6215, "step": 7930 }, { "epoch": 2.486206896551724, "grad_norm": 6.364945888519287, "learning_rate": 3.9655e-06, "loss": 2.0928, "step": 7931 }, { "epoch": 2.4865203761755486, "grad_norm": 9.065265655517578, "learning_rate": 3.966000000000001e-06, "loss": 2.2012, "step": 7932 }, { "epoch": 2.486833855799373, "grad_norm": 9.903694152832031, "learning_rate": 3.9665e-06, "loss": 2.0001, "step": 7933 }, { "epoch": 2.4871473354231974, "grad_norm": 32.66429138183594, "learning_rate": 3.967e-06, "loss": 2.4208, "step": 7934 }, { "epoch": 2.487460815047022, "grad_norm": 30.170425415039062, "learning_rate": 3.9675000000000006e-06, "loss": 2.9884, "step": 7935 }, { "epoch": 2.4877742946708463, "grad_norm": 25.184734344482422, "learning_rate": 3.968e-06, "loss": 2.3474, "step": 7936 }, { "epoch": 2.4880877742946708, "grad_norm": 16.326433181762695, "learning_rate": 3.9685e-06, "loss": 2.1261, "step": 7937 }, { "epoch": 2.488401253918495, "grad_norm": 38.6700553894043, "learning_rate": 3.9690000000000005e-06, "loss": 2.2317, "step": 7938 }, { "epoch": 2.4887147335423196, "grad_norm": 15.540034294128418, "learning_rate": 3.969500000000001e-06, "loss": 2.4282, "step": 7939 }, { "epoch": 2.489028213166144, "grad_norm": 10.576045989990234, "learning_rate": 3.97e-06, "loss": 2.2038, "step": 7940 }, { "epoch": 2.4893416927899685, "grad_norm": 16.077131271362305, "learning_rate": 3.9705e-06, "loss": 1.8669, "step": 7941 }, { "epoch": 2.489655172413793, "grad_norm": 15.25275707244873, "learning_rate": 3.971000000000001e-06, "loss": 2.104, "step": 7942 }, { "epoch": 2.4899686520376174, "grad_norm": 22.44020652770996, "learning_rate": 3.9715e-06, "loss": 1.9914, "step": 7943 }, { "epoch": 2.490282131661442, "grad_norm": 5.067534923553467, "learning_rate": 3.972e-06, "loss": 2.0518, "step": 7944 }, { "epoch": 2.4905956112852663, "grad_norm": 16.034116744995117, "learning_rate": 3.9725000000000005e-06, "loss": 2.1224, "step": 7945 }, { "epoch": 2.4909090909090907, "grad_norm": 14.29828929901123, "learning_rate": 3.973e-06, "loss": 2.226, "step": 7946 }, { "epoch": 2.491222570532915, "grad_norm": 20.10262680053711, "learning_rate": 3.9735e-06, "loss": 2.1935, "step": 7947 }, { "epoch": 2.4915360501567396, "grad_norm": 118.35943603515625, "learning_rate": 3.974e-06, "loss": 2.2985, "step": 7948 }, { "epoch": 2.4918495297805645, "grad_norm": 198.3577423095703, "learning_rate": 3.974500000000001e-06, "loss": 2.1735, "step": 7949 }, { "epoch": 2.492163009404389, "grad_norm": 9.204904556274414, "learning_rate": 3.975000000000001e-06, "loss": 1.9673, "step": 7950 }, { "epoch": 2.4924764890282134, "grad_norm": 10.24039363861084, "learning_rate": 3.9755e-06, "loss": 2.2772, "step": 7951 }, { "epoch": 2.492789968652038, "grad_norm": 11.183293342590332, "learning_rate": 3.9760000000000006e-06, "loss": 2.0437, "step": 7952 }, { "epoch": 2.4931034482758623, "grad_norm": 13.324175834655762, "learning_rate": 3.9765e-06, "loss": 2.1169, "step": 7953 }, { "epoch": 2.4934169278996867, "grad_norm": 42.77281951904297, "learning_rate": 3.977e-06, "loss": 2.2251, "step": 7954 }, { "epoch": 2.493730407523511, "grad_norm": 11.70572566986084, "learning_rate": 3.9775000000000005e-06, "loss": 2.2319, "step": 7955 }, { "epoch": 2.4940438871473356, "grad_norm": 9.434075355529785, "learning_rate": 3.978e-06, "loss": 2.0437, "step": 7956 }, { "epoch": 2.49435736677116, "grad_norm": 59.38090896606445, "learning_rate": 3.9785e-06, "loss": 2.1806, "step": 7957 }, { "epoch": 2.4946708463949845, "grad_norm": 12.585026741027832, "learning_rate": 3.979e-06, "loss": 2.1482, "step": 7958 }, { "epoch": 2.494984326018809, "grad_norm": 16.244274139404297, "learning_rate": 3.979500000000001e-06, "loss": 2.1909, "step": 7959 }, { "epoch": 2.4952978056426334, "grad_norm": 8.943524360656738, "learning_rate": 3.980000000000001e-06, "loss": 2.2167, "step": 7960 }, { "epoch": 2.495611285266458, "grad_norm": 21.497127532958984, "learning_rate": 3.9805e-06, "loss": 2.1816, "step": 7961 }, { "epoch": 2.4959247648902823, "grad_norm": 74.79740142822266, "learning_rate": 3.9810000000000005e-06, "loss": 2.8729, "step": 7962 }, { "epoch": 2.4962382445141067, "grad_norm": 17.44632911682129, "learning_rate": 3.981500000000001e-06, "loss": 1.8672, "step": 7963 }, { "epoch": 2.496551724137931, "grad_norm": 13.862711906433105, "learning_rate": 3.982e-06, "loss": 2.1079, "step": 7964 }, { "epoch": 2.4968652037617556, "grad_norm": 10.682537078857422, "learning_rate": 3.9825e-06, "loss": 2.0226, "step": 7965 }, { "epoch": 2.49717868338558, "grad_norm": 6.049378395080566, "learning_rate": 3.983e-06, "loss": 2.0664, "step": 7966 }, { "epoch": 2.4974921630094045, "grad_norm": 37.682891845703125, "learning_rate": 3.9835e-06, "loss": 2.5832, "step": 7967 }, { "epoch": 2.497805642633229, "grad_norm": 12.364509582519531, "learning_rate": 3.984e-06, "loss": 2.0997, "step": 7968 }, { "epoch": 2.4981191222570533, "grad_norm": 9.614590644836426, "learning_rate": 3.9845000000000006e-06, "loss": 2.2431, "step": 7969 }, { "epoch": 2.498432601880878, "grad_norm": 9.42934513092041, "learning_rate": 3.985000000000001e-06, "loss": 2.333, "step": 7970 }, { "epoch": 2.4987460815047022, "grad_norm": 10.112471580505371, "learning_rate": 3.9855e-06, "loss": 2.0337, "step": 7971 }, { "epoch": 2.4990595611285267, "grad_norm": 10.008768081665039, "learning_rate": 3.9860000000000005e-06, "loss": 2.0406, "step": 7972 }, { "epoch": 2.499373040752351, "grad_norm": 8.429323196411133, "learning_rate": 3.986500000000001e-06, "loss": 2.505, "step": 7973 }, { "epoch": 2.4996865203761756, "grad_norm": 164.6767578125, "learning_rate": 3.987e-06, "loss": 2.6305, "step": 7974 }, { "epoch": 2.5, "grad_norm": 13.313830375671387, "learning_rate": 3.9875e-06, "loss": 2.2268, "step": 7975 }, { "epoch": 2.5003134796238244, "grad_norm": 13.262435913085938, "learning_rate": 3.988000000000001e-06, "loss": 1.9174, "step": 7976 }, { "epoch": 2.500626959247649, "grad_norm": 14.187859535217285, "learning_rate": 3.9885e-06, "loss": 2.0781, "step": 7977 }, { "epoch": 2.5009404388714733, "grad_norm": 10.560633659362793, "learning_rate": 3.989e-06, "loss": 1.9178, "step": 7978 }, { "epoch": 2.5012539184952978, "grad_norm": 15.659086227416992, "learning_rate": 3.9895000000000005e-06, "loss": 2.2199, "step": 7979 }, { "epoch": 2.501567398119122, "grad_norm": 10.836274147033691, "learning_rate": 3.990000000000001e-06, "loss": 2.3964, "step": 7980 }, { "epoch": 2.501567398119122, "eval_loss": 2.391023635864258, "eval_runtime": 21.0674, "eval_samples_per_second": 127.543, "eval_steps_per_second": 7.974, "step": 7980 }, { "epoch": 2.5018808777429467, "grad_norm": 58.7084846496582, "learning_rate": 3.9905e-06, "loss": 2.4134, "step": 7981 }, { "epoch": 2.502194357366771, "grad_norm": 10.498461723327637, "learning_rate": 3.991e-06, "loss": 2.0795, "step": 7982 }, { "epoch": 2.5025078369905955, "grad_norm": 11.97678279876709, "learning_rate": 3.991500000000001e-06, "loss": 2.2052, "step": 7983 }, { "epoch": 2.50282131661442, "grad_norm": 9.905413627624512, "learning_rate": 3.992e-06, "loss": 1.9496, "step": 7984 }, { "epoch": 2.5031347962382444, "grad_norm": 20.838537216186523, "learning_rate": 3.9925e-06, "loss": 2.0481, "step": 7985 }, { "epoch": 2.503448275862069, "grad_norm": 12.945446968078613, "learning_rate": 3.9930000000000006e-06, "loss": 2.512, "step": 7986 }, { "epoch": 2.5037617554858933, "grad_norm": 11.733619689941406, "learning_rate": 3.9935e-06, "loss": 2.1203, "step": 7987 }, { "epoch": 2.5040752351097177, "grad_norm": 114.54523468017578, "learning_rate": 3.994e-06, "loss": 2.3957, "step": 7988 }, { "epoch": 2.504388714733542, "grad_norm": 18.717134475708008, "learning_rate": 3.9945000000000005e-06, "loss": 2.7515, "step": 7989 }, { "epoch": 2.5047021943573666, "grad_norm": 10.485302925109863, "learning_rate": 3.995000000000001e-06, "loss": 2.2795, "step": 7990 }, { "epoch": 2.505015673981191, "grad_norm": 10.899035453796387, "learning_rate": 3.9955e-06, "loss": 2.0513, "step": 7991 }, { "epoch": 2.5053291536050155, "grad_norm": 28.7159423828125, "learning_rate": 3.996e-06, "loss": 2.1722, "step": 7992 }, { "epoch": 2.50564263322884, "grad_norm": 12.646659851074219, "learning_rate": 3.996500000000001e-06, "loss": 1.9325, "step": 7993 }, { "epoch": 2.5059561128526644, "grad_norm": 10.466874122619629, "learning_rate": 3.997e-06, "loss": 2.1259, "step": 7994 }, { "epoch": 2.506269592476489, "grad_norm": 7.8314738273620605, "learning_rate": 3.9975e-06, "loss": 2.1982, "step": 7995 }, { "epoch": 2.5065830721003133, "grad_norm": 13.87162971496582, "learning_rate": 3.9980000000000005e-06, "loss": 1.9479, "step": 7996 }, { "epoch": 2.5068965517241377, "grad_norm": 15.63772964477539, "learning_rate": 3.9985e-06, "loss": 2.0469, "step": 7997 }, { "epoch": 2.507210031347962, "grad_norm": 13.099417686462402, "learning_rate": 3.999e-06, "loss": 2.2229, "step": 7998 }, { "epoch": 2.5075235109717866, "grad_norm": 9.772567749023438, "learning_rate": 3.9995e-06, "loss": 2.1554, "step": 7999 }, { "epoch": 2.507836990595611, "grad_norm": 14.402785301208496, "learning_rate": 4.000000000000001e-06, "loss": 2.2147, "step": 8000 }, { "epoch": 2.5081504702194355, "grad_norm": 13.171868324279785, "learning_rate": 4.000500000000001e-06, "loss": 2.1934, "step": 8001 }, { "epoch": 2.50846394984326, "grad_norm": 23.49083709716797, "learning_rate": 4.001e-06, "loss": 2.4032, "step": 8002 }, { "epoch": 2.5087774294670844, "grad_norm": 21.6545352935791, "learning_rate": 4.0015000000000006e-06, "loss": 2.2659, "step": 8003 }, { "epoch": 2.509090909090909, "grad_norm": 13.846197128295898, "learning_rate": 4.002e-06, "loss": 2.0774, "step": 8004 }, { "epoch": 2.5094043887147337, "grad_norm": 5.960597515106201, "learning_rate": 4.0025e-06, "loss": 1.9789, "step": 8005 }, { "epoch": 2.509717868338558, "grad_norm": 53.3001708984375, "learning_rate": 4.0030000000000005e-06, "loss": 2.1669, "step": 8006 }, { "epoch": 2.5100313479623826, "grad_norm": 16.969310760498047, "learning_rate": 4.0035e-06, "loss": 3.3468, "step": 8007 }, { "epoch": 2.510344827586207, "grad_norm": 12.277050018310547, "learning_rate": 4.004e-06, "loss": 2.2082, "step": 8008 }, { "epoch": 2.5106583072100315, "grad_norm": 14.029617309570312, "learning_rate": 4.0045e-06, "loss": 1.8568, "step": 8009 }, { "epoch": 2.510971786833856, "grad_norm": 8.591123580932617, "learning_rate": 4.005000000000001e-06, "loss": 2.1302, "step": 8010 }, { "epoch": 2.5112852664576804, "grad_norm": 12.393437385559082, "learning_rate": 4.005500000000001e-06, "loss": 2.2532, "step": 8011 }, { "epoch": 2.511598746081505, "grad_norm": 6.509521484375, "learning_rate": 4.006e-06, "loss": 2.2521, "step": 8012 }, { "epoch": 2.5119122257053292, "grad_norm": 8.343978881835938, "learning_rate": 4.0065000000000005e-06, "loss": 2.3241, "step": 8013 }, { "epoch": 2.5122257053291537, "grad_norm": 11.071076393127441, "learning_rate": 4.007e-06, "loss": 2.2336, "step": 8014 }, { "epoch": 2.512539184952978, "grad_norm": 123.21747589111328, "learning_rate": 4.0075e-06, "loss": 2.346, "step": 8015 }, { "epoch": 2.5128526645768026, "grad_norm": 7.459530830383301, "learning_rate": 4.008e-06, "loss": 1.9053, "step": 8016 }, { "epoch": 2.513166144200627, "grad_norm": 8.925275802612305, "learning_rate": 4.0085e-06, "loss": 2.4643, "step": 8017 }, { "epoch": 2.5134796238244514, "grad_norm": 91.87667083740234, "learning_rate": 4.009e-06, "loss": 2.6044, "step": 8018 }, { "epoch": 2.513793103448276, "grad_norm": 11.647605895996094, "learning_rate": 4.0095e-06, "loss": 2.0206, "step": 8019 }, { "epoch": 2.5141065830721003, "grad_norm": 21.304723739624023, "learning_rate": 4.0100000000000006e-06, "loss": 1.9225, "step": 8020 }, { "epoch": 2.5144200626959248, "grad_norm": 6.793734073638916, "learning_rate": 4.010500000000001e-06, "loss": 2.0431, "step": 8021 }, { "epoch": 2.514733542319749, "grad_norm": 32.75812911987305, "learning_rate": 4.011e-06, "loss": 2.1625, "step": 8022 }, { "epoch": 2.5150470219435737, "grad_norm": 13.880643844604492, "learning_rate": 4.0115000000000005e-06, "loss": 2.3, "step": 8023 }, { "epoch": 2.515360501567398, "grad_norm": 7.174182415008545, "learning_rate": 4.012000000000001e-06, "loss": 2.0949, "step": 8024 }, { "epoch": 2.5156739811912225, "grad_norm": 41.7447509765625, "learning_rate": 4.0125e-06, "loss": 2.1306, "step": 8025 }, { "epoch": 2.515987460815047, "grad_norm": 163.42405700683594, "learning_rate": 4.013e-06, "loss": 2.3177, "step": 8026 }, { "epoch": 2.5163009404388714, "grad_norm": 13.993059158325195, "learning_rate": 4.0135e-06, "loss": 2.2062, "step": 8027 }, { "epoch": 2.516614420062696, "grad_norm": 9.977289199829102, "learning_rate": 4.014e-06, "loss": 2.2695, "step": 8028 }, { "epoch": 2.5169278996865203, "grad_norm": 14.920466423034668, "learning_rate": 4.0145e-06, "loss": 2.0994, "step": 8029 }, { "epoch": 2.5172413793103448, "grad_norm": 8.766217231750488, "learning_rate": 4.0150000000000005e-06, "loss": 2.2947, "step": 8030 }, { "epoch": 2.517554858934169, "grad_norm": 12.6229887008667, "learning_rate": 4.015500000000001e-06, "loss": 2.0564, "step": 8031 }, { "epoch": 2.5178683385579936, "grad_norm": 8.838521003723145, "learning_rate": 4.016e-06, "loss": 2.0935, "step": 8032 }, { "epoch": 2.518181818181818, "grad_norm": 18.84675407409668, "learning_rate": 4.0165e-06, "loss": 2.0082, "step": 8033 }, { "epoch": 2.5184952978056425, "grad_norm": 13.945859909057617, "learning_rate": 4.017000000000001e-06, "loss": 2.1996, "step": 8034 }, { "epoch": 2.518808777429467, "grad_norm": 10.350574493408203, "learning_rate": 4.0175e-06, "loss": 2.2201, "step": 8035 }, { "epoch": 2.5191222570532914, "grad_norm": 15.972736358642578, "learning_rate": 4.018e-06, "loss": 2.0198, "step": 8036 }, { "epoch": 2.519435736677116, "grad_norm": 41.97819900512695, "learning_rate": 4.0185000000000006e-06, "loss": 2.3875, "step": 8037 }, { "epoch": 2.5197492163009403, "grad_norm": 13.823982238769531, "learning_rate": 4.019e-06, "loss": 2.3397, "step": 8038 }, { "epoch": 2.5200626959247647, "grad_norm": 13.188786506652832, "learning_rate": 4.0195e-06, "loss": 1.9894, "step": 8039 }, { "epoch": 2.5203761755485896, "grad_norm": 23.64742660522461, "learning_rate": 4.0200000000000005e-06, "loss": 2.1372, "step": 8040 }, { "epoch": 2.520689655172414, "grad_norm": 12.344711303710938, "learning_rate": 4.020500000000001e-06, "loss": 2.1608, "step": 8041 }, { "epoch": 2.5210031347962385, "grad_norm": 7.13676118850708, "learning_rate": 4.021e-06, "loss": 2.2052, "step": 8042 }, { "epoch": 2.521316614420063, "grad_norm": 12.095306396484375, "learning_rate": 4.0215e-06, "loss": 2.1556, "step": 8043 }, { "epoch": 2.5216300940438874, "grad_norm": 9.025074005126953, "learning_rate": 4.022000000000001e-06, "loss": 2.2232, "step": 8044 }, { "epoch": 2.521943573667712, "grad_norm": 106.57372283935547, "learning_rate": 4.0225e-06, "loss": 2.6447, "step": 8045 }, { "epoch": 2.5222570532915363, "grad_norm": 7.148740291595459, "learning_rate": 4.023e-06, "loss": 2.0018, "step": 8046 }, { "epoch": 2.5225705329153607, "grad_norm": 11.567998886108398, "learning_rate": 4.0235000000000005e-06, "loss": 2.205, "step": 8047 }, { "epoch": 2.522884012539185, "grad_norm": 20.14932632446289, "learning_rate": 4.024e-06, "loss": 2.1369, "step": 8048 }, { "epoch": 2.5231974921630096, "grad_norm": 7.9424238204956055, "learning_rate": 4.0245e-06, "loss": 2.2844, "step": 8049 }, { "epoch": 2.523510971786834, "grad_norm": 88.3066635131836, "learning_rate": 4.0250000000000004e-06, "loss": 2.4786, "step": 8050 }, { "epoch": 2.5238244514106585, "grad_norm": 7.915011882781982, "learning_rate": 4.025500000000001e-06, "loss": 2.041, "step": 8051 }, { "epoch": 2.524137931034483, "grad_norm": 11.020607948303223, "learning_rate": 4.026e-06, "loss": 2.0473, "step": 8052 }, { "epoch": 2.5244514106583074, "grad_norm": 60.73957061767578, "learning_rate": 4.0265e-06, "loss": 2.5664, "step": 8053 }, { "epoch": 2.524764890282132, "grad_norm": 8.79599380493164, "learning_rate": 4.027000000000001e-06, "loss": 2.0998, "step": 8054 }, { "epoch": 2.5250783699059562, "grad_norm": 9.447891235351562, "learning_rate": 4.0275e-06, "loss": 2.0458, "step": 8055 }, { "epoch": 2.5253918495297807, "grad_norm": 81.61319732666016, "learning_rate": 4.028e-06, "loss": 2.6504, "step": 8056 }, { "epoch": 2.525705329153605, "grad_norm": 18.912580490112305, "learning_rate": 4.0285000000000005e-06, "loss": 2.3592, "step": 8057 }, { "epoch": 2.5260188087774296, "grad_norm": 12.761161804199219, "learning_rate": 4.029e-06, "loss": 2.2341, "step": 8058 }, { "epoch": 2.526332288401254, "grad_norm": 9.130534172058105, "learning_rate": 4.0295e-06, "loss": 2.6537, "step": 8059 }, { "epoch": 2.5266457680250785, "grad_norm": 7.268718242645264, "learning_rate": 4.03e-06, "loss": 2.3193, "step": 8060 }, { "epoch": 2.526959247648903, "grad_norm": 22.320903778076172, "learning_rate": 4.030500000000001e-06, "loss": 2.171, "step": 8061 }, { "epoch": 2.5272727272727273, "grad_norm": 7.862981796264648, "learning_rate": 4.031000000000001e-06, "loss": 2.1972, "step": 8062 }, { "epoch": 2.527586206896552, "grad_norm": 69.889404296875, "learning_rate": 4.0315e-06, "loss": 2.6708, "step": 8063 }, { "epoch": 2.527899686520376, "grad_norm": 68.6167984008789, "learning_rate": 4.0320000000000005e-06, "loss": 2.5389, "step": 8064 }, { "epoch": 2.5282131661442007, "grad_norm": 9.618413925170898, "learning_rate": 4.0325e-06, "loss": 1.993, "step": 8065 }, { "epoch": 2.528526645768025, "grad_norm": 200.17262268066406, "learning_rate": 4.033e-06, "loss": 2.3001, "step": 8066 }, { "epoch": 2.5288401253918495, "grad_norm": 14.443070411682129, "learning_rate": 4.0335000000000004e-06, "loss": 2.2056, "step": 8067 }, { "epoch": 2.529153605015674, "grad_norm": 13.994797706604004, "learning_rate": 4.034e-06, "loss": 2.6699, "step": 8068 }, { "epoch": 2.5294670846394984, "grad_norm": 16.71522331237793, "learning_rate": 4.0345e-06, "loss": 2.0892, "step": 8069 }, { "epoch": 2.529780564263323, "grad_norm": 7.302309989929199, "learning_rate": 4.035e-06, "loss": 1.9192, "step": 8070 }, { "epoch": 2.5300940438871473, "grad_norm": 10.238965034484863, "learning_rate": 4.035500000000001e-06, "loss": 2.0095, "step": 8071 }, { "epoch": 2.5304075235109718, "grad_norm": 9.30863094329834, "learning_rate": 4.036000000000001e-06, "loss": 2.1205, "step": 8072 }, { "epoch": 2.530721003134796, "grad_norm": 16.60915184020996, "learning_rate": 4.0365e-06, "loss": 2.4925, "step": 8073 }, { "epoch": 2.5310344827586206, "grad_norm": 29.450918197631836, "learning_rate": 4.0370000000000005e-06, "loss": 2.6953, "step": 8074 }, { "epoch": 2.531347962382445, "grad_norm": 12.318449020385742, "learning_rate": 4.037500000000001e-06, "loss": 2.0765, "step": 8075 }, { "epoch": 2.5316614420062695, "grad_norm": 11.640390396118164, "learning_rate": 4.038e-06, "loss": 2.6422, "step": 8076 }, { "epoch": 2.531974921630094, "grad_norm": 8.642571449279785, "learning_rate": 4.0385e-06, "loss": 2.0575, "step": 8077 }, { "epoch": 2.5322884012539184, "grad_norm": 7.810276508331299, "learning_rate": 4.039e-06, "loss": 2.0451, "step": 8078 }, { "epoch": 2.532601880877743, "grad_norm": 12.933732032775879, "learning_rate": 4.0395e-06, "loss": 2.1733, "step": 8079 }, { "epoch": 2.5329153605015673, "grad_norm": 8.662042617797852, "learning_rate": 4.04e-06, "loss": 2.0392, "step": 8080 }, { "epoch": 2.5332288401253917, "grad_norm": 9.610318183898926, "learning_rate": 4.0405000000000005e-06, "loss": 2.1714, "step": 8081 }, { "epoch": 2.533542319749216, "grad_norm": 12.181833267211914, "learning_rate": 4.041000000000001e-06, "loss": 2.1629, "step": 8082 }, { "epoch": 2.5338557993730406, "grad_norm": 5.754836559295654, "learning_rate": 4.0415e-06, "loss": 2.1116, "step": 8083 }, { "epoch": 2.534169278996865, "grad_norm": 32.80358123779297, "learning_rate": 4.0420000000000004e-06, "loss": 2.2835, "step": 8084 }, { "epoch": 2.5344827586206895, "grad_norm": 15.29365348815918, "learning_rate": 4.042500000000001e-06, "loss": 2.1407, "step": 8085 }, { "epoch": 2.534796238244514, "grad_norm": 8.040053367614746, "learning_rate": 4.043e-06, "loss": 2.0654, "step": 8086 }, { "epoch": 2.5351097178683384, "grad_norm": 11.311137199401855, "learning_rate": 4.0435e-06, "loss": 2.0594, "step": 8087 }, { "epoch": 2.535423197492163, "grad_norm": 35.71397018432617, "learning_rate": 4.044e-06, "loss": 3.0128, "step": 8088 }, { "epoch": 2.5357366771159873, "grad_norm": 13.440254211425781, "learning_rate": 4.0445e-06, "loss": 1.9888, "step": 8089 }, { "epoch": 2.5360501567398117, "grad_norm": 12.319478034973145, "learning_rate": 4.045e-06, "loss": 2.1037, "step": 8090 }, { "epoch": 2.536363636363636, "grad_norm": 9.520621299743652, "learning_rate": 4.0455000000000005e-06, "loss": 2.1544, "step": 8091 }, { "epoch": 2.5366771159874606, "grad_norm": 17.017839431762695, "learning_rate": 4.046000000000001e-06, "loss": 2.0337, "step": 8092 }, { "epoch": 2.536990595611285, "grad_norm": 10.644001960754395, "learning_rate": 4.0465e-06, "loss": 1.9253, "step": 8093 }, { "epoch": 2.5373040752351095, "grad_norm": 15.452028274536133, "learning_rate": 4.047e-06, "loss": 2.1825, "step": 8094 }, { "epoch": 2.537617554858934, "grad_norm": 27.162477493286133, "learning_rate": 4.047500000000001e-06, "loss": 2.3837, "step": 8095 }, { "epoch": 2.5379310344827584, "grad_norm": 16.162139892578125, "learning_rate": 4.048e-06, "loss": 2.1234, "step": 8096 }, { "epoch": 2.538244514106583, "grad_norm": 11.38028335571289, "learning_rate": 4.0485e-06, "loss": 2.1261, "step": 8097 }, { "epoch": 2.5385579937304072, "grad_norm": 10.215377807617188, "learning_rate": 4.0490000000000005e-06, "loss": 2.1792, "step": 8098 }, { "epoch": 2.538871473354232, "grad_norm": 50.97389221191406, "learning_rate": 4.0495e-06, "loss": 2.0727, "step": 8099 }, { "epoch": 2.5391849529780566, "grad_norm": 34.17228698730469, "learning_rate": 4.05e-06, "loss": 2.0819, "step": 8100 }, { "epoch": 2.539498432601881, "grad_norm": 8.434781074523926, "learning_rate": 4.0505000000000004e-06, "loss": 2.543, "step": 8101 }, { "epoch": 2.5398119122257055, "grad_norm": 67.30236053466797, "learning_rate": 4.051000000000001e-06, "loss": 2.383, "step": 8102 }, { "epoch": 2.54012539184953, "grad_norm": 9.708230972290039, "learning_rate": 4.0515e-06, "loss": 1.9708, "step": 8103 }, { "epoch": 2.5404388714733543, "grad_norm": 113.73563385009766, "learning_rate": 4.052e-06, "loss": 2.2777, "step": 8104 }, { "epoch": 2.540752351097179, "grad_norm": 9.123244285583496, "learning_rate": 4.052500000000001e-06, "loss": 2.5942, "step": 8105 }, { "epoch": 2.5410658307210032, "grad_norm": 10.319272994995117, "learning_rate": 4.053e-06, "loss": 2.0247, "step": 8106 }, { "epoch": 2.5413793103448277, "grad_norm": 15.69609260559082, "learning_rate": 4.0535e-06, "loss": 2.182, "step": 8107 }, { "epoch": 2.541692789968652, "grad_norm": 36.223411560058594, "learning_rate": 4.0540000000000005e-06, "loss": 2.5037, "step": 8108 }, { "epoch": 2.5420062695924766, "grad_norm": 7.516111850738525, "learning_rate": 4.0545e-06, "loss": 2.1368, "step": 8109 }, { "epoch": 2.542319749216301, "grad_norm": 6.841305732727051, "learning_rate": 4.055000000000001e-06, "loss": 2.0854, "step": 8110 }, { "epoch": 2.5426332288401254, "grad_norm": 15.166303634643555, "learning_rate": 4.0555e-06, "loss": 2.4191, "step": 8111 }, { "epoch": 2.54294670846395, "grad_norm": 11.26488971710205, "learning_rate": 4.056000000000001e-06, "loss": 2.2859, "step": 8112 }, { "epoch": 2.5432601880877743, "grad_norm": 7.863527297973633, "learning_rate": 4.0565e-06, "loss": 2.2661, "step": 8113 }, { "epoch": 2.5435736677115988, "grad_norm": 7.9421892166137695, "learning_rate": 4.057e-06, "loss": 2.0505, "step": 8114 }, { "epoch": 2.543887147335423, "grad_norm": 11.674912452697754, "learning_rate": 4.0575000000000005e-06, "loss": 2.0665, "step": 8115 }, { "epoch": 2.5442006269592476, "grad_norm": 89.61035919189453, "learning_rate": 4.058e-06, "loss": 2.5718, "step": 8116 }, { "epoch": 2.544514106583072, "grad_norm": 28.6441593170166, "learning_rate": 4.0585e-06, "loss": 2.5268, "step": 8117 }, { "epoch": 2.5448275862068965, "grad_norm": 9.622467041015625, "learning_rate": 4.0590000000000004e-06, "loss": 2.2, "step": 8118 }, { "epoch": 2.545141065830721, "grad_norm": 7.657288074493408, "learning_rate": 4.0595e-06, "loss": 2.1883, "step": 8119 }, { "epoch": 2.5454545454545454, "grad_norm": 6.957535743713379, "learning_rate": 4.060000000000001e-06, "loss": 2.0607, "step": 8120 }, { "epoch": 2.54576802507837, "grad_norm": 12.58651351928711, "learning_rate": 4.0605e-06, "loss": 2.2076, "step": 8121 }, { "epoch": 2.5460815047021943, "grad_norm": 20.456663131713867, "learning_rate": 4.061000000000001e-06, "loss": 2.0495, "step": 8122 }, { "epoch": 2.5463949843260187, "grad_norm": 10.933107376098633, "learning_rate": 4.061500000000001e-06, "loss": 2.1725, "step": 8123 }, { "epoch": 2.546708463949843, "grad_norm": 9.598475456237793, "learning_rate": 4.062e-06, "loss": 2.1198, "step": 8124 }, { "epoch": 2.5470219435736676, "grad_norm": 10.045226097106934, "learning_rate": 4.0625000000000005e-06, "loss": 2.2901, "step": 8125 }, { "epoch": 2.547335423197492, "grad_norm": 99.91444396972656, "learning_rate": 4.063e-06, "loss": 2.5085, "step": 8126 }, { "epoch": 2.5476489028213165, "grad_norm": 35.88787841796875, "learning_rate": 4.0635e-06, "loss": 1.9435, "step": 8127 }, { "epoch": 2.547962382445141, "grad_norm": 5.425751686096191, "learning_rate": 4.064e-06, "loss": 1.9634, "step": 8128 }, { "epoch": 2.5482758620689654, "grad_norm": 8.353808403015137, "learning_rate": 4.0645e-06, "loss": 2.1249, "step": 8129 }, { "epoch": 2.54858934169279, "grad_norm": 8.584625244140625, "learning_rate": 4.065e-06, "loss": 2.1499, "step": 8130 }, { "epoch": 2.5489028213166143, "grad_norm": 17.858436584472656, "learning_rate": 4.0655e-06, "loss": 2.2014, "step": 8131 }, { "epoch": 2.5492163009404387, "grad_norm": 10.603771209716797, "learning_rate": 4.0660000000000005e-06, "loss": 2.1907, "step": 8132 }, { "epoch": 2.549529780564263, "grad_norm": 15.071435928344727, "learning_rate": 4.066500000000001e-06, "loss": 2.2845, "step": 8133 }, { "epoch": 2.549843260188088, "grad_norm": 7.560123920440674, "learning_rate": 4.067e-06, "loss": 2.1754, "step": 8134 }, { "epoch": 2.5501567398119125, "grad_norm": 19.460298538208008, "learning_rate": 4.0675000000000004e-06, "loss": 2.5934, "step": 8135 }, { "epoch": 2.550470219435737, "grad_norm": 88.80795288085938, "learning_rate": 4.068000000000001e-06, "loss": 2.2202, "step": 8136 }, { "epoch": 2.5507836990595614, "grad_norm": 12.220767974853516, "learning_rate": 4.0685e-06, "loss": 2.1745, "step": 8137 }, { "epoch": 2.551097178683386, "grad_norm": 15.343364715576172, "learning_rate": 4.069e-06, "loss": 2.6924, "step": 8138 }, { "epoch": 2.5514106583072103, "grad_norm": 8.319578170776367, "learning_rate": 4.0695e-06, "loss": 2.0149, "step": 8139 }, { "epoch": 2.5517241379310347, "grad_norm": 27.43376922607422, "learning_rate": 4.07e-06, "loss": 2.7113, "step": 8140 }, { "epoch": 2.552037617554859, "grad_norm": 8.542135238647461, "learning_rate": 4.0705e-06, "loss": 2.0816, "step": 8141 }, { "epoch": 2.5523510971786836, "grad_norm": 7.414121627807617, "learning_rate": 4.0710000000000005e-06, "loss": 2.0104, "step": 8142 }, { "epoch": 2.552664576802508, "grad_norm": 9.500592231750488, "learning_rate": 4.071500000000001e-06, "loss": 2.0282, "step": 8143 }, { "epoch": 2.5529780564263325, "grad_norm": 7.668811798095703, "learning_rate": 4.072e-06, "loss": 2.1187, "step": 8144 }, { "epoch": 2.553291536050157, "grad_norm": 10.514863014221191, "learning_rate": 4.0725e-06, "loss": 2.0162, "step": 8145 }, { "epoch": 2.5536050156739813, "grad_norm": 8.46315860748291, "learning_rate": 4.073000000000001e-06, "loss": 2.1688, "step": 8146 }, { "epoch": 2.553918495297806, "grad_norm": 17.803491592407227, "learning_rate": 4.0735e-06, "loss": 1.9491, "step": 8147 }, { "epoch": 2.5542319749216302, "grad_norm": 17.352296829223633, "learning_rate": 4.074e-06, "loss": 2.3017, "step": 8148 }, { "epoch": 2.5545454545454547, "grad_norm": 29.93524742126465, "learning_rate": 4.0745e-06, "loss": 2.1831, "step": 8149 }, { "epoch": 2.554858934169279, "grad_norm": 11.592525482177734, "learning_rate": 4.075e-06, "loss": 2.2412, "step": 8150 }, { "epoch": 2.5551724137931036, "grad_norm": 8.025237083435059, "learning_rate": 4.0755e-06, "loss": 2.1289, "step": 8151 }, { "epoch": 2.555485893416928, "grad_norm": 22.710350036621094, "learning_rate": 4.0760000000000004e-06, "loss": 2.1923, "step": 8152 }, { "epoch": 2.5557993730407524, "grad_norm": 7.470148086547852, "learning_rate": 4.076500000000001e-06, "loss": 2.1507, "step": 8153 }, { "epoch": 2.556112852664577, "grad_norm": 10.3600435256958, "learning_rate": 4.077e-06, "loss": 2.1169, "step": 8154 }, { "epoch": 2.5564263322884013, "grad_norm": 13.835014343261719, "learning_rate": 4.0775e-06, "loss": 2.241, "step": 8155 }, { "epoch": 2.5567398119122258, "grad_norm": 9.720179557800293, "learning_rate": 4.078000000000001e-06, "loss": 2.3644, "step": 8156 }, { "epoch": 2.55705329153605, "grad_norm": 8.383511543273926, "learning_rate": 4.0785e-06, "loss": 2.245, "step": 8157 }, { "epoch": 2.5573667711598747, "grad_norm": 17.129615783691406, "learning_rate": 4.079e-06, "loss": 2.03, "step": 8158 }, { "epoch": 2.557680250783699, "grad_norm": 24.29751968383789, "learning_rate": 4.0795000000000005e-06, "loss": 2.311, "step": 8159 }, { "epoch": 2.5579937304075235, "grad_norm": 11.373943328857422, "learning_rate": 4.08e-06, "loss": 2.0034, "step": 8160 }, { "epoch": 2.558307210031348, "grad_norm": 9.97189712524414, "learning_rate": 4.0805e-06, "loss": 2.1931, "step": 8161 }, { "epoch": 2.5586206896551724, "grad_norm": 9.933320045471191, "learning_rate": 4.081e-06, "loss": 1.9313, "step": 8162 }, { "epoch": 2.558934169278997, "grad_norm": 12.304068565368652, "learning_rate": 4.081500000000001e-06, "loss": 2.2206, "step": 8163 }, { "epoch": 2.5592476489028213, "grad_norm": 17.47506332397461, "learning_rate": 4.082e-06, "loss": 1.9051, "step": 8164 }, { "epoch": 2.5595611285266457, "grad_norm": 12.494292259216309, "learning_rate": 4.0825e-06, "loss": 2.1742, "step": 8165 }, { "epoch": 2.55987460815047, "grad_norm": 10.735198020935059, "learning_rate": 4.0830000000000005e-06, "loss": 2.2102, "step": 8166 }, { "epoch": 2.5601880877742946, "grad_norm": 7.649764060974121, "learning_rate": 4.0835e-06, "loss": 2.3185, "step": 8167 }, { "epoch": 2.560501567398119, "grad_norm": 17.26466178894043, "learning_rate": 4.084e-06, "loss": 2.0808, "step": 8168 }, { "epoch": 2.5608150470219435, "grad_norm": 79.97025299072266, "learning_rate": 4.0845000000000004e-06, "loss": 2.2742, "step": 8169 }, { "epoch": 2.561128526645768, "grad_norm": 8.160067558288574, "learning_rate": 4.085e-06, "loss": 2.0214, "step": 8170 }, { "epoch": 2.5614420062695924, "grad_norm": 28.319028854370117, "learning_rate": 4.085500000000001e-06, "loss": 2.2699, "step": 8171 }, { "epoch": 2.561755485893417, "grad_norm": 24.294252395629883, "learning_rate": 4.086e-06, "loss": 2.4206, "step": 8172 }, { "epoch": 2.5620689655172413, "grad_norm": 10.982461929321289, "learning_rate": 4.086500000000001e-06, "loss": 2.3321, "step": 8173 }, { "epoch": 2.5623824451410657, "grad_norm": 7.100314140319824, "learning_rate": 4.087e-06, "loss": 2.1242, "step": 8174 }, { "epoch": 2.56269592476489, "grad_norm": 9.78661060333252, "learning_rate": 4.0875e-06, "loss": 2.3951, "step": 8175 }, { "epoch": 2.5630094043887146, "grad_norm": 10.735273361206055, "learning_rate": 4.0880000000000005e-06, "loss": 2.0593, "step": 8176 }, { "epoch": 2.563322884012539, "grad_norm": 21.8847713470459, "learning_rate": 4.0885e-06, "loss": 2.177, "step": 8177 }, { "epoch": 2.5636363636363635, "grad_norm": 10.890393257141113, "learning_rate": 4.089e-06, "loss": 2.358, "step": 8178 }, { "epoch": 2.563949843260188, "grad_norm": 31.5275936126709, "learning_rate": 4.0895e-06, "loss": 2.2002, "step": 8179 }, { "epoch": 2.5642633228840124, "grad_norm": 46.149898529052734, "learning_rate": 4.09e-06, "loss": 2.5792, "step": 8180 }, { "epoch": 2.564576802507837, "grad_norm": 8.118950843811035, "learning_rate": 4.090500000000001e-06, "loss": 1.9171, "step": 8181 }, { "epoch": 2.5648902821316613, "grad_norm": 11.35036849975586, "learning_rate": 4.091e-06, "loss": 2.1263, "step": 8182 }, { "epoch": 2.5652037617554857, "grad_norm": 81.98320770263672, "learning_rate": 4.0915000000000006e-06, "loss": 2.5172, "step": 8183 }, { "epoch": 2.56551724137931, "grad_norm": 26.404911041259766, "learning_rate": 4.092000000000001e-06, "loss": 2.198, "step": 8184 }, { "epoch": 2.5658307210031346, "grad_norm": 44.18767166137695, "learning_rate": 4.0925e-06, "loss": 2.3621, "step": 8185 }, { "epoch": 2.566144200626959, "grad_norm": 15.990948677062988, "learning_rate": 4.0930000000000005e-06, "loss": 2.5489, "step": 8186 }, { "epoch": 2.5664576802507835, "grad_norm": 52.44417953491211, "learning_rate": 4.0935e-06, "loss": 2.434, "step": 8187 }, { "epoch": 2.566771159874608, "grad_norm": 9.85326099395752, "learning_rate": 4.094e-06, "loss": 2.0593, "step": 8188 }, { "epoch": 2.5670846394984324, "grad_norm": 13.923583030700684, "learning_rate": 4.0945e-06, "loss": 2.2647, "step": 8189 }, { "epoch": 2.567398119122257, "grad_norm": 15.340763092041016, "learning_rate": 4.095e-06, "loss": 2.3633, "step": 8190 }, { "epoch": 2.5677115987460812, "grad_norm": 13.438297271728516, "learning_rate": 4.095500000000001e-06, "loss": 2.4, "step": 8191 }, { "epoch": 2.568025078369906, "grad_norm": 16.5212459564209, "learning_rate": 4.096e-06, "loss": 1.8535, "step": 8192 }, { "epoch": 2.5683385579937306, "grad_norm": 13.712716102600098, "learning_rate": 4.0965000000000005e-06, "loss": 2.3463, "step": 8193 }, { "epoch": 2.568652037617555, "grad_norm": 8.175334930419922, "learning_rate": 4.097000000000001e-06, "loss": 2.1563, "step": 8194 }, { "epoch": 2.5689655172413794, "grad_norm": 9.215187072753906, "learning_rate": 4.0975e-06, "loss": 2.6425, "step": 8195 }, { "epoch": 2.569278996865204, "grad_norm": 8.00756549835205, "learning_rate": 4.098e-06, "loss": 2.1191, "step": 8196 }, { "epoch": 2.5695924764890283, "grad_norm": 20.623464584350586, "learning_rate": 4.098500000000001e-06, "loss": 2.1906, "step": 8197 }, { "epoch": 2.5699059561128528, "grad_norm": 8.194958686828613, "learning_rate": 4.099e-06, "loss": 2.2689, "step": 8198 }, { "epoch": 2.570219435736677, "grad_norm": 16.225160598754883, "learning_rate": 4.0995e-06, "loss": 2.1272, "step": 8199 }, { "epoch": 2.5705329153605017, "grad_norm": 8.179239273071289, "learning_rate": 4.1e-06, "loss": 2.2536, "step": 8200 }, { "epoch": 2.570846394984326, "grad_norm": 22.631364822387695, "learning_rate": 4.100500000000001e-06, "loss": 2.9476, "step": 8201 }, { "epoch": 2.5711598746081505, "grad_norm": 56.786895751953125, "learning_rate": 4.101e-06, "loss": 1.9553, "step": 8202 }, { "epoch": 2.571473354231975, "grad_norm": 8.74102783203125, "learning_rate": 4.1015000000000005e-06, "loss": 1.9751, "step": 8203 }, { "epoch": 2.5717868338557994, "grad_norm": 7.660512924194336, "learning_rate": 4.102000000000001e-06, "loss": 2.0875, "step": 8204 }, { "epoch": 2.572100313479624, "grad_norm": 8.298091888427734, "learning_rate": 4.1025e-06, "loss": 2.2124, "step": 8205 }, { "epoch": 2.5724137931034483, "grad_norm": 8.302192687988281, "learning_rate": 4.103e-06, "loss": 2.0961, "step": 8206 }, { "epoch": 2.5727272727272728, "grad_norm": 15.752297401428223, "learning_rate": 4.103500000000001e-06, "loss": 2.4504, "step": 8207 }, { "epoch": 2.573040752351097, "grad_norm": 30.765972137451172, "learning_rate": 4.104e-06, "loss": 2.1763, "step": 8208 }, { "epoch": 2.5733542319749216, "grad_norm": 11.383828163146973, "learning_rate": 4.1045e-06, "loss": 1.8748, "step": 8209 }, { "epoch": 2.573667711598746, "grad_norm": 19.24489402770996, "learning_rate": 4.1050000000000005e-06, "loss": 1.917, "step": 8210 }, { "epoch": 2.5739811912225705, "grad_norm": 11.525765419006348, "learning_rate": 4.105500000000001e-06, "loss": 2.4316, "step": 8211 }, { "epoch": 2.574294670846395, "grad_norm": 9.98393440246582, "learning_rate": 4.106e-06, "loss": 2.2639, "step": 8212 }, { "epoch": 2.5746081504702194, "grad_norm": 7.089967727661133, "learning_rate": 4.1065e-06, "loss": 2.1142, "step": 8213 }, { "epoch": 2.574921630094044, "grad_norm": 16.791013717651367, "learning_rate": 4.107000000000001e-06, "loss": 2.2392, "step": 8214 }, { "epoch": 2.5752351097178683, "grad_norm": 12.266351699829102, "learning_rate": 4.1075e-06, "loss": 2.0372, "step": 8215 }, { "epoch": 2.5755485893416927, "grad_norm": 9.812451362609863, "learning_rate": 4.108e-06, "loss": 2.2624, "step": 8216 }, { "epoch": 2.575862068965517, "grad_norm": 13.189496040344238, "learning_rate": 4.1085000000000006e-06, "loss": 2.039, "step": 8217 }, { "epoch": 2.5761755485893416, "grad_norm": 39.981353759765625, "learning_rate": 4.109e-06, "loss": 2.3199, "step": 8218 }, { "epoch": 2.576489028213166, "grad_norm": 8.123067855834961, "learning_rate": 4.1095e-06, "loss": 2.1402, "step": 8219 }, { "epoch": 2.5768025078369905, "grad_norm": 11.104731559753418, "learning_rate": 4.1100000000000005e-06, "loss": 2.2441, "step": 8220 }, { "epoch": 2.577115987460815, "grad_norm": 9.119978904724121, "learning_rate": 4.110500000000001e-06, "loss": 2.093, "step": 8221 }, { "epoch": 2.5774294670846394, "grad_norm": 6.167422771453857, "learning_rate": 4.111000000000001e-06, "loss": 2.1249, "step": 8222 }, { "epoch": 2.577742946708464, "grad_norm": 10.455440521240234, "learning_rate": 4.1115e-06, "loss": 2.285, "step": 8223 }, { "epoch": 2.5780564263322883, "grad_norm": 19.300506591796875, "learning_rate": 4.112000000000001e-06, "loss": 2.3727, "step": 8224 }, { "epoch": 2.5783699059561127, "grad_norm": 9.30392074584961, "learning_rate": 4.1125e-06, "loss": 2.0403, "step": 8225 }, { "epoch": 2.578683385579937, "grad_norm": 20.748441696166992, "learning_rate": 4.113e-06, "loss": 2.0922, "step": 8226 }, { "epoch": 2.5789968652037616, "grad_norm": 11.308359146118164, "learning_rate": 4.1135000000000005e-06, "loss": 2.105, "step": 8227 }, { "epoch": 2.5793103448275865, "grad_norm": 22.46990394592285, "learning_rate": 4.114e-06, "loss": 2.0555, "step": 8228 }, { "epoch": 2.579623824451411, "grad_norm": 76.13819885253906, "learning_rate": 4.1145e-06, "loss": 2.4406, "step": 8229 }, { "epoch": 2.5799373040752354, "grad_norm": 14.803643226623535, "learning_rate": 4.115e-06, "loss": 2.2998, "step": 8230 }, { "epoch": 2.58025078369906, "grad_norm": 12.239740371704102, "learning_rate": 4.115500000000001e-06, "loss": 2.1172, "step": 8231 }, { "epoch": 2.5805642633228842, "grad_norm": 11.718621253967285, "learning_rate": 4.116000000000001e-06, "loss": 2.1346, "step": 8232 }, { "epoch": 2.5808777429467087, "grad_norm": 12.111909866333008, "learning_rate": 4.1165e-06, "loss": 2.0387, "step": 8233 }, { "epoch": 2.581191222570533, "grad_norm": 15.876483917236328, "learning_rate": 4.1170000000000006e-06, "loss": 2.7176, "step": 8234 }, { "epoch": 2.5815047021943576, "grad_norm": 15.144723892211914, "learning_rate": 4.1175e-06, "loss": 2.6972, "step": 8235 }, { "epoch": 2.581818181818182, "grad_norm": 11.002240180969238, "learning_rate": 4.118e-06, "loss": 2.1327, "step": 8236 }, { "epoch": 2.5821316614420065, "grad_norm": 24.561195373535156, "learning_rate": 4.1185000000000005e-06, "loss": 2.211, "step": 8237 }, { "epoch": 2.582445141065831, "grad_norm": 9.62204647064209, "learning_rate": 4.119e-06, "loss": 2.246, "step": 8238 }, { "epoch": 2.5827586206896553, "grad_norm": 9.03166389465332, "learning_rate": 4.1195e-06, "loss": 1.9755, "step": 8239 }, { "epoch": 2.58307210031348, "grad_norm": 15.370232582092285, "learning_rate": 4.12e-06, "loss": 2.3042, "step": 8240 }, { "epoch": 2.583385579937304, "grad_norm": 13.837961196899414, "learning_rate": 4.120500000000001e-06, "loss": 2.1086, "step": 8241 }, { "epoch": 2.5836990595611287, "grad_norm": 56.6268196105957, "learning_rate": 4.121000000000001e-06, "loss": 2.3763, "step": 8242 }, { "epoch": 2.584012539184953, "grad_norm": 66.34306335449219, "learning_rate": 4.1215e-06, "loss": 2.5498, "step": 8243 }, { "epoch": 2.5843260188087775, "grad_norm": 9.78904914855957, "learning_rate": 4.1220000000000005e-06, "loss": 2.2488, "step": 8244 }, { "epoch": 2.584639498432602, "grad_norm": 8.285097122192383, "learning_rate": 4.122500000000001e-06, "loss": 2.0399, "step": 8245 }, { "epoch": 2.5849529780564264, "grad_norm": 13.0098876953125, "learning_rate": 4.123e-06, "loss": 2.3571, "step": 8246 }, { "epoch": 2.585266457680251, "grad_norm": 8.850469589233398, "learning_rate": 4.1235e-06, "loss": 2.3183, "step": 8247 }, { "epoch": 2.5855799373040753, "grad_norm": 34.077693939208984, "learning_rate": 4.124e-06, "loss": 2.1661, "step": 8248 }, { "epoch": 2.5858934169278998, "grad_norm": 13.503628730773926, "learning_rate": 4.1245e-06, "loss": 2.0544, "step": 8249 }, { "epoch": 2.586206896551724, "grad_norm": 9.38256549835205, "learning_rate": 4.125e-06, "loss": 1.9256, "step": 8250 }, { "epoch": 2.5865203761755486, "grad_norm": 8.2373628616333, "learning_rate": 4.1255e-06, "loss": 2.29, "step": 8251 }, { "epoch": 2.586833855799373, "grad_norm": 21.7275390625, "learning_rate": 4.126000000000001e-06, "loss": 2.2198, "step": 8252 }, { "epoch": 2.5871473354231975, "grad_norm": 8.765398979187012, "learning_rate": 4.1265e-06, "loss": 2.3538, "step": 8253 }, { "epoch": 2.587460815047022, "grad_norm": 9.432440757751465, "learning_rate": 4.1270000000000005e-06, "loss": 2.3259, "step": 8254 }, { "epoch": 2.5877742946708464, "grad_norm": 10.251505851745605, "learning_rate": 4.127500000000001e-06, "loss": 2.2047, "step": 8255 }, { "epoch": 2.588087774294671, "grad_norm": 23.77132797241211, "learning_rate": 4.128e-06, "loss": 2.3561, "step": 8256 }, { "epoch": 2.5884012539184953, "grad_norm": 10.63782787322998, "learning_rate": 4.1285e-06, "loss": 1.9518, "step": 8257 }, { "epoch": 2.5887147335423197, "grad_norm": 67.7890853881836, "learning_rate": 4.129000000000001e-06, "loss": 2.6671, "step": 8258 }, { "epoch": 2.589028213166144, "grad_norm": 9.824442863464355, "learning_rate": 4.1295e-06, "loss": 2.2139, "step": 8259 }, { "epoch": 2.5893416927899686, "grad_norm": 8.935091018676758, "learning_rate": 4.13e-06, "loss": 2.268, "step": 8260 }, { "epoch": 2.589655172413793, "grad_norm": 21.136892318725586, "learning_rate": 4.1305e-06, "loss": 2.184, "step": 8261 }, { "epoch": 2.5899686520376175, "grad_norm": 28.960996627807617, "learning_rate": 4.131000000000001e-06, "loss": 2.2074, "step": 8262 }, { "epoch": 2.590282131661442, "grad_norm": 11.432787895202637, "learning_rate": 4.1315e-06, "loss": 2.0248, "step": 8263 }, { "epoch": 2.5905956112852664, "grad_norm": 26.409940719604492, "learning_rate": 4.132e-06, "loss": 2.2092, "step": 8264 }, { "epoch": 2.590909090909091, "grad_norm": 10.69225788116455, "learning_rate": 4.132500000000001e-06, "loss": 2.1662, "step": 8265 }, { "epoch": 2.5912225705329153, "grad_norm": 9.840431213378906, "learning_rate": 4.133e-06, "loss": 2.3937, "step": 8266 }, { "epoch": 2.5915360501567397, "grad_norm": 8.000717163085938, "learning_rate": 4.1335e-06, "loss": 2.6257, "step": 8267 }, { "epoch": 2.591849529780564, "grad_norm": 66.77828979492188, "learning_rate": 4.1340000000000006e-06, "loss": 2.2849, "step": 8268 }, { "epoch": 2.5921630094043886, "grad_norm": 7.377851963043213, "learning_rate": 4.1345e-06, "loss": 2.2119, "step": 8269 }, { "epoch": 2.592476489028213, "grad_norm": 14.871600151062012, "learning_rate": 4.135e-06, "loss": 2.1306, "step": 8270 }, { "epoch": 2.5927899686520375, "grad_norm": 12.94255256652832, "learning_rate": 4.1355000000000005e-06, "loss": 2.2755, "step": 8271 }, { "epoch": 2.593103448275862, "grad_norm": 14.233874320983887, "learning_rate": 4.136000000000001e-06, "loss": 2.092, "step": 8272 }, { "epoch": 2.5934169278996864, "grad_norm": 9.9481201171875, "learning_rate": 4.1365e-06, "loss": 2.1215, "step": 8273 }, { "epoch": 2.593730407523511, "grad_norm": 8.752227783203125, "learning_rate": 4.137e-06, "loss": 2.1772, "step": 8274 }, { "epoch": 2.5940438871473352, "grad_norm": 41.305946350097656, "learning_rate": 4.137500000000001e-06, "loss": 2.2434, "step": 8275 }, { "epoch": 2.5943573667711597, "grad_norm": 9.648118019104004, "learning_rate": 4.138e-06, "loss": 2.1772, "step": 8276 }, { "epoch": 2.594670846394984, "grad_norm": 7.743842124938965, "learning_rate": 4.1385e-06, "loss": 2.5681, "step": 8277 }, { "epoch": 2.5949843260188086, "grad_norm": 21.656124114990234, "learning_rate": 4.1390000000000005e-06, "loss": 2.2791, "step": 8278 }, { "epoch": 2.595297805642633, "grad_norm": 29.323015213012695, "learning_rate": 4.1395e-06, "loss": 2.1858, "step": 8279 }, { "epoch": 2.5956112852664575, "grad_norm": 8.343573570251465, "learning_rate": 4.14e-06, "loss": 2.1422, "step": 8280 }, { "epoch": 2.595924764890282, "grad_norm": 5.896507740020752, "learning_rate": 4.1405e-06, "loss": 1.9595, "step": 8281 }, { "epoch": 2.5962382445141063, "grad_norm": 8.935463905334473, "learning_rate": 4.141000000000001e-06, "loss": 1.7961, "step": 8282 }, { "epoch": 2.596551724137931, "grad_norm": 9.092925071716309, "learning_rate": 4.141500000000001e-06, "loss": 2.0055, "step": 8283 }, { "epoch": 2.5968652037617552, "grad_norm": 24.03822898864746, "learning_rate": 4.142e-06, "loss": 2.2324, "step": 8284 }, { "epoch": 2.5971786833855797, "grad_norm": 8.95445728302002, "learning_rate": 4.1425000000000006e-06, "loss": 1.9684, "step": 8285 }, { "epoch": 2.5974921630094046, "grad_norm": 38.807498931884766, "learning_rate": 4.143e-06, "loss": 2.3632, "step": 8286 }, { "epoch": 2.597805642633229, "grad_norm": 8.724321365356445, "learning_rate": 4.1435e-06, "loss": 2.148, "step": 8287 }, { "epoch": 2.5981191222570534, "grad_norm": 44.358131408691406, "learning_rate": 4.1440000000000005e-06, "loss": 2.0427, "step": 8288 }, { "epoch": 2.598432601880878, "grad_norm": 11.453152656555176, "learning_rate": 4.1445e-06, "loss": 2.4092, "step": 8289 }, { "epoch": 2.5987460815047023, "grad_norm": 15.903132438659668, "learning_rate": 4.145e-06, "loss": 2.2588, "step": 8290 }, { "epoch": 2.5990595611285268, "grad_norm": 28.336408615112305, "learning_rate": 4.1455e-06, "loss": 2.0542, "step": 8291 }, { "epoch": 2.599373040752351, "grad_norm": 7.553894996643066, "learning_rate": 4.146000000000001e-06, "loss": 2.2797, "step": 8292 }, { "epoch": 2.5996865203761756, "grad_norm": 54.837589263916016, "learning_rate": 4.146500000000001e-06, "loss": 2.8817, "step": 8293 }, { "epoch": 2.6, "grad_norm": 10.172371864318848, "learning_rate": 4.147e-06, "loss": 2.0674, "step": 8294 }, { "epoch": 2.6003134796238245, "grad_norm": 9.880024909973145, "learning_rate": 4.1475000000000005e-06, "loss": 2.1057, "step": 8295 }, { "epoch": 2.600626959247649, "grad_norm": 23.346281051635742, "learning_rate": 4.148000000000001e-06, "loss": 2.2679, "step": 8296 }, { "epoch": 2.6009404388714734, "grad_norm": 8.534043312072754, "learning_rate": 4.1485e-06, "loss": 1.9373, "step": 8297 }, { "epoch": 2.601253918495298, "grad_norm": 6.999251365661621, "learning_rate": 4.1490000000000004e-06, "loss": 2.1854, "step": 8298 }, { "epoch": 2.6015673981191223, "grad_norm": 12.397709846496582, "learning_rate": 4.1495e-06, "loss": 2.0382, "step": 8299 }, { "epoch": 2.6018808777429467, "grad_norm": 14.955727577209473, "learning_rate": 4.15e-06, "loss": 2.1716, "step": 8300 }, { "epoch": 2.602194357366771, "grad_norm": 10.29279613494873, "learning_rate": 4.1505e-06, "loss": 2.3876, "step": 8301 }, { "epoch": 2.6025078369905956, "grad_norm": 17.181440353393555, "learning_rate": 4.1510000000000006e-06, "loss": 2.1746, "step": 8302 }, { "epoch": 2.60282131661442, "grad_norm": 11.074188232421875, "learning_rate": 4.151500000000001e-06, "loss": 2.319, "step": 8303 }, { "epoch": 2.6031347962382445, "grad_norm": 11.742688179016113, "learning_rate": 4.152e-06, "loss": 1.9238, "step": 8304 }, { "epoch": 2.603448275862069, "grad_norm": 9.570671081542969, "learning_rate": 4.1525000000000005e-06, "loss": 2.2312, "step": 8305 }, { "epoch": 2.6037617554858934, "grad_norm": 14.536748886108398, "learning_rate": 4.153000000000001e-06, "loss": 2.0083, "step": 8306 }, { "epoch": 2.604075235109718, "grad_norm": 35.707210540771484, "learning_rate": 4.1535e-06, "loss": 2.9085, "step": 8307 }, { "epoch": 2.6043887147335423, "grad_norm": 15.03381061553955, "learning_rate": 4.154e-06, "loss": 2.3617, "step": 8308 }, { "epoch": 2.6047021943573667, "grad_norm": 49.565616607666016, "learning_rate": 4.1545e-06, "loss": 2.2874, "step": 8309 }, { "epoch": 2.605015673981191, "grad_norm": 14.380892753601074, "learning_rate": 4.155e-06, "loss": 2.1433, "step": 8310 }, { "epoch": 2.6053291536050156, "grad_norm": 8.938322067260742, "learning_rate": 4.1555e-06, "loss": 2.1393, "step": 8311 }, { "epoch": 2.60564263322884, "grad_norm": 9.698333740234375, "learning_rate": 4.1560000000000005e-06, "loss": 2.2165, "step": 8312 }, { "epoch": 2.6059561128526645, "grad_norm": 20.239227294921875, "learning_rate": 4.156500000000001e-06, "loss": 2.3463, "step": 8313 }, { "epoch": 2.606269592476489, "grad_norm": 12.427850723266602, "learning_rate": 4.157e-06, "loss": 2.2235, "step": 8314 }, { "epoch": 2.6065830721003134, "grad_norm": 6.410386085510254, "learning_rate": 4.1575000000000004e-06, "loss": 2.194, "step": 8315 }, { "epoch": 2.606896551724138, "grad_norm": 6.156004428863525, "learning_rate": 4.158000000000001e-06, "loss": 2.0916, "step": 8316 }, { "epoch": 2.6072100313479623, "grad_norm": 8.54709243774414, "learning_rate": 4.1585e-06, "loss": 2.1952, "step": 8317 }, { "epoch": 2.6075235109717867, "grad_norm": 8.950973510742188, "learning_rate": 4.159e-06, "loss": 2.1056, "step": 8318 }, { "epoch": 2.607836990595611, "grad_norm": 9.264294624328613, "learning_rate": 4.159500000000001e-06, "loss": 2.7437, "step": 8319 }, { "epoch": 2.6081504702194356, "grad_norm": 18.964914321899414, "learning_rate": 4.16e-06, "loss": 2.0996, "step": 8320 }, { "epoch": 2.6084639498432605, "grad_norm": 12.056415557861328, "learning_rate": 4.1605e-06, "loss": 2.1652, "step": 8321 }, { "epoch": 2.608777429467085, "grad_norm": 19.019515991210938, "learning_rate": 4.1610000000000005e-06, "loss": 2.2533, "step": 8322 }, { "epoch": 2.6090909090909093, "grad_norm": 9.825438499450684, "learning_rate": 4.161500000000001e-06, "loss": 2.0973, "step": 8323 }, { "epoch": 2.609404388714734, "grad_norm": 9.150247573852539, "learning_rate": 4.162e-06, "loss": 2.4399, "step": 8324 }, { "epoch": 2.6097178683385582, "grad_norm": 16.464221954345703, "learning_rate": 4.1625e-06, "loss": 2.2283, "step": 8325 }, { "epoch": 2.6100313479623827, "grad_norm": 37.46894073486328, "learning_rate": 4.163000000000001e-06, "loss": 2.2384, "step": 8326 }, { "epoch": 2.610344827586207, "grad_norm": 9.69559383392334, "learning_rate": 4.1635e-06, "loss": 2.1118, "step": 8327 }, { "epoch": 2.6106583072100316, "grad_norm": 7.265785217285156, "learning_rate": 4.164e-06, "loss": 2.3269, "step": 8328 }, { "epoch": 2.610971786833856, "grad_norm": 7.528814315795898, "learning_rate": 4.1645000000000005e-06, "loss": 2.0235, "step": 8329 }, { "epoch": 2.6112852664576804, "grad_norm": 17.32964515686035, "learning_rate": 4.165e-06, "loss": 2.2659, "step": 8330 }, { "epoch": 2.611598746081505, "grad_norm": 31.277774810791016, "learning_rate": 4.1655e-06, "loss": 2.2757, "step": 8331 }, { "epoch": 2.6119122257053293, "grad_norm": 9.275627136230469, "learning_rate": 4.1660000000000004e-06, "loss": 2.1844, "step": 8332 }, { "epoch": 2.6122257053291538, "grad_norm": 7.434737205505371, "learning_rate": 4.166500000000001e-06, "loss": 2.0292, "step": 8333 }, { "epoch": 2.612539184952978, "grad_norm": 14.723471641540527, "learning_rate": 4.167e-06, "loss": 2.1213, "step": 8334 }, { "epoch": 2.6128526645768027, "grad_norm": 12.374506950378418, "learning_rate": 4.1675e-06, "loss": 2.3284, "step": 8335 }, { "epoch": 2.613166144200627, "grad_norm": 30.19709014892578, "learning_rate": 4.168000000000001e-06, "loss": 2.1647, "step": 8336 }, { "epoch": 2.6134796238244515, "grad_norm": 8.046975135803223, "learning_rate": 4.1685e-06, "loss": 1.9632, "step": 8337 }, { "epoch": 2.613793103448276, "grad_norm": 11.449087142944336, "learning_rate": 4.169e-06, "loss": 2.1301, "step": 8338 }, { "epoch": 2.6141065830721004, "grad_norm": 12.102714538574219, "learning_rate": 4.1695000000000005e-06, "loss": 2.0929, "step": 8339 }, { "epoch": 2.614420062695925, "grad_norm": 31.399158477783203, "learning_rate": 4.17e-06, "loss": 2.2725, "step": 8340 }, { "epoch": 2.6147335423197493, "grad_norm": 15.629863739013672, "learning_rate": 4.1705e-06, "loss": 2.0198, "step": 8341 }, { "epoch": 2.6150470219435737, "grad_norm": 10.33690071105957, "learning_rate": 4.171e-06, "loss": 1.947, "step": 8342 }, { "epoch": 2.615360501567398, "grad_norm": 43.66338348388672, "learning_rate": 4.171500000000001e-06, "loss": 2.3471, "step": 8343 }, { "epoch": 2.6156739811912226, "grad_norm": 10.542743682861328, "learning_rate": 4.172000000000001e-06, "loss": 2.1413, "step": 8344 }, { "epoch": 2.615987460815047, "grad_norm": 10.060778617858887, "learning_rate": 4.1725e-06, "loss": 2.1483, "step": 8345 }, { "epoch": 2.6163009404388715, "grad_norm": 20.161468505859375, "learning_rate": 4.1730000000000005e-06, "loss": 2.3394, "step": 8346 }, { "epoch": 2.616614420062696, "grad_norm": 8.513916969299316, "learning_rate": 4.1735e-06, "loss": 1.8955, "step": 8347 }, { "epoch": 2.6169278996865204, "grad_norm": 25.299470901489258, "learning_rate": 4.174e-06, "loss": 2.1776, "step": 8348 }, { "epoch": 2.617241379310345, "grad_norm": 14.288599014282227, "learning_rate": 4.1745000000000004e-06, "loss": 2.0739, "step": 8349 }, { "epoch": 2.6175548589341693, "grad_norm": 40.74323272705078, "learning_rate": 4.175e-06, "loss": 1.9644, "step": 8350 }, { "epoch": 2.6178683385579937, "grad_norm": 12.229666709899902, "learning_rate": 4.1755e-06, "loss": 2.1388, "step": 8351 }, { "epoch": 2.618181818181818, "grad_norm": 41.0211296081543, "learning_rate": 4.176e-06, "loss": 2.3576, "step": 8352 }, { "epoch": 2.6184952978056426, "grad_norm": 10.737504005432129, "learning_rate": 4.176500000000001e-06, "loss": 2.1398, "step": 8353 }, { "epoch": 2.618808777429467, "grad_norm": 34.50044250488281, "learning_rate": 4.177000000000001e-06, "loss": 2.2027, "step": 8354 }, { "epoch": 2.6191222570532915, "grad_norm": 15.533110618591309, "learning_rate": 4.1775e-06, "loss": 2.6283, "step": 8355 }, { "epoch": 2.619435736677116, "grad_norm": 40.24463653564453, "learning_rate": 4.1780000000000005e-06, "loss": 2.3389, "step": 8356 }, { "epoch": 2.6197492163009404, "grad_norm": 19.797321319580078, "learning_rate": 4.178500000000001e-06, "loss": 2.3079, "step": 8357 }, { "epoch": 2.620062695924765, "grad_norm": 10.649677276611328, "learning_rate": 4.179e-06, "loss": 2.3513, "step": 8358 }, { "epoch": 2.6203761755485893, "grad_norm": 10.62403678894043, "learning_rate": 4.1795e-06, "loss": 2.1037, "step": 8359 }, { "epoch": 2.6206896551724137, "grad_norm": 29.208194732666016, "learning_rate": 4.18e-06, "loss": 2.3904, "step": 8360 }, { "epoch": 2.621003134796238, "grad_norm": 7.979300022125244, "learning_rate": 4.1805e-06, "loss": 2.1199, "step": 8361 }, { "epoch": 2.6213166144200626, "grad_norm": 7.917868614196777, "learning_rate": 4.181e-06, "loss": 1.8131, "step": 8362 }, { "epoch": 2.621630094043887, "grad_norm": 6.296743392944336, "learning_rate": 4.1815000000000005e-06, "loss": 1.9427, "step": 8363 }, { "epoch": 2.6219435736677115, "grad_norm": 20.250572204589844, "learning_rate": 4.182000000000001e-06, "loss": 2.1859, "step": 8364 }, { "epoch": 2.622257053291536, "grad_norm": 21.719078063964844, "learning_rate": 4.1825e-06, "loss": 2.1713, "step": 8365 }, { "epoch": 2.6225705329153604, "grad_norm": 6.516077518463135, "learning_rate": 4.1830000000000004e-06, "loss": 2.0094, "step": 8366 }, { "epoch": 2.622884012539185, "grad_norm": 12.200885772705078, "learning_rate": 4.183500000000001e-06, "loss": 1.9532, "step": 8367 }, { "epoch": 2.6231974921630092, "grad_norm": 8.400837898254395, "learning_rate": 4.184e-06, "loss": 2.1127, "step": 8368 }, { "epoch": 2.6235109717868337, "grad_norm": 11.864903450012207, "learning_rate": 4.1845e-06, "loss": 2.0255, "step": 8369 }, { "epoch": 2.623824451410658, "grad_norm": 204.66696166992188, "learning_rate": 4.185000000000001e-06, "loss": 2.6827, "step": 8370 }, { "epoch": 2.6241379310344826, "grad_norm": 9.072761535644531, "learning_rate": 4.1855e-06, "loss": 2.2119, "step": 8371 }, { "epoch": 2.624451410658307, "grad_norm": 23.512405395507812, "learning_rate": 4.186e-06, "loss": 2.0022, "step": 8372 }, { "epoch": 2.6247648902821314, "grad_norm": 8.300741195678711, "learning_rate": 4.1865000000000005e-06, "loss": 2.024, "step": 8373 }, { "epoch": 2.625078369905956, "grad_norm": 7.966546058654785, "learning_rate": 4.187000000000001e-06, "loss": 2.1199, "step": 8374 }, { "epoch": 2.6253918495297803, "grad_norm": 12.485355377197266, "learning_rate": 4.1875e-06, "loss": 2.11, "step": 8375 }, { "epoch": 2.6257053291536048, "grad_norm": 12.69727611541748, "learning_rate": 4.188e-06, "loss": 2.2185, "step": 8376 }, { "epoch": 2.626018808777429, "grad_norm": 32.500362396240234, "learning_rate": 4.188500000000001e-06, "loss": 2.0866, "step": 8377 }, { "epoch": 2.6263322884012537, "grad_norm": 19.909923553466797, "learning_rate": 4.189e-06, "loss": 2.1419, "step": 8378 }, { "epoch": 2.626645768025078, "grad_norm": 12.31171703338623, "learning_rate": 4.1895e-06, "loss": 2.2196, "step": 8379 }, { "epoch": 2.626959247648903, "grad_norm": 35.06986618041992, "learning_rate": 4.1900000000000005e-06, "loss": 2.2653, "step": 8380 }, { "epoch": 2.6272727272727274, "grad_norm": 15.9246826171875, "learning_rate": 4.1905e-06, "loss": 2.0954, "step": 8381 }, { "epoch": 2.627586206896552, "grad_norm": 8.85057544708252, "learning_rate": 4.191e-06, "loss": 2.0356, "step": 8382 }, { "epoch": 2.6278996865203763, "grad_norm": 9.108238220214844, "learning_rate": 4.1915000000000004e-06, "loss": 2.1692, "step": 8383 }, { "epoch": 2.6282131661442008, "grad_norm": 7.710699081420898, "learning_rate": 4.192000000000001e-06, "loss": 2.0578, "step": 8384 }, { "epoch": 2.628526645768025, "grad_norm": 10.245841026306152, "learning_rate": 4.1925e-06, "loss": 2.0849, "step": 8385 }, { "epoch": 2.6288401253918496, "grad_norm": 10.73650074005127, "learning_rate": 4.193e-06, "loss": 2.0568, "step": 8386 }, { "epoch": 2.629153605015674, "grad_norm": 14.61035442352295, "learning_rate": 4.193500000000001e-06, "loss": 2.17, "step": 8387 }, { "epoch": 2.6294670846394985, "grad_norm": 12.7315034866333, "learning_rate": 4.194e-06, "loss": 2.0286, "step": 8388 }, { "epoch": 2.629780564263323, "grad_norm": 21.808212280273438, "learning_rate": 4.1945e-06, "loss": 2.1631, "step": 8389 }, { "epoch": 2.6300940438871474, "grad_norm": 8.714883804321289, "learning_rate": 4.1950000000000005e-06, "loss": 2.174, "step": 8390 }, { "epoch": 2.630407523510972, "grad_norm": 9.587726593017578, "learning_rate": 4.1955e-06, "loss": 2.0854, "step": 8391 }, { "epoch": 2.6307210031347963, "grad_norm": 13.802691459655762, "learning_rate": 4.196e-06, "loss": 2.2901, "step": 8392 }, { "epoch": 2.6310344827586207, "grad_norm": 10.187498092651367, "learning_rate": 4.1965e-06, "loss": 2.0632, "step": 8393 }, { "epoch": 2.631347962382445, "grad_norm": 15.831109046936035, "learning_rate": 4.197000000000001e-06, "loss": 1.9084, "step": 8394 }, { "epoch": 2.6316614420062696, "grad_norm": 8.753616333007812, "learning_rate": 4.1975e-06, "loss": 2.1953, "step": 8395 }, { "epoch": 2.631974921630094, "grad_norm": 8.83792495727539, "learning_rate": 4.198e-06, "loss": 2.1549, "step": 8396 }, { "epoch": 2.6322884012539185, "grad_norm": 10.381096839904785, "learning_rate": 4.1985000000000005e-06, "loss": 2.02, "step": 8397 }, { "epoch": 2.632601880877743, "grad_norm": 12.265913963317871, "learning_rate": 4.199e-06, "loss": 2.1735, "step": 8398 }, { "epoch": 2.6329153605015674, "grad_norm": 10.290362358093262, "learning_rate": 4.1995e-06, "loss": 2.0504, "step": 8399 }, { "epoch": 2.633228840125392, "grad_norm": 12.413207054138184, "learning_rate": 4.2000000000000004e-06, "loss": 2.339, "step": 8400 }, { "epoch": 2.6335423197492163, "grad_norm": 31.935354232788086, "learning_rate": 4.2005e-06, "loss": 2.0376, "step": 8401 }, { "epoch": 2.6338557993730407, "grad_norm": 57.332950592041016, "learning_rate": 4.201e-06, "loss": 2.4084, "step": 8402 }, { "epoch": 2.634169278996865, "grad_norm": 9.95567798614502, "learning_rate": 4.2015e-06, "loss": 2.0783, "step": 8403 }, { "epoch": 2.6344827586206896, "grad_norm": 9.191915512084961, "learning_rate": 4.202000000000001e-06, "loss": 1.9481, "step": 8404 }, { "epoch": 2.634796238244514, "grad_norm": 136.2806396484375, "learning_rate": 4.202500000000001e-06, "loss": 2.2298, "step": 8405 }, { "epoch": 2.6351097178683385, "grad_norm": 26.957263946533203, "learning_rate": 4.203e-06, "loss": 2.2142, "step": 8406 }, { "epoch": 2.635423197492163, "grad_norm": 69.08128356933594, "learning_rate": 4.2035000000000005e-06, "loss": 2.7577, "step": 8407 }, { "epoch": 2.6357366771159874, "grad_norm": 25.558231353759766, "learning_rate": 4.204e-06, "loss": 2.3011, "step": 8408 }, { "epoch": 2.636050156739812, "grad_norm": 9.680856704711914, "learning_rate": 4.2045e-06, "loss": 2.2343, "step": 8409 }, { "epoch": 2.6363636363636362, "grad_norm": 7.282766342163086, "learning_rate": 4.205e-06, "loss": 2.1086, "step": 8410 }, { "epoch": 2.6366771159874607, "grad_norm": 12.967825889587402, "learning_rate": 4.2055e-06, "loss": 2.0128, "step": 8411 }, { "epoch": 2.636990595611285, "grad_norm": 46.111629486083984, "learning_rate": 4.206e-06, "loss": 2.5771, "step": 8412 }, { "epoch": 2.6373040752351096, "grad_norm": 16.87361717224121, "learning_rate": 4.2065e-06, "loss": 2.0832, "step": 8413 }, { "epoch": 2.637617554858934, "grad_norm": 83.00596618652344, "learning_rate": 4.2070000000000005e-06, "loss": 2.3681, "step": 8414 }, { "epoch": 2.637931034482759, "grad_norm": 6.852484226226807, "learning_rate": 4.207500000000001e-06, "loss": 2.0815, "step": 8415 }, { "epoch": 2.6382445141065833, "grad_norm": 31.999250411987305, "learning_rate": 4.208e-06, "loss": 2.5717, "step": 8416 }, { "epoch": 2.638557993730408, "grad_norm": 76.99288940429688, "learning_rate": 4.2085000000000004e-06, "loss": 2.3171, "step": 8417 }, { "epoch": 2.6388714733542322, "grad_norm": 86.08372497558594, "learning_rate": 4.209000000000001e-06, "loss": 1.895, "step": 8418 }, { "epoch": 2.6391849529780567, "grad_norm": 57.6121711730957, "learning_rate": 4.2095e-06, "loss": 2.2193, "step": 8419 }, { "epoch": 2.639498432601881, "grad_norm": 7.138797760009766, "learning_rate": 4.21e-06, "loss": 1.895, "step": 8420 }, { "epoch": 2.6398119122257055, "grad_norm": 12.953635215759277, "learning_rate": 4.2105e-06, "loss": 2.1682, "step": 8421 }, { "epoch": 2.64012539184953, "grad_norm": 11.97398567199707, "learning_rate": 4.211e-06, "loss": 2.0887, "step": 8422 }, { "epoch": 2.6404388714733544, "grad_norm": 16.1533260345459, "learning_rate": 4.2115e-06, "loss": 1.9837, "step": 8423 }, { "epoch": 2.640752351097179, "grad_norm": 10.425740242004395, "learning_rate": 4.2120000000000005e-06, "loss": 2.5789, "step": 8424 }, { "epoch": 2.6410658307210033, "grad_norm": 7.5437140464782715, "learning_rate": 4.212500000000001e-06, "loss": 2.4158, "step": 8425 }, { "epoch": 2.6413793103448278, "grad_norm": 12.495124816894531, "learning_rate": 4.213e-06, "loss": 1.9595, "step": 8426 }, { "epoch": 2.641692789968652, "grad_norm": 29.08905029296875, "learning_rate": 4.2135e-06, "loss": 2.6104, "step": 8427 }, { "epoch": 2.6420062695924766, "grad_norm": 8.886795997619629, "learning_rate": 4.214000000000001e-06, "loss": 1.9153, "step": 8428 }, { "epoch": 2.642319749216301, "grad_norm": 24.72938346862793, "learning_rate": 4.2145e-06, "loss": 2.2358, "step": 8429 }, { "epoch": 2.6426332288401255, "grad_norm": 11.306891441345215, "learning_rate": 4.215e-06, "loss": 1.9679, "step": 8430 }, { "epoch": 2.64294670846395, "grad_norm": 12.032949447631836, "learning_rate": 4.2155000000000006e-06, "loss": 2.0856, "step": 8431 }, { "epoch": 2.6432601880877744, "grad_norm": 77.22188568115234, "learning_rate": 4.216e-06, "loss": 2.3835, "step": 8432 }, { "epoch": 2.643573667711599, "grad_norm": 5.402823448181152, "learning_rate": 4.2165e-06, "loss": 2.2042, "step": 8433 }, { "epoch": 2.6438871473354233, "grad_norm": 11.13780689239502, "learning_rate": 4.2170000000000005e-06, "loss": 2.0214, "step": 8434 }, { "epoch": 2.6442006269592477, "grad_norm": 30.66798210144043, "learning_rate": 4.217500000000001e-06, "loss": 2.2739, "step": 8435 }, { "epoch": 2.644514106583072, "grad_norm": 6.225828170776367, "learning_rate": 4.218e-06, "loss": 2.1119, "step": 8436 }, { "epoch": 2.6448275862068966, "grad_norm": 253.38861083984375, "learning_rate": 4.2185e-06, "loss": 2.8377, "step": 8437 }, { "epoch": 2.645141065830721, "grad_norm": 13.203694343566895, "learning_rate": 4.219000000000001e-06, "loss": 2.4006, "step": 8438 }, { "epoch": 2.6454545454545455, "grad_norm": 9.309622764587402, "learning_rate": 4.2195e-06, "loss": 2.3324, "step": 8439 }, { "epoch": 2.64576802507837, "grad_norm": 76.7125473022461, "learning_rate": 4.22e-06, "loss": 2.3917, "step": 8440 }, { "epoch": 2.6460815047021944, "grad_norm": 10.378171920776367, "learning_rate": 4.2205000000000005e-06, "loss": 2.7759, "step": 8441 }, { "epoch": 2.646394984326019, "grad_norm": 77.28691101074219, "learning_rate": 4.221e-06, "loss": 2.7859, "step": 8442 }, { "epoch": 2.6467084639498433, "grad_norm": 8.362545013427734, "learning_rate": 4.221500000000001e-06, "loss": 1.9511, "step": 8443 }, { "epoch": 2.6470219435736677, "grad_norm": 12.6875581741333, "learning_rate": 4.222e-06, "loss": 1.8518, "step": 8444 }, { "epoch": 2.647335423197492, "grad_norm": 5.32600212097168, "learning_rate": 4.222500000000001e-06, "loss": 2.2246, "step": 8445 }, { "epoch": 2.6476489028213166, "grad_norm": 11.357797622680664, "learning_rate": 4.223e-06, "loss": 2.1165, "step": 8446 }, { "epoch": 2.647962382445141, "grad_norm": 10.440389633178711, "learning_rate": 4.2235e-06, "loss": 2.1616, "step": 8447 }, { "epoch": 2.6482758620689655, "grad_norm": 46.6624755859375, "learning_rate": 4.2240000000000006e-06, "loss": 2.1594, "step": 8448 }, { "epoch": 2.64858934169279, "grad_norm": 28.59185218811035, "learning_rate": 4.2245e-06, "loss": 2.1961, "step": 8449 }, { "epoch": 2.6489028213166144, "grad_norm": 7.238767147064209, "learning_rate": 4.225e-06, "loss": 2.1646, "step": 8450 }, { "epoch": 2.649216300940439, "grad_norm": 12.251442909240723, "learning_rate": 4.2255000000000005e-06, "loss": 2.3611, "step": 8451 }, { "epoch": 2.6495297805642632, "grad_norm": 9.600554466247559, "learning_rate": 4.226e-06, "loss": 2.2533, "step": 8452 }, { "epoch": 2.6498432601880877, "grad_norm": 13.789568901062012, "learning_rate": 4.226500000000001e-06, "loss": 2.0451, "step": 8453 }, { "epoch": 2.650156739811912, "grad_norm": 11.725865364074707, "learning_rate": 4.227e-06, "loss": 2.2033, "step": 8454 }, { "epoch": 2.6504702194357366, "grad_norm": 12.232000350952148, "learning_rate": 4.227500000000001e-06, "loss": 2.1767, "step": 8455 }, { "epoch": 2.650783699059561, "grad_norm": 16.56768035888672, "learning_rate": 4.228000000000001e-06, "loss": 2.5445, "step": 8456 }, { "epoch": 2.6510971786833855, "grad_norm": 11.237393379211426, "learning_rate": 4.2285e-06, "loss": 2.175, "step": 8457 }, { "epoch": 2.65141065830721, "grad_norm": 8.288626670837402, "learning_rate": 4.2290000000000005e-06, "loss": 2.1363, "step": 8458 }, { "epoch": 2.6517241379310343, "grad_norm": 17.3511962890625, "learning_rate": 4.2295e-06, "loss": 2.4532, "step": 8459 }, { "epoch": 2.652037617554859, "grad_norm": 9.099507331848145, "learning_rate": 4.23e-06, "loss": 1.9713, "step": 8460 }, { "epoch": 2.6523510971786832, "grad_norm": 15.459739685058594, "learning_rate": 4.2305e-06, "loss": 3.1978, "step": 8461 }, { "epoch": 2.6526645768025077, "grad_norm": 9.58663558959961, "learning_rate": 4.231e-06, "loss": 2.2674, "step": 8462 }, { "epoch": 2.652978056426332, "grad_norm": 11.586244583129883, "learning_rate": 4.231500000000001e-06, "loss": 2.3106, "step": 8463 }, { "epoch": 2.6532915360501566, "grad_norm": 9.487924575805664, "learning_rate": 4.232e-06, "loss": 2.2845, "step": 8464 }, { "epoch": 2.653605015673981, "grad_norm": 21.017959594726562, "learning_rate": 4.2325000000000006e-06, "loss": 2.4504, "step": 8465 }, { "epoch": 2.6539184952978054, "grad_norm": 16.91088104248047, "learning_rate": 4.233000000000001e-06, "loss": 2.1835, "step": 8466 }, { "epoch": 2.65423197492163, "grad_norm": 9.140762329101562, "learning_rate": 4.2335e-06, "loss": 2.1734, "step": 8467 }, { "epoch": 2.6545454545454543, "grad_norm": 12.38077163696289, "learning_rate": 4.2340000000000005e-06, "loss": 1.9597, "step": 8468 }, { "epoch": 2.6548589341692788, "grad_norm": 11.388206481933594, "learning_rate": 4.2345e-06, "loss": 2.1022, "step": 8469 }, { "epoch": 2.655172413793103, "grad_norm": 9.735517501831055, "learning_rate": 4.235e-06, "loss": 2.044, "step": 8470 }, { "epoch": 2.6554858934169276, "grad_norm": 44.2464485168457, "learning_rate": 4.2355e-06, "loss": 2.4353, "step": 8471 }, { "epoch": 2.655799373040752, "grad_norm": 12.253683090209961, "learning_rate": 4.236e-06, "loss": 1.9684, "step": 8472 }, { "epoch": 2.6561128526645765, "grad_norm": 11.716306686401367, "learning_rate": 4.236500000000001e-06, "loss": 1.9503, "step": 8473 }, { "epoch": 2.6564263322884014, "grad_norm": 37.37388610839844, "learning_rate": 4.237e-06, "loss": 2.1369, "step": 8474 }, { "epoch": 2.656739811912226, "grad_norm": 68.51834106445312, "learning_rate": 4.2375000000000005e-06, "loss": 2.206, "step": 8475 }, { "epoch": 2.6570532915360503, "grad_norm": 70.04552459716797, "learning_rate": 4.238000000000001e-06, "loss": 2.3221, "step": 8476 }, { "epoch": 2.6573667711598747, "grad_norm": 10.652451515197754, "learning_rate": 4.2385e-06, "loss": 1.9998, "step": 8477 }, { "epoch": 2.657680250783699, "grad_norm": 7.90840482711792, "learning_rate": 4.239e-06, "loss": 2.1822, "step": 8478 }, { "epoch": 2.6579937304075236, "grad_norm": 20.211496353149414, "learning_rate": 4.239500000000001e-06, "loss": 2.6216, "step": 8479 }, { "epoch": 2.658307210031348, "grad_norm": 14.542661666870117, "learning_rate": 4.24e-06, "loss": 2.0958, "step": 8480 }, { "epoch": 2.6586206896551725, "grad_norm": 8.018304824829102, "learning_rate": 4.2405e-06, "loss": 2.1679, "step": 8481 }, { "epoch": 2.658934169278997, "grad_norm": 18.81037139892578, "learning_rate": 4.241e-06, "loss": 2.327, "step": 8482 }, { "epoch": 2.6592476489028214, "grad_norm": 8.859085083007812, "learning_rate": 4.241500000000001e-06, "loss": 2.1656, "step": 8483 }, { "epoch": 2.659561128526646, "grad_norm": 10.291755676269531, "learning_rate": 4.242e-06, "loss": 1.9884, "step": 8484 }, { "epoch": 2.6598746081504703, "grad_norm": 7.871489524841309, "learning_rate": 4.2425000000000005e-06, "loss": 1.9957, "step": 8485 }, { "epoch": 2.6601880877742947, "grad_norm": 13.625410079956055, "learning_rate": 4.243000000000001e-06, "loss": 2.0703, "step": 8486 }, { "epoch": 2.660501567398119, "grad_norm": 9.744209289550781, "learning_rate": 4.2435e-06, "loss": 2.1506, "step": 8487 }, { "epoch": 2.6608150470219436, "grad_norm": 7.022274494171143, "learning_rate": 4.244e-06, "loss": 2.1919, "step": 8488 }, { "epoch": 2.661128526645768, "grad_norm": 35.43350601196289, "learning_rate": 4.244500000000001e-06, "loss": 3.2454, "step": 8489 }, { "epoch": 2.6614420062695925, "grad_norm": 9.877697944641113, "learning_rate": 4.245e-06, "loss": 1.8437, "step": 8490 }, { "epoch": 2.661755485893417, "grad_norm": 8.767868041992188, "learning_rate": 4.2455e-06, "loss": 2.2134, "step": 8491 }, { "epoch": 2.6620689655172414, "grad_norm": 13.66367244720459, "learning_rate": 4.2460000000000005e-06, "loss": 1.9112, "step": 8492 }, { "epoch": 2.662382445141066, "grad_norm": 7.585222244262695, "learning_rate": 4.246500000000001e-06, "loss": 2.054, "step": 8493 }, { "epoch": 2.6626959247648903, "grad_norm": 13.316164016723633, "learning_rate": 4.247e-06, "loss": 2.2477, "step": 8494 }, { "epoch": 2.6630094043887147, "grad_norm": 18.74990463256836, "learning_rate": 4.2475e-06, "loss": 2.06, "step": 8495 }, { "epoch": 2.663322884012539, "grad_norm": 7.873201370239258, "learning_rate": 4.248000000000001e-06, "loss": 2.1076, "step": 8496 }, { "epoch": 2.6636363636363636, "grad_norm": 54.38701629638672, "learning_rate": 4.2485e-06, "loss": 2.0524, "step": 8497 }, { "epoch": 2.663949843260188, "grad_norm": 11.607866287231445, "learning_rate": 4.249e-06, "loss": 2.2318, "step": 8498 }, { "epoch": 2.6642633228840125, "grad_norm": 44.299312591552734, "learning_rate": 4.2495000000000006e-06, "loss": 2.1318, "step": 8499 }, { "epoch": 2.664576802507837, "grad_norm": 67.33507537841797, "learning_rate": 4.25e-06, "loss": 2.2319, "step": 8500 }, { "epoch": 2.6648902821316613, "grad_norm": 9.74003791809082, "learning_rate": 4.2505e-06, "loss": 1.8654, "step": 8501 }, { "epoch": 2.665203761755486, "grad_norm": 10.623579978942871, "learning_rate": 4.2510000000000005e-06, "loss": 2.2366, "step": 8502 }, { "epoch": 2.6655172413793102, "grad_norm": 7.370429039001465, "learning_rate": 4.2515e-06, "loss": 1.9137, "step": 8503 }, { "epoch": 2.6658307210031347, "grad_norm": 30.498140335083008, "learning_rate": 4.252000000000001e-06, "loss": 2.1474, "step": 8504 }, { "epoch": 2.666144200626959, "grad_norm": 14.312164306640625, "learning_rate": 4.2525e-06, "loss": 2.0816, "step": 8505 }, { "epoch": 2.6664576802507836, "grad_norm": 10.828268051147461, "learning_rate": 4.253000000000001e-06, "loss": 2.1952, "step": 8506 }, { "epoch": 2.666771159874608, "grad_norm": 11.983382225036621, "learning_rate": 4.2535e-06, "loss": 2.2261, "step": 8507 }, { "epoch": 2.6670846394984324, "grad_norm": 8.043965339660645, "learning_rate": 4.254e-06, "loss": 2.1705, "step": 8508 }, { "epoch": 2.6673981191222573, "grad_norm": 9.69068431854248, "learning_rate": 4.2545000000000005e-06, "loss": 2.1912, "step": 8509 }, { "epoch": 2.6677115987460818, "grad_norm": 11.125262260437012, "learning_rate": 4.255e-06, "loss": 2.1757, "step": 8510 }, { "epoch": 2.668025078369906, "grad_norm": 11.922386169433594, "learning_rate": 4.2555e-06, "loss": 2.2662, "step": 8511 }, { "epoch": 2.6683385579937307, "grad_norm": 18.725645065307617, "learning_rate": 4.256e-06, "loss": 2.3176, "step": 8512 }, { "epoch": 2.668652037617555, "grad_norm": 15.327168464660645, "learning_rate": 4.2565e-06, "loss": 2.2601, "step": 8513 }, { "epoch": 2.6689655172413795, "grad_norm": 30.386491775512695, "learning_rate": 4.257000000000001e-06, "loss": 2.0679, "step": 8514 }, { "epoch": 2.669278996865204, "grad_norm": 8.546671867370605, "learning_rate": 4.2575e-06, "loss": 2.2153, "step": 8515 }, { "epoch": 2.6695924764890284, "grad_norm": 12.696763038635254, "learning_rate": 4.2580000000000006e-06, "loss": 2.7258, "step": 8516 }, { "epoch": 2.669905956112853, "grad_norm": 14.519058227539062, "learning_rate": 4.258500000000001e-06, "loss": 2.2206, "step": 8517 }, { "epoch": 2.6702194357366773, "grad_norm": 7.0597243309021, "learning_rate": 4.259e-06, "loss": 2.216, "step": 8518 }, { "epoch": 2.6705329153605017, "grad_norm": 7.697901248931885, "learning_rate": 4.2595000000000005e-06, "loss": 2.1547, "step": 8519 }, { "epoch": 2.670846394984326, "grad_norm": 10.198399543762207, "learning_rate": 4.26e-06, "loss": 2.2459, "step": 8520 }, { "epoch": 2.6711598746081506, "grad_norm": 11.078211784362793, "learning_rate": 4.2605e-06, "loss": 2.0599, "step": 8521 }, { "epoch": 2.671473354231975, "grad_norm": 11.405905723571777, "learning_rate": 4.261e-06, "loss": 2.6211, "step": 8522 }, { "epoch": 2.6717868338557995, "grad_norm": 12.333938598632812, "learning_rate": 4.2615e-06, "loss": 2.062, "step": 8523 }, { "epoch": 2.672100313479624, "grad_norm": 6.181169509887695, "learning_rate": 4.262000000000001e-06, "loss": 2.212, "step": 8524 }, { "epoch": 2.6724137931034484, "grad_norm": 6.129042625427246, "learning_rate": 4.2625e-06, "loss": 2.1543, "step": 8525 }, { "epoch": 2.672727272727273, "grad_norm": 7.347682476043701, "learning_rate": 4.2630000000000005e-06, "loss": 1.9271, "step": 8526 }, { "epoch": 2.6730407523510973, "grad_norm": 18.46999740600586, "learning_rate": 4.263500000000001e-06, "loss": 2.1368, "step": 8527 }, { "epoch": 2.6733542319749217, "grad_norm": 15.65620231628418, "learning_rate": 4.264e-06, "loss": 2.3864, "step": 8528 }, { "epoch": 2.673667711598746, "grad_norm": 26.222217559814453, "learning_rate": 4.2645e-06, "loss": 2.2943, "step": 8529 }, { "epoch": 2.6739811912225706, "grad_norm": 10.884167671203613, "learning_rate": 4.265000000000001e-06, "loss": 2.0762, "step": 8530 }, { "epoch": 2.674294670846395, "grad_norm": 14.360615730285645, "learning_rate": 4.2655e-06, "loss": 2.1774, "step": 8531 }, { "epoch": 2.6746081504702195, "grad_norm": 12.414725303649902, "learning_rate": 4.266e-06, "loss": 2.2415, "step": 8532 }, { "epoch": 2.674921630094044, "grad_norm": 9.171478271484375, "learning_rate": 4.2665e-06, "loss": 2.0019, "step": 8533 }, { "epoch": 2.6752351097178684, "grad_norm": 9.179827690124512, "learning_rate": 4.267000000000001e-06, "loss": 1.992, "step": 8534 }, { "epoch": 2.675548589341693, "grad_norm": 50.13624954223633, "learning_rate": 4.2675e-06, "loss": 2.3966, "step": 8535 }, { "epoch": 2.6758620689655173, "grad_norm": 10.653006553649902, "learning_rate": 4.2680000000000005e-06, "loss": 2.3781, "step": 8536 }, { "epoch": 2.6761755485893417, "grad_norm": 6.012962818145752, "learning_rate": 4.268500000000001e-06, "loss": 2.2595, "step": 8537 }, { "epoch": 2.676489028213166, "grad_norm": 13.12833309173584, "learning_rate": 4.269e-06, "loss": 2.0102, "step": 8538 }, { "epoch": 2.6768025078369906, "grad_norm": 27.81395149230957, "learning_rate": 4.2695e-06, "loss": 2.3541, "step": 8539 }, { "epoch": 2.677115987460815, "grad_norm": 9.969282150268555, "learning_rate": 4.270000000000001e-06, "loss": 1.8566, "step": 8540 }, { "epoch": 2.6774294670846395, "grad_norm": 95.49557495117188, "learning_rate": 4.2705e-06, "loss": 2.5611, "step": 8541 }, { "epoch": 2.677742946708464, "grad_norm": 25.597890853881836, "learning_rate": 4.271e-06, "loss": 2.4399, "step": 8542 }, { "epoch": 2.6780564263322884, "grad_norm": 93.12142944335938, "learning_rate": 4.2715e-06, "loss": 2.7557, "step": 8543 }, { "epoch": 2.678369905956113, "grad_norm": 13.254643440246582, "learning_rate": 4.272000000000001e-06, "loss": 2.1742, "step": 8544 }, { "epoch": 2.6786833855799372, "grad_norm": 18.312170028686523, "learning_rate": 4.2725e-06, "loss": 2.0138, "step": 8545 }, { "epoch": 2.6789968652037617, "grad_norm": 22.449153900146484, "learning_rate": 4.2730000000000004e-06, "loss": 2.6176, "step": 8546 }, { "epoch": 2.679310344827586, "grad_norm": 11.108996391296387, "learning_rate": 4.273500000000001e-06, "loss": 2.0596, "step": 8547 }, { "epoch": 2.6796238244514106, "grad_norm": 116.44094848632812, "learning_rate": 4.274e-06, "loss": 2.3821, "step": 8548 }, { "epoch": 2.679937304075235, "grad_norm": 6.7298407554626465, "learning_rate": 4.2745e-06, "loss": 1.8883, "step": 8549 }, { "epoch": 2.6802507836990594, "grad_norm": 154.56558227539062, "learning_rate": 4.2750000000000006e-06, "loss": 2.1256, "step": 8550 }, { "epoch": 2.680564263322884, "grad_norm": 12.542025566101074, "learning_rate": 4.2755e-06, "loss": 2.4984, "step": 8551 }, { "epoch": 2.6808777429467083, "grad_norm": 8.535301208496094, "learning_rate": 4.276e-06, "loss": 2.0068, "step": 8552 }, { "epoch": 2.6811912225705328, "grad_norm": 9.630982398986816, "learning_rate": 4.2765000000000005e-06, "loss": 2.1175, "step": 8553 }, { "epoch": 2.681504702194357, "grad_norm": 25.329782485961914, "learning_rate": 4.277000000000001e-06, "loss": 2.2309, "step": 8554 }, { "epoch": 2.6818181818181817, "grad_norm": 13.15017318725586, "learning_rate": 4.2775e-06, "loss": 2.346, "step": 8555 }, { "epoch": 2.682131661442006, "grad_norm": 16.22644805908203, "learning_rate": 4.278e-06, "loss": 2.0663, "step": 8556 }, { "epoch": 2.6824451410658305, "grad_norm": 11.391193389892578, "learning_rate": 4.278500000000001e-06, "loss": 2.0258, "step": 8557 }, { "epoch": 2.682758620689655, "grad_norm": 8.333171844482422, "learning_rate": 4.279e-06, "loss": 2.0668, "step": 8558 }, { "epoch": 2.6830721003134794, "grad_norm": 59.27569580078125, "learning_rate": 4.2795e-06, "loss": 2.024, "step": 8559 }, { "epoch": 2.683385579937304, "grad_norm": 12.23531436920166, "learning_rate": 4.2800000000000005e-06, "loss": 2.2791, "step": 8560 }, { "epoch": 2.6836990595611283, "grad_norm": 13.814367294311523, "learning_rate": 4.2805e-06, "loss": 2.4726, "step": 8561 }, { "epoch": 2.6840125391849528, "grad_norm": 9.134698867797852, "learning_rate": 4.281e-06, "loss": 2.2874, "step": 8562 }, { "epoch": 2.684326018808777, "grad_norm": 17.347841262817383, "learning_rate": 4.2815000000000004e-06, "loss": 1.9884, "step": 8563 }, { "epoch": 2.6846394984326016, "grad_norm": 10.795247077941895, "learning_rate": 4.282000000000001e-06, "loss": 1.9836, "step": 8564 }, { "epoch": 2.684952978056426, "grad_norm": 8.303966522216797, "learning_rate": 4.282500000000001e-06, "loss": 1.9429, "step": 8565 }, { "epoch": 2.6852664576802505, "grad_norm": 5.694846153259277, "learning_rate": 4.283e-06, "loss": 2.0371, "step": 8566 }, { "epoch": 2.685579937304075, "grad_norm": 9.72620677947998, "learning_rate": 4.283500000000001e-06, "loss": 2.0962, "step": 8567 }, { "epoch": 2.6858934169279, "grad_norm": 6.453702449798584, "learning_rate": 4.284e-06, "loss": 2.0239, "step": 8568 }, { "epoch": 2.6862068965517243, "grad_norm": 38.59422302246094, "learning_rate": 4.2845e-06, "loss": 2.2824, "step": 8569 }, { "epoch": 2.6865203761755487, "grad_norm": 15.985084533691406, "learning_rate": 4.2850000000000005e-06, "loss": 2.0061, "step": 8570 }, { "epoch": 2.686833855799373, "grad_norm": 45.90361785888672, "learning_rate": 4.2855e-06, "loss": 2.2337, "step": 8571 }, { "epoch": 2.6871473354231976, "grad_norm": 7.742265224456787, "learning_rate": 4.286e-06, "loss": 2.1058, "step": 8572 }, { "epoch": 2.687460815047022, "grad_norm": 10.369441986083984, "learning_rate": 4.2865e-06, "loss": 1.9956, "step": 8573 }, { "epoch": 2.6877742946708465, "grad_norm": 13.09179401397705, "learning_rate": 4.287000000000001e-06, "loss": 2.1497, "step": 8574 }, { "epoch": 2.688087774294671, "grad_norm": 121.22309875488281, "learning_rate": 4.287500000000001e-06, "loss": 2.3517, "step": 8575 }, { "epoch": 2.6884012539184954, "grad_norm": 26.52750587463379, "learning_rate": 4.288e-06, "loss": 2.0848, "step": 8576 }, { "epoch": 2.68871473354232, "grad_norm": 9.859085083007812, "learning_rate": 4.2885000000000005e-06, "loss": 2.0346, "step": 8577 }, { "epoch": 2.6890282131661443, "grad_norm": 9.688697814941406, "learning_rate": 4.289000000000001e-06, "loss": 2.1091, "step": 8578 }, { "epoch": 2.6893416927899687, "grad_norm": 26.372373580932617, "learning_rate": 4.2895e-06, "loss": 2.2485, "step": 8579 }, { "epoch": 2.689655172413793, "grad_norm": 8.525503158569336, "learning_rate": 4.2900000000000004e-06, "loss": 2.2176, "step": 8580 }, { "epoch": 2.6899686520376176, "grad_norm": 7.837364196777344, "learning_rate": 4.2905e-06, "loss": 1.9473, "step": 8581 }, { "epoch": 2.690282131661442, "grad_norm": 11.756917953491211, "learning_rate": 4.291e-06, "loss": 2.4857, "step": 8582 }, { "epoch": 2.6905956112852665, "grad_norm": 5.485519886016846, "learning_rate": 4.2915e-06, "loss": 2.0958, "step": 8583 }, { "epoch": 2.690909090909091, "grad_norm": 9.29818344116211, "learning_rate": 4.292000000000001e-06, "loss": 3.1814, "step": 8584 }, { "epoch": 2.6912225705329154, "grad_norm": 31.793415069580078, "learning_rate": 4.292500000000001e-06, "loss": 2.3712, "step": 8585 }, { "epoch": 2.69153605015674, "grad_norm": 8.784714698791504, "learning_rate": 4.293e-06, "loss": 2.3084, "step": 8586 }, { "epoch": 2.6918495297805642, "grad_norm": 7.539299964904785, "learning_rate": 4.2935000000000005e-06, "loss": 2.0587, "step": 8587 }, { "epoch": 2.6921630094043887, "grad_norm": 61.362159729003906, "learning_rate": 4.294000000000001e-06, "loss": 1.9447, "step": 8588 }, { "epoch": 2.692476489028213, "grad_norm": 24.04331398010254, "learning_rate": 4.2945e-06, "loss": 2.1055, "step": 8589 }, { "epoch": 2.6927899686520376, "grad_norm": 17.749061584472656, "learning_rate": 4.295e-06, "loss": 2.1728, "step": 8590 }, { "epoch": 2.693103448275862, "grad_norm": 40.12238311767578, "learning_rate": 4.295500000000001e-06, "loss": 1.9741, "step": 8591 }, { "epoch": 2.6934169278996865, "grad_norm": 41.994327545166016, "learning_rate": 4.296e-06, "loss": 2.3411, "step": 8592 }, { "epoch": 2.693730407523511, "grad_norm": 13.70588207244873, "learning_rate": 4.2965e-06, "loss": 1.718, "step": 8593 }, { "epoch": 2.6940438871473353, "grad_norm": 40.76674270629883, "learning_rate": 4.2970000000000005e-06, "loss": 2.2609, "step": 8594 }, { "epoch": 2.69435736677116, "grad_norm": 18.902467727661133, "learning_rate": 4.297500000000001e-06, "loss": 2.0859, "step": 8595 }, { "epoch": 2.694670846394984, "grad_norm": 7.330137252807617, "learning_rate": 4.298e-06, "loss": 1.9776, "step": 8596 }, { "epoch": 2.6949843260188087, "grad_norm": 73.83060455322266, "learning_rate": 4.2985000000000004e-06, "loss": 2.2463, "step": 8597 }, { "epoch": 2.695297805642633, "grad_norm": 49.29768371582031, "learning_rate": 4.299000000000001e-06, "loss": 2.0542, "step": 8598 }, { "epoch": 2.6956112852664575, "grad_norm": 9.857036590576172, "learning_rate": 4.2995e-06, "loss": 2.0516, "step": 8599 }, { "epoch": 2.695924764890282, "grad_norm": 9.115958213806152, "learning_rate": 4.3e-06, "loss": 1.9455, "step": 8600 }, { "epoch": 2.6962382445141064, "grad_norm": 19.109947204589844, "learning_rate": 4.300500000000001e-06, "loss": 1.9323, "step": 8601 }, { "epoch": 2.696551724137931, "grad_norm": 12.084345817565918, "learning_rate": 4.301e-06, "loss": 2.5554, "step": 8602 }, { "epoch": 2.6968652037617558, "grad_norm": 8.36884593963623, "learning_rate": 4.3015e-06, "loss": 2.2779, "step": 8603 }, { "epoch": 2.69717868338558, "grad_norm": 8.88129711151123, "learning_rate": 4.3020000000000005e-06, "loss": 2.3545, "step": 8604 }, { "epoch": 2.6974921630094046, "grad_norm": 8.011765480041504, "learning_rate": 4.302500000000001e-06, "loss": 2.0524, "step": 8605 }, { "epoch": 2.697805642633229, "grad_norm": 7.164164066314697, "learning_rate": 4.303e-06, "loss": 2.0373, "step": 8606 }, { "epoch": 2.6981191222570535, "grad_norm": 14.840588569641113, "learning_rate": 4.3035e-06, "loss": 2.1852, "step": 8607 }, { "epoch": 2.698432601880878, "grad_norm": 7.729261875152588, "learning_rate": 4.304000000000001e-06, "loss": 2.2271, "step": 8608 }, { "epoch": 2.6987460815047024, "grad_norm": 25.98223876953125, "learning_rate": 4.3045e-06, "loss": 2.0386, "step": 8609 }, { "epoch": 2.699059561128527, "grad_norm": 13.872118949890137, "learning_rate": 4.305e-06, "loss": 2.2624, "step": 8610 }, { "epoch": 2.6993730407523513, "grad_norm": 10.246224403381348, "learning_rate": 4.3055000000000005e-06, "loss": 2.3997, "step": 8611 }, { "epoch": 2.6996865203761757, "grad_norm": 8.267459869384766, "learning_rate": 4.306e-06, "loss": 2.2797, "step": 8612 }, { "epoch": 2.7, "grad_norm": 9.931675910949707, "learning_rate": 4.3065e-06, "loss": 2.0446, "step": 8613 }, { "epoch": 2.7003134796238246, "grad_norm": 60.30348205566406, "learning_rate": 4.3070000000000004e-06, "loss": 2.3267, "step": 8614 }, { "epoch": 2.700626959247649, "grad_norm": 7.0497307777404785, "learning_rate": 4.307500000000001e-06, "loss": 2.146, "step": 8615 }, { "epoch": 2.7009404388714735, "grad_norm": 11.400117874145508, "learning_rate": 4.308000000000001e-06, "loss": 2.8523, "step": 8616 }, { "epoch": 2.701253918495298, "grad_norm": 14.064170837402344, "learning_rate": 4.3085e-06, "loss": 2.0216, "step": 8617 }, { "epoch": 2.7015673981191224, "grad_norm": 15.306402206420898, "learning_rate": 4.309000000000001e-06, "loss": 2.1553, "step": 8618 }, { "epoch": 2.701880877742947, "grad_norm": 11.356364250183105, "learning_rate": 4.3095e-06, "loss": 2.2419, "step": 8619 }, { "epoch": 2.7021943573667713, "grad_norm": 7.852354526519775, "learning_rate": 4.31e-06, "loss": 2.1157, "step": 8620 }, { "epoch": 2.7025078369905957, "grad_norm": 12.175455093383789, "learning_rate": 4.3105000000000005e-06, "loss": 2.0605, "step": 8621 }, { "epoch": 2.70282131661442, "grad_norm": 10.941705703735352, "learning_rate": 4.311e-06, "loss": 2.5041, "step": 8622 }, { "epoch": 2.7031347962382446, "grad_norm": 45.262271881103516, "learning_rate": 4.3115e-06, "loss": 2.2373, "step": 8623 }, { "epoch": 2.703448275862069, "grad_norm": 12.629764556884766, "learning_rate": 4.312e-06, "loss": 2.2937, "step": 8624 }, { "epoch": 2.7037617554858935, "grad_norm": 14.933594703674316, "learning_rate": 4.312500000000001e-06, "loss": 2.2205, "step": 8625 }, { "epoch": 2.704075235109718, "grad_norm": 6.434473037719727, "learning_rate": 4.313000000000001e-06, "loss": 2.0958, "step": 8626 }, { "epoch": 2.7043887147335424, "grad_norm": 17.502376556396484, "learning_rate": 4.3135e-06, "loss": 2.0704, "step": 8627 }, { "epoch": 2.704702194357367, "grad_norm": 19.888286590576172, "learning_rate": 4.3140000000000005e-06, "loss": 2.3895, "step": 8628 }, { "epoch": 2.7050156739811912, "grad_norm": 7.369458198547363, "learning_rate": 4.3145e-06, "loss": 2.0044, "step": 8629 }, { "epoch": 2.7053291536050157, "grad_norm": 9.804317474365234, "learning_rate": 4.315e-06, "loss": 2.2172, "step": 8630 }, { "epoch": 2.70564263322884, "grad_norm": 8.422747611999512, "learning_rate": 4.3155000000000004e-06, "loss": 2.079, "step": 8631 }, { "epoch": 2.7059561128526646, "grad_norm": 9.839719772338867, "learning_rate": 4.316e-06, "loss": 2.2312, "step": 8632 }, { "epoch": 2.706269592476489, "grad_norm": 5.688465595245361, "learning_rate": 4.3165e-06, "loss": 2.0721, "step": 8633 }, { "epoch": 2.7065830721003135, "grad_norm": 13.522320747375488, "learning_rate": 4.317e-06, "loss": 2.1458, "step": 8634 }, { "epoch": 2.706896551724138, "grad_norm": 16.43364715576172, "learning_rate": 4.317500000000001e-06, "loss": 2.2155, "step": 8635 }, { "epoch": 2.7072100313479623, "grad_norm": 11.751569747924805, "learning_rate": 4.318000000000001e-06, "loss": 2.7277, "step": 8636 }, { "epoch": 2.707523510971787, "grad_norm": 18.3969783782959, "learning_rate": 4.3185e-06, "loss": 1.9763, "step": 8637 }, { "epoch": 2.7078369905956112, "grad_norm": 9.816944122314453, "learning_rate": 4.3190000000000005e-06, "loss": 2.0095, "step": 8638 }, { "epoch": 2.7081504702194357, "grad_norm": 8.698689460754395, "learning_rate": 4.319500000000001e-06, "loss": 2.0658, "step": 8639 }, { "epoch": 2.70846394984326, "grad_norm": 11.822721481323242, "learning_rate": 4.32e-06, "loss": 2.0797, "step": 8640 }, { "epoch": 2.7087774294670846, "grad_norm": 8.341374397277832, "learning_rate": 4.3205e-06, "loss": 2.1799, "step": 8641 }, { "epoch": 2.709090909090909, "grad_norm": 12.805225372314453, "learning_rate": 4.321e-06, "loss": 2.0983, "step": 8642 }, { "epoch": 2.7094043887147334, "grad_norm": 11.128278732299805, "learning_rate": 4.3215e-06, "loss": 2.2715, "step": 8643 }, { "epoch": 2.709717868338558, "grad_norm": 7.545168876647949, "learning_rate": 4.322e-06, "loss": 2.2586, "step": 8644 }, { "epoch": 2.7100313479623823, "grad_norm": 8.739923477172852, "learning_rate": 4.3225000000000005e-06, "loss": 2.1059, "step": 8645 }, { "epoch": 2.7103448275862068, "grad_norm": 9.637774467468262, "learning_rate": 4.323000000000001e-06, "loss": 2.4953, "step": 8646 }, { "epoch": 2.710658307210031, "grad_norm": 7.650784969329834, "learning_rate": 4.3235e-06, "loss": 2.2114, "step": 8647 }, { "epoch": 2.7109717868338556, "grad_norm": 41.22919464111328, "learning_rate": 4.3240000000000004e-06, "loss": 2.0783, "step": 8648 }, { "epoch": 2.71128526645768, "grad_norm": 6.860087871551514, "learning_rate": 4.324500000000001e-06, "loss": 2.1928, "step": 8649 }, { "epoch": 2.7115987460815045, "grad_norm": 18.400096893310547, "learning_rate": 4.325e-06, "loss": 2.2576, "step": 8650 }, { "epoch": 2.711912225705329, "grad_norm": 36.66834259033203, "learning_rate": 4.3255e-06, "loss": 1.7385, "step": 8651 }, { "epoch": 2.7122257053291534, "grad_norm": 7.550413131713867, "learning_rate": 4.326000000000001e-06, "loss": 2.1936, "step": 8652 }, { "epoch": 2.712539184952978, "grad_norm": 9.701518058776855, "learning_rate": 4.3265e-06, "loss": 2.2775, "step": 8653 }, { "epoch": 2.7128526645768023, "grad_norm": 64.71399688720703, "learning_rate": 4.327e-06, "loss": 1.8781, "step": 8654 }, { "epoch": 2.7131661442006267, "grad_norm": 30.144121170043945, "learning_rate": 4.3275000000000005e-06, "loss": 2.0702, "step": 8655 }, { "epoch": 2.713479623824451, "grad_norm": 16.76169776916504, "learning_rate": 4.328000000000001e-06, "loss": 2.2867, "step": 8656 }, { "epoch": 2.7137931034482756, "grad_norm": 13.014815330505371, "learning_rate": 4.3285e-06, "loss": 2.1327, "step": 8657 }, { "epoch": 2.7141065830721, "grad_norm": 10.662934303283691, "learning_rate": 4.329e-06, "loss": 2.3145, "step": 8658 }, { "epoch": 2.7144200626959245, "grad_norm": 7.78823184967041, "learning_rate": 4.329500000000001e-06, "loss": 2.1713, "step": 8659 }, { "epoch": 2.714733542319749, "grad_norm": 7.604531764984131, "learning_rate": 4.33e-06, "loss": 2.2995, "step": 8660 }, { "epoch": 2.7150470219435734, "grad_norm": 6.970873832702637, "learning_rate": 4.3305e-06, "loss": 2.2681, "step": 8661 }, { "epoch": 2.7153605015673983, "grad_norm": 11.495867729187012, "learning_rate": 4.3310000000000005e-06, "loss": 2.2671, "step": 8662 }, { "epoch": 2.7156739811912227, "grad_norm": 7.3347978591918945, "learning_rate": 4.3315e-06, "loss": 2.2394, "step": 8663 }, { "epoch": 2.715987460815047, "grad_norm": 79.1141128540039, "learning_rate": 4.332e-06, "loss": 2.4582, "step": 8664 }, { "epoch": 2.7163009404388716, "grad_norm": 6.373804092407227, "learning_rate": 4.3325000000000004e-06, "loss": 2.1792, "step": 8665 }, { "epoch": 2.716614420062696, "grad_norm": 10.75312614440918, "learning_rate": 4.333000000000001e-06, "loss": 2.1263, "step": 8666 }, { "epoch": 2.7169278996865205, "grad_norm": 8.839099884033203, "learning_rate": 4.3335e-06, "loss": 1.978, "step": 8667 }, { "epoch": 2.717241379310345, "grad_norm": 11.39769172668457, "learning_rate": 4.334e-06, "loss": 2.2133, "step": 8668 }, { "epoch": 2.7175548589341694, "grad_norm": 44.96700668334961, "learning_rate": 4.334500000000001e-06, "loss": 2.2572, "step": 8669 }, { "epoch": 2.717868338557994, "grad_norm": 6.971530437469482, "learning_rate": 4.335e-06, "loss": 2.3769, "step": 8670 }, { "epoch": 2.7181818181818183, "grad_norm": 47.08424758911133, "learning_rate": 4.3355e-06, "loss": 2.8763, "step": 8671 }, { "epoch": 2.7184952978056427, "grad_norm": 88.6609115600586, "learning_rate": 4.3360000000000005e-06, "loss": 2.1695, "step": 8672 }, { "epoch": 2.718808777429467, "grad_norm": 22.023578643798828, "learning_rate": 4.3365e-06, "loss": 2.2713, "step": 8673 }, { "epoch": 2.7191222570532916, "grad_norm": 12.42693042755127, "learning_rate": 4.337e-06, "loss": 2.2449, "step": 8674 }, { "epoch": 2.719435736677116, "grad_norm": 10.886938095092773, "learning_rate": 4.3375e-06, "loss": 2.1388, "step": 8675 }, { "epoch": 2.7197492163009405, "grad_norm": 16.41759490966797, "learning_rate": 4.338000000000001e-06, "loss": 2.8295, "step": 8676 }, { "epoch": 2.720062695924765, "grad_norm": 6.647755146026611, "learning_rate": 4.338500000000001e-06, "loss": 1.8883, "step": 8677 }, { "epoch": 2.7203761755485893, "grad_norm": 19.876205444335938, "learning_rate": 4.339e-06, "loss": 2.3329, "step": 8678 }, { "epoch": 2.720689655172414, "grad_norm": 69.12499237060547, "learning_rate": 4.3395000000000005e-06, "loss": 3.1044, "step": 8679 }, { "epoch": 2.7210031347962382, "grad_norm": 10.235759735107422, "learning_rate": 4.34e-06, "loss": 1.9482, "step": 8680 }, { "epoch": 2.7213166144200627, "grad_norm": 6.963871479034424, "learning_rate": 4.3405e-06, "loss": 2.0843, "step": 8681 }, { "epoch": 2.721630094043887, "grad_norm": 9.091208457946777, "learning_rate": 4.3410000000000005e-06, "loss": 2.2127, "step": 8682 }, { "epoch": 2.7219435736677116, "grad_norm": 109.02669525146484, "learning_rate": 4.3415e-06, "loss": 2.4547, "step": 8683 }, { "epoch": 2.722257053291536, "grad_norm": 12.508950233459473, "learning_rate": 4.342e-06, "loss": 2.2022, "step": 8684 }, { "epoch": 2.7225705329153604, "grad_norm": 10.736917495727539, "learning_rate": 4.3425e-06, "loss": 2.455, "step": 8685 }, { "epoch": 2.722884012539185, "grad_norm": 12.652810096740723, "learning_rate": 4.343000000000001e-06, "loss": 2.4223, "step": 8686 }, { "epoch": 2.7231974921630093, "grad_norm": 18.127893447875977, "learning_rate": 4.343500000000001e-06, "loss": 2.4075, "step": 8687 }, { "epoch": 2.7235109717868338, "grad_norm": 7.180667400360107, "learning_rate": 4.344e-06, "loss": 1.9277, "step": 8688 }, { "epoch": 2.723824451410658, "grad_norm": 13.35740852355957, "learning_rate": 4.3445000000000005e-06, "loss": 2.0567, "step": 8689 }, { "epoch": 2.7241379310344827, "grad_norm": 12.085336685180664, "learning_rate": 4.345000000000001e-06, "loss": 1.9378, "step": 8690 }, { "epoch": 2.724451410658307, "grad_norm": 6.162071228027344, "learning_rate": 4.3455e-06, "loss": 2.2512, "step": 8691 }, { "epoch": 2.7247648902821315, "grad_norm": 25.19955062866211, "learning_rate": 4.346e-06, "loss": 1.9054, "step": 8692 }, { "epoch": 2.725078369905956, "grad_norm": 8.612536430358887, "learning_rate": 4.3465e-06, "loss": 2.1829, "step": 8693 }, { "epoch": 2.7253918495297804, "grad_norm": 19.819421768188477, "learning_rate": 4.347e-06, "loss": 2.3507, "step": 8694 }, { "epoch": 2.725705329153605, "grad_norm": 118.35186004638672, "learning_rate": 4.3475e-06, "loss": 2.324, "step": 8695 }, { "epoch": 2.7260188087774293, "grad_norm": 11.278396606445312, "learning_rate": 4.3480000000000006e-06, "loss": 2.0185, "step": 8696 }, { "epoch": 2.726332288401254, "grad_norm": 10.387089729309082, "learning_rate": 4.348500000000001e-06, "loss": 1.9099, "step": 8697 }, { "epoch": 2.7266457680250786, "grad_norm": 16.367502212524414, "learning_rate": 4.349e-06, "loss": 2.0369, "step": 8698 }, { "epoch": 2.726959247648903, "grad_norm": 16.660776138305664, "learning_rate": 4.3495000000000005e-06, "loss": 2.2871, "step": 8699 }, { "epoch": 2.7272727272727275, "grad_norm": 15.27818775177002, "learning_rate": 4.350000000000001e-06, "loss": 2.1391, "step": 8700 }, { "epoch": 2.727586206896552, "grad_norm": 8.89338493347168, "learning_rate": 4.3505e-06, "loss": 2.1689, "step": 8701 }, { "epoch": 2.7278996865203764, "grad_norm": 17.436250686645508, "learning_rate": 4.351e-06, "loss": 2.9337, "step": 8702 }, { "epoch": 2.728213166144201, "grad_norm": 35.67147445678711, "learning_rate": 4.3515e-06, "loss": 2.0088, "step": 8703 }, { "epoch": 2.7285266457680253, "grad_norm": 36.881534576416016, "learning_rate": 4.352e-06, "loss": 1.9105, "step": 8704 }, { "epoch": 2.7288401253918497, "grad_norm": 9.210811614990234, "learning_rate": 4.3525e-06, "loss": 2.16, "step": 8705 }, { "epoch": 2.729153605015674, "grad_norm": 9.225309371948242, "learning_rate": 4.3530000000000005e-06, "loss": 2.1316, "step": 8706 }, { "epoch": 2.7294670846394986, "grad_norm": 14.6263427734375, "learning_rate": 4.353500000000001e-06, "loss": 2.4285, "step": 8707 }, { "epoch": 2.729780564263323, "grad_norm": 6.241610527038574, "learning_rate": 4.354e-06, "loss": 2.1806, "step": 8708 }, { "epoch": 2.7300940438871475, "grad_norm": 10.107898712158203, "learning_rate": 4.3545e-06, "loss": 2.6691, "step": 8709 }, { "epoch": 2.730407523510972, "grad_norm": 32.53678512573242, "learning_rate": 4.355000000000001e-06, "loss": 2.5688, "step": 8710 }, { "epoch": 2.7307210031347964, "grad_norm": 6.81553840637207, "learning_rate": 4.3555e-06, "loss": 2.2365, "step": 8711 }, { "epoch": 2.731034482758621, "grad_norm": 9.018298149108887, "learning_rate": 4.356e-06, "loss": 2.1913, "step": 8712 }, { "epoch": 2.7313479623824453, "grad_norm": 11.691299438476562, "learning_rate": 4.3565000000000006e-06, "loss": 2.3005, "step": 8713 }, { "epoch": 2.7316614420062697, "grad_norm": 11.494564056396484, "learning_rate": 4.357e-06, "loss": 2.2097, "step": 8714 }, { "epoch": 2.731974921630094, "grad_norm": 8.151761054992676, "learning_rate": 4.3575e-06, "loss": 2.0032, "step": 8715 }, { "epoch": 2.7322884012539186, "grad_norm": 5.431791305541992, "learning_rate": 4.3580000000000005e-06, "loss": 2.1136, "step": 8716 }, { "epoch": 2.732601880877743, "grad_norm": 5.966331958770752, "learning_rate": 4.358500000000001e-06, "loss": 2.1214, "step": 8717 }, { "epoch": 2.7329153605015675, "grad_norm": 9.783186912536621, "learning_rate": 4.359e-06, "loss": 1.8205, "step": 8718 }, { "epoch": 2.733228840125392, "grad_norm": 10.54366397857666, "learning_rate": 4.3595e-06, "loss": 2.4545, "step": 8719 }, { "epoch": 2.7335423197492164, "grad_norm": 8.567094802856445, "learning_rate": 4.360000000000001e-06, "loss": 2.1128, "step": 8720 }, { "epoch": 2.733855799373041, "grad_norm": 6.157483100891113, "learning_rate": 4.3605e-06, "loss": 2.3013, "step": 8721 }, { "epoch": 2.7341692789968652, "grad_norm": 10.722454071044922, "learning_rate": 4.361e-06, "loss": 2.2814, "step": 8722 }, { "epoch": 2.7344827586206897, "grad_norm": 67.67513275146484, "learning_rate": 4.3615000000000005e-06, "loss": 2.4819, "step": 8723 }, { "epoch": 2.734796238244514, "grad_norm": 24.104970932006836, "learning_rate": 4.362e-06, "loss": 2.376, "step": 8724 }, { "epoch": 2.7351097178683386, "grad_norm": 39.21940231323242, "learning_rate": 4.362500000000001e-06, "loss": 2.5344, "step": 8725 }, { "epoch": 2.735423197492163, "grad_norm": 25.463512420654297, "learning_rate": 4.363e-06, "loss": 2.2022, "step": 8726 }, { "epoch": 2.7357366771159874, "grad_norm": 7.078239917755127, "learning_rate": 4.363500000000001e-06, "loss": 1.9901, "step": 8727 }, { "epoch": 2.736050156739812, "grad_norm": 12.24302864074707, "learning_rate": 4.364e-06, "loss": 2.0838, "step": 8728 }, { "epoch": 2.7363636363636363, "grad_norm": 9.073789596557617, "learning_rate": 4.3645e-06, "loss": 1.9942, "step": 8729 }, { "epoch": 2.7366771159874608, "grad_norm": 7.676347255706787, "learning_rate": 4.3650000000000006e-06, "loss": 1.956, "step": 8730 }, { "epoch": 2.736990595611285, "grad_norm": 67.66808319091797, "learning_rate": 4.3655e-06, "loss": 2.0912, "step": 8731 }, { "epoch": 2.7373040752351097, "grad_norm": 8.18461799621582, "learning_rate": 4.366e-06, "loss": 2.1976, "step": 8732 }, { "epoch": 2.737617554858934, "grad_norm": 15.176887512207031, "learning_rate": 4.3665000000000005e-06, "loss": 1.8471, "step": 8733 }, { "epoch": 2.7379310344827585, "grad_norm": 26.231393814086914, "learning_rate": 4.367e-06, "loss": 2.1518, "step": 8734 }, { "epoch": 2.738244514106583, "grad_norm": 15.792627334594727, "learning_rate": 4.367500000000001e-06, "loss": 2.2551, "step": 8735 }, { "epoch": 2.7385579937304074, "grad_norm": 9.082563400268555, "learning_rate": 4.368e-06, "loss": 2.1368, "step": 8736 }, { "epoch": 2.738871473354232, "grad_norm": 12.056462287902832, "learning_rate": 4.368500000000001e-06, "loss": 2.836, "step": 8737 }, { "epoch": 2.7391849529780563, "grad_norm": 15.846611976623535, "learning_rate": 4.369000000000001e-06, "loss": 2.3139, "step": 8738 }, { "epoch": 2.7394984326018808, "grad_norm": 10.829202651977539, "learning_rate": 4.3695e-06, "loss": 1.9912, "step": 8739 }, { "epoch": 2.739811912225705, "grad_norm": 10.6749906539917, "learning_rate": 4.3700000000000005e-06, "loss": 2.36, "step": 8740 }, { "epoch": 2.7401253918495296, "grad_norm": 12.234994888305664, "learning_rate": 4.3705e-06, "loss": 2.0213, "step": 8741 }, { "epoch": 2.740438871473354, "grad_norm": 73.86055755615234, "learning_rate": 4.371e-06, "loss": 2.2849, "step": 8742 }, { "epoch": 2.7407523510971785, "grad_norm": 10.176410675048828, "learning_rate": 4.3715e-06, "loss": 2.1276, "step": 8743 }, { "epoch": 2.741065830721003, "grad_norm": 20.30948829650879, "learning_rate": 4.372e-06, "loss": 2.1619, "step": 8744 }, { "epoch": 2.7413793103448274, "grad_norm": 10.40246295928955, "learning_rate": 4.372500000000001e-06, "loss": 1.9926, "step": 8745 }, { "epoch": 2.741692789968652, "grad_norm": 13.326762199401855, "learning_rate": 4.373e-06, "loss": 2.0367, "step": 8746 }, { "epoch": 2.7420062695924763, "grad_norm": 42.64368438720703, "learning_rate": 4.3735000000000006e-06, "loss": 2.4097, "step": 8747 }, { "epoch": 2.7423197492163007, "grad_norm": 6.306972503662109, "learning_rate": 4.374000000000001e-06, "loss": 2.3297, "step": 8748 }, { "epoch": 2.742633228840125, "grad_norm": 13.6160306930542, "learning_rate": 4.3745e-06, "loss": 2.2999, "step": 8749 }, { "epoch": 2.7429467084639496, "grad_norm": 64.26676177978516, "learning_rate": 4.3750000000000005e-06, "loss": 2.3328, "step": 8750 }, { "epoch": 2.743260188087774, "grad_norm": 15.906298637390137, "learning_rate": 4.375500000000001e-06, "loss": 2.0968, "step": 8751 }, { "epoch": 2.7435736677115985, "grad_norm": 9.482664108276367, "learning_rate": 4.376e-06, "loss": 2.1003, "step": 8752 }, { "epoch": 2.743887147335423, "grad_norm": 10.362493515014648, "learning_rate": 4.3765e-06, "loss": 1.8751, "step": 8753 }, { "epoch": 2.7442006269592474, "grad_norm": 14.289345741271973, "learning_rate": 4.377e-06, "loss": 1.9342, "step": 8754 }, { "epoch": 2.7445141065830723, "grad_norm": 11.082356452941895, "learning_rate": 4.3775e-06, "loss": 2.0513, "step": 8755 }, { "epoch": 2.7448275862068967, "grad_norm": 9.805420875549316, "learning_rate": 4.378e-06, "loss": 2.2192, "step": 8756 }, { "epoch": 2.745141065830721, "grad_norm": 149.66529846191406, "learning_rate": 4.3785000000000005e-06, "loss": 2.3515, "step": 8757 }, { "epoch": 2.7454545454545456, "grad_norm": 10.692829132080078, "learning_rate": 4.379000000000001e-06, "loss": 2.2262, "step": 8758 }, { "epoch": 2.74576802507837, "grad_norm": 11.389817237854004, "learning_rate": 4.3795e-06, "loss": 1.9763, "step": 8759 }, { "epoch": 2.7460815047021945, "grad_norm": 7.813292503356934, "learning_rate": 4.38e-06, "loss": 1.9229, "step": 8760 }, { "epoch": 2.746394984326019, "grad_norm": 11.586114883422852, "learning_rate": 4.380500000000001e-06, "loss": 2.2859, "step": 8761 }, { "epoch": 2.7467084639498434, "grad_norm": 6.598621368408203, "learning_rate": 4.381e-06, "loss": 2.0855, "step": 8762 }, { "epoch": 2.747021943573668, "grad_norm": 11.488838195800781, "learning_rate": 4.3815e-06, "loss": 2.2518, "step": 8763 }, { "epoch": 2.7473354231974922, "grad_norm": 8.629611015319824, "learning_rate": 4.382e-06, "loss": 2.1344, "step": 8764 }, { "epoch": 2.7476489028213167, "grad_norm": 7.51088809967041, "learning_rate": 4.3825e-06, "loss": 2.1943, "step": 8765 }, { "epoch": 2.747962382445141, "grad_norm": 15.733593940734863, "learning_rate": 4.383e-06, "loss": 2.3817, "step": 8766 }, { "epoch": 2.7482758620689656, "grad_norm": 5.379866600036621, "learning_rate": 4.3835000000000005e-06, "loss": 1.9565, "step": 8767 }, { "epoch": 2.74858934169279, "grad_norm": 8.458078384399414, "learning_rate": 4.384000000000001e-06, "loss": 2.0798, "step": 8768 }, { "epoch": 2.7489028213166145, "grad_norm": 16.590438842773438, "learning_rate": 4.3845e-06, "loss": 2.1464, "step": 8769 }, { "epoch": 2.749216300940439, "grad_norm": 11.81791877746582, "learning_rate": 4.385e-06, "loss": 2.3889, "step": 8770 }, { "epoch": 2.7495297805642633, "grad_norm": 8.347785949707031, "learning_rate": 4.385500000000001e-06, "loss": 2.249, "step": 8771 }, { "epoch": 2.749843260188088, "grad_norm": 57.46329879760742, "learning_rate": 4.386e-06, "loss": 2.1428, "step": 8772 }, { "epoch": 2.750156739811912, "grad_norm": 6.635656833648682, "learning_rate": 4.3865e-06, "loss": 2.1542, "step": 8773 }, { "epoch": 2.7504702194357367, "grad_norm": 10.658465385437012, "learning_rate": 4.3870000000000005e-06, "loss": 2.4706, "step": 8774 }, { "epoch": 2.750783699059561, "grad_norm": 6.2919440269470215, "learning_rate": 4.3875e-06, "loss": 2.1062, "step": 8775 }, { "epoch": 2.7510971786833855, "grad_norm": 35.24327087402344, "learning_rate": 4.388e-06, "loss": 2.4187, "step": 8776 }, { "epoch": 2.75141065830721, "grad_norm": 9.974841117858887, "learning_rate": 4.3885e-06, "loss": 2.4681, "step": 8777 }, { "epoch": 2.7517241379310344, "grad_norm": 9.542842864990234, "learning_rate": 4.389000000000001e-06, "loss": 2.1141, "step": 8778 }, { "epoch": 2.7517241379310344, "eval_loss": 2.3617959022521973, "eval_runtime": 21.3368, "eval_samples_per_second": 125.933, "eval_steps_per_second": 7.874, "step": 8778 }, { "epoch": 2.752037617554859, "grad_norm": 6.141866683959961, "learning_rate": 4.3895e-06, "loss": 2.0021, "step": 8779 }, { "epoch": 2.7523510971786833, "grad_norm": 9.911970138549805, "learning_rate": 4.39e-06, "loss": 1.9035, "step": 8780 }, { "epoch": 2.7526645768025078, "grad_norm": 11.476873397827148, "learning_rate": 4.3905000000000006e-06, "loss": 2.0896, "step": 8781 }, { "epoch": 2.752978056426332, "grad_norm": 10.003170013427734, "learning_rate": 4.391e-06, "loss": 2.4521, "step": 8782 }, { "epoch": 2.7532915360501566, "grad_norm": 12.520665168762207, "learning_rate": 4.3915e-06, "loss": 2.0142, "step": 8783 }, { "epoch": 2.753605015673981, "grad_norm": 23.982580184936523, "learning_rate": 4.3920000000000005e-06, "loss": 1.9249, "step": 8784 }, { "epoch": 2.7539184952978055, "grad_norm": 20.520601272583008, "learning_rate": 4.3925e-06, "loss": 2.4014, "step": 8785 }, { "epoch": 2.75423197492163, "grad_norm": 11.348067283630371, "learning_rate": 4.393000000000001e-06, "loss": 2.0867, "step": 8786 }, { "epoch": 2.7545454545454544, "grad_norm": 11.997109413146973, "learning_rate": 4.3935e-06, "loss": 1.8523, "step": 8787 }, { "epoch": 2.754858934169279, "grad_norm": 13.382556915283203, "learning_rate": 4.394000000000001e-06, "loss": 2.2265, "step": 8788 }, { "epoch": 2.7551724137931033, "grad_norm": 7.337337970733643, "learning_rate": 4.3945e-06, "loss": 2.1969, "step": 8789 }, { "epoch": 2.7554858934169277, "grad_norm": 10.263776779174805, "learning_rate": 4.395e-06, "loss": 2.0197, "step": 8790 }, { "epoch": 2.7557993730407526, "grad_norm": 16.822704315185547, "learning_rate": 4.3955000000000005e-06, "loss": 2.0548, "step": 8791 }, { "epoch": 2.756112852664577, "grad_norm": 7.352022647857666, "learning_rate": 4.396e-06, "loss": 2.5021, "step": 8792 }, { "epoch": 2.7564263322884015, "grad_norm": 17.032766342163086, "learning_rate": 4.3965e-06, "loss": 2.3737, "step": 8793 }, { "epoch": 2.756739811912226, "grad_norm": 6.446961879730225, "learning_rate": 4.397e-06, "loss": 2.1699, "step": 8794 }, { "epoch": 2.7570532915360504, "grad_norm": 9.741681098937988, "learning_rate": 4.3975e-06, "loss": 2.0394, "step": 8795 }, { "epoch": 2.757366771159875, "grad_norm": 7.377527713775635, "learning_rate": 4.398000000000001e-06, "loss": 2.1179, "step": 8796 }, { "epoch": 2.7576802507836993, "grad_norm": 12.159534454345703, "learning_rate": 4.3985e-06, "loss": 2.303, "step": 8797 }, { "epoch": 2.7579937304075237, "grad_norm": 9.59463119506836, "learning_rate": 4.3990000000000006e-06, "loss": 2.114, "step": 8798 }, { "epoch": 2.758307210031348, "grad_norm": 16.09486961364746, "learning_rate": 4.399500000000001e-06, "loss": 2.1531, "step": 8799 }, { "epoch": 2.7586206896551726, "grad_norm": 11.489596366882324, "learning_rate": 4.4e-06, "loss": 2.2233, "step": 8800 }, { "epoch": 2.758934169278997, "grad_norm": 7.64333963394165, "learning_rate": 4.4005000000000005e-06, "loss": 1.9012, "step": 8801 }, { "epoch": 2.7592476489028215, "grad_norm": 5.046221733093262, "learning_rate": 4.401e-06, "loss": 2.0765, "step": 8802 }, { "epoch": 2.759561128526646, "grad_norm": 10.52009391784668, "learning_rate": 4.4015e-06, "loss": 2.2994, "step": 8803 }, { "epoch": 2.7598746081504704, "grad_norm": 45.82973861694336, "learning_rate": 4.402e-06, "loss": 2.5796, "step": 8804 }, { "epoch": 2.760188087774295, "grad_norm": 9.874204635620117, "learning_rate": 4.4025e-06, "loss": 2.1705, "step": 8805 }, { "epoch": 2.7605015673981192, "grad_norm": 19.64636993408203, "learning_rate": 4.403000000000001e-06, "loss": 2.0047, "step": 8806 }, { "epoch": 2.7608150470219437, "grad_norm": 10.80135726928711, "learning_rate": 4.4035e-06, "loss": 2.2131, "step": 8807 }, { "epoch": 2.761128526645768, "grad_norm": 20.790512084960938, "learning_rate": 4.4040000000000005e-06, "loss": 2.1861, "step": 8808 }, { "epoch": 2.7614420062695926, "grad_norm": 10.548378944396973, "learning_rate": 4.404500000000001e-06, "loss": 2.2352, "step": 8809 }, { "epoch": 2.761755485893417, "grad_norm": 13.071473121643066, "learning_rate": 4.405e-06, "loss": 2.2005, "step": 8810 }, { "epoch": 2.7620689655172415, "grad_norm": 7.709262847900391, "learning_rate": 4.4055000000000004e-06, "loss": 1.9974, "step": 8811 }, { "epoch": 2.762382445141066, "grad_norm": 8.52371883392334, "learning_rate": 4.406000000000001e-06, "loss": 2.2967, "step": 8812 }, { "epoch": 2.7626959247648903, "grad_norm": 13.496247291564941, "learning_rate": 4.4065e-06, "loss": 2.0976, "step": 8813 }, { "epoch": 2.763009404388715, "grad_norm": 29.376224517822266, "learning_rate": 4.407e-06, "loss": 1.904, "step": 8814 }, { "epoch": 2.7633228840125392, "grad_norm": 11.797502517700195, "learning_rate": 4.4075e-06, "loss": 2.1544, "step": 8815 }, { "epoch": 2.7636363636363637, "grad_norm": 11.623775482177734, "learning_rate": 4.408000000000001e-06, "loss": 2.032, "step": 8816 }, { "epoch": 2.763949843260188, "grad_norm": 44.81899642944336, "learning_rate": 4.4085e-06, "loss": 2.0823, "step": 8817 }, { "epoch": 2.7642633228840126, "grad_norm": 10.980966567993164, "learning_rate": 4.4090000000000005e-06, "loss": 1.9466, "step": 8818 }, { "epoch": 2.764576802507837, "grad_norm": 14.396851539611816, "learning_rate": 4.409500000000001e-06, "loss": 2.1866, "step": 8819 }, { "epoch": 2.7648902821316614, "grad_norm": 10.871788024902344, "learning_rate": 4.41e-06, "loss": 2.4454, "step": 8820 }, { "epoch": 2.765203761755486, "grad_norm": 7.55269718170166, "learning_rate": 4.4105e-06, "loss": 2.2037, "step": 8821 }, { "epoch": 2.7655172413793103, "grad_norm": 12.933215141296387, "learning_rate": 4.411000000000001e-06, "loss": 1.9373, "step": 8822 }, { "epoch": 2.7658307210031348, "grad_norm": 50.21009826660156, "learning_rate": 4.4115e-06, "loss": 1.9855, "step": 8823 }, { "epoch": 2.766144200626959, "grad_norm": 7.311219215393066, "learning_rate": 4.412e-06, "loss": 2.042, "step": 8824 }, { "epoch": 2.7664576802507836, "grad_norm": 64.9020767211914, "learning_rate": 4.4125000000000005e-06, "loss": 2.0924, "step": 8825 }, { "epoch": 2.766771159874608, "grad_norm": 50.68932342529297, "learning_rate": 4.413000000000001e-06, "loss": 2.0722, "step": 8826 }, { "epoch": 2.7670846394984325, "grad_norm": 7.230908393859863, "learning_rate": 4.4135e-06, "loss": 2.176, "step": 8827 }, { "epoch": 2.767398119122257, "grad_norm": 14.81300163269043, "learning_rate": 4.4140000000000004e-06, "loss": 2.2068, "step": 8828 }, { "epoch": 2.7677115987460814, "grad_norm": 13.327717781066895, "learning_rate": 4.414500000000001e-06, "loss": 2.3298, "step": 8829 }, { "epoch": 2.768025078369906, "grad_norm": 15.708708763122559, "learning_rate": 4.415e-06, "loss": 2.3346, "step": 8830 }, { "epoch": 2.7683385579937303, "grad_norm": 9.580621719360352, "learning_rate": 4.4155e-06, "loss": 2.085, "step": 8831 }, { "epoch": 2.7686520376175547, "grad_norm": 8.304826736450195, "learning_rate": 4.416000000000001e-06, "loss": 1.8869, "step": 8832 }, { "epoch": 2.768965517241379, "grad_norm": 13.014691352844238, "learning_rate": 4.4165e-06, "loss": 1.8898, "step": 8833 }, { "epoch": 2.7692789968652036, "grad_norm": 40.26395034790039, "learning_rate": 4.417e-06, "loss": 2.5348, "step": 8834 }, { "epoch": 2.769592476489028, "grad_norm": 29.082359313964844, "learning_rate": 4.4175000000000005e-06, "loss": 2.2525, "step": 8835 }, { "epoch": 2.7699059561128525, "grad_norm": 9.791305541992188, "learning_rate": 4.418000000000001e-06, "loss": 2.1287, "step": 8836 }, { "epoch": 2.770219435736677, "grad_norm": 7.087009906768799, "learning_rate": 4.418500000000001e-06, "loss": 2.131, "step": 8837 }, { "epoch": 2.7705329153605014, "grad_norm": 9.832541465759277, "learning_rate": 4.419e-06, "loss": 1.8359, "step": 8838 }, { "epoch": 2.770846394984326, "grad_norm": 12.309270858764648, "learning_rate": 4.419500000000001e-06, "loss": 2.3067, "step": 8839 }, { "epoch": 2.7711598746081503, "grad_norm": 59.421016693115234, "learning_rate": 4.42e-06, "loss": 2.4335, "step": 8840 }, { "epoch": 2.7714733542319747, "grad_norm": 23.043010711669922, "learning_rate": 4.4205e-06, "loss": 2.2378, "step": 8841 }, { "epoch": 2.771786833855799, "grad_norm": 12.809768676757812, "learning_rate": 4.4210000000000005e-06, "loss": 2.3007, "step": 8842 }, { "epoch": 2.7721003134796236, "grad_norm": 8.925813674926758, "learning_rate": 4.4215e-06, "loss": 2.1803, "step": 8843 }, { "epoch": 2.772413793103448, "grad_norm": 10.098374366760254, "learning_rate": 4.422e-06, "loss": 2.2867, "step": 8844 }, { "epoch": 2.7727272727272725, "grad_norm": 22.964994430541992, "learning_rate": 4.4225000000000004e-06, "loss": 2.4681, "step": 8845 }, { "epoch": 2.773040752351097, "grad_norm": 21.515104293823242, "learning_rate": 4.423000000000001e-06, "loss": 1.8973, "step": 8846 }, { "epoch": 2.7733542319749214, "grad_norm": 7.05153751373291, "learning_rate": 4.423500000000001e-06, "loss": 2.1681, "step": 8847 }, { "epoch": 2.773667711598746, "grad_norm": 7.453849792480469, "learning_rate": 4.424e-06, "loss": 2.3064, "step": 8848 }, { "epoch": 2.7739811912225707, "grad_norm": 11.146045684814453, "learning_rate": 4.424500000000001e-06, "loss": 2.0165, "step": 8849 }, { "epoch": 2.774294670846395, "grad_norm": 15.386848449707031, "learning_rate": 4.425e-06, "loss": 2.2881, "step": 8850 }, { "epoch": 2.7746081504702196, "grad_norm": 9.545374870300293, "learning_rate": 4.4255e-06, "loss": 2.473, "step": 8851 }, { "epoch": 2.774921630094044, "grad_norm": 13.441974639892578, "learning_rate": 4.4260000000000005e-06, "loss": 2.0285, "step": 8852 }, { "epoch": 2.7752351097178685, "grad_norm": 9.309048652648926, "learning_rate": 4.4265e-06, "loss": 2.0226, "step": 8853 }, { "epoch": 2.775548589341693, "grad_norm": 9.009796142578125, "learning_rate": 4.427e-06, "loss": 2.1872, "step": 8854 }, { "epoch": 2.7758620689655173, "grad_norm": 8.012994766235352, "learning_rate": 4.4275e-06, "loss": 2.2668, "step": 8855 }, { "epoch": 2.776175548589342, "grad_norm": 15.51158332824707, "learning_rate": 4.428000000000001e-06, "loss": 2.2528, "step": 8856 }, { "epoch": 2.7764890282131662, "grad_norm": 8.40670394897461, "learning_rate": 4.428500000000001e-06, "loss": 2.2335, "step": 8857 }, { "epoch": 2.7768025078369907, "grad_norm": 64.97605895996094, "learning_rate": 4.429e-06, "loss": 2.5415, "step": 8858 }, { "epoch": 2.777115987460815, "grad_norm": 14.467185974121094, "learning_rate": 4.4295000000000005e-06, "loss": 2.0422, "step": 8859 }, { "epoch": 2.7774294670846396, "grad_norm": 34.767578125, "learning_rate": 4.430000000000001e-06, "loss": 3.126, "step": 8860 }, { "epoch": 2.777742946708464, "grad_norm": 15.762866020202637, "learning_rate": 4.4305e-06, "loss": 2.1705, "step": 8861 }, { "epoch": 2.7780564263322884, "grad_norm": 8.35089111328125, "learning_rate": 4.4310000000000004e-06, "loss": 2.2454, "step": 8862 }, { "epoch": 2.778369905956113, "grad_norm": 46.07892608642578, "learning_rate": 4.4315e-06, "loss": 2.3334, "step": 8863 }, { "epoch": 2.7786833855799373, "grad_norm": 118.09009552001953, "learning_rate": 4.432e-06, "loss": 2.5326, "step": 8864 }, { "epoch": 2.7789968652037618, "grad_norm": 118.22344970703125, "learning_rate": 4.4325e-06, "loss": 2.3335, "step": 8865 }, { "epoch": 2.779310344827586, "grad_norm": 11.5182523727417, "learning_rate": 4.433000000000001e-06, "loss": 1.9205, "step": 8866 }, { "epoch": 2.7796238244514107, "grad_norm": 9.261908531188965, "learning_rate": 4.433500000000001e-06, "loss": 2.1941, "step": 8867 }, { "epoch": 2.779937304075235, "grad_norm": 6.157930374145508, "learning_rate": 4.434e-06, "loss": 2.1004, "step": 8868 }, { "epoch": 2.7802507836990595, "grad_norm": 39.35654067993164, "learning_rate": 4.4345000000000005e-06, "loss": 2.4497, "step": 8869 }, { "epoch": 2.780564263322884, "grad_norm": 8.866540908813477, "learning_rate": 4.435000000000001e-06, "loss": 2.1817, "step": 8870 }, { "epoch": 2.7808777429467084, "grad_norm": 8.943115234375, "learning_rate": 4.4355e-06, "loss": 2.2464, "step": 8871 }, { "epoch": 2.781191222570533, "grad_norm": 12.45877456665039, "learning_rate": 4.436e-06, "loss": 2.3073, "step": 8872 }, { "epoch": 2.7815047021943573, "grad_norm": 87.09159088134766, "learning_rate": 4.436500000000001e-06, "loss": 2.3789, "step": 8873 }, { "epoch": 2.7818181818181817, "grad_norm": 13.245519638061523, "learning_rate": 4.437e-06, "loss": 2.0128, "step": 8874 }, { "epoch": 2.782131661442006, "grad_norm": 10.666522979736328, "learning_rate": 4.4375e-06, "loss": 2.2148, "step": 8875 }, { "epoch": 2.7824451410658306, "grad_norm": 9.369501113891602, "learning_rate": 4.438e-06, "loss": 2.2129, "step": 8876 }, { "epoch": 2.782758620689655, "grad_norm": 13.705166816711426, "learning_rate": 4.438500000000001e-06, "loss": 2.3283, "step": 8877 }, { "epoch": 2.7830721003134795, "grad_norm": 42.696937561035156, "learning_rate": 4.439e-06, "loss": 2.2972, "step": 8878 }, { "epoch": 2.783385579937304, "grad_norm": 6.780346870422363, "learning_rate": 4.4395000000000004e-06, "loss": 1.8835, "step": 8879 }, { "epoch": 2.7836990595611284, "grad_norm": 6.855872631072998, "learning_rate": 4.440000000000001e-06, "loss": 2.0603, "step": 8880 }, { "epoch": 2.784012539184953, "grad_norm": 23.189697265625, "learning_rate": 4.4405e-06, "loss": 2.6584, "step": 8881 }, { "epoch": 2.7843260188087773, "grad_norm": 19.165937423706055, "learning_rate": 4.441e-06, "loss": 2.0947, "step": 8882 }, { "epoch": 2.7846394984326017, "grad_norm": 12.241186141967773, "learning_rate": 4.441500000000001e-06, "loss": 2.1631, "step": 8883 }, { "epoch": 2.7849529780564266, "grad_norm": 29.361146926879883, "learning_rate": 4.442e-06, "loss": 2.064, "step": 8884 }, { "epoch": 2.785266457680251, "grad_norm": 12.325139045715332, "learning_rate": 4.4425e-06, "loss": 2.4826, "step": 8885 }, { "epoch": 2.7855799373040755, "grad_norm": 101.8729476928711, "learning_rate": 4.4430000000000005e-06, "loss": 2.4885, "step": 8886 }, { "epoch": 2.7858934169279, "grad_norm": 6.408265113830566, "learning_rate": 4.443500000000001e-06, "loss": 2.0984, "step": 8887 }, { "epoch": 2.7862068965517244, "grad_norm": 10.48788833618164, "learning_rate": 4.444e-06, "loss": 2.4544, "step": 8888 }, { "epoch": 2.786520376175549, "grad_norm": 11.668582916259766, "learning_rate": 4.4445e-06, "loss": 2.0868, "step": 8889 }, { "epoch": 2.7868338557993733, "grad_norm": 7.010461807250977, "learning_rate": 4.445000000000001e-06, "loss": 2.1672, "step": 8890 }, { "epoch": 2.7871473354231977, "grad_norm": 10.394134521484375, "learning_rate": 4.4455e-06, "loss": 1.7329, "step": 8891 }, { "epoch": 2.787460815047022, "grad_norm": 9.518021583557129, "learning_rate": 4.446e-06, "loss": 2.0006, "step": 8892 }, { "epoch": 2.7877742946708466, "grad_norm": 8.575425148010254, "learning_rate": 4.4465000000000005e-06, "loss": 2.0128, "step": 8893 }, { "epoch": 2.788087774294671, "grad_norm": 15.147712707519531, "learning_rate": 4.447e-06, "loss": 2.024, "step": 8894 }, { "epoch": 2.7884012539184955, "grad_norm": 8.381316184997559, "learning_rate": 4.4475e-06, "loss": 2.0709, "step": 8895 }, { "epoch": 2.78871473354232, "grad_norm": 17.02097511291504, "learning_rate": 4.4480000000000004e-06, "loss": 2.2298, "step": 8896 }, { "epoch": 2.7890282131661444, "grad_norm": 9.578057289123535, "learning_rate": 4.448500000000001e-06, "loss": 2.1958, "step": 8897 }, { "epoch": 2.789341692789969, "grad_norm": 19.883148193359375, "learning_rate": 4.449000000000001e-06, "loss": 2.6898, "step": 8898 }, { "epoch": 2.7896551724137932, "grad_norm": 9.921863555908203, "learning_rate": 4.4495e-06, "loss": 2.3476, "step": 8899 }, { "epoch": 2.7899686520376177, "grad_norm": 8.587871551513672, "learning_rate": 4.450000000000001e-06, "loss": 2.0288, "step": 8900 }, { "epoch": 2.790282131661442, "grad_norm": 10.42520809173584, "learning_rate": 4.4505e-06, "loss": 2.0105, "step": 8901 }, { "epoch": 2.7905956112852666, "grad_norm": 6.413170337677002, "learning_rate": 4.451e-06, "loss": 1.9845, "step": 8902 }, { "epoch": 2.790909090909091, "grad_norm": 14.913331985473633, "learning_rate": 4.4515000000000005e-06, "loss": 2.2859, "step": 8903 }, { "epoch": 2.7912225705329154, "grad_norm": 7.387153148651123, "learning_rate": 4.452e-06, "loss": 2.157, "step": 8904 }, { "epoch": 2.79153605015674, "grad_norm": 8.039084434509277, "learning_rate": 4.4525e-06, "loss": 2.3603, "step": 8905 }, { "epoch": 2.7918495297805643, "grad_norm": 9.784404754638672, "learning_rate": 4.453e-06, "loss": 2.0533, "step": 8906 }, { "epoch": 2.7921630094043888, "grad_norm": 6.479061603546143, "learning_rate": 4.453500000000001e-06, "loss": 2.0598, "step": 8907 }, { "epoch": 2.792476489028213, "grad_norm": 25.488035202026367, "learning_rate": 4.454000000000001e-06, "loss": 2.4351, "step": 8908 }, { "epoch": 2.7927899686520377, "grad_norm": 10.290550231933594, "learning_rate": 4.4545e-06, "loss": 2.1876, "step": 8909 }, { "epoch": 2.793103448275862, "grad_norm": 92.74346160888672, "learning_rate": 4.4550000000000005e-06, "loss": 2.5315, "step": 8910 }, { "epoch": 2.7934169278996865, "grad_norm": 56.78178024291992, "learning_rate": 4.455500000000001e-06, "loss": 2.6847, "step": 8911 }, { "epoch": 2.793730407523511, "grad_norm": 10.42226791381836, "learning_rate": 4.456e-06, "loss": 1.8981, "step": 8912 }, { "epoch": 2.7940438871473354, "grad_norm": 9.551287651062012, "learning_rate": 4.4565000000000004e-06, "loss": 1.9468, "step": 8913 }, { "epoch": 2.79435736677116, "grad_norm": 8.744683265686035, "learning_rate": 4.457e-06, "loss": 2.0248, "step": 8914 }, { "epoch": 2.7946708463949843, "grad_norm": 7.037728786468506, "learning_rate": 4.4575e-06, "loss": 1.9707, "step": 8915 }, { "epoch": 2.7949843260188088, "grad_norm": 14.505931854248047, "learning_rate": 4.458e-06, "loss": 1.898, "step": 8916 }, { "epoch": 2.795297805642633, "grad_norm": 11.145400047302246, "learning_rate": 4.458500000000001e-06, "loss": 1.774, "step": 8917 }, { "epoch": 2.7956112852664576, "grad_norm": 17.71766471862793, "learning_rate": 4.459000000000001e-06, "loss": 2.0299, "step": 8918 }, { "epoch": 2.795924764890282, "grad_norm": 11.049284934997559, "learning_rate": 4.4595e-06, "loss": 2.2841, "step": 8919 }, { "epoch": 2.7962382445141065, "grad_norm": 6.764120101928711, "learning_rate": 4.4600000000000005e-06, "loss": 2.28, "step": 8920 }, { "epoch": 2.796551724137931, "grad_norm": 8.527141571044922, "learning_rate": 4.460500000000001e-06, "loss": 2.1364, "step": 8921 }, { "epoch": 2.7968652037617554, "grad_norm": 27.075227737426758, "learning_rate": 4.461e-06, "loss": 2.5891, "step": 8922 }, { "epoch": 2.79717868338558, "grad_norm": 11.936616897583008, "learning_rate": 4.4615e-06, "loss": 1.9973, "step": 8923 }, { "epoch": 2.7974921630094043, "grad_norm": 81.39875793457031, "learning_rate": 4.462e-06, "loss": 2.2896, "step": 8924 }, { "epoch": 2.7978056426332287, "grad_norm": 11.963140487670898, "learning_rate": 4.4625e-06, "loss": 2.0503, "step": 8925 }, { "epoch": 2.798119122257053, "grad_norm": 7.974978446960449, "learning_rate": 4.463e-06, "loss": 2.0642, "step": 8926 }, { "epoch": 2.7984326018808776, "grad_norm": 9.932296752929688, "learning_rate": 4.4635000000000005e-06, "loss": 1.8638, "step": 8927 }, { "epoch": 2.798746081504702, "grad_norm": 10.0237455368042, "learning_rate": 4.464000000000001e-06, "loss": 1.98, "step": 8928 }, { "epoch": 2.7990595611285265, "grad_norm": 7.270865440368652, "learning_rate": 4.4645e-06, "loss": 2.088, "step": 8929 }, { "epoch": 2.799373040752351, "grad_norm": 6.515256881713867, "learning_rate": 4.4650000000000004e-06, "loss": 2.1266, "step": 8930 }, { "epoch": 2.7996865203761754, "grad_norm": 11.318952560424805, "learning_rate": 4.465500000000001e-06, "loss": 2.0915, "step": 8931 }, { "epoch": 2.8, "grad_norm": 71.32071685791016, "learning_rate": 4.466e-06, "loss": 2.1624, "step": 8932 }, { "epoch": 2.8003134796238243, "grad_norm": 5.972036838531494, "learning_rate": 4.4665e-06, "loss": 1.9366, "step": 8933 }, { "epoch": 2.8006269592476487, "grad_norm": 10.313633918762207, "learning_rate": 4.467000000000001e-06, "loss": 2.1382, "step": 8934 }, { "epoch": 2.800940438871473, "grad_norm": 12.486124992370605, "learning_rate": 4.4675e-06, "loss": 2.3986, "step": 8935 }, { "epoch": 2.8012539184952976, "grad_norm": 9.801366806030273, "learning_rate": 4.468e-06, "loss": 2.5913, "step": 8936 }, { "epoch": 2.801567398119122, "grad_norm": 10.750311851501465, "learning_rate": 4.4685000000000005e-06, "loss": 2.0242, "step": 8937 }, { "epoch": 2.8018808777429465, "grad_norm": 11.844935417175293, "learning_rate": 4.469000000000001e-06, "loss": 2.2552, "step": 8938 }, { "epoch": 2.802194357366771, "grad_norm": 13.268489837646484, "learning_rate": 4.4695e-06, "loss": 2.0179, "step": 8939 }, { "epoch": 2.8025078369905954, "grad_norm": 8.299356460571289, "learning_rate": 4.47e-06, "loss": 2.3727, "step": 8940 }, { "epoch": 2.80282131661442, "grad_norm": 5.049087047576904, "learning_rate": 4.470500000000001e-06, "loss": 2.1255, "step": 8941 }, { "epoch": 2.8031347962382442, "grad_norm": 10.893635749816895, "learning_rate": 4.471e-06, "loss": 2.9099, "step": 8942 }, { "epoch": 2.803448275862069, "grad_norm": 11.529300689697266, "learning_rate": 4.4715e-06, "loss": 2.5637, "step": 8943 }, { "epoch": 2.8037617554858936, "grad_norm": 9.241459846496582, "learning_rate": 4.4720000000000006e-06, "loss": 1.9661, "step": 8944 }, { "epoch": 2.804075235109718, "grad_norm": 10.427563667297363, "learning_rate": 4.4725e-06, "loss": 2.2509, "step": 8945 }, { "epoch": 2.8043887147335425, "grad_norm": 12.335206985473633, "learning_rate": 4.473e-06, "loss": 2.2733, "step": 8946 }, { "epoch": 2.804702194357367, "grad_norm": 35.176483154296875, "learning_rate": 4.4735000000000005e-06, "loss": 2.2794, "step": 8947 }, { "epoch": 2.8050156739811913, "grad_norm": 10.523015022277832, "learning_rate": 4.474000000000001e-06, "loss": 2.1059, "step": 8948 }, { "epoch": 2.805329153605016, "grad_norm": 8.507455825805664, "learning_rate": 4.4745e-06, "loss": 2.158, "step": 8949 }, { "epoch": 2.80564263322884, "grad_norm": 10.326408386230469, "learning_rate": 4.475e-06, "loss": 2.6805, "step": 8950 }, { "epoch": 2.8059561128526647, "grad_norm": 10.275668144226074, "learning_rate": 4.475500000000001e-06, "loss": 2.0613, "step": 8951 }, { "epoch": 2.806269592476489, "grad_norm": 9.237815856933594, "learning_rate": 4.476e-06, "loss": 2.2169, "step": 8952 }, { "epoch": 2.8065830721003135, "grad_norm": 26.920560836791992, "learning_rate": 4.4765e-06, "loss": 2.3189, "step": 8953 }, { "epoch": 2.806896551724138, "grad_norm": 45.41976547241211, "learning_rate": 4.4770000000000005e-06, "loss": 2.1531, "step": 8954 }, { "epoch": 2.8072100313479624, "grad_norm": 57.38456344604492, "learning_rate": 4.4775e-06, "loss": 2.0104, "step": 8955 }, { "epoch": 2.807523510971787, "grad_norm": 13.814677238464355, "learning_rate": 4.478e-06, "loss": 2.209, "step": 8956 }, { "epoch": 2.8078369905956113, "grad_norm": 12.040349960327148, "learning_rate": 4.4785e-06, "loss": 2.3702, "step": 8957 }, { "epoch": 2.8081504702194358, "grad_norm": 12.343782424926758, "learning_rate": 4.479000000000001e-06, "loss": 2.334, "step": 8958 }, { "epoch": 2.80846394984326, "grad_norm": 11.390047073364258, "learning_rate": 4.479500000000001e-06, "loss": 2.2881, "step": 8959 }, { "epoch": 2.8087774294670846, "grad_norm": 9.788714408874512, "learning_rate": 4.48e-06, "loss": 2.0453, "step": 8960 }, { "epoch": 2.809090909090909, "grad_norm": 6.3288493156433105, "learning_rate": 4.4805000000000006e-06, "loss": 2.2622, "step": 8961 }, { "epoch": 2.8094043887147335, "grad_norm": 40.452964782714844, "learning_rate": 4.481e-06, "loss": 2.6254, "step": 8962 }, { "epoch": 2.809717868338558, "grad_norm": 14.963085174560547, "learning_rate": 4.4815e-06, "loss": 2.4563, "step": 8963 }, { "epoch": 2.8100313479623824, "grad_norm": 11.784871101379395, "learning_rate": 4.4820000000000005e-06, "loss": 2.2176, "step": 8964 }, { "epoch": 2.810344827586207, "grad_norm": 15.158439636230469, "learning_rate": 4.4825e-06, "loss": 2.0336, "step": 8965 }, { "epoch": 2.8106583072100313, "grad_norm": 9.447674751281738, "learning_rate": 4.483e-06, "loss": 2.2189, "step": 8966 }, { "epoch": 2.8109717868338557, "grad_norm": 11.19941234588623, "learning_rate": 4.4835e-06, "loss": 2.1274, "step": 8967 }, { "epoch": 2.81128526645768, "grad_norm": 7.923600196838379, "learning_rate": 4.484000000000001e-06, "loss": 2.1434, "step": 8968 }, { "epoch": 2.8115987460815046, "grad_norm": 10.660632133483887, "learning_rate": 4.484500000000001e-06, "loss": 2.0567, "step": 8969 }, { "epoch": 2.811912225705329, "grad_norm": 9.11224365234375, "learning_rate": 4.485e-06, "loss": 2.1105, "step": 8970 }, { "epoch": 2.8122257053291535, "grad_norm": 12.176597595214844, "learning_rate": 4.4855000000000005e-06, "loss": 1.7323, "step": 8971 }, { "epoch": 2.812539184952978, "grad_norm": 21.556398391723633, "learning_rate": 4.486000000000001e-06, "loss": 2.003, "step": 8972 }, { "epoch": 2.8128526645768024, "grad_norm": 7.698760032653809, "learning_rate": 4.4865e-06, "loss": 2.1953, "step": 8973 }, { "epoch": 2.813166144200627, "grad_norm": 61.005977630615234, "learning_rate": 4.487e-06, "loss": 2.2959, "step": 8974 }, { "epoch": 2.8134796238244513, "grad_norm": 14.767295837402344, "learning_rate": 4.4875e-06, "loss": 2.104, "step": 8975 }, { "epoch": 2.8137931034482757, "grad_norm": 43.7757568359375, "learning_rate": 4.488e-06, "loss": 2.2464, "step": 8976 }, { "epoch": 2.8141065830721, "grad_norm": 11.51502513885498, "learning_rate": 4.4885e-06, "loss": 2.3415, "step": 8977 }, { "epoch": 2.814420062695925, "grad_norm": 14.869585990905762, "learning_rate": 4.4890000000000006e-06, "loss": 1.9543, "step": 8978 }, { "epoch": 2.8147335423197495, "grad_norm": 8.134459495544434, "learning_rate": 4.489500000000001e-06, "loss": 2.2213, "step": 8979 }, { "epoch": 2.815047021943574, "grad_norm": 9.880690574645996, "learning_rate": 4.49e-06, "loss": 2.2259, "step": 8980 }, { "epoch": 2.8153605015673984, "grad_norm": 6.740392208099365, "learning_rate": 4.4905000000000005e-06, "loss": 2.0825, "step": 8981 }, { "epoch": 2.815673981191223, "grad_norm": 7.842469692230225, "learning_rate": 4.491000000000001e-06, "loss": 2.0853, "step": 8982 }, { "epoch": 2.8159874608150472, "grad_norm": 14.057806015014648, "learning_rate": 4.4915e-06, "loss": 2.3229, "step": 8983 }, { "epoch": 2.8163009404388717, "grad_norm": 73.59526062011719, "learning_rate": 4.492e-06, "loss": 2.7831, "step": 8984 }, { "epoch": 2.816614420062696, "grad_norm": 35.451011657714844, "learning_rate": 4.492500000000001e-06, "loss": 2.5415, "step": 8985 }, { "epoch": 2.8169278996865206, "grad_norm": 13.504325866699219, "learning_rate": 4.493e-06, "loss": 1.951, "step": 8986 }, { "epoch": 2.817241379310345, "grad_norm": 22.659767150878906, "learning_rate": 4.4935e-06, "loss": 2.3785, "step": 8987 }, { "epoch": 2.8175548589341695, "grad_norm": 7.116614818572998, "learning_rate": 4.4940000000000005e-06, "loss": 2.1183, "step": 8988 }, { "epoch": 2.817868338557994, "grad_norm": 10.698858261108398, "learning_rate": 4.494500000000001e-06, "loss": 2.0648, "step": 8989 }, { "epoch": 2.8181818181818183, "grad_norm": 16.16118812561035, "learning_rate": 4.495e-06, "loss": 2.3168, "step": 8990 }, { "epoch": 2.818495297805643, "grad_norm": 8.268321990966797, "learning_rate": 4.4955e-06, "loss": 2.0292, "step": 8991 }, { "epoch": 2.8188087774294672, "grad_norm": 6.772553443908691, "learning_rate": 4.496000000000001e-06, "loss": 2.0487, "step": 8992 }, { "epoch": 2.8191222570532917, "grad_norm": 13.590089797973633, "learning_rate": 4.4965e-06, "loss": 2.3, "step": 8993 }, { "epoch": 2.819435736677116, "grad_norm": 9.376701354980469, "learning_rate": 4.497e-06, "loss": 2.2071, "step": 8994 }, { "epoch": 2.8197492163009406, "grad_norm": 9.256257057189941, "learning_rate": 4.4975000000000006e-06, "loss": 2.0739, "step": 8995 }, { "epoch": 2.820062695924765, "grad_norm": 7.156892776489258, "learning_rate": 4.498e-06, "loss": 2.2141, "step": 8996 }, { "epoch": 2.8203761755485894, "grad_norm": 8.66650104522705, "learning_rate": 4.498500000000001e-06, "loss": 2.2167, "step": 8997 }, { "epoch": 2.820689655172414, "grad_norm": 8.071707725524902, "learning_rate": 4.4990000000000005e-06, "loss": 2.0001, "step": 8998 }, { "epoch": 2.8210031347962383, "grad_norm": 12.297947883605957, "learning_rate": 4.499500000000001e-06, "loss": 2.0467, "step": 8999 }, { "epoch": 2.8213166144200628, "grad_norm": 8.34979248046875, "learning_rate": 4.5e-06, "loss": 2.1129, "step": 9000 }, { "epoch": 2.821630094043887, "grad_norm": 11.1865816116333, "learning_rate": 4.5005e-06, "loss": 2.1643, "step": 9001 }, { "epoch": 2.8219435736677116, "grad_norm": 103.330810546875, "learning_rate": 4.501000000000001e-06, "loss": 2.2524, "step": 9002 }, { "epoch": 2.822257053291536, "grad_norm": 7.96821403503418, "learning_rate": 4.5015e-06, "loss": 2.3146, "step": 9003 }, { "epoch": 2.8225705329153605, "grad_norm": 10.36768627166748, "learning_rate": 4.502e-06, "loss": 2.2553, "step": 9004 }, { "epoch": 2.822884012539185, "grad_norm": 15.110746383666992, "learning_rate": 4.5025000000000005e-06, "loss": 1.9758, "step": 9005 }, { "epoch": 2.8231974921630094, "grad_norm": 10.323888778686523, "learning_rate": 4.503e-06, "loss": 2.4217, "step": 9006 }, { "epoch": 2.823510971786834, "grad_norm": 19.16963768005371, "learning_rate": 4.5035e-06, "loss": 1.7378, "step": 9007 }, { "epoch": 2.8238244514106583, "grad_norm": 13.708786010742188, "learning_rate": 4.504e-06, "loss": 2.2223, "step": 9008 }, { "epoch": 2.8241379310344827, "grad_norm": 12.663023948669434, "learning_rate": 4.504500000000001e-06, "loss": 2.1878, "step": 9009 }, { "epoch": 2.824451410658307, "grad_norm": 16.468090057373047, "learning_rate": 4.505e-06, "loss": 2.3482, "step": 9010 }, { "epoch": 2.8247648902821316, "grad_norm": 5.656088352203369, "learning_rate": 4.5055e-06, "loss": 1.9894, "step": 9011 }, { "epoch": 2.825078369905956, "grad_norm": 16.82462501525879, "learning_rate": 4.5060000000000006e-06, "loss": 2.1139, "step": 9012 }, { "epoch": 2.8253918495297805, "grad_norm": 5.906308650970459, "learning_rate": 4.5065e-06, "loss": 2.5359, "step": 9013 }, { "epoch": 2.825705329153605, "grad_norm": 9.538488388061523, "learning_rate": 4.507e-06, "loss": 2.1039, "step": 9014 }, { "epoch": 2.8260188087774294, "grad_norm": 7.262725830078125, "learning_rate": 4.5075000000000005e-06, "loss": 2.165, "step": 9015 }, { "epoch": 2.826332288401254, "grad_norm": 7.189306259155273, "learning_rate": 4.508e-06, "loss": 1.8715, "step": 9016 }, { "epoch": 2.8266457680250783, "grad_norm": 6.765747547149658, "learning_rate": 4.5085e-06, "loss": 1.9635, "step": 9017 }, { "epoch": 2.8269592476489027, "grad_norm": 8.568934440612793, "learning_rate": 4.509e-06, "loss": 2.0622, "step": 9018 }, { "epoch": 2.827272727272727, "grad_norm": 31.09558868408203, "learning_rate": 4.509500000000001e-06, "loss": 2.3204, "step": 9019 }, { "epoch": 2.8275862068965516, "grad_norm": 6.971149921417236, "learning_rate": 4.510000000000001e-06, "loss": 2.1637, "step": 9020 }, { "epoch": 2.827899686520376, "grad_norm": 9.02498722076416, "learning_rate": 4.5105e-06, "loss": 2.0255, "step": 9021 }, { "epoch": 2.8282131661442005, "grad_norm": 6.301297187805176, "learning_rate": 4.5110000000000005e-06, "loss": 2.3035, "step": 9022 }, { "epoch": 2.828526645768025, "grad_norm": 28.766643524169922, "learning_rate": 4.5115e-06, "loss": 2.1431, "step": 9023 }, { "epoch": 2.8288401253918494, "grad_norm": 8.59941291809082, "learning_rate": 4.512e-06, "loss": 2.1766, "step": 9024 }, { "epoch": 2.829153605015674, "grad_norm": 12.546303749084473, "learning_rate": 4.5125e-06, "loss": 2.1225, "step": 9025 }, { "epoch": 2.8294670846394983, "grad_norm": 67.53726959228516, "learning_rate": 4.513e-06, "loss": 1.9965, "step": 9026 }, { "epoch": 2.8297805642633227, "grad_norm": 48.090511322021484, "learning_rate": 4.5135e-06, "loss": 1.8382, "step": 9027 }, { "epoch": 2.830094043887147, "grad_norm": 60.43867492675781, "learning_rate": 4.514e-06, "loss": 2.412, "step": 9028 }, { "epoch": 2.8304075235109716, "grad_norm": 11.452451705932617, "learning_rate": 4.5145000000000006e-06, "loss": 2.3742, "step": 9029 }, { "epoch": 2.830721003134796, "grad_norm": 143.02935791015625, "learning_rate": 4.515000000000001e-06, "loss": 1.9552, "step": 9030 }, { "epoch": 2.8310344827586205, "grad_norm": 11.755420684814453, "learning_rate": 4.5155e-06, "loss": 2.2418, "step": 9031 }, { "epoch": 2.831347962382445, "grad_norm": 7.70328950881958, "learning_rate": 4.5160000000000005e-06, "loss": 2.106, "step": 9032 }, { "epoch": 2.8316614420062693, "grad_norm": 13.105820655822754, "learning_rate": 4.516500000000001e-06, "loss": 2.2302, "step": 9033 }, { "epoch": 2.831974921630094, "grad_norm": 10.807720184326172, "learning_rate": 4.517e-06, "loss": 2.0056, "step": 9034 }, { "epoch": 2.8322884012539182, "grad_norm": 15.04885482788086, "learning_rate": 4.5175e-06, "loss": 2.1006, "step": 9035 }, { "epoch": 2.8326018808777427, "grad_norm": 31.516225814819336, "learning_rate": 4.518e-06, "loss": 1.9643, "step": 9036 }, { "epoch": 2.8329153605015676, "grad_norm": 8.040714263916016, "learning_rate": 4.5185e-06, "loss": 2.0746, "step": 9037 }, { "epoch": 2.833228840125392, "grad_norm": 18.60951805114746, "learning_rate": 4.519e-06, "loss": 2.1059, "step": 9038 }, { "epoch": 2.8335423197492164, "grad_norm": 55.438716888427734, "learning_rate": 4.5195000000000005e-06, "loss": 1.5143, "step": 9039 }, { "epoch": 2.833855799373041, "grad_norm": 10.365557670593262, "learning_rate": 4.520000000000001e-06, "loss": 2.3775, "step": 9040 }, { "epoch": 2.8341692789968653, "grad_norm": 67.2275619506836, "learning_rate": 4.5205e-06, "loss": 2.4593, "step": 9041 }, { "epoch": 2.8344827586206898, "grad_norm": 10.001243591308594, "learning_rate": 4.521e-06, "loss": 2.1687, "step": 9042 }, { "epoch": 2.834796238244514, "grad_norm": 30.088306427001953, "learning_rate": 4.521500000000001e-06, "loss": 2.008, "step": 9043 }, { "epoch": 2.8351097178683387, "grad_norm": 25.20590591430664, "learning_rate": 4.522e-06, "loss": 2.3407, "step": 9044 }, { "epoch": 2.835423197492163, "grad_norm": 6.711122989654541, "learning_rate": 4.5225e-06, "loss": 2.0922, "step": 9045 }, { "epoch": 2.8357366771159875, "grad_norm": 47.22773361206055, "learning_rate": 4.5230000000000006e-06, "loss": 2.1342, "step": 9046 }, { "epoch": 2.836050156739812, "grad_norm": 6.254315376281738, "learning_rate": 4.5235e-06, "loss": 1.9072, "step": 9047 }, { "epoch": 2.8363636363636364, "grad_norm": 29.29647445678711, "learning_rate": 4.524e-06, "loss": 2.4178, "step": 9048 }, { "epoch": 2.836677115987461, "grad_norm": 23.642534255981445, "learning_rate": 4.5245000000000005e-06, "loss": 2.4983, "step": 9049 }, { "epoch": 2.8369905956112853, "grad_norm": 9.380431175231934, "learning_rate": 4.525000000000001e-06, "loss": 2.0672, "step": 9050 }, { "epoch": 2.8373040752351097, "grad_norm": 8.24057674407959, "learning_rate": 4.5255e-06, "loss": 2.1255, "step": 9051 }, { "epoch": 2.837617554858934, "grad_norm": 8.004820823669434, "learning_rate": 4.526e-06, "loss": 1.8642, "step": 9052 }, { "epoch": 2.8379310344827586, "grad_norm": 16.403759002685547, "learning_rate": 4.526500000000001e-06, "loss": 2.7953, "step": 9053 }, { "epoch": 2.838244514106583, "grad_norm": 8.52223014831543, "learning_rate": 4.527e-06, "loss": 2.2693, "step": 9054 }, { "epoch": 2.8385579937304075, "grad_norm": 20.790197372436523, "learning_rate": 4.5275e-06, "loss": 2.21, "step": 9055 }, { "epoch": 2.838871473354232, "grad_norm": 7.040374279022217, "learning_rate": 4.5280000000000005e-06, "loss": 2.2705, "step": 9056 }, { "epoch": 2.8391849529780564, "grad_norm": 7.638535499572754, "learning_rate": 4.5285e-06, "loss": 2.1726, "step": 9057 }, { "epoch": 2.839498432601881, "grad_norm": 12.026571273803711, "learning_rate": 4.529000000000001e-06, "loss": 2.0591, "step": 9058 }, { "epoch": 2.8398119122257053, "grad_norm": 13.010356903076172, "learning_rate": 4.5295000000000004e-06, "loss": 2.1869, "step": 9059 }, { "epoch": 2.8401253918495297, "grad_norm": 17.154539108276367, "learning_rate": 4.530000000000001e-06, "loss": 2.1676, "step": 9060 }, { "epoch": 2.840438871473354, "grad_norm": 30.21226692199707, "learning_rate": 4.5305e-06, "loss": 2.3421, "step": 9061 }, { "epoch": 2.8407523510971786, "grad_norm": 15.545875549316406, "learning_rate": 4.531e-06, "loss": 2.6347, "step": 9062 }, { "epoch": 2.841065830721003, "grad_norm": 83.15999603271484, "learning_rate": 4.5315000000000006e-06, "loss": 2.7607, "step": 9063 }, { "epoch": 2.8413793103448275, "grad_norm": 10.217707633972168, "learning_rate": 4.532e-06, "loss": 2.2033, "step": 9064 }, { "epoch": 2.841692789968652, "grad_norm": 7.062702178955078, "learning_rate": 4.5325e-06, "loss": 2.4543, "step": 9065 }, { "epoch": 2.8420062695924764, "grad_norm": 5.571096420288086, "learning_rate": 4.5330000000000005e-06, "loss": 2.1564, "step": 9066 }, { "epoch": 2.842319749216301, "grad_norm": 16.34004783630371, "learning_rate": 4.5335e-06, "loss": 2.3607, "step": 9067 }, { "epoch": 2.8426332288401253, "grad_norm": 8.674178123474121, "learning_rate": 4.534000000000001e-06, "loss": 2.1637, "step": 9068 }, { "epoch": 2.8429467084639497, "grad_norm": 12.969141006469727, "learning_rate": 4.5345e-06, "loss": 2.2545, "step": 9069 }, { "epoch": 2.843260188087774, "grad_norm": 9.04113483428955, "learning_rate": 4.535000000000001e-06, "loss": 1.9826, "step": 9070 }, { "epoch": 2.8435736677115986, "grad_norm": 11.02372932434082, "learning_rate": 4.535500000000001e-06, "loss": 2.1107, "step": 9071 }, { "epoch": 2.8438871473354235, "grad_norm": 12.608735084533691, "learning_rate": 4.536e-06, "loss": 2.268, "step": 9072 }, { "epoch": 2.844200626959248, "grad_norm": 15.756956100463867, "learning_rate": 4.5365000000000005e-06, "loss": 2.0319, "step": 9073 }, { "epoch": 2.8445141065830724, "grad_norm": 20.114643096923828, "learning_rate": 4.537e-06, "loss": 2.1377, "step": 9074 }, { "epoch": 2.844827586206897, "grad_norm": 19.924602508544922, "learning_rate": 4.5375e-06, "loss": 2.1289, "step": 9075 }, { "epoch": 2.8451410658307212, "grad_norm": 11.955313682556152, "learning_rate": 4.5380000000000004e-06, "loss": 2.183, "step": 9076 }, { "epoch": 2.8454545454545457, "grad_norm": 57.07658386230469, "learning_rate": 4.5385e-06, "loss": 2.1821, "step": 9077 }, { "epoch": 2.84576802507837, "grad_norm": 10.38676929473877, "learning_rate": 4.539000000000001e-06, "loss": 2.1546, "step": 9078 }, { "epoch": 2.8460815047021946, "grad_norm": 22.609840393066406, "learning_rate": 4.5395e-06, "loss": 2.5771, "step": 9079 }, { "epoch": 2.846394984326019, "grad_norm": 9.782651901245117, "learning_rate": 4.540000000000001e-06, "loss": 2.2867, "step": 9080 }, { "epoch": 2.8467084639498434, "grad_norm": 11.213095664978027, "learning_rate": 4.540500000000001e-06, "loss": 2.1748, "step": 9081 }, { "epoch": 2.847021943573668, "grad_norm": 8.247215270996094, "learning_rate": 4.541e-06, "loss": 2.2832, "step": 9082 }, { "epoch": 2.8473354231974923, "grad_norm": 7.0169548988342285, "learning_rate": 4.5415000000000005e-06, "loss": 2.2501, "step": 9083 }, { "epoch": 2.8476489028213168, "grad_norm": 10.02524471282959, "learning_rate": 4.542e-06, "loss": 2.0875, "step": 9084 }, { "epoch": 2.847962382445141, "grad_norm": 7.164865970611572, "learning_rate": 4.5425e-06, "loss": 2.1558, "step": 9085 }, { "epoch": 2.8482758620689657, "grad_norm": 6.424918174743652, "learning_rate": 4.543e-06, "loss": 2.1467, "step": 9086 }, { "epoch": 2.84858934169279, "grad_norm": 6.685555458068848, "learning_rate": 4.5435e-06, "loss": 2.0468, "step": 9087 }, { "epoch": 2.8489028213166145, "grad_norm": 32.3623161315918, "learning_rate": 4.544000000000001e-06, "loss": 1.8598, "step": 9088 }, { "epoch": 2.849216300940439, "grad_norm": 64.27349853515625, "learning_rate": 4.5445e-06, "loss": 2.2256, "step": 9089 }, { "epoch": 2.8495297805642634, "grad_norm": 50.32256317138672, "learning_rate": 4.5450000000000005e-06, "loss": 2.2095, "step": 9090 }, { "epoch": 2.849843260188088, "grad_norm": 13.999059677124023, "learning_rate": 4.545500000000001e-06, "loss": 2.5409, "step": 9091 }, { "epoch": 2.8501567398119123, "grad_norm": 12.45600700378418, "learning_rate": 4.546e-06, "loss": 2.1083, "step": 9092 }, { "epoch": 2.8504702194357368, "grad_norm": 10.817824363708496, "learning_rate": 4.5465000000000004e-06, "loss": 2.2242, "step": 9093 }, { "epoch": 2.850783699059561, "grad_norm": 106.28800201416016, "learning_rate": 4.547000000000001e-06, "loss": 2.4932, "step": 9094 }, { "epoch": 2.8510971786833856, "grad_norm": 7.137877941131592, "learning_rate": 4.5475e-06, "loss": 2.2756, "step": 9095 }, { "epoch": 2.85141065830721, "grad_norm": 15.857748985290527, "learning_rate": 4.548e-06, "loss": 1.9946, "step": 9096 }, { "epoch": 2.8517241379310345, "grad_norm": 74.98131561279297, "learning_rate": 4.5485e-06, "loss": 1.6695, "step": 9097 }, { "epoch": 2.852037617554859, "grad_norm": 13.657424926757812, "learning_rate": 4.549000000000001e-06, "loss": 2.2767, "step": 9098 }, { "epoch": 2.8523510971786834, "grad_norm": 12.505475997924805, "learning_rate": 4.5495e-06, "loss": 2.0142, "step": 9099 }, { "epoch": 2.852664576802508, "grad_norm": 24.43087387084961, "learning_rate": 4.5500000000000005e-06, "loss": 2.5818, "step": 9100 }, { "epoch": 2.8529780564263323, "grad_norm": 8.38932991027832, "learning_rate": 4.550500000000001e-06, "loss": 2.1954, "step": 9101 }, { "epoch": 2.8532915360501567, "grad_norm": 9.744585037231445, "learning_rate": 4.551e-06, "loss": 2.213, "step": 9102 }, { "epoch": 2.853605015673981, "grad_norm": 10.532763481140137, "learning_rate": 4.5515e-06, "loss": 1.9457, "step": 9103 }, { "epoch": 2.8539184952978056, "grad_norm": 9.644865989685059, "learning_rate": 4.552000000000001e-06, "loss": 2.3883, "step": 9104 }, { "epoch": 2.85423197492163, "grad_norm": 52.522911071777344, "learning_rate": 4.5525e-06, "loss": 2.232, "step": 9105 }, { "epoch": 2.8545454545454545, "grad_norm": 11.825468063354492, "learning_rate": 4.553e-06, "loss": 2.3362, "step": 9106 }, { "epoch": 2.854858934169279, "grad_norm": 18.06389808654785, "learning_rate": 4.5535000000000005e-06, "loss": 2.5161, "step": 9107 }, { "epoch": 2.8551724137931034, "grad_norm": 102.82127380371094, "learning_rate": 4.554000000000001e-06, "loss": 2.3949, "step": 9108 }, { "epoch": 2.855485893416928, "grad_norm": 6.679622173309326, "learning_rate": 4.5545e-06, "loss": 1.858, "step": 9109 }, { "epoch": 2.8557993730407523, "grad_norm": 31.4483585357666, "learning_rate": 4.5550000000000004e-06, "loss": 2.4502, "step": 9110 }, { "epoch": 2.8561128526645767, "grad_norm": 11.145098686218262, "learning_rate": 4.555500000000001e-06, "loss": 1.8126, "step": 9111 }, { "epoch": 2.856426332288401, "grad_norm": 11.879880905151367, "learning_rate": 4.556e-06, "loss": 2.2943, "step": 9112 }, { "epoch": 2.8567398119122256, "grad_norm": 103.81322479248047, "learning_rate": 4.5565e-06, "loss": 1.9622, "step": 9113 }, { "epoch": 2.85705329153605, "grad_norm": 14.415818214416504, "learning_rate": 4.557000000000001e-06, "loss": 1.9572, "step": 9114 }, { "epoch": 2.8573667711598745, "grad_norm": 14.442065238952637, "learning_rate": 4.5575e-06, "loss": 2.1071, "step": 9115 }, { "epoch": 2.857680250783699, "grad_norm": 5.627063751220703, "learning_rate": 4.558e-06, "loss": 2.2578, "step": 9116 }, { "epoch": 2.8579937304075234, "grad_norm": 10.174920082092285, "learning_rate": 4.5585000000000005e-06, "loss": 1.8368, "step": 9117 }, { "epoch": 2.858307210031348, "grad_norm": 8.824463844299316, "learning_rate": 4.559000000000001e-06, "loss": 1.9739, "step": 9118 }, { "epoch": 2.8586206896551722, "grad_norm": 8.743282318115234, "learning_rate": 4.559500000000001e-06, "loss": 2.0754, "step": 9119 }, { "epoch": 2.8589341692789967, "grad_norm": 9.912028312683105, "learning_rate": 4.56e-06, "loss": 1.9982, "step": 9120 }, { "epoch": 2.859247648902821, "grad_norm": 10.031474113464355, "learning_rate": 4.560500000000001e-06, "loss": 2.0205, "step": 9121 }, { "epoch": 2.8595611285266456, "grad_norm": 14.197060585021973, "learning_rate": 4.561e-06, "loss": 2.7967, "step": 9122 }, { "epoch": 2.85987460815047, "grad_norm": 41.413272857666016, "learning_rate": 4.5615e-06, "loss": 2.1999, "step": 9123 }, { "epoch": 2.8601880877742945, "grad_norm": 11.720254898071289, "learning_rate": 4.5620000000000005e-06, "loss": 2.2363, "step": 9124 }, { "epoch": 2.860501567398119, "grad_norm": 6.382421016693115, "learning_rate": 4.5625e-06, "loss": 2.1473, "step": 9125 }, { "epoch": 2.8608150470219433, "grad_norm": 9.175348281860352, "learning_rate": 4.563e-06, "loss": 2.2581, "step": 9126 }, { "epoch": 2.8611285266457678, "grad_norm": 13.904825210571289, "learning_rate": 4.5635000000000004e-06, "loss": 2.0562, "step": 9127 }, { "epoch": 2.861442006269592, "grad_norm": 12.167485237121582, "learning_rate": 4.564e-06, "loss": 2.1633, "step": 9128 }, { "epoch": 2.8617554858934167, "grad_norm": 14.469586372375488, "learning_rate": 4.564500000000001e-06, "loss": 2.1778, "step": 9129 }, { "epoch": 2.862068965517241, "grad_norm": 27.839563369750977, "learning_rate": 4.565e-06, "loss": 2.1651, "step": 9130 }, { "epoch": 2.862382445141066, "grad_norm": 13.52730655670166, "learning_rate": 4.565500000000001e-06, "loss": 1.9427, "step": 9131 }, { "epoch": 2.8626959247648904, "grad_norm": 12.060941696166992, "learning_rate": 4.566000000000001e-06, "loss": 2.2066, "step": 9132 }, { "epoch": 2.863009404388715, "grad_norm": 7.97295618057251, "learning_rate": 4.5665e-06, "loss": 1.9775, "step": 9133 }, { "epoch": 2.8633228840125393, "grad_norm": 27.993362426757812, "learning_rate": 4.5670000000000005e-06, "loss": 2.4793, "step": 9134 }, { "epoch": 2.8636363636363638, "grad_norm": 26.5953426361084, "learning_rate": 4.5675e-06, "loss": 2.2419, "step": 9135 }, { "epoch": 2.863949843260188, "grad_norm": 27.93085289001465, "learning_rate": 4.568e-06, "loss": 2.0404, "step": 9136 }, { "epoch": 2.8642633228840126, "grad_norm": 8.528057098388672, "learning_rate": 4.5685e-06, "loss": 2.1696, "step": 9137 }, { "epoch": 2.864576802507837, "grad_norm": 10.730964660644531, "learning_rate": 4.569e-06, "loss": 2.1756, "step": 9138 }, { "epoch": 2.8648902821316615, "grad_norm": 15.276510238647461, "learning_rate": 4.569500000000001e-06, "loss": 3.3717, "step": 9139 }, { "epoch": 2.865203761755486, "grad_norm": 11.65485668182373, "learning_rate": 4.57e-06, "loss": 2.3657, "step": 9140 }, { "epoch": 2.8655172413793104, "grad_norm": 9.183744430541992, "learning_rate": 4.5705000000000005e-06, "loss": 2.3365, "step": 9141 }, { "epoch": 2.865830721003135, "grad_norm": 24.150142669677734, "learning_rate": 4.571000000000001e-06, "loss": 2.3236, "step": 9142 }, { "epoch": 2.8661442006269593, "grad_norm": 66.39964294433594, "learning_rate": 4.5715e-06, "loss": 2.4438, "step": 9143 }, { "epoch": 2.8664576802507837, "grad_norm": 12.52354621887207, "learning_rate": 4.5720000000000004e-06, "loss": 2.2167, "step": 9144 }, { "epoch": 2.866771159874608, "grad_norm": 11.77927303314209, "learning_rate": 4.572500000000001e-06, "loss": 1.935, "step": 9145 }, { "epoch": 2.8670846394984326, "grad_norm": 12.650484085083008, "learning_rate": 4.573e-06, "loss": 2.2191, "step": 9146 }, { "epoch": 2.867398119122257, "grad_norm": 49.40640640258789, "learning_rate": 4.5735e-06, "loss": 2.6976, "step": 9147 }, { "epoch": 2.8677115987460815, "grad_norm": 7.6341753005981445, "learning_rate": 4.574e-06, "loss": 2.0367, "step": 9148 }, { "epoch": 2.868025078369906, "grad_norm": 10.144675254821777, "learning_rate": 4.574500000000001e-06, "loss": 2.1442, "step": 9149 }, { "epoch": 2.8683385579937304, "grad_norm": 10.940845489501953, "learning_rate": 4.575e-06, "loss": 2.1009, "step": 9150 }, { "epoch": 2.868652037617555, "grad_norm": 27.509716033935547, "learning_rate": 4.5755000000000005e-06, "loss": 2.33, "step": 9151 }, { "epoch": 2.8689655172413793, "grad_norm": 7.187202453613281, "learning_rate": 4.576000000000001e-06, "loss": 2.0537, "step": 9152 }, { "epoch": 2.8692789968652037, "grad_norm": 10.156408309936523, "learning_rate": 4.5765e-06, "loss": 2.1267, "step": 9153 }, { "epoch": 2.869592476489028, "grad_norm": 9.965381622314453, "learning_rate": 4.577e-06, "loss": 1.8071, "step": 9154 }, { "epoch": 2.8699059561128526, "grad_norm": 8.296344757080078, "learning_rate": 4.577500000000001e-06, "loss": 2.2944, "step": 9155 }, { "epoch": 2.870219435736677, "grad_norm": 119.81610870361328, "learning_rate": 4.578e-06, "loss": 2.2659, "step": 9156 }, { "epoch": 2.8705329153605015, "grad_norm": 43.86922836303711, "learning_rate": 4.5785e-06, "loss": 2.5447, "step": 9157 }, { "epoch": 2.870846394984326, "grad_norm": 7.4885454177856445, "learning_rate": 4.579e-06, "loss": 2.5539, "step": 9158 }, { "epoch": 2.8711598746081504, "grad_norm": 5.773125171661377, "learning_rate": 4.579500000000001e-06, "loss": 1.9726, "step": 9159 }, { "epoch": 2.871473354231975, "grad_norm": 8.475078582763672, "learning_rate": 4.58e-06, "loss": 2.1332, "step": 9160 }, { "epoch": 2.8717868338557992, "grad_norm": 15.827296257019043, "learning_rate": 4.5805000000000004e-06, "loss": 1.8159, "step": 9161 }, { "epoch": 2.8721003134796237, "grad_norm": 16.365901947021484, "learning_rate": 4.581000000000001e-06, "loss": 1.8841, "step": 9162 }, { "epoch": 2.872413793103448, "grad_norm": 20.792438507080078, "learning_rate": 4.5815e-06, "loss": 2.1684, "step": 9163 }, { "epoch": 2.8727272727272726, "grad_norm": 6.1416473388671875, "learning_rate": 4.582e-06, "loss": 2.3206, "step": 9164 }, { "epoch": 2.873040752351097, "grad_norm": 36.0546989440918, "learning_rate": 4.582500000000001e-06, "loss": 2.0461, "step": 9165 }, { "epoch": 2.873354231974922, "grad_norm": 11.975510597229004, "learning_rate": 4.583e-06, "loss": 2.682, "step": 9166 }, { "epoch": 2.8736677115987463, "grad_norm": 12.212711334228516, "learning_rate": 4.5835e-06, "loss": 2.118, "step": 9167 }, { "epoch": 2.873981191222571, "grad_norm": 11.151400566101074, "learning_rate": 4.5840000000000005e-06, "loss": 2.1423, "step": 9168 }, { "epoch": 2.8742946708463952, "grad_norm": 15.207825660705566, "learning_rate": 4.584500000000001e-06, "loss": 2.2028, "step": 9169 }, { "epoch": 2.8746081504702197, "grad_norm": 7.056772232055664, "learning_rate": 4.585e-06, "loss": 2.3567, "step": 9170 }, { "epoch": 2.874921630094044, "grad_norm": 48.902801513671875, "learning_rate": 4.5855e-06, "loss": 1.9994, "step": 9171 }, { "epoch": 2.8752351097178686, "grad_norm": 14.9516019821167, "learning_rate": 4.586000000000001e-06, "loss": 2.3134, "step": 9172 }, { "epoch": 2.875548589341693, "grad_norm": 6.07791805267334, "learning_rate": 4.5865e-06, "loss": 2.0215, "step": 9173 }, { "epoch": 2.8758620689655174, "grad_norm": 6.6773858070373535, "learning_rate": 4.587e-06, "loss": 2.0969, "step": 9174 }, { "epoch": 2.876175548589342, "grad_norm": 9.941762924194336, "learning_rate": 4.5875000000000005e-06, "loss": 2.2075, "step": 9175 }, { "epoch": 2.8764890282131663, "grad_norm": 8.060948371887207, "learning_rate": 4.588e-06, "loss": 2.3522, "step": 9176 }, { "epoch": 2.8768025078369908, "grad_norm": 9.07833194732666, "learning_rate": 4.5885e-06, "loss": 2.2517, "step": 9177 }, { "epoch": 2.877115987460815, "grad_norm": 37.12391662597656, "learning_rate": 4.5890000000000004e-06, "loss": 2.9356, "step": 9178 }, { "epoch": 2.8774294670846396, "grad_norm": 14.64284610748291, "learning_rate": 4.589500000000001e-06, "loss": 2.1122, "step": 9179 }, { "epoch": 2.877742946708464, "grad_norm": 5.455367088317871, "learning_rate": 4.590000000000001e-06, "loss": 1.9373, "step": 9180 }, { "epoch": 2.8780564263322885, "grad_norm": 9.653911590576172, "learning_rate": 4.5905e-06, "loss": 1.9401, "step": 9181 }, { "epoch": 2.878369905956113, "grad_norm": 7.2161784172058105, "learning_rate": 4.591000000000001e-06, "loss": 2.0775, "step": 9182 }, { "epoch": 2.8786833855799374, "grad_norm": 9.823296546936035, "learning_rate": 4.5915e-06, "loss": 2.098, "step": 9183 }, { "epoch": 2.878996865203762, "grad_norm": 6.756521224975586, "learning_rate": 4.592e-06, "loss": 2.0816, "step": 9184 }, { "epoch": 2.8793103448275863, "grad_norm": 5.684502601623535, "learning_rate": 4.5925000000000005e-06, "loss": 2.2074, "step": 9185 }, { "epoch": 2.8796238244514107, "grad_norm": 10.061199188232422, "learning_rate": 4.593e-06, "loss": 2.12, "step": 9186 }, { "epoch": 2.879937304075235, "grad_norm": 9.058341979980469, "learning_rate": 4.5935e-06, "loss": 2.0825, "step": 9187 }, { "epoch": 2.8802507836990596, "grad_norm": 8.205059051513672, "learning_rate": 4.594e-06, "loss": 2.1764, "step": 9188 }, { "epoch": 2.880564263322884, "grad_norm": 11.3743257522583, "learning_rate": 4.594500000000001e-06, "loss": 2.2273, "step": 9189 }, { "epoch": 2.8808777429467085, "grad_norm": 9.229462623596191, "learning_rate": 4.595000000000001e-06, "loss": 2.1912, "step": 9190 }, { "epoch": 2.881191222570533, "grad_norm": 7.820175647735596, "learning_rate": 4.5955e-06, "loss": 2.2154, "step": 9191 }, { "epoch": 2.8815047021943574, "grad_norm": 27.639631271362305, "learning_rate": 4.5960000000000006e-06, "loss": 2.3217, "step": 9192 }, { "epoch": 2.881818181818182, "grad_norm": 15.355805397033691, "learning_rate": 4.596500000000001e-06, "loss": 2.072, "step": 9193 }, { "epoch": 2.8821316614420063, "grad_norm": 66.39558410644531, "learning_rate": 4.597e-06, "loss": 1.9694, "step": 9194 }, { "epoch": 2.8824451410658307, "grad_norm": 13.9536771774292, "learning_rate": 4.5975000000000005e-06, "loss": 1.9031, "step": 9195 }, { "epoch": 2.882758620689655, "grad_norm": 17.938701629638672, "learning_rate": 4.598e-06, "loss": 2.6071, "step": 9196 }, { "epoch": 2.8830721003134796, "grad_norm": 6.168633460998535, "learning_rate": 4.5985e-06, "loss": 1.9514, "step": 9197 }, { "epoch": 2.883385579937304, "grad_norm": 6.665476322174072, "learning_rate": 4.599e-06, "loss": 2.1223, "step": 9198 }, { "epoch": 2.8836990595611285, "grad_norm": 65.79302978515625, "learning_rate": 4.599500000000001e-06, "loss": 2.4955, "step": 9199 }, { "epoch": 2.884012539184953, "grad_norm": 9.85312271118164, "learning_rate": 4.600000000000001e-06, "loss": 2.0559, "step": 9200 }, { "epoch": 2.8843260188087774, "grad_norm": 10.498476028442383, "learning_rate": 4.6005e-06, "loss": 2.3918, "step": 9201 }, { "epoch": 2.884639498432602, "grad_norm": 12.171208381652832, "learning_rate": 4.6010000000000005e-06, "loss": 1.7917, "step": 9202 }, { "epoch": 2.8849529780564263, "grad_norm": 9.090826988220215, "learning_rate": 4.601500000000001e-06, "loss": 2.0333, "step": 9203 }, { "epoch": 2.8852664576802507, "grad_norm": 22.084802627563477, "learning_rate": 4.602e-06, "loss": 2.1144, "step": 9204 }, { "epoch": 2.885579937304075, "grad_norm": 16.068744659423828, "learning_rate": 4.6025e-06, "loss": 2.1351, "step": 9205 }, { "epoch": 2.8858934169278996, "grad_norm": 20.864887237548828, "learning_rate": 4.603000000000001e-06, "loss": 2.1774, "step": 9206 }, { "epoch": 2.886206896551724, "grad_norm": 8.459478378295898, "learning_rate": 4.6035e-06, "loss": 2.1552, "step": 9207 }, { "epoch": 2.8865203761755485, "grad_norm": 6.956141948699951, "learning_rate": 4.604e-06, "loss": 2.0743, "step": 9208 }, { "epoch": 2.886833855799373, "grad_norm": 107.7117691040039, "learning_rate": 4.6045000000000006e-06, "loss": 1.7796, "step": 9209 }, { "epoch": 2.8871473354231973, "grad_norm": 8.948983192443848, "learning_rate": 4.605000000000001e-06, "loss": 2.2547, "step": 9210 }, { "epoch": 2.887460815047022, "grad_norm": 5.027045726776123, "learning_rate": 4.6055e-06, "loss": 2.1068, "step": 9211 }, { "epoch": 2.8877742946708462, "grad_norm": 8.643916130065918, "learning_rate": 4.6060000000000005e-06, "loss": 2.2155, "step": 9212 }, { "epoch": 2.8880877742946707, "grad_norm": 11.14082145690918, "learning_rate": 4.606500000000001e-06, "loss": 2.4367, "step": 9213 }, { "epoch": 2.888401253918495, "grad_norm": 52.892059326171875, "learning_rate": 4.607e-06, "loss": 2.6772, "step": 9214 }, { "epoch": 2.8887147335423196, "grad_norm": 92.553466796875, "learning_rate": 4.6075e-06, "loss": 2.8352, "step": 9215 }, { "epoch": 2.889028213166144, "grad_norm": 16.943626403808594, "learning_rate": 4.608000000000001e-06, "loss": 2.0578, "step": 9216 }, { "epoch": 2.8893416927899684, "grad_norm": 11.302656173706055, "learning_rate": 4.6085e-06, "loss": 2.0451, "step": 9217 }, { "epoch": 2.889655172413793, "grad_norm": 13.117034912109375, "learning_rate": 4.609e-06, "loss": 2.1073, "step": 9218 }, { "epoch": 2.8899686520376173, "grad_norm": 10.4174222946167, "learning_rate": 4.6095000000000005e-06, "loss": 2.4406, "step": 9219 }, { "epoch": 2.8902821316614418, "grad_norm": 7.523171424865723, "learning_rate": 4.610000000000001e-06, "loss": 2.0209, "step": 9220 }, { "epoch": 2.890595611285266, "grad_norm": 166.037353515625, "learning_rate": 4.6105e-06, "loss": 2.5283, "step": 9221 }, { "epoch": 2.8909090909090907, "grad_norm": 7.849234580993652, "learning_rate": 4.611e-06, "loss": 2.5436, "step": 9222 }, { "epoch": 2.891222570532915, "grad_norm": 12.063940048217773, "learning_rate": 4.611500000000001e-06, "loss": 2.1211, "step": 9223 }, { "epoch": 2.8915360501567395, "grad_norm": 47.45783615112305, "learning_rate": 4.612e-06, "loss": 2.0851, "step": 9224 }, { "epoch": 2.8918495297805644, "grad_norm": 9.472049713134766, "learning_rate": 4.6125e-06, "loss": 2.1665, "step": 9225 }, { "epoch": 2.892163009404389, "grad_norm": 5.293562412261963, "learning_rate": 4.6130000000000006e-06, "loss": 2.2031, "step": 9226 }, { "epoch": 2.8924764890282133, "grad_norm": 17.39164924621582, "learning_rate": 4.6135e-06, "loss": 2.4076, "step": 9227 }, { "epoch": 2.8927899686520377, "grad_norm": 6.936942100524902, "learning_rate": 4.614e-06, "loss": 2.1009, "step": 9228 }, { "epoch": 2.893103448275862, "grad_norm": 20.060213088989258, "learning_rate": 4.6145000000000005e-06, "loss": 2.1933, "step": 9229 }, { "epoch": 2.8934169278996866, "grad_norm": 9.701957702636719, "learning_rate": 4.615000000000001e-06, "loss": 2.0589, "step": 9230 }, { "epoch": 2.893730407523511, "grad_norm": 9.421142578125, "learning_rate": 4.615500000000001e-06, "loss": 2.1063, "step": 9231 }, { "epoch": 2.8940438871473355, "grad_norm": 13.232415199279785, "learning_rate": 4.616e-06, "loss": 1.9591, "step": 9232 }, { "epoch": 2.89435736677116, "grad_norm": 6.596160888671875, "learning_rate": 4.616500000000001e-06, "loss": 2.3227, "step": 9233 }, { "epoch": 2.8946708463949844, "grad_norm": 7.292934894561768, "learning_rate": 4.617e-06, "loss": 2.0696, "step": 9234 }, { "epoch": 2.894984326018809, "grad_norm": 9.982529640197754, "learning_rate": 4.6175e-06, "loss": 2.1774, "step": 9235 }, { "epoch": 2.8952978056426333, "grad_norm": 5.620611667633057, "learning_rate": 4.6180000000000005e-06, "loss": 1.9459, "step": 9236 }, { "epoch": 2.8956112852664577, "grad_norm": 8.985682487487793, "learning_rate": 4.6185e-06, "loss": 2.1289, "step": 9237 }, { "epoch": 2.895924764890282, "grad_norm": 34.218360900878906, "learning_rate": 4.619e-06, "loss": 2.3738, "step": 9238 }, { "epoch": 2.8962382445141066, "grad_norm": 13.593003273010254, "learning_rate": 4.6195e-06, "loss": 1.8685, "step": 9239 }, { "epoch": 2.896551724137931, "grad_norm": 8.38582992553711, "learning_rate": 4.620000000000001e-06, "loss": 1.7873, "step": 9240 }, { "epoch": 2.8968652037617555, "grad_norm": 7.708801746368408, "learning_rate": 4.620500000000001e-06, "loss": 1.8508, "step": 9241 }, { "epoch": 2.89717868338558, "grad_norm": 13.71678638458252, "learning_rate": 4.621e-06, "loss": 2.3705, "step": 9242 }, { "epoch": 2.8974921630094044, "grad_norm": 12.179491996765137, "learning_rate": 4.6215000000000006e-06, "loss": 2.15, "step": 9243 }, { "epoch": 2.897805642633229, "grad_norm": 10.628372192382812, "learning_rate": 4.622e-06, "loss": 2.1835, "step": 9244 }, { "epoch": 2.8981191222570533, "grad_norm": 12.030497550964355, "learning_rate": 4.6225e-06, "loss": 2.1982, "step": 9245 }, { "epoch": 2.8984326018808777, "grad_norm": 9.762775421142578, "learning_rate": 4.6230000000000005e-06, "loss": 1.9228, "step": 9246 }, { "epoch": 2.898746081504702, "grad_norm": 11.008511543273926, "learning_rate": 4.6235e-06, "loss": 2.4114, "step": 9247 }, { "epoch": 2.8990595611285266, "grad_norm": 13.40495777130127, "learning_rate": 4.624e-06, "loss": 2.2685, "step": 9248 }, { "epoch": 2.899373040752351, "grad_norm": 20.526884078979492, "learning_rate": 4.6245e-06, "loss": 3.6866, "step": 9249 }, { "epoch": 2.8996865203761755, "grad_norm": 8.61413860321045, "learning_rate": 4.625000000000001e-06, "loss": 1.8628, "step": 9250 }, { "epoch": 2.9, "grad_norm": 7.046261310577393, "learning_rate": 4.625500000000001e-06, "loss": 1.974, "step": 9251 }, { "epoch": 2.9003134796238244, "grad_norm": 9.90766429901123, "learning_rate": 4.626e-06, "loss": 2.0509, "step": 9252 }, { "epoch": 2.900626959247649, "grad_norm": 10.7845458984375, "learning_rate": 4.6265000000000005e-06, "loss": 2.113, "step": 9253 }, { "epoch": 2.9009404388714732, "grad_norm": 67.67601776123047, "learning_rate": 4.627000000000001e-06, "loss": 1.9962, "step": 9254 }, { "epoch": 2.9012539184952977, "grad_norm": 11.186670303344727, "learning_rate": 4.6275e-06, "loss": 2.0223, "step": 9255 }, { "epoch": 2.901567398119122, "grad_norm": 6.537659645080566, "learning_rate": 4.628e-06, "loss": 1.9912, "step": 9256 }, { "epoch": 2.9018808777429466, "grad_norm": 9.417458534240723, "learning_rate": 4.6285e-06, "loss": 2.1272, "step": 9257 }, { "epoch": 2.902194357366771, "grad_norm": 14.114328384399414, "learning_rate": 4.629e-06, "loss": 2.3396, "step": 9258 }, { "epoch": 2.9025078369905954, "grad_norm": 140.81790161132812, "learning_rate": 4.6295e-06, "loss": 2.4372, "step": 9259 }, { "epoch": 2.9028213166144203, "grad_norm": 13.968887329101562, "learning_rate": 4.6300000000000006e-06, "loss": 2.0617, "step": 9260 }, { "epoch": 2.9031347962382448, "grad_norm": 12.569509506225586, "learning_rate": 4.630500000000001e-06, "loss": 2.2661, "step": 9261 }, { "epoch": 2.903448275862069, "grad_norm": 5.047782897949219, "learning_rate": 4.631e-06, "loss": 2.1446, "step": 9262 }, { "epoch": 2.9037617554858937, "grad_norm": 155.54925537109375, "learning_rate": 4.6315000000000005e-06, "loss": 2.3908, "step": 9263 }, { "epoch": 2.904075235109718, "grad_norm": 13.980104446411133, "learning_rate": 4.632000000000001e-06, "loss": 2.273, "step": 9264 }, { "epoch": 2.9043887147335425, "grad_norm": 9.930087089538574, "learning_rate": 4.6325e-06, "loss": 2.1328, "step": 9265 }, { "epoch": 2.904702194357367, "grad_norm": 9.19558334350586, "learning_rate": 4.633e-06, "loss": 2.0437, "step": 9266 }, { "epoch": 2.9050156739811914, "grad_norm": 27.977828979492188, "learning_rate": 4.633500000000001e-06, "loss": 2.0686, "step": 9267 }, { "epoch": 2.905329153605016, "grad_norm": 12.206077575683594, "learning_rate": 4.634e-06, "loss": 2.2973, "step": 9268 }, { "epoch": 2.9056426332288403, "grad_norm": 49.46797180175781, "learning_rate": 4.6345e-06, "loss": 1.9769, "step": 9269 }, { "epoch": 2.9059561128526648, "grad_norm": 13.761004447937012, "learning_rate": 4.6350000000000005e-06, "loss": 1.9131, "step": 9270 }, { "epoch": 2.906269592476489, "grad_norm": 22.17747688293457, "learning_rate": 4.635500000000001e-06, "loss": 2.0663, "step": 9271 }, { "epoch": 2.9065830721003136, "grad_norm": 10.937813758850098, "learning_rate": 4.636e-06, "loss": 2.0508, "step": 9272 }, { "epoch": 2.906896551724138, "grad_norm": 9.165519714355469, "learning_rate": 4.6365e-06, "loss": 2.3587, "step": 9273 }, { "epoch": 2.9072100313479625, "grad_norm": 17.458711624145508, "learning_rate": 4.637000000000001e-06, "loss": 2.2505, "step": 9274 }, { "epoch": 2.907523510971787, "grad_norm": 6.0234856605529785, "learning_rate": 4.6375e-06, "loss": 2.2864, "step": 9275 }, { "epoch": 2.9078369905956114, "grad_norm": 10.620022773742676, "learning_rate": 4.638e-06, "loss": 2.3136, "step": 9276 }, { "epoch": 2.908150470219436, "grad_norm": 6.939550399780273, "learning_rate": 4.6385000000000006e-06, "loss": 1.9694, "step": 9277 }, { "epoch": 2.9084639498432603, "grad_norm": 9.526083946228027, "learning_rate": 4.639e-06, "loss": 1.9902, "step": 9278 }, { "epoch": 2.9087774294670847, "grad_norm": 6.371539115905762, "learning_rate": 4.6395e-06, "loss": 2.0567, "step": 9279 }, { "epoch": 2.909090909090909, "grad_norm": 6.703181743621826, "learning_rate": 4.6400000000000005e-06, "loss": 1.89, "step": 9280 }, { "epoch": 2.9094043887147336, "grad_norm": 13.142341613769531, "learning_rate": 4.640500000000001e-06, "loss": 2.4348, "step": 9281 }, { "epoch": 2.909717868338558, "grad_norm": 12.439029693603516, "learning_rate": 4.641e-06, "loss": 2.1945, "step": 9282 }, { "epoch": 2.9100313479623825, "grad_norm": 59.09040069580078, "learning_rate": 4.6415e-06, "loss": 2.1321, "step": 9283 }, { "epoch": 2.910344827586207, "grad_norm": 11.633599281311035, "learning_rate": 4.642000000000001e-06, "loss": 1.9187, "step": 9284 }, { "epoch": 2.9106583072100314, "grad_norm": 6.521206378936768, "learning_rate": 4.6425e-06, "loss": 2.0374, "step": 9285 }, { "epoch": 2.910971786833856, "grad_norm": 9.743823051452637, "learning_rate": 4.643e-06, "loss": 1.9645, "step": 9286 }, { "epoch": 2.9112852664576803, "grad_norm": 10.241686820983887, "learning_rate": 4.6435000000000005e-06, "loss": 2.3193, "step": 9287 }, { "epoch": 2.9115987460815047, "grad_norm": 7.893749237060547, "learning_rate": 4.644e-06, "loss": 1.9509, "step": 9288 }, { "epoch": 2.911912225705329, "grad_norm": 10.710136413574219, "learning_rate": 4.6445e-06, "loss": 2.1633, "step": 9289 }, { "epoch": 2.9122257053291536, "grad_norm": 7.827268600463867, "learning_rate": 4.645e-06, "loss": 2.1655, "step": 9290 }, { "epoch": 2.912539184952978, "grad_norm": 6.5134992599487305, "learning_rate": 4.645500000000001e-06, "loss": 2.274, "step": 9291 }, { "epoch": 2.9128526645768025, "grad_norm": 21.032333374023438, "learning_rate": 4.646000000000001e-06, "loss": 2.3859, "step": 9292 }, { "epoch": 2.913166144200627, "grad_norm": 26.863637924194336, "learning_rate": 4.6465e-06, "loss": 2.9791, "step": 9293 }, { "epoch": 2.9134796238244514, "grad_norm": 48.64115524291992, "learning_rate": 4.6470000000000006e-06, "loss": 2.3378, "step": 9294 }, { "epoch": 2.913793103448276, "grad_norm": 85.30345153808594, "learning_rate": 4.6475e-06, "loss": 2.7848, "step": 9295 }, { "epoch": 2.9141065830721002, "grad_norm": 18.767473220825195, "learning_rate": 4.648e-06, "loss": 2.1301, "step": 9296 }, { "epoch": 2.9144200626959247, "grad_norm": 12.547724723815918, "learning_rate": 4.6485000000000005e-06, "loss": 2.0649, "step": 9297 }, { "epoch": 2.914733542319749, "grad_norm": 6.537840366363525, "learning_rate": 4.649e-06, "loss": 1.8543, "step": 9298 }, { "epoch": 2.9150470219435736, "grad_norm": 5.3381123542785645, "learning_rate": 4.6495e-06, "loss": 2.2279, "step": 9299 }, { "epoch": 2.915360501567398, "grad_norm": 8.2509765625, "learning_rate": 4.65e-06, "loss": 1.9733, "step": 9300 }, { "epoch": 2.9156739811912225, "grad_norm": 9.803706169128418, "learning_rate": 4.650500000000001e-06, "loss": 2.1291, "step": 9301 }, { "epoch": 2.915987460815047, "grad_norm": 10.162749290466309, "learning_rate": 4.651000000000001e-06, "loss": 2.2336, "step": 9302 }, { "epoch": 2.9163009404388713, "grad_norm": 12.483780860900879, "learning_rate": 4.6515e-06, "loss": 2.1877, "step": 9303 }, { "epoch": 2.916614420062696, "grad_norm": 9.938650131225586, "learning_rate": 4.6520000000000005e-06, "loss": 2.277, "step": 9304 }, { "epoch": 2.91692789968652, "grad_norm": 11.724613189697266, "learning_rate": 4.652500000000001e-06, "loss": 2.2447, "step": 9305 }, { "epoch": 2.9172413793103447, "grad_norm": 14.443130493164062, "learning_rate": 4.653e-06, "loss": 2.0948, "step": 9306 }, { "epoch": 2.917554858934169, "grad_norm": 41.508201599121094, "learning_rate": 4.6535e-06, "loss": 2.2625, "step": 9307 }, { "epoch": 2.9178683385579935, "grad_norm": 7.287661075592041, "learning_rate": 4.654e-06, "loss": 2.2982, "step": 9308 }, { "epoch": 2.918181818181818, "grad_norm": 12.363125801086426, "learning_rate": 4.6545e-06, "loss": 2.2304, "step": 9309 }, { "epoch": 2.9184952978056424, "grad_norm": 5.786609172821045, "learning_rate": 4.655e-06, "loss": 2.1634, "step": 9310 }, { "epoch": 2.918808777429467, "grad_norm": 17.33257484436035, "learning_rate": 4.6555000000000006e-06, "loss": 2.35, "step": 9311 }, { "epoch": 2.9191222570532913, "grad_norm": 34.08716583251953, "learning_rate": 4.656000000000001e-06, "loss": 2.622, "step": 9312 }, { "epoch": 2.9194357366771158, "grad_norm": 6.9964919090271, "learning_rate": 4.6565e-06, "loss": 1.9741, "step": 9313 }, { "epoch": 2.91974921630094, "grad_norm": 57.03401565551758, "learning_rate": 4.6570000000000005e-06, "loss": 2.1685, "step": 9314 }, { "epoch": 2.9200626959247646, "grad_norm": 52.216156005859375, "learning_rate": 4.657500000000001e-06, "loss": 2.1133, "step": 9315 }, { "epoch": 2.920376175548589, "grad_norm": 9.194500923156738, "learning_rate": 4.658e-06, "loss": 2.1174, "step": 9316 }, { "epoch": 2.9206896551724135, "grad_norm": 10.391203880310059, "learning_rate": 4.6585e-06, "loss": 1.9255, "step": 9317 }, { "epoch": 2.9210031347962384, "grad_norm": 17.786378860473633, "learning_rate": 4.659e-06, "loss": 2.1736, "step": 9318 }, { "epoch": 2.921316614420063, "grad_norm": 9.851551055908203, "learning_rate": 4.6595e-06, "loss": 2.1299, "step": 9319 }, { "epoch": 2.9216300940438873, "grad_norm": 9.640374183654785, "learning_rate": 4.66e-06, "loss": 2.0928, "step": 9320 }, { "epoch": 2.9219435736677117, "grad_norm": 8.944016456604004, "learning_rate": 4.6605000000000005e-06, "loss": 1.8544, "step": 9321 }, { "epoch": 2.922257053291536, "grad_norm": 11.810482025146484, "learning_rate": 4.661000000000001e-06, "loss": 2.1252, "step": 9322 }, { "epoch": 2.9225705329153606, "grad_norm": 8.933077812194824, "learning_rate": 4.6615e-06, "loss": 2.0464, "step": 9323 }, { "epoch": 2.922884012539185, "grad_norm": 13.77690601348877, "learning_rate": 4.6620000000000004e-06, "loss": 2.1203, "step": 9324 }, { "epoch": 2.9231974921630095, "grad_norm": 16.029823303222656, "learning_rate": 4.662500000000001e-06, "loss": 1.998, "step": 9325 }, { "epoch": 2.923510971786834, "grad_norm": 11.494406700134277, "learning_rate": 4.663e-06, "loss": 2.4935, "step": 9326 }, { "epoch": 2.9238244514106584, "grad_norm": 6.371428966522217, "learning_rate": 4.6635e-06, "loss": 2.5327, "step": 9327 }, { "epoch": 2.924137931034483, "grad_norm": 7.54836368560791, "learning_rate": 4.664000000000001e-06, "loss": 1.964, "step": 9328 }, { "epoch": 2.9244514106583073, "grad_norm": 145.6495819091797, "learning_rate": 4.6645e-06, "loss": 2.357, "step": 9329 }, { "epoch": 2.9247648902821317, "grad_norm": 79.84230041503906, "learning_rate": 4.665e-06, "loss": 2.1085, "step": 9330 }, { "epoch": 2.925078369905956, "grad_norm": 17.38780403137207, "learning_rate": 4.6655000000000005e-06, "loss": 2.166, "step": 9331 }, { "epoch": 2.9253918495297806, "grad_norm": 8.548654556274414, "learning_rate": 4.666000000000001e-06, "loss": 2.0849, "step": 9332 }, { "epoch": 2.925705329153605, "grad_norm": 7.906950950622559, "learning_rate": 4.6665e-06, "loss": 1.8115, "step": 9333 }, { "epoch": 2.9260188087774295, "grad_norm": 6.34598445892334, "learning_rate": 4.667e-06, "loss": 2.1244, "step": 9334 }, { "epoch": 2.926332288401254, "grad_norm": 7.06488037109375, "learning_rate": 4.667500000000001e-06, "loss": 2.0213, "step": 9335 }, { "epoch": 2.9266457680250784, "grad_norm": 6.777124404907227, "learning_rate": 4.668e-06, "loss": 2.1502, "step": 9336 }, { "epoch": 2.926959247648903, "grad_norm": 8.838664054870605, "learning_rate": 4.6685e-06, "loss": 2.2453, "step": 9337 }, { "epoch": 2.9272727272727272, "grad_norm": 7.748332977294922, "learning_rate": 4.6690000000000005e-06, "loss": 2.3988, "step": 9338 }, { "epoch": 2.9275862068965517, "grad_norm": 9.702028274536133, "learning_rate": 4.6695e-06, "loss": 2.3568, "step": 9339 }, { "epoch": 2.927899686520376, "grad_norm": 47.05451202392578, "learning_rate": 4.670000000000001e-06, "loss": 2.329, "step": 9340 }, { "epoch": 2.9282131661442006, "grad_norm": 7.066833972930908, "learning_rate": 4.6705000000000004e-06, "loss": 1.924, "step": 9341 }, { "epoch": 2.928526645768025, "grad_norm": 7.492162704467773, "learning_rate": 4.671000000000001e-06, "loss": 2.0874, "step": 9342 }, { "epoch": 2.9288401253918495, "grad_norm": 9.780634880065918, "learning_rate": 4.6715e-06, "loss": 2.2537, "step": 9343 }, { "epoch": 2.929153605015674, "grad_norm": 6.84855318069458, "learning_rate": 4.672e-06, "loss": 2.1118, "step": 9344 }, { "epoch": 2.9294670846394983, "grad_norm": 10.00306224822998, "learning_rate": 4.672500000000001e-06, "loss": 2.1389, "step": 9345 }, { "epoch": 2.929780564263323, "grad_norm": 47.93046188354492, "learning_rate": 4.673e-06, "loss": 2.688, "step": 9346 }, { "epoch": 2.9300940438871472, "grad_norm": 9.102670669555664, "learning_rate": 4.6735e-06, "loss": 2.0213, "step": 9347 }, { "epoch": 2.9304075235109717, "grad_norm": 197.1841583251953, "learning_rate": 4.6740000000000005e-06, "loss": 2.6314, "step": 9348 }, { "epoch": 2.930721003134796, "grad_norm": 23.26283836364746, "learning_rate": 4.6745e-06, "loss": 2.3045, "step": 9349 }, { "epoch": 2.9310344827586206, "grad_norm": 8.816997528076172, "learning_rate": 4.675000000000001e-06, "loss": 1.8492, "step": 9350 }, { "epoch": 2.931347962382445, "grad_norm": 7.223552703857422, "learning_rate": 4.6755e-06, "loss": 2.4236, "step": 9351 }, { "epoch": 2.9316614420062694, "grad_norm": 15.561086654663086, "learning_rate": 4.676000000000001e-06, "loss": 2.2442, "step": 9352 }, { "epoch": 2.931974921630094, "grad_norm": 13.795266151428223, "learning_rate": 4.676500000000001e-06, "loss": 2.6152, "step": 9353 }, { "epoch": 2.9322884012539188, "grad_norm": 42.73121643066406, "learning_rate": 4.677e-06, "loss": 2.135, "step": 9354 }, { "epoch": 2.932601880877743, "grad_norm": 8.189810752868652, "learning_rate": 4.6775000000000005e-06, "loss": 1.9799, "step": 9355 }, { "epoch": 2.9329153605015676, "grad_norm": 10.083736419677734, "learning_rate": 4.678e-06, "loss": 2.0533, "step": 9356 }, { "epoch": 2.933228840125392, "grad_norm": 5.65605354309082, "learning_rate": 4.6785e-06, "loss": 2.0252, "step": 9357 }, { "epoch": 2.9335423197492165, "grad_norm": 8.045872688293457, "learning_rate": 4.6790000000000004e-06, "loss": 1.9526, "step": 9358 }, { "epoch": 2.933855799373041, "grad_norm": 9.793978691101074, "learning_rate": 4.6795e-06, "loss": 2.0379, "step": 9359 }, { "epoch": 2.9341692789968654, "grad_norm": 13.365437507629395, "learning_rate": 4.680000000000001e-06, "loss": 2.0766, "step": 9360 }, { "epoch": 2.93448275862069, "grad_norm": 164.11392211914062, "learning_rate": 4.6805e-06, "loss": 2.2605, "step": 9361 }, { "epoch": 2.9347962382445143, "grad_norm": 6.863374710083008, "learning_rate": 4.681000000000001e-06, "loss": 2.1252, "step": 9362 }, { "epoch": 2.9351097178683387, "grad_norm": 14.481553077697754, "learning_rate": 4.681500000000001e-06, "loss": 2.1371, "step": 9363 }, { "epoch": 2.935423197492163, "grad_norm": 7.506541728973389, "learning_rate": 4.682e-06, "loss": 2.3317, "step": 9364 }, { "epoch": 2.9357366771159876, "grad_norm": 8.37493896484375, "learning_rate": 4.6825000000000005e-06, "loss": 2.4845, "step": 9365 }, { "epoch": 2.936050156739812, "grad_norm": 57.40079879760742, "learning_rate": 4.683000000000001e-06, "loss": 2.2381, "step": 9366 }, { "epoch": 2.9363636363636365, "grad_norm": 31.36387062072754, "learning_rate": 4.6835e-06, "loss": 1.9137, "step": 9367 }, { "epoch": 2.936677115987461, "grad_norm": 67.0276870727539, "learning_rate": 4.684e-06, "loss": 2.0697, "step": 9368 }, { "epoch": 2.9369905956112854, "grad_norm": 81.57054138183594, "learning_rate": 4.6845e-06, "loss": 2.2325, "step": 9369 }, { "epoch": 2.93730407523511, "grad_norm": 18.141416549682617, "learning_rate": 4.685000000000001e-06, "loss": 2.2386, "step": 9370 }, { "epoch": 2.9376175548589343, "grad_norm": 9.082082748413086, "learning_rate": 4.6855e-06, "loss": 2.092, "step": 9371 }, { "epoch": 2.9379310344827587, "grad_norm": 16.048583984375, "learning_rate": 4.6860000000000005e-06, "loss": 1.9213, "step": 9372 }, { "epoch": 2.938244514106583, "grad_norm": 8.304791450500488, "learning_rate": 4.686500000000001e-06, "loss": 2.2165, "step": 9373 }, { "epoch": 2.9385579937304076, "grad_norm": 13.300721168518066, "learning_rate": 4.687e-06, "loss": 2.3541, "step": 9374 }, { "epoch": 2.938871473354232, "grad_norm": 16.850629806518555, "learning_rate": 4.6875000000000004e-06, "loss": 2.2726, "step": 9375 }, { "epoch": 2.9391849529780565, "grad_norm": 14.40479850769043, "learning_rate": 4.688000000000001e-06, "loss": 1.8351, "step": 9376 }, { "epoch": 2.939498432601881, "grad_norm": 14.313511848449707, "learning_rate": 4.6885e-06, "loss": 2.2308, "step": 9377 }, { "epoch": 2.9398119122257054, "grad_norm": 17.993722915649414, "learning_rate": 4.689e-06, "loss": 2.038, "step": 9378 }, { "epoch": 2.94012539184953, "grad_norm": 6.146407127380371, "learning_rate": 4.689500000000001e-06, "loss": 2.3219, "step": 9379 }, { "epoch": 2.9404388714733543, "grad_norm": 8.184374809265137, "learning_rate": 4.69e-06, "loss": 1.9887, "step": 9380 }, { "epoch": 2.9407523510971787, "grad_norm": 13.649717330932617, "learning_rate": 4.6905e-06, "loss": 1.8415, "step": 9381 }, { "epoch": 2.941065830721003, "grad_norm": 9.674529075622559, "learning_rate": 4.6910000000000005e-06, "loss": 2.1291, "step": 9382 }, { "epoch": 2.9413793103448276, "grad_norm": 21.27426528930664, "learning_rate": 4.691500000000001e-06, "loss": 2.0091, "step": 9383 }, { "epoch": 2.941692789968652, "grad_norm": 27.80304527282715, "learning_rate": 4.692e-06, "loss": 2.9071, "step": 9384 }, { "epoch": 2.9420062695924765, "grad_norm": 9.116766929626465, "learning_rate": 4.6925e-06, "loss": 2.2819, "step": 9385 }, { "epoch": 2.942319749216301, "grad_norm": 9.223663330078125, "learning_rate": 4.693000000000001e-06, "loss": 2.1769, "step": 9386 }, { "epoch": 2.9426332288401253, "grad_norm": 6.417978286743164, "learning_rate": 4.6935e-06, "loss": 2.1089, "step": 9387 }, { "epoch": 2.94294670846395, "grad_norm": 8.301006317138672, "learning_rate": 4.694e-06, "loss": 2.0138, "step": 9388 }, { "epoch": 2.9432601880877742, "grad_norm": 7.5382280349731445, "learning_rate": 4.6945000000000005e-06, "loss": 2.6735, "step": 9389 }, { "epoch": 2.9435736677115987, "grad_norm": 10.022287368774414, "learning_rate": 4.695e-06, "loss": 2.3202, "step": 9390 }, { "epoch": 2.943887147335423, "grad_norm": 25.814146041870117, "learning_rate": 4.6955e-06, "loss": 2.3381, "step": 9391 }, { "epoch": 2.9442006269592476, "grad_norm": 8.916748046875, "learning_rate": 4.6960000000000004e-06, "loss": 2.614, "step": 9392 }, { "epoch": 2.944514106583072, "grad_norm": 24.86414337158203, "learning_rate": 4.696500000000001e-06, "loss": 2.2349, "step": 9393 }, { "epoch": 2.9448275862068964, "grad_norm": 22.695783615112305, "learning_rate": 4.697e-06, "loss": 2.14, "step": 9394 }, { "epoch": 2.945141065830721, "grad_norm": 10.422751426696777, "learning_rate": 4.6975e-06, "loss": 2.0287, "step": 9395 }, { "epoch": 2.9454545454545453, "grad_norm": 9.932967185974121, "learning_rate": 4.698000000000001e-06, "loss": 1.9999, "step": 9396 }, { "epoch": 2.9457680250783698, "grad_norm": 23.31585693359375, "learning_rate": 4.6985e-06, "loss": 2.1977, "step": 9397 }, { "epoch": 2.946081504702194, "grad_norm": 7.152619361877441, "learning_rate": 4.699e-06, "loss": 1.9195, "step": 9398 }, { "epoch": 2.9463949843260187, "grad_norm": 11.567204475402832, "learning_rate": 4.6995000000000005e-06, "loss": 2.241, "step": 9399 }, { "epoch": 2.946708463949843, "grad_norm": 35.511104583740234, "learning_rate": 4.7e-06, "loss": 2.5646, "step": 9400 }, { "epoch": 2.9470219435736675, "grad_norm": 16.156341552734375, "learning_rate": 4.700500000000001e-06, "loss": 2.1259, "step": 9401 }, { "epoch": 2.947335423197492, "grad_norm": 28.503211975097656, "learning_rate": 4.701e-06, "loss": 2.0662, "step": 9402 }, { "epoch": 2.9476489028213164, "grad_norm": 9.611867904663086, "learning_rate": 4.701500000000001e-06, "loss": 2.0387, "step": 9403 }, { "epoch": 2.947962382445141, "grad_norm": 13.767318725585938, "learning_rate": 4.702e-06, "loss": 2.1414, "step": 9404 }, { "epoch": 2.9482758620689653, "grad_norm": 6.023205280303955, "learning_rate": 4.7025e-06, "loss": 1.9418, "step": 9405 }, { "epoch": 2.9485893416927897, "grad_norm": 7.297669410705566, "learning_rate": 4.7030000000000005e-06, "loss": 2.4959, "step": 9406 }, { "epoch": 2.948902821316614, "grad_norm": 10.526333808898926, "learning_rate": 4.7035e-06, "loss": 2.3774, "step": 9407 }, { "epoch": 2.9492163009404386, "grad_norm": 16.836910247802734, "learning_rate": 4.704e-06, "loss": 2.3027, "step": 9408 }, { "epoch": 2.949529780564263, "grad_norm": 13.978313446044922, "learning_rate": 4.7045000000000004e-06, "loss": 2.2607, "step": 9409 }, { "epoch": 2.9498432601880875, "grad_norm": 25.487403869628906, "learning_rate": 4.705e-06, "loss": 2.9073, "step": 9410 }, { "epoch": 2.950156739811912, "grad_norm": 5.864297866821289, "learning_rate": 4.705500000000001e-06, "loss": 2.2117, "step": 9411 }, { "epoch": 2.950470219435737, "grad_norm": 9.06821060180664, "learning_rate": 4.706e-06, "loss": 2.6777, "step": 9412 }, { "epoch": 2.9507836990595613, "grad_norm": 8.11877727508545, "learning_rate": 4.706500000000001e-06, "loss": 2.0353, "step": 9413 }, { "epoch": 2.9510971786833857, "grad_norm": 10.383318901062012, "learning_rate": 4.707000000000001e-06, "loss": 2.7202, "step": 9414 }, { "epoch": 2.95141065830721, "grad_norm": 8.18381404876709, "learning_rate": 4.7075e-06, "loss": 2.4542, "step": 9415 }, { "epoch": 2.9517241379310346, "grad_norm": 28.680334091186523, "learning_rate": 4.7080000000000005e-06, "loss": 2.3897, "step": 9416 }, { "epoch": 2.952037617554859, "grad_norm": 11.344430923461914, "learning_rate": 4.7085e-06, "loss": 2.3191, "step": 9417 }, { "epoch": 2.9523510971786835, "grad_norm": 12.884014129638672, "learning_rate": 4.709e-06, "loss": 2.12, "step": 9418 }, { "epoch": 2.952664576802508, "grad_norm": 8.016007423400879, "learning_rate": 4.7095e-06, "loss": 2.2327, "step": 9419 }, { "epoch": 2.9529780564263324, "grad_norm": 20.812702178955078, "learning_rate": 4.71e-06, "loss": 3.4071, "step": 9420 }, { "epoch": 2.953291536050157, "grad_norm": 7.646617889404297, "learning_rate": 4.710500000000001e-06, "loss": 1.9446, "step": 9421 }, { "epoch": 2.9536050156739813, "grad_norm": 7.737258434295654, "learning_rate": 4.711e-06, "loss": 2.1809, "step": 9422 }, { "epoch": 2.9539184952978057, "grad_norm": 30.939794540405273, "learning_rate": 4.7115000000000005e-06, "loss": 2.5008, "step": 9423 }, { "epoch": 2.95423197492163, "grad_norm": 9.922723770141602, "learning_rate": 4.712000000000001e-06, "loss": 2.3384, "step": 9424 }, { "epoch": 2.9545454545454546, "grad_norm": 9.688403129577637, "learning_rate": 4.7125e-06, "loss": 1.9085, "step": 9425 }, { "epoch": 2.954858934169279, "grad_norm": 49.84483337402344, "learning_rate": 4.7130000000000004e-06, "loss": 2.4595, "step": 9426 }, { "epoch": 2.9551724137931035, "grad_norm": 8.037410736083984, "learning_rate": 4.713500000000001e-06, "loss": 2.2021, "step": 9427 }, { "epoch": 2.955485893416928, "grad_norm": 17.965370178222656, "learning_rate": 4.714e-06, "loss": 2.1182, "step": 9428 }, { "epoch": 2.9557993730407524, "grad_norm": 8.034486770629883, "learning_rate": 4.7145e-06, "loss": 2.2249, "step": 9429 }, { "epoch": 2.956112852664577, "grad_norm": 15.204883575439453, "learning_rate": 4.715e-06, "loss": 1.947, "step": 9430 }, { "epoch": 2.9564263322884012, "grad_norm": 74.3429946899414, "learning_rate": 4.715500000000001e-06, "loss": 2.1213, "step": 9431 }, { "epoch": 2.9567398119122257, "grad_norm": 8.02340030670166, "learning_rate": 4.716e-06, "loss": 2.0167, "step": 9432 }, { "epoch": 2.95705329153605, "grad_norm": 41.44495391845703, "learning_rate": 4.7165000000000005e-06, "loss": 2.2161, "step": 9433 }, { "epoch": 2.9573667711598746, "grad_norm": 8.7019681930542, "learning_rate": 4.717000000000001e-06, "loss": 2.0961, "step": 9434 }, { "epoch": 2.957680250783699, "grad_norm": 44.380615234375, "learning_rate": 4.7175e-06, "loss": 2.2989, "step": 9435 }, { "epoch": 2.9579937304075234, "grad_norm": 59.554927825927734, "learning_rate": 4.718e-06, "loss": 2.5265, "step": 9436 }, { "epoch": 2.958307210031348, "grad_norm": 11.365534782409668, "learning_rate": 4.718500000000001e-06, "loss": 2.104, "step": 9437 }, { "epoch": 2.9586206896551723, "grad_norm": 8.621533393859863, "learning_rate": 4.719e-06, "loss": 1.986, "step": 9438 }, { "epoch": 2.9589341692789968, "grad_norm": 10.279128074645996, "learning_rate": 4.7195e-06, "loss": 2.2569, "step": 9439 }, { "epoch": 2.959247648902821, "grad_norm": 12.237083435058594, "learning_rate": 4.7200000000000005e-06, "loss": 2.3127, "step": 9440 }, { "epoch": 2.9595611285266457, "grad_norm": 7.4483795166015625, "learning_rate": 4.720500000000001e-06, "loss": 2.1353, "step": 9441 }, { "epoch": 2.95987460815047, "grad_norm": 10.915425300598145, "learning_rate": 4.721e-06, "loss": 2.0551, "step": 9442 }, { "epoch": 2.9601880877742945, "grad_norm": 6.573949337005615, "learning_rate": 4.7215000000000004e-06, "loss": 1.8536, "step": 9443 }, { "epoch": 2.960501567398119, "grad_norm": 9.041077613830566, "learning_rate": 4.722000000000001e-06, "loss": 2.1444, "step": 9444 }, { "epoch": 2.9608150470219434, "grad_norm": 9.121186256408691, "learning_rate": 4.7225e-06, "loss": 2.261, "step": 9445 }, { "epoch": 2.961128526645768, "grad_norm": 16.1588191986084, "learning_rate": 4.723e-06, "loss": 1.9159, "step": 9446 }, { "epoch": 2.9614420062695928, "grad_norm": 13.04690170288086, "learning_rate": 4.723500000000001e-06, "loss": 1.9505, "step": 9447 }, { "epoch": 2.961755485893417, "grad_norm": 45.714481353759766, "learning_rate": 4.724e-06, "loss": 2.1915, "step": 9448 }, { "epoch": 2.9620689655172416, "grad_norm": 69.0879135131836, "learning_rate": 4.7245e-06, "loss": 2.8395, "step": 9449 }, { "epoch": 2.962382445141066, "grad_norm": 8.945505142211914, "learning_rate": 4.7250000000000005e-06, "loss": 2.273, "step": 9450 }, { "epoch": 2.9626959247648905, "grad_norm": 8.749455451965332, "learning_rate": 4.725500000000001e-06, "loss": 2.0319, "step": 9451 }, { "epoch": 2.963009404388715, "grad_norm": 7.7648515701293945, "learning_rate": 4.726000000000001e-06, "loss": 2.0326, "step": 9452 }, { "epoch": 2.9633228840125394, "grad_norm": 7.449599742889404, "learning_rate": 4.7265e-06, "loss": 2.2354, "step": 9453 }, { "epoch": 2.963636363636364, "grad_norm": 8.937540054321289, "learning_rate": 4.727000000000001e-06, "loss": 2.2646, "step": 9454 }, { "epoch": 2.9639498432601883, "grad_norm": 7.455075740814209, "learning_rate": 4.7275e-06, "loss": 1.988, "step": 9455 }, { "epoch": 2.9642633228840127, "grad_norm": 28.006298065185547, "learning_rate": 4.728e-06, "loss": 2.0926, "step": 9456 }, { "epoch": 2.964576802507837, "grad_norm": 9.41918659210205, "learning_rate": 4.7285000000000006e-06, "loss": 1.9229, "step": 9457 }, { "epoch": 2.9648902821316616, "grad_norm": 8.421971321105957, "learning_rate": 4.729e-06, "loss": 2.0894, "step": 9458 }, { "epoch": 2.965203761755486, "grad_norm": 10.615500450134277, "learning_rate": 4.7295e-06, "loss": 2.2238, "step": 9459 }, { "epoch": 2.9655172413793105, "grad_norm": 7.271868705749512, "learning_rate": 4.7300000000000005e-06, "loss": 2.0563, "step": 9460 }, { "epoch": 2.965830721003135, "grad_norm": 20.375560760498047, "learning_rate": 4.730500000000001e-06, "loss": 2.1397, "step": 9461 }, { "epoch": 2.9661442006269594, "grad_norm": 20.549915313720703, "learning_rate": 4.731000000000001e-06, "loss": 1.8666, "step": 9462 }, { "epoch": 2.966457680250784, "grad_norm": 10.327241897583008, "learning_rate": 4.7315e-06, "loss": 2.0368, "step": 9463 }, { "epoch": 2.9667711598746083, "grad_norm": 8.2266206741333, "learning_rate": 4.732000000000001e-06, "loss": 2.2902, "step": 9464 }, { "epoch": 2.9670846394984327, "grad_norm": 7.388582706451416, "learning_rate": 4.7325e-06, "loss": 2.1283, "step": 9465 }, { "epoch": 2.967398119122257, "grad_norm": 38.946529388427734, "learning_rate": 4.733e-06, "loss": 2.5009, "step": 9466 }, { "epoch": 2.9677115987460816, "grad_norm": 13.569231986999512, "learning_rate": 4.7335000000000005e-06, "loss": 2.0731, "step": 9467 }, { "epoch": 2.968025078369906, "grad_norm": 10.80717658996582, "learning_rate": 4.734e-06, "loss": 2.0886, "step": 9468 }, { "epoch": 2.9683385579937305, "grad_norm": 12.45492935180664, "learning_rate": 4.7345e-06, "loss": 2.1393, "step": 9469 }, { "epoch": 2.968652037617555, "grad_norm": 5.224359512329102, "learning_rate": 4.735e-06, "loss": 2.1769, "step": 9470 }, { "epoch": 2.9689655172413794, "grad_norm": 5.110821723937988, "learning_rate": 4.735500000000001e-06, "loss": 1.987, "step": 9471 }, { "epoch": 2.969278996865204, "grad_norm": 7.895551681518555, "learning_rate": 4.736000000000001e-06, "loss": 2.14, "step": 9472 }, { "epoch": 2.9695924764890282, "grad_norm": 6.788575649261475, "learning_rate": 4.7365e-06, "loss": 1.838, "step": 9473 }, { "epoch": 2.9699059561128527, "grad_norm": 11.671319007873535, "learning_rate": 4.7370000000000006e-06, "loss": 2.0954, "step": 9474 }, { "epoch": 2.970219435736677, "grad_norm": 172.9235076904297, "learning_rate": 4.737500000000001e-06, "loss": 2.0517, "step": 9475 }, { "epoch": 2.9705329153605016, "grad_norm": 7.782164096832275, "learning_rate": 4.738e-06, "loss": 2.2107, "step": 9476 }, { "epoch": 2.970846394984326, "grad_norm": 16.09248161315918, "learning_rate": 4.7385000000000005e-06, "loss": 1.928, "step": 9477 }, { "epoch": 2.9711598746081505, "grad_norm": 7.648118495941162, "learning_rate": 4.739e-06, "loss": 2.0897, "step": 9478 }, { "epoch": 2.971473354231975, "grad_norm": 6.92603063583374, "learning_rate": 4.7395e-06, "loss": 2.2197, "step": 9479 }, { "epoch": 2.9717868338557993, "grad_norm": 8.581344604492188, "learning_rate": 4.74e-06, "loss": 2.3306, "step": 9480 }, { "epoch": 2.972100313479624, "grad_norm": 13.093215942382812, "learning_rate": 4.740500000000001e-06, "loss": 2.0675, "step": 9481 }, { "epoch": 2.972413793103448, "grad_norm": 7.767763614654541, "learning_rate": 4.741000000000001e-06, "loss": 2.1263, "step": 9482 }, { "epoch": 2.9727272727272727, "grad_norm": 7.693376541137695, "learning_rate": 4.7415e-06, "loss": 2.0568, "step": 9483 }, { "epoch": 2.973040752351097, "grad_norm": 19.743440628051758, "learning_rate": 4.7420000000000005e-06, "loss": 2.0659, "step": 9484 }, { "epoch": 2.9733542319749215, "grad_norm": 9.665297508239746, "learning_rate": 4.742500000000001e-06, "loss": 2.2835, "step": 9485 }, { "epoch": 2.973667711598746, "grad_norm": 12.232024192810059, "learning_rate": 4.743e-06, "loss": 2.0861, "step": 9486 }, { "epoch": 2.9739811912225704, "grad_norm": 8.619750022888184, "learning_rate": 4.7435e-06, "loss": 1.8808, "step": 9487 }, { "epoch": 2.974294670846395, "grad_norm": 20.196224212646484, "learning_rate": 4.744000000000001e-06, "loss": 2.0878, "step": 9488 }, { "epoch": 2.9746081504702193, "grad_norm": 19.604290008544922, "learning_rate": 4.7445e-06, "loss": 2.2231, "step": 9489 }, { "epoch": 2.9749216300940438, "grad_norm": 10.714457511901855, "learning_rate": 4.745e-06, "loss": 2.1569, "step": 9490 }, { "epoch": 2.975235109717868, "grad_norm": 37.0594367980957, "learning_rate": 4.7455000000000006e-06, "loss": 2.1578, "step": 9491 }, { "epoch": 2.9755485893416926, "grad_norm": 198.78399658203125, "learning_rate": 4.746000000000001e-06, "loss": 2.3641, "step": 9492 }, { "epoch": 2.975862068965517, "grad_norm": 102.64945983886719, "learning_rate": 4.7465e-06, "loss": 2.5818, "step": 9493 }, { "epoch": 2.9761755485893415, "grad_norm": 28.096132278442383, "learning_rate": 4.7470000000000005e-06, "loss": 2.2159, "step": 9494 }, { "epoch": 2.976489028213166, "grad_norm": 12.7365140914917, "learning_rate": 4.747500000000001e-06, "loss": 2.4381, "step": 9495 }, { "epoch": 2.9768025078369904, "grad_norm": 129.6446075439453, "learning_rate": 4.748e-06, "loss": 2.3956, "step": 9496 }, { "epoch": 2.977115987460815, "grad_norm": 9.241852760314941, "learning_rate": 4.7485e-06, "loss": 2.0901, "step": 9497 }, { "epoch": 2.9774294670846393, "grad_norm": 8.458335876464844, "learning_rate": 4.749000000000001e-06, "loss": 2.0769, "step": 9498 }, { "epoch": 2.9777429467084637, "grad_norm": 15.9458589553833, "learning_rate": 4.7495e-06, "loss": 2.9178, "step": 9499 }, { "epoch": 2.978056426332288, "grad_norm": 6.00883674621582, "learning_rate": 4.75e-06, "loss": 2.1179, "step": 9500 }, { "epoch": 2.9783699059561126, "grad_norm": 17.76551628112793, "learning_rate": 4.7505000000000005e-06, "loss": 2.0117, "step": 9501 }, { "epoch": 2.978683385579937, "grad_norm": 9.227352142333984, "learning_rate": 4.751000000000001e-06, "loss": 1.8803, "step": 9502 }, { "epoch": 2.9789968652037615, "grad_norm": 9.37998104095459, "learning_rate": 4.7515e-06, "loss": 2.2083, "step": 9503 }, { "epoch": 2.979310344827586, "grad_norm": 7.4627532958984375, "learning_rate": 4.752e-06, "loss": 1.8116, "step": 9504 }, { "epoch": 2.9796238244514104, "grad_norm": 6.722050666809082, "learning_rate": 4.752500000000001e-06, "loss": 2.6186, "step": 9505 }, { "epoch": 2.9799373040752353, "grad_norm": 6.502493381500244, "learning_rate": 4.753e-06, "loss": 2.1572, "step": 9506 }, { "epoch": 2.9802507836990597, "grad_norm": 8.201019287109375, "learning_rate": 4.7535e-06, "loss": 2.1498, "step": 9507 }, { "epoch": 2.980564263322884, "grad_norm": 74.14036560058594, "learning_rate": 4.7540000000000006e-06, "loss": 2.281, "step": 9508 }, { "epoch": 2.9808777429467086, "grad_norm": 49.861148834228516, "learning_rate": 4.7545e-06, "loss": 1.854, "step": 9509 }, { "epoch": 2.981191222570533, "grad_norm": 16.236431121826172, "learning_rate": 4.755e-06, "loss": 2.1212, "step": 9510 }, { "epoch": 2.9815047021943575, "grad_norm": 7.515372276306152, "learning_rate": 4.7555000000000005e-06, "loss": 1.9797, "step": 9511 }, { "epoch": 2.981818181818182, "grad_norm": 7.8705735206604, "learning_rate": 4.756000000000001e-06, "loss": 2.3871, "step": 9512 }, { "epoch": 2.9821316614420064, "grad_norm": 6.880337715148926, "learning_rate": 4.756500000000001e-06, "loss": 2.2361, "step": 9513 }, { "epoch": 2.982445141065831, "grad_norm": 12.13143253326416, "learning_rate": 4.757e-06, "loss": 1.8843, "step": 9514 }, { "epoch": 2.9827586206896552, "grad_norm": 33.823631286621094, "learning_rate": 4.757500000000001e-06, "loss": 2.5258, "step": 9515 }, { "epoch": 2.9830721003134797, "grad_norm": 8.110567092895508, "learning_rate": 4.758e-06, "loss": 2.5943, "step": 9516 }, { "epoch": 2.983385579937304, "grad_norm": 9.85314655303955, "learning_rate": 4.7585e-06, "loss": 2.3138, "step": 9517 }, { "epoch": 2.9836990595611286, "grad_norm": 100.14962768554688, "learning_rate": 4.7590000000000005e-06, "loss": 2.4222, "step": 9518 }, { "epoch": 2.984012539184953, "grad_norm": 9.922337532043457, "learning_rate": 4.7595e-06, "loss": 2.1085, "step": 9519 }, { "epoch": 2.9843260188087775, "grad_norm": 105.21549987792969, "learning_rate": 4.76e-06, "loss": 2.2196, "step": 9520 }, { "epoch": 2.984639498432602, "grad_norm": 6.230706214904785, "learning_rate": 4.7605e-06, "loss": 1.9822, "step": 9521 }, { "epoch": 2.9849529780564263, "grad_norm": 17.318565368652344, "learning_rate": 4.761000000000001e-06, "loss": 2.0543, "step": 9522 }, { "epoch": 2.985266457680251, "grad_norm": 10.2170991897583, "learning_rate": 4.761500000000001e-06, "loss": 2.1099, "step": 9523 }, { "epoch": 2.9855799373040752, "grad_norm": 6.637144088745117, "learning_rate": 4.762e-06, "loss": 2.0936, "step": 9524 }, { "epoch": 2.9858934169278997, "grad_norm": 9.456480026245117, "learning_rate": 4.7625000000000006e-06, "loss": 2.1869, "step": 9525 }, { "epoch": 2.986206896551724, "grad_norm": 5.885220527648926, "learning_rate": 4.763000000000001e-06, "loss": 1.9146, "step": 9526 }, { "epoch": 2.9865203761755486, "grad_norm": 218.4715118408203, "learning_rate": 4.7635e-06, "loss": 2.7327, "step": 9527 }, { "epoch": 2.986833855799373, "grad_norm": 18.660215377807617, "learning_rate": 4.7640000000000005e-06, "loss": 3.2755, "step": 9528 }, { "epoch": 2.9871473354231974, "grad_norm": 13.571133613586426, "learning_rate": 4.7645e-06, "loss": 2.2309, "step": 9529 }, { "epoch": 2.987460815047022, "grad_norm": 8.085278511047363, "learning_rate": 4.765e-06, "loss": 2.1601, "step": 9530 }, { "epoch": 2.9877742946708463, "grad_norm": 14.36074161529541, "learning_rate": 4.7655e-06, "loss": 2.078, "step": 9531 }, { "epoch": 2.9880877742946708, "grad_norm": 8.615720748901367, "learning_rate": 4.766000000000001e-06, "loss": 2.2335, "step": 9532 }, { "epoch": 2.988401253918495, "grad_norm": 11.980807304382324, "learning_rate": 4.766500000000001e-06, "loss": 1.978, "step": 9533 }, { "epoch": 2.9887147335423196, "grad_norm": 29.99042320251465, "learning_rate": 4.767e-06, "loss": 2.3915, "step": 9534 }, { "epoch": 2.989028213166144, "grad_norm": 7.426431179046631, "learning_rate": 4.7675000000000005e-06, "loss": 2.1384, "step": 9535 }, { "epoch": 2.9893416927899685, "grad_norm": 8.940190315246582, "learning_rate": 4.768000000000001e-06, "loss": 2.0381, "step": 9536 }, { "epoch": 2.989655172413793, "grad_norm": 19.570085525512695, "learning_rate": 4.7685e-06, "loss": 1.951, "step": 9537 }, { "epoch": 2.9899686520376174, "grad_norm": 8.389922142028809, "learning_rate": 4.769e-06, "loss": 2.2639, "step": 9538 }, { "epoch": 2.990282131661442, "grad_norm": 9.735523223876953, "learning_rate": 4.7695e-06, "loss": 2.1704, "step": 9539 }, { "epoch": 2.9905956112852663, "grad_norm": 7.256843090057373, "learning_rate": 4.77e-06, "loss": 1.9766, "step": 9540 }, { "epoch": 2.990909090909091, "grad_norm": 18.244461059570312, "learning_rate": 4.7705e-06, "loss": 1.9512, "step": 9541 }, { "epoch": 2.9912225705329156, "grad_norm": 10.988954544067383, "learning_rate": 4.7710000000000006e-06, "loss": 2.1113, "step": 9542 }, { "epoch": 2.99153605015674, "grad_norm": 5.592411994934082, "learning_rate": 4.771500000000001e-06, "loss": 2.1959, "step": 9543 }, { "epoch": 2.9918495297805645, "grad_norm": 43.091888427734375, "learning_rate": 4.772e-06, "loss": 2.0321, "step": 9544 }, { "epoch": 2.992163009404389, "grad_norm": 80.94044494628906, "learning_rate": 4.7725000000000005e-06, "loss": 2.1723, "step": 9545 }, { "epoch": 2.9924764890282134, "grad_norm": 9.173052787780762, "learning_rate": 4.773000000000001e-06, "loss": 2.0507, "step": 9546 }, { "epoch": 2.992789968652038, "grad_norm": 9.058065414428711, "learning_rate": 4.7735e-06, "loss": 2.1809, "step": 9547 }, { "epoch": 2.9931034482758623, "grad_norm": 9.534920692443848, "learning_rate": 4.774e-06, "loss": 2.3156, "step": 9548 }, { "epoch": 2.9934169278996867, "grad_norm": 6.235188007354736, "learning_rate": 4.774500000000001e-06, "loss": 2.0472, "step": 9549 }, { "epoch": 2.993730407523511, "grad_norm": 13.413022994995117, "learning_rate": 4.775e-06, "loss": 2.4863, "step": 9550 }, { "epoch": 2.9940438871473356, "grad_norm": 5.9052886962890625, "learning_rate": 4.7755e-06, "loss": 2.1659, "step": 9551 }, { "epoch": 2.99435736677116, "grad_norm": 17.41887855529785, "learning_rate": 4.7760000000000005e-06, "loss": 1.9686, "step": 9552 }, { "epoch": 2.9946708463949845, "grad_norm": 8.941686630249023, "learning_rate": 4.776500000000001e-06, "loss": 2.2308, "step": 9553 }, { "epoch": 2.994984326018809, "grad_norm": 5.957551956176758, "learning_rate": 4.777e-06, "loss": 2.1032, "step": 9554 }, { "epoch": 2.9952978056426334, "grad_norm": 6.600414752960205, "learning_rate": 4.7775e-06, "loss": 2.2067, "step": 9555 }, { "epoch": 2.995611285266458, "grad_norm": 6.964043617248535, "learning_rate": 4.778000000000001e-06, "loss": 2.0554, "step": 9556 }, { "epoch": 2.9959247648902823, "grad_norm": 6.6755805015563965, "learning_rate": 4.7785e-06, "loss": 2.1694, "step": 9557 }, { "epoch": 2.9962382445141067, "grad_norm": 10.168830871582031, "learning_rate": 4.779e-06, "loss": 2.075, "step": 9558 }, { "epoch": 2.996551724137931, "grad_norm": 47.286983489990234, "learning_rate": 4.7795000000000006e-06, "loss": 2.6181, "step": 9559 }, { "epoch": 2.9968652037617556, "grad_norm": 23.60767936706543, "learning_rate": 4.78e-06, "loss": 2.1055, "step": 9560 }, { "epoch": 2.99717868338558, "grad_norm": 9.204062461853027, "learning_rate": 4.7805e-06, "loss": 2.1491, "step": 9561 }, { "epoch": 2.9974921630094045, "grad_norm": 10.915852546691895, "learning_rate": 4.7810000000000005e-06, "loss": 1.9908, "step": 9562 }, { "epoch": 2.997805642633229, "grad_norm": 7.748519420623779, "learning_rate": 4.781500000000001e-06, "loss": 2.1005, "step": 9563 }, { "epoch": 2.9981191222570533, "grad_norm": 5.714078903198242, "learning_rate": 4.782e-06, "loss": 1.9584, "step": 9564 }, { "epoch": 2.998432601880878, "grad_norm": 8.184148788452148, "learning_rate": 4.7825e-06, "loss": 2.2521, "step": 9565 }, { "epoch": 2.9987460815047022, "grad_norm": 8.809863090515137, "learning_rate": 4.783000000000001e-06, "loss": 2.2099, "step": 9566 }, { "epoch": 2.9990595611285267, "grad_norm": 9.872733116149902, "learning_rate": 4.7835e-06, "loss": 2.2035, "step": 9567 }, { "epoch": 2.999373040752351, "grad_norm": 16.570030212402344, "learning_rate": 4.784e-06, "loss": 2.0555, "step": 9568 }, { "epoch": 2.9996865203761756, "grad_norm": 8.70211410522461, "learning_rate": 4.7845000000000005e-06, "loss": 2.0201, "step": 9569 }, { "epoch": 3.0, "grad_norm": 10.009987831115723, "learning_rate": 4.785e-06, "loss": 1.9691, "step": 9570 }, { "epoch": 3.0003134796238244, "grad_norm": 8.36247444152832, "learning_rate": 4.7855e-06, "loss": 2.1575, "step": 9571 }, { "epoch": 3.000626959247649, "grad_norm": 8.633979797363281, "learning_rate": 4.7860000000000004e-06, "loss": 2.186, "step": 9572 }, { "epoch": 3.0009404388714733, "grad_norm": 12.12990665435791, "learning_rate": 4.786500000000001e-06, "loss": 2.1615, "step": 9573 }, { "epoch": 3.0012539184952978, "grad_norm": 41.975990295410156, "learning_rate": 4.787000000000001e-06, "loss": 1.9593, "step": 9574 }, { "epoch": 3.001567398119122, "grad_norm": 6.331096172332764, "learning_rate": 4.7875e-06, "loss": 2.0643, "step": 9575 }, { "epoch": 3.0018808777429467, "grad_norm": 23.415904998779297, "learning_rate": 4.7880000000000006e-06, "loss": 2.7874, "step": 9576 }, { "epoch": 3.0018808777429467, "eval_loss": 2.3029861450195312, "eval_runtime": 20.8046, "eval_samples_per_second": 129.154, "eval_steps_per_second": 8.075, "step": 9576 }, { "epoch": 3.002194357366771, "grad_norm": 11.364919662475586, "learning_rate": 4.7885e-06, "loss": 2.1737, "step": 9577 }, { "epoch": 3.0025078369905955, "grad_norm": 10.438787460327148, "learning_rate": 4.789e-06, "loss": 1.9203, "step": 9578 }, { "epoch": 3.00282131661442, "grad_norm": 5.614959716796875, "learning_rate": 4.7895000000000005e-06, "loss": 2.1453, "step": 9579 }, { "epoch": 3.0031347962382444, "grad_norm": 12.433500289916992, "learning_rate": 4.79e-06, "loss": 2.1198, "step": 9580 }, { "epoch": 3.003448275862069, "grad_norm": 6.100986480712891, "learning_rate": 4.7905e-06, "loss": 2.1649, "step": 9581 }, { "epoch": 3.0037617554858933, "grad_norm": 23.833477020263672, "learning_rate": 4.791e-06, "loss": 2.3952, "step": 9582 }, { "epoch": 3.0040752351097177, "grad_norm": 7.875548839569092, "learning_rate": 4.791500000000001e-06, "loss": 1.9524, "step": 9583 }, { "epoch": 3.004388714733542, "grad_norm": 16.73760414123535, "learning_rate": 4.792000000000001e-06, "loss": 2.1136, "step": 9584 }, { "epoch": 3.0047021943573666, "grad_norm": 6.842322826385498, "learning_rate": 4.7925e-06, "loss": 1.9462, "step": 9585 }, { "epoch": 3.005015673981191, "grad_norm": 8.689022064208984, "learning_rate": 4.7930000000000005e-06, "loss": 2.0003, "step": 9586 }, { "epoch": 3.0053291536050155, "grad_norm": 7.339883327484131, "learning_rate": 4.793500000000001e-06, "loss": 2.0737, "step": 9587 }, { "epoch": 3.00564263322884, "grad_norm": 7.914515018463135, "learning_rate": 4.794e-06, "loss": 1.8574, "step": 9588 }, { "epoch": 3.0059561128526644, "grad_norm": 12.989086151123047, "learning_rate": 4.7945000000000004e-06, "loss": 2.165, "step": 9589 }, { "epoch": 3.006269592476489, "grad_norm": 7.834178924560547, "learning_rate": 4.795e-06, "loss": 1.9589, "step": 9590 }, { "epoch": 3.0065830721003133, "grad_norm": 15.043722152709961, "learning_rate": 4.7955e-06, "loss": 2.2711, "step": 9591 }, { "epoch": 3.0068965517241377, "grad_norm": 22.429834365844727, "learning_rate": 4.796e-06, "loss": 2.0765, "step": 9592 }, { "epoch": 3.007210031347962, "grad_norm": 7.104536056518555, "learning_rate": 4.796500000000001e-06, "loss": 2.3116, "step": 9593 }, { "epoch": 3.007523510971787, "grad_norm": 11.817303657531738, "learning_rate": 4.797000000000001e-06, "loss": 2.2483, "step": 9594 }, { "epoch": 3.0078369905956115, "grad_norm": 13.12219524383545, "learning_rate": 4.7975e-06, "loss": 2.4215, "step": 9595 }, { "epoch": 3.008150470219436, "grad_norm": 21.51630401611328, "learning_rate": 4.7980000000000005e-06, "loss": 2.3032, "step": 9596 }, { "epoch": 3.0084639498432604, "grad_norm": 17.585336685180664, "learning_rate": 4.798500000000001e-06, "loss": 2.1945, "step": 9597 }, { "epoch": 3.008777429467085, "grad_norm": 7.516443729400635, "learning_rate": 4.799e-06, "loss": 2.1832, "step": 9598 }, { "epoch": 3.0090909090909093, "grad_norm": 16.574399948120117, "learning_rate": 4.7995e-06, "loss": 2.2174, "step": 9599 }, { "epoch": 3.0094043887147337, "grad_norm": 7.617154598236084, "learning_rate": 4.800000000000001e-06, "loss": 2.0327, "step": 9600 }, { "epoch": 3.009717868338558, "grad_norm": 14.824995994567871, "learning_rate": 4.8005e-06, "loss": 1.7606, "step": 9601 }, { "epoch": 3.0100313479623826, "grad_norm": 6.91386079788208, "learning_rate": 4.801e-06, "loss": 2.209, "step": 9602 }, { "epoch": 3.010344827586207, "grad_norm": 8.046427726745605, "learning_rate": 4.8015000000000005e-06, "loss": 2.0615, "step": 9603 }, { "epoch": 3.0106583072100315, "grad_norm": 25.725475311279297, "learning_rate": 4.802000000000001e-06, "loss": 2.0191, "step": 9604 }, { "epoch": 3.010971786833856, "grad_norm": 8.534031867980957, "learning_rate": 4.8025e-06, "loss": 2.0703, "step": 9605 }, { "epoch": 3.0112852664576804, "grad_norm": 10.530138969421387, "learning_rate": 4.8030000000000004e-06, "loss": 2.2729, "step": 9606 }, { "epoch": 3.011598746081505, "grad_norm": 10.902108192443848, "learning_rate": 4.803500000000001e-06, "loss": 2.5151, "step": 9607 }, { "epoch": 3.0119122257053292, "grad_norm": 13.0545015335083, "learning_rate": 4.804e-06, "loss": 2.1198, "step": 9608 }, { "epoch": 3.0122257053291537, "grad_norm": 19.0584659576416, "learning_rate": 4.8045e-06, "loss": 1.7489, "step": 9609 }, { "epoch": 3.012539184952978, "grad_norm": 7.175193786621094, "learning_rate": 4.805000000000001e-06, "loss": 2.0744, "step": 9610 }, { "epoch": 3.0128526645768026, "grad_norm": 7.771414279937744, "learning_rate": 4.8055e-06, "loss": 1.739, "step": 9611 }, { "epoch": 3.013166144200627, "grad_norm": 10.549572944641113, "learning_rate": 4.806000000000001e-06, "loss": 2.1789, "step": 9612 }, { "epoch": 3.0134796238244514, "grad_norm": 32.086849212646484, "learning_rate": 4.8065000000000005e-06, "loss": 2.3912, "step": 9613 }, { "epoch": 3.013793103448276, "grad_norm": 21.184261322021484, "learning_rate": 4.807000000000001e-06, "loss": 2.3728, "step": 9614 }, { "epoch": 3.0141065830721003, "grad_norm": 5.17852258682251, "learning_rate": 4.8075e-06, "loss": 2.0193, "step": 9615 }, { "epoch": 3.0144200626959248, "grad_norm": 12.410475730895996, "learning_rate": 4.808e-06, "loss": 2.1515, "step": 9616 }, { "epoch": 3.014733542319749, "grad_norm": 18.111955642700195, "learning_rate": 4.808500000000001e-06, "loss": 2.4146, "step": 9617 }, { "epoch": 3.0150470219435737, "grad_norm": 42.61017990112305, "learning_rate": 4.809e-06, "loss": 2.3013, "step": 9618 }, { "epoch": 3.015360501567398, "grad_norm": 9.8378324508667, "learning_rate": 4.8095e-06, "loss": 2.1613, "step": 9619 }, { "epoch": 3.0156739811912225, "grad_norm": 6.261403560638428, "learning_rate": 4.8100000000000005e-06, "loss": 2.1786, "step": 9620 }, { "epoch": 3.015987460815047, "grad_norm": 10.620481491088867, "learning_rate": 4.8105e-06, "loss": 2.1587, "step": 9621 }, { "epoch": 3.0163009404388714, "grad_norm": 22.7048397064209, "learning_rate": 4.811000000000001e-06, "loss": 2.2198, "step": 9622 }, { "epoch": 3.016614420062696, "grad_norm": 5.913995265960693, "learning_rate": 4.8115000000000004e-06, "loss": 2.0562, "step": 9623 }, { "epoch": 3.0169278996865203, "grad_norm": 12.190495491027832, "learning_rate": 4.812000000000001e-06, "loss": 2.103, "step": 9624 }, { "epoch": 3.0172413793103448, "grad_norm": 12.885970115661621, "learning_rate": 4.8125e-06, "loss": 2.4961, "step": 9625 }, { "epoch": 3.017554858934169, "grad_norm": 8.773770332336426, "learning_rate": 4.813e-06, "loss": 2.1748, "step": 9626 }, { "epoch": 3.0178683385579936, "grad_norm": 8.804131507873535, "learning_rate": 4.813500000000001e-06, "loss": 2.1157, "step": 9627 }, { "epoch": 3.018181818181818, "grad_norm": 7.811643600463867, "learning_rate": 4.814e-06, "loss": 2.259, "step": 9628 }, { "epoch": 3.0184952978056425, "grad_norm": 8.567361831665039, "learning_rate": 4.8145e-06, "loss": 2.1517, "step": 9629 }, { "epoch": 3.018808777429467, "grad_norm": 9.984244346618652, "learning_rate": 4.8150000000000005e-06, "loss": 1.9947, "step": 9630 }, { "epoch": 3.0191222570532914, "grad_norm": 9.422271728515625, "learning_rate": 4.8155e-06, "loss": 1.8166, "step": 9631 }, { "epoch": 3.019435736677116, "grad_norm": 87.44471740722656, "learning_rate": 4.816e-06, "loss": 2.0439, "step": 9632 }, { "epoch": 3.0197492163009403, "grad_norm": 8.883208274841309, "learning_rate": 4.8165e-06, "loss": 2.601, "step": 9633 }, { "epoch": 3.0200626959247647, "grad_norm": 15.995524406433105, "learning_rate": 4.817000000000001e-06, "loss": 2.2617, "step": 9634 }, { "epoch": 3.020376175548589, "grad_norm": 57.90908432006836, "learning_rate": 4.817500000000001e-06, "loss": 2.2734, "step": 9635 }, { "epoch": 3.0206896551724136, "grad_norm": 61.1854248046875, "learning_rate": 4.818e-06, "loss": 2.39, "step": 9636 }, { "epoch": 3.021003134796238, "grad_norm": 6.006817817687988, "learning_rate": 4.8185000000000005e-06, "loss": 1.9233, "step": 9637 }, { "epoch": 3.0213166144200625, "grad_norm": 6.158638954162598, "learning_rate": 4.819e-06, "loss": 1.7945, "step": 9638 }, { "epoch": 3.021630094043887, "grad_norm": 7.1804704666137695, "learning_rate": 4.8195e-06, "loss": 2.1548, "step": 9639 }, { "epoch": 3.0219435736677114, "grad_norm": 38.70308303833008, "learning_rate": 4.8200000000000004e-06, "loss": 2.3152, "step": 9640 }, { "epoch": 3.0222570532915363, "grad_norm": 11.656840324401855, "learning_rate": 4.8205e-06, "loss": 2.0132, "step": 9641 }, { "epoch": 3.0225705329153607, "grad_norm": 7.369824409484863, "learning_rate": 4.821e-06, "loss": 1.8127, "step": 9642 }, { "epoch": 3.022884012539185, "grad_norm": 13.488260269165039, "learning_rate": 4.8215e-06, "loss": 2.1152, "step": 9643 }, { "epoch": 3.0231974921630096, "grad_norm": 5.308684349060059, "learning_rate": 4.822000000000001e-06, "loss": 1.8972, "step": 9644 }, { "epoch": 3.023510971786834, "grad_norm": 35.12510299682617, "learning_rate": 4.822500000000001e-06, "loss": 1.9296, "step": 9645 }, { "epoch": 3.0238244514106585, "grad_norm": 13.186325073242188, "learning_rate": 4.823e-06, "loss": 2.182, "step": 9646 }, { "epoch": 3.024137931034483, "grad_norm": 8.352350234985352, "learning_rate": 4.8235000000000005e-06, "loss": 2.3163, "step": 9647 }, { "epoch": 3.0244514106583074, "grad_norm": 11.199338912963867, "learning_rate": 4.824000000000001e-06, "loss": 2.1714, "step": 9648 }, { "epoch": 3.024764890282132, "grad_norm": 9.293842315673828, "learning_rate": 4.8245e-06, "loss": 2.1022, "step": 9649 }, { "epoch": 3.0250783699059562, "grad_norm": 14.901413917541504, "learning_rate": 4.825e-06, "loss": 1.9706, "step": 9650 }, { "epoch": 3.0253918495297807, "grad_norm": 11.885848045349121, "learning_rate": 4.8255e-06, "loss": 2.0088, "step": 9651 }, { "epoch": 3.025705329153605, "grad_norm": 10.663497924804688, "learning_rate": 4.826e-06, "loss": 2.266, "step": 9652 }, { "epoch": 3.0260188087774296, "grad_norm": 10.967124938964844, "learning_rate": 4.8265e-06, "loss": 2.1294, "step": 9653 }, { "epoch": 3.026332288401254, "grad_norm": 8.360021591186523, "learning_rate": 4.8270000000000005e-06, "loss": 2.1174, "step": 9654 }, { "epoch": 3.0266457680250785, "grad_norm": 243.59866333007812, "learning_rate": 4.827500000000001e-06, "loss": 2.6076, "step": 9655 }, { "epoch": 3.026959247648903, "grad_norm": 7.718048095703125, "learning_rate": 4.828e-06, "loss": 1.8706, "step": 9656 }, { "epoch": 3.0272727272727273, "grad_norm": 11.432957649230957, "learning_rate": 4.8285000000000004e-06, "loss": 2.0047, "step": 9657 }, { "epoch": 3.027586206896552, "grad_norm": 9.70063304901123, "learning_rate": 4.829000000000001e-06, "loss": 1.9112, "step": 9658 }, { "epoch": 3.027899686520376, "grad_norm": 9.284975051879883, "learning_rate": 4.8295e-06, "loss": 1.9603, "step": 9659 }, { "epoch": 3.0282131661442007, "grad_norm": 15.696516036987305, "learning_rate": 4.83e-06, "loss": 2.2704, "step": 9660 }, { "epoch": 3.028526645768025, "grad_norm": 10.030242919921875, "learning_rate": 4.830500000000001e-06, "loss": 2.2335, "step": 9661 }, { "epoch": 3.0288401253918495, "grad_norm": 5.712493419647217, "learning_rate": 4.831e-06, "loss": 1.9627, "step": 9662 }, { "epoch": 3.029153605015674, "grad_norm": 7.336620807647705, "learning_rate": 4.8315e-06, "loss": 2.3243, "step": 9663 }, { "epoch": 3.0294670846394984, "grad_norm": 8.985504150390625, "learning_rate": 4.8320000000000005e-06, "loss": 2.1798, "step": 9664 }, { "epoch": 3.029780564263323, "grad_norm": 13.471259117126465, "learning_rate": 4.832500000000001e-06, "loss": 2.182, "step": 9665 }, { "epoch": 3.0300940438871473, "grad_norm": 9.325427055358887, "learning_rate": 4.833e-06, "loss": 2.3699, "step": 9666 }, { "epoch": 3.0304075235109718, "grad_norm": 9.555724143981934, "learning_rate": 4.8335e-06, "loss": 2.2426, "step": 9667 }, { "epoch": 3.030721003134796, "grad_norm": 6.257678985595703, "learning_rate": 4.834000000000001e-06, "loss": 1.9044, "step": 9668 }, { "epoch": 3.0310344827586206, "grad_norm": 7.913098335266113, "learning_rate": 4.8345e-06, "loss": 1.9233, "step": 9669 }, { "epoch": 3.031347962382445, "grad_norm": 10.691987037658691, "learning_rate": 4.835e-06, "loss": 2.8355, "step": 9670 }, { "epoch": 3.0316614420062695, "grad_norm": 8.5975980758667, "learning_rate": 4.8355000000000005e-06, "loss": 1.9638, "step": 9671 }, { "epoch": 3.031974921630094, "grad_norm": 9.735017776489258, "learning_rate": 4.836e-06, "loss": 2.1568, "step": 9672 }, { "epoch": 3.0322884012539184, "grad_norm": 32.45009231567383, "learning_rate": 4.836500000000001e-06, "loss": 2.3648, "step": 9673 }, { "epoch": 3.032601880877743, "grad_norm": 9.889928817749023, "learning_rate": 4.8370000000000004e-06, "loss": 2.1773, "step": 9674 }, { "epoch": 3.0329153605015673, "grad_norm": 7.009579658508301, "learning_rate": 4.837500000000001e-06, "loss": 1.687, "step": 9675 }, { "epoch": 3.0332288401253917, "grad_norm": 16.89594268798828, "learning_rate": 4.838e-06, "loss": 2.1093, "step": 9676 }, { "epoch": 3.033542319749216, "grad_norm": 5.963584899902344, "learning_rate": 4.8385e-06, "loss": 1.9583, "step": 9677 }, { "epoch": 3.0338557993730406, "grad_norm": 10.985859870910645, "learning_rate": 4.839000000000001e-06, "loss": 1.9426, "step": 9678 }, { "epoch": 3.034169278996865, "grad_norm": 5.285604000091553, "learning_rate": 4.8395e-06, "loss": 2.0601, "step": 9679 }, { "epoch": 3.0344827586206895, "grad_norm": 30.6269588470459, "learning_rate": 4.84e-06, "loss": 2.4097, "step": 9680 }, { "epoch": 3.034796238244514, "grad_norm": 12.806248664855957, "learning_rate": 4.8405000000000005e-06, "loss": 2.2387, "step": 9681 }, { "epoch": 3.0351097178683384, "grad_norm": 11.392602920532227, "learning_rate": 4.841e-06, "loss": 1.9226, "step": 9682 }, { "epoch": 3.035423197492163, "grad_norm": 11.961045265197754, "learning_rate": 4.841500000000001e-06, "loss": 2.3498, "step": 9683 }, { "epoch": 3.0357366771159873, "grad_norm": 14.382877349853516, "learning_rate": 4.842e-06, "loss": 2.5838, "step": 9684 }, { "epoch": 3.0360501567398117, "grad_norm": 54.11642074584961, "learning_rate": 4.842500000000001e-06, "loss": 2.2078, "step": 9685 }, { "epoch": 3.036363636363636, "grad_norm": 39.23794937133789, "learning_rate": 4.843000000000001e-06, "loss": 2.3465, "step": 9686 }, { "epoch": 3.0366771159874606, "grad_norm": 10.329286575317383, "learning_rate": 4.8435e-06, "loss": 1.9982, "step": 9687 }, { "epoch": 3.0369905956112855, "grad_norm": 9.26152229309082, "learning_rate": 4.8440000000000005e-06, "loss": 1.9557, "step": 9688 }, { "epoch": 3.03730407523511, "grad_norm": 8.399603843688965, "learning_rate": 4.8445e-06, "loss": 1.7598, "step": 9689 }, { "epoch": 3.0376175548589344, "grad_norm": 8.128475189208984, "learning_rate": 4.845e-06, "loss": 2.0594, "step": 9690 }, { "epoch": 3.037931034482759, "grad_norm": 20.706310272216797, "learning_rate": 4.8455000000000004e-06, "loss": 2.0807, "step": 9691 }, { "epoch": 3.0382445141065832, "grad_norm": 6.798467636108398, "learning_rate": 4.846e-06, "loss": 2.207, "step": 9692 }, { "epoch": 3.0385579937304077, "grad_norm": 8.023409843444824, "learning_rate": 4.846500000000001e-06, "loss": 2.3793, "step": 9693 }, { "epoch": 3.038871473354232, "grad_norm": 9.03481388092041, "learning_rate": 4.847e-06, "loss": 2.0057, "step": 9694 }, { "epoch": 3.0391849529780566, "grad_norm": 8.092584609985352, "learning_rate": 4.847500000000001e-06, "loss": 2.0009, "step": 9695 }, { "epoch": 3.039498432601881, "grad_norm": 9.910582542419434, "learning_rate": 4.848000000000001e-06, "loss": 2.0671, "step": 9696 }, { "epoch": 3.0398119122257055, "grad_norm": 11.584993362426758, "learning_rate": 4.8485e-06, "loss": 1.9203, "step": 9697 }, { "epoch": 3.04012539184953, "grad_norm": 8.482593536376953, "learning_rate": 4.8490000000000005e-06, "loss": 2.1253, "step": 9698 }, { "epoch": 3.0404388714733543, "grad_norm": 23.642993927001953, "learning_rate": 4.8495e-06, "loss": 2.1881, "step": 9699 }, { "epoch": 3.040752351097179, "grad_norm": 28.937408447265625, "learning_rate": 4.85e-06, "loss": 2.4801, "step": 9700 }, { "epoch": 3.0410658307210032, "grad_norm": 97.05280303955078, "learning_rate": 4.8505e-06, "loss": 2.3679, "step": 9701 }, { "epoch": 3.0413793103448277, "grad_norm": 5.877405643463135, "learning_rate": 4.851e-06, "loss": 2.0249, "step": 9702 }, { "epoch": 3.041692789968652, "grad_norm": 5.934315204620361, "learning_rate": 4.851500000000001e-06, "loss": 2.254, "step": 9703 }, { "epoch": 3.0420062695924766, "grad_norm": 5.877453327178955, "learning_rate": 4.852e-06, "loss": 1.7261, "step": 9704 }, { "epoch": 3.042319749216301, "grad_norm": 10.667168617248535, "learning_rate": 4.8525000000000006e-06, "loss": 1.9509, "step": 9705 }, { "epoch": 3.0426332288401254, "grad_norm": 11.455928802490234, "learning_rate": 4.853000000000001e-06, "loss": 1.7807, "step": 9706 }, { "epoch": 3.04294670846395, "grad_norm": 60.460269927978516, "learning_rate": 4.8535e-06, "loss": 2.215, "step": 9707 }, { "epoch": 3.0432601880877743, "grad_norm": 15.084699630737305, "learning_rate": 4.8540000000000005e-06, "loss": 1.8466, "step": 9708 }, { "epoch": 3.0435736677115988, "grad_norm": 13.102865219116211, "learning_rate": 4.854500000000001e-06, "loss": 2.1286, "step": 9709 }, { "epoch": 3.043887147335423, "grad_norm": 12.340470314025879, "learning_rate": 4.855e-06, "loss": 2.0208, "step": 9710 }, { "epoch": 3.0442006269592476, "grad_norm": 12.026216506958008, "learning_rate": 4.8555e-06, "loss": 2.1644, "step": 9711 }, { "epoch": 3.044514106583072, "grad_norm": 7.588474750518799, "learning_rate": 4.856e-06, "loss": 2.2654, "step": 9712 }, { "epoch": 3.0448275862068965, "grad_norm": 24.675811767578125, "learning_rate": 4.856500000000001e-06, "loss": 2.0972, "step": 9713 }, { "epoch": 3.045141065830721, "grad_norm": 9.338077545166016, "learning_rate": 4.857e-06, "loss": 2.0848, "step": 9714 }, { "epoch": 3.0454545454545454, "grad_norm": 10.304922103881836, "learning_rate": 4.8575000000000005e-06, "loss": 2.3907, "step": 9715 }, { "epoch": 3.04576802507837, "grad_norm": 40.66522216796875, "learning_rate": 4.858000000000001e-06, "loss": 2.2723, "step": 9716 }, { "epoch": 3.0460815047021943, "grad_norm": 9.999760627746582, "learning_rate": 4.8585e-06, "loss": 2.1049, "step": 9717 }, { "epoch": 3.0463949843260187, "grad_norm": 14.931526184082031, "learning_rate": 4.859e-06, "loss": 2.2613, "step": 9718 }, { "epoch": 3.046708463949843, "grad_norm": 19.11994171142578, "learning_rate": 4.859500000000001e-06, "loss": 3.3183, "step": 9719 }, { "epoch": 3.0470219435736676, "grad_norm": 5.8735737800598145, "learning_rate": 4.86e-06, "loss": 1.8347, "step": 9720 }, { "epoch": 3.047335423197492, "grad_norm": 8.620322227478027, "learning_rate": 4.8605e-06, "loss": 2.0844, "step": 9721 }, { "epoch": 3.0476489028213165, "grad_norm": 8.041973114013672, "learning_rate": 4.8610000000000006e-06, "loss": 1.8606, "step": 9722 }, { "epoch": 3.047962382445141, "grad_norm": 16.238475799560547, "learning_rate": 4.861500000000001e-06, "loss": 2.4644, "step": 9723 }, { "epoch": 3.0482758620689654, "grad_norm": 58.11094284057617, "learning_rate": 4.862e-06, "loss": 2.7665, "step": 9724 }, { "epoch": 3.04858934169279, "grad_norm": 9.297150611877441, "learning_rate": 4.8625000000000005e-06, "loss": 2.1894, "step": 9725 }, { "epoch": 3.0489028213166143, "grad_norm": 15.647204399108887, "learning_rate": 4.863000000000001e-06, "loss": 1.9063, "step": 9726 }, { "epoch": 3.0492163009404387, "grad_norm": 10.52225399017334, "learning_rate": 4.8635e-06, "loss": 2.0784, "step": 9727 }, { "epoch": 3.049529780564263, "grad_norm": 175.61654663085938, "learning_rate": 4.864e-06, "loss": 2.2871, "step": 9728 }, { "epoch": 3.0498432601880876, "grad_norm": 11.77319622039795, "learning_rate": 4.864500000000001e-06, "loss": 2.12, "step": 9729 }, { "epoch": 3.050156739811912, "grad_norm": 9.945347785949707, "learning_rate": 4.865e-06, "loss": 2.1219, "step": 9730 }, { "epoch": 3.0504702194357365, "grad_norm": 8.918695449829102, "learning_rate": 4.8655e-06, "loss": 2.2608, "step": 9731 }, { "epoch": 3.050783699059561, "grad_norm": 23.0948543548584, "learning_rate": 4.8660000000000005e-06, "loss": 2.1471, "step": 9732 }, { "epoch": 3.0510971786833854, "grad_norm": 8.93450927734375, "learning_rate": 4.866500000000001e-06, "loss": 2.0317, "step": 9733 }, { "epoch": 3.0514106583072103, "grad_norm": 46.83767318725586, "learning_rate": 4.867000000000001e-06, "loss": 2.1102, "step": 9734 }, { "epoch": 3.0517241379310347, "grad_norm": 55.116912841796875, "learning_rate": 4.8675e-06, "loss": 2.0737, "step": 9735 }, { "epoch": 3.052037617554859, "grad_norm": 14.163415908813477, "learning_rate": 4.868000000000001e-06, "loss": 2.1689, "step": 9736 }, { "epoch": 3.0523510971786836, "grad_norm": 14.96715259552002, "learning_rate": 4.8685e-06, "loss": 2.2389, "step": 9737 }, { "epoch": 3.052664576802508, "grad_norm": 35.29718780517578, "learning_rate": 4.869e-06, "loss": 2.0, "step": 9738 }, { "epoch": 3.0529780564263325, "grad_norm": 12.71137809753418, "learning_rate": 4.8695000000000006e-06, "loss": 2.3323, "step": 9739 }, { "epoch": 3.053291536050157, "grad_norm": 14.414074897766113, "learning_rate": 4.87e-06, "loss": 2.1626, "step": 9740 }, { "epoch": 3.0536050156739813, "grad_norm": 8.662321090698242, "learning_rate": 4.8705e-06, "loss": 2.2165, "step": 9741 }, { "epoch": 3.053918495297806, "grad_norm": 13.033429145812988, "learning_rate": 4.8710000000000005e-06, "loss": 2.5886, "step": 9742 }, { "epoch": 3.0542319749216302, "grad_norm": 15.023706436157227, "learning_rate": 4.871500000000001e-06, "loss": 1.7711, "step": 9743 }, { "epoch": 3.0545454545454547, "grad_norm": 7.542405128479004, "learning_rate": 4.872000000000001e-06, "loss": 2.1319, "step": 9744 }, { "epoch": 3.054858934169279, "grad_norm": 6.37819242477417, "learning_rate": 4.8725e-06, "loss": 2.1317, "step": 9745 }, { "epoch": 3.0551724137931036, "grad_norm": 8.80975341796875, "learning_rate": 4.873000000000001e-06, "loss": 2.0179, "step": 9746 }, { "epoch": 3.055485893416928, "grad_norm": 17.454038619995117, "learning_rate": 4.873500000000001e-06, "loss": 2.218, "step": 9747 }, { "epoch": 3.0557993730407524, "grad_norm": 17.700443267822266, "learning_rate": 4.874e-06, "loss": 2.2745, "step": 9748 }, { "epoch": 3.056112852664577, "grad_norm": 12.918540000915527, "learning_rate": 4.8745000000000005e-06, "loss": 2.1175, "step": 9749 }, { "epoch": 3.0564263322884013, "grad_norm": 10.829713821411133, "learning_rate": 4.875e-06, "loss": 2.0602, "step": 9750 }, { "epoch": 3.0567398119122258, "grad_norm": 10.335433006286621, "learning_rate": 4.8755e-06, "loss": 2.0706, "step": 9751 }, { "epoch": 3.05705329153605, "grad_norm": 70.36122131347656, "learning_rate": 4.876e-06, "loss": 2.1989, "step": 9752 }, { "epoch": 3.0573667711598747, "grad_norm": 7.512118339538574, "learning_rate": 4.8765e-06, "loss": 2.1567, "step": 9753 }, { "epoch": 3.057680250783699, "grad_norm": 9.538561820983887, "learning_rate": 4.877000000000001e-06, "loss": 1.8591, "step": 9754 }, { "epoch": 3.0579937304075235, "grad_norm": 55.565250396728516, "learning_rate": 4.8775e-06, "loss": 2.3316, "step": 9755 }, { "epoch": 3.058307210031348, "grad_norm": 14.550153732299805, "learning_rate": 4.8780000000000006e-06, "loss": 2.5421, "step": 9756 }, { "epoch": 3.0586206896551724, "grad_norm": 6.09670352935791, "learning_rate": 4.878500000000001e-06, "loss": 2.1257, "step": 9757 }, { "epoch": 3.058934169278997, "grad_norm": 11.489277839660645, "learning_rate": 4.879e-06, "loss": 2.3132, "step": 9758 }, { "epoch": 3.0592476489028213, "grad_norm": 10.346170425415039, "learning_rate": 4.8795000000000005e-06, "loss": 2.0037, "step": 9759 }, { "epoch": 3.0595611285266457, "grad_norm": 10.96655559539795, "learning_rate": 4.880000000000001e-06, "loss": 2.1513, "step": 9760 }, { "epoch": 3.05987460815047, "grad_norm": 52.38270950317383, "learning_rate": 4.8805e-06, "loss": 2.0191, "step": 9761 }, { "epoch": 3.0601880877742946, "grad_norm": 9.03677749633789, "learning_rate": 4.881e-06, "loss": 2.0018, "step": 9762 }, { "epoch": 3.060501567398119, "grad_norm": 6.009624481201172, "learning_rate": 4.8815e-06, "loss": 1.8745, "step": 9763 }, { "epoch": 3.0608150470219435, "grad_norm": 31.539417266845703, "learning_rate": 4.882000000000001e-06, "loss": 2.8971, "step": 9764 }, { "epoch": 3.061128526645768, "grad_norm": 10.484251022338867, "learning_rate": 4.8825e-06, "loss": 2.0742, "step": 9765 }, { "epoch": 3.0614420062695924, "grad_norm": 10.369336128234863, "learning_rate": 4.8830000000000005e-06, "loss": 2.0345, "step": 9766 }, { "epoch": 3.061755485893417, "grad_norm": 5.438061237335205, "learning_rate": 4.883500000000001e-06, "loss": 2.074, "step": 9767 }, { "epoch": 3.0620689655172413, "grad_norm": 7.912973403930664, "learning_rate": 4.884e-06, "loss": 2.2623, "step": 9768 }, { "epoch": 3.0623824451410657, "grad_norm": 27.36191749572754, "learning_rate": 4.8845e-06, "loss": 2.2003, "step": 9769 }, { "epoch": 3.06269592476489, "grad_norm": 7.364097595214844, "learning_rate": 4.885000000000001e-06, "loss": 2.1396, "step": 9770 }, { "epoch": 3.0630094043887146, "grad_norm": 6.197089672088623, "learning_rate": 4.8855e-06, "loss": 2.0697, "step": 9771 }, { "epoch": 3.063322884012539, "grad_norm": 11.331680297851562, "learning_rate": 4.886e-06, "loss": 1.9978, "step": 9772 }, { "epoch": 3.0636363636363635, "grad_norm": 10.14016342163086, "learning_rate": 4.8865e-06, "loss": 2.0886, "step": 9773 }, { "epoch": 3.063949843260188, "grad_norm": 9.808079719543457, "learning_rate": 4.887000000000001e-06, "loss": 2.1533, "step": 9774 }, { "epoch": 3.0642633228840124, "grad_norm": 5.3018598556518555, "learning_rate": 4.8875e-06, "loss": 2.2143, "step": 9775 }, { "epoch": 3.064576802507837, "grad_norm": 7.584348678588867, "learning_rate": 4.8880000000000005e-06, "loss": 2.1517, "step": 9776 }, { "epoch": 3.0648902821316613, "grad_norm": 9.004948616027832, "learning_rate": 4.888500000000001e-06, "loss": 2.1759, "step": 9777 }, { "epoch": 3.0652037617554857, "grad_norm": 8.949450492858887, "learning_rate": 4.889e-06, "loss": 2.0321, "step": 9778 }, { "epoch": 3.06551724137931, "grad_norm": 20.952369689941406, "learning_rate": 4.8895e-06, "loss": 2.1433, "step": 9779 }, { "epoch": 3.0658307210031346, "grad_norm": 6.663933277130127, "learning_rate": 4.890000000000001e-06, "loss": 2.0441, "step": 9780 }, { "epoch": 3.066144200626959, "grad_norm": 9.422451972961426, "learning_rate": 4.8905e-06, "loss": 2.2451, "step": 9781 }, { "epoch": 3.066457680250784, "grad_norm": 12.67189884185791, "learning_rate": 4.891e-06, "loss": 2.038, "step": 9782 }, { "epoch": 3.0667711598746084, "grad_norm": 15.784832000732422, "learning_rate": 4.8915000000000005e-06, "loss": 2.1509, "step": 9783 }, { "epoch": 3.067084639498433, "grad_norm": 7.525753021240234, "learning_rate": 4.892000000000001e-06, "loss": 2.2983, "step": 9784 }, { "epoch": 3.0673981191222572, "grad_norm": 10.03464126586914, "learning_rate": 4.8925e-06, "loss": 1.7098, "step": 9785 }, { "epoch": 3.0677115987460817, "grad_norm": 11.838010787963867, "learning_rate": 4.893e-06, "loss": 2.2992, "step": 9786 }, { "epoch": 3.068025078369906, "grad_norm": 14.0263671875, "learning_rate": 4.893500000000001e-06, "loss": 2.3842, "step": 9787 }, { "epoch": 3.0683385579937306, "grad_norm": 7.498491287231445, "learning_rate": 4.894e-06, "loss": 2.0737, "step": 9788 }, { "epoch": 3.068652037617555, "grad_norm": 9.578131675720215, "learning_rate": 4.8945e-06, "loss": 2.2724, "step": 9789 }, { "epoch": 3.0689655172413794, "grad_norm": 38.842185974121094, "learning_rate": 4.8950000000000006e-06, "loss": 1.9705, "step": 9790 }, { "epoch": 3.069278996865204, "grad_norm": 9.945923805236816, "learning_rate": 4.8955e-06, "loss": 1.8413, "step": 9791 }, { "epoch": 3.0695924764890283, "grad_norm": 20.9439697265625, "learning_rate": 4.896e-06, "loss": 2.8266, "step": 9792 }, { "epoch": 3.0699059561128528, "grad_norm": 10.95971393585205, "learning_rate": 4.8965000000000005e-06, "loss": 2.4049, "step": 9793 }, { "epoch": 3.070219435736677, "grad_norm": 24.807018280029297, "learning_rate": 4.897000000000001e-06, "loss": 2.0616, "step": 9794 }, { "epoch": 3.0705329153605017, "grad_norm": 10.41948127746582, "learning_rate": 4.897500000000001e-06, "loss": 2.1478, "step": 9795 }, { "epoch": 3.070846394984326, "grad_norm": 15.943039894104004, "learning_rate": 4.898e-06, "loss": 2.2266, "step": 9796 }, { "epoch": 3.0711598746081505, "grad_norm": 7.002707004547119, "learning_rate": 4.898500000000001e-06, "loss": 1.6419, "step": 9797 }, { "epoch": 3.071473354231975, "grad_norm": 13.97940444946289, "learning_rate": 4.899e-06, "loss": 1.9562, "step": 9798 }, { "epoch": 3.0717868338557994, "grad_norm": 7.476836681365967, "learning_rate": 4.8995e-06, "loss": 2.0545, "step": 9799 }, { "epoch": 3.072100313479624, "grad_norm": 14.416303634643555, "learning_rate": 4.9000000000000005e-06, "loss": 1.7386, "step": 9800 }, { "epoch": 3.0724137931034483, "grad_norm": 7.000453948974609, "learning_rate": 4.9005e-06, "loss": 1.977, "step": 9801 }, { "epoch": 3.0727272727272728, "grad_norm": 24.11768913269043, "learning_rate": 4.901e-06, "loss": 2.2955, "step": 9802 }, { "epoch": 3.073040752351097, "grad_norm": 12.105878829956055, "learning_rate": 4.9015e-06, "loss": 2.1743, "step": 9803 }, { "epoch": 3.0733542319749216, "grad_norm": 18.79386329650879, "learning_rate": 4.902000000000001e-06, "loss": 2.1926, "step": 9804 }, { "epoch": 3.073667711598746, "grad_norm": 5.913775444030762, "learning_rate": 4.902500000000001e-06, "loss": 2.1077, "step": 9805 }, { "epoch": 3.0739811912225705, "grad_norm": 14.90146255493164, "learning_rate": 4.903e-06, "loss": 2.4819, "step": 9806 }, { "epoch": 3.074294670846395, "grad_norm": 23.476722717285156, "learning_rate": 4.9035000000000006e-06, "loss": 2.1495, "step": 9807 }, { "epoch": 3.0746081504702194, "grad_norm": 44.62371826171875, "learning_rate": 4.904000000000001e-06, "loss": 2.3396, "step": 9808 }, { "epoch": 3.074921630094044, "grad_norm": 48.950313568115234, "learning_rate": 4.9045e-06, "loss": 2.7886, "step": 9809 }, { "epoch": 3.0752351097178683, "grad_norm": 31.194311141967773, "learning_rate": 4.9050000000000005e-06, "loss": 1.8162, "step": 9810 }, { "epoch": 3.0755485893416927, "grad_norm": 5.3813934326171875, "learning_rate": 4.9055e-06, "loss": 2.1557, "step": 9811 }, { "epoch": 3.075862068965517, "grad_norm": 15.960404396057129, "learning_rate": 4.906e-06, "loss": 2.3035, "step": 9812 }, { "epoch": 3.0761755485893416, "grad_norm": 12.124314308166504, "learning_rate": 4.9065e-06, "loss": 2.0501, "step": 9813 }, { "epoch": 3.076489028213166, "grad_norm": 9.751871109008789, "learning_rate": 4.907000000000001e-06, "loss": 2.0039, "step": 9814 }, { "epoch": 3.0768025078369905, "grad_norm": 6.956913948059082, "learning_rate": 4.907500000000001e-06, "loss": 2.157, "step": 9815 }, { "epoch": 3.077115987460815, "grad_norm": 8.757210731506348, "learning_rate": 4.908e-06, "loss": 2.1428, "step": 9816 }, { "epoch": 3.0774294670846394, "grad_norm": 15.783744812011719, "learning_rate": 4.9085000000000005e-06, "loss": 2.3823, "step": 9817 }, { "epoch": 3.077742946708464, "grad_norm": 8.385710716247559, "learning_rate": 4.909000000000001e-06, "loss": 2.24, "step": 9818 }, { "epoch": 3.0780564263322883, "grad_norm": 6.7168707847595215, "learning_rate": 4.9095e-06, "loss": 2.0369, "step": 9819 }, { "epoch": 3.0783699059561127, "grad_norm": 6.9227614402771, "learning_rate": 4.9100000000000004e-06, "loss": 2.1438, "step": 9820 }, { "epoch": 3.078683385579937, "grad_norm": 8.871114730834961, "learning_rate": 4.910500000000001e-06, "loss": 2.8965, "step": 9821 }, { "epoch": 3.0789968652037616, "grad_norm": 15.077126502990723, "learning_rate": 4.911e-06, "loss": 2.0417, "step": 9822 }, { "epoch": 3.079310344827586, "grad_norm": 19.72150993347168, "learning_rate": 4.9115e-06, "loss": 2.6255, "step": 9823 }, { "epoch": 3.0796238244514105, "grad_norm": 8.059910774230957, "learning_rate": 4.9120000000000006e-06, "loss": 2.1185, "step": 9824 }, { "epoch": 3.079937304075235, "grad_norm": 25.7774600982666, "learning_rate": 4.912500000000001e-06, "loss": 2.4799, "step": 9825 }, { "epoch": 3.0802507836990594, "grad_norm": 144.56661987304688, "learning_rate": 4.913e-06, "loss": 2.3008, "step": 9826 }, { "epoch": 3.080564263322884, "grad_norm": 11.638893127441406, "learning_rate": 4.9135000000000005e-06, "loss": 2.1672, "step": 9827 }, { "epoch": 3.0808777429467087, "grad_norm": 14.369083404541016, "learning_rate": 4.914000000000001e-06, "loss": 2.0765, "step": 9828 }, { "epoch": 3.081191222570533, "grad_norm": 8.76433277130127, "learning_rate": 4.9145e-06, "loss": 2.1652, "step": 9829 }, { "epoch": 3.0815047021943576, "grad_norm": 9.549826622009277, "learning_rate": 4.915e-06, "loss": 1.8765, "step": 9830 }, { "epoch": 3.081818181818182, "grad_norm": 8.204703330993652, "learning_rate": 4.915500000000001e-06, "loss": 2.2977, "step": 9831 }, { "epoch": 3.0821316614420065, "grad_norm": 7.688598155975342, "learning_rate": 4.916e-06, "loss": 2.0463, "step": 9832 }, { "epoch": 3.082445141065831, "grad_norm": 12.627134323120117, "learning_rate": 4.9165e-06, "loss": 1.9903, "step": 9833 }, { "epoch": 3.0827586206896553, "grad_norm": 5.122074127197266, "learning_rate": 4.9170000000000005e-06, "loss": 2.0495, "step": 9834 }, { "epoch": 3.08307210031348, "grad_norm": 7.45738410949707, "learning_rate": 4.917500000000001e-06, "loss": 2.2233, "step": 9835 }, { "epoch": 3.083385579937304, "grad_norm": 66.76324462890625, "learning_rate": 4.918e-06, "loss": 2.0425, "step": 9836 }, { "epoch": 3.0836990595611287, "grad_norm": 11.138689041137695, "learning_rate": 4.9185000000000004e-06, "loss": 1.9883, "step": 9837 }, { "epoch": 3.084012539184953, "grad_norm": 8.100717544555664, "learning_rate": 4.919000000000001e-06, "loss": 2.1425, "step": 9838 }, { "epoch": 3.0843260188087775, "grad_norm": 78.62770080566406, "learning_rate": 4.9195e-06, "loss": 2.2242, "step": 9839 }, { "epoch": 3.084639498432602, "grad_norm": 6.230515480041504, "learning_rate": 4.92e-06, "loss": 1.9937, "step": 9840 }, { "epoch": 3.0849529780564264, "grad_norm": 84.15657043457031, "learning_rate": 4.920500000000001e-06, "loss": 2.7476, "step": 9841 }, { "epoch": 3.085266457680251, "grad_norm": 104.37451171875, "learning_rate": 4.921e-06, "loss": 1.9644, "step": 9842 }, { "epoch": 3.0855799373040753, "grad_norm": 11.789735794067383, "learning_rate": 4.9215e-06, "loss": 1.957, "step": 9843 }, { "epoch": 3.0858934169278998, "grad_norm": 9.657096862792969, "learning_rate": 4.9220000000000005e-06, "loss": 2.1026, "step": 9844 }, { "epoch": 3.086206896551724, "grad_norm": 93.80183410644531, "learning_rate": 4.922500000000001e-06, "loss": 2.4012, "step": 9845 }, { "epoch": 3.0865203761755486, "grad_norm": 8.737438201904297, "learning_rate": 4.923000000000001e-06, "loss": 2.1393, "step": 9846 }, { "epoch": 3.086833855799373, "grad_norm": 34.829593658447266, "learning_rate": 4.9235e-06, "loss": 2.4256, "step": 9847 }, { "epoch": 3.0871473354231975, "grad_norm": 57.01554870605469, "learning_rate": 4.924000000000001e-06, "loss": 2.3224, "step": 9848 }, { "epoch": 3.087460815047022, "grad_norm": 29.754079818725586, "learning_rate": 4.9245e-06, "loss": 2.2403, "step": 9849 }, { "epoch": 3.0877742946708464, "grad_norm": 9.498865127563477, "learning_rate": 4.925e-06, "loss": 2.0926, "step": 9850 }, { "epoch": 3.088087774294671, "grad_norm": 12.558642387390137, "learning_rate": 4.9255000000000005e-06, "loss": 2.4338, "step": 9851 }, { "epoch": 3.0884012539184953, "grad_norm": 32.99839401245117, "learning_rate": 4.926e-06, "loss": 1.8569, "step": 9852 }, { "epoch": 3.0887147335423197, "grad_norm": 9.631637573242188, "learning_rate": 4.9265e-06, "loss": 2.1864, "step": 9853 }, { "epoch": 3.089028213166144, "grad_norm": 10.247138023376465, "learning_rate": 4.9270000000000004e-06, "loss": 2.1798, "step": 9854 }, { "epoch": 3.0893416927899686, "grad_norm": 5.919518947601318, "learning_rate": 4.927500000000001e-06, "loss": 2.0948, "step": 9855 }, { "epoch": 3.089655172413793, "grad_norm": 55.71178436279297, "learning_rate": 4.928000000000001e-06, "loss": 2.3, "step": 9856 }, { "epoch": 3.0899686520376175, "grad_norm": 16.308521270751953, "learning_rate": 4.9285e-06, "loss": 2.2252, "step": 9857 }, { "epoch": 3.090282131661442, "grad_norm": 12.91220760345459, "learning_rate": 4.929000000000001e-06, "loss": 2.2088, "step": 9858 }, { "epoch": 3.0905956112852664, "grad_norm": 7.857017993927002, "learning_rate": 4.9295e-06, "loss": 2.4743, "step": 9859 }, { "epoch": 3.090909090909091, "grad_norm": 7.526261329650879, "learning_rate": 4.93e-06, "loss": 2.1065, "step": 9860 }, { "epoch": 3.0912225705329153, "grad_norm": 8.914759635925293, "learning_rate": 4.9305000000000005e-06, "loss": 2.1379, "step": 9861 }, { "epoch": 3.0915360501567397, "grad_norm": 17.24327278137207, "learning_rate": 4.931e-06, "loss": 2.6773, "step": 9862 }, { "epoch": 3.091849529780564, "grad_norm": 29.173786163330078, "learning_rate": 4.9315e-06, "loss": 2.5098, "step": 9863 }, { "epoch": 3.0921630094043886, "grad_norm": 6.834757328033447, "learning_rate": 4.932e-06, "loss": 2.2227, "step": 9864 }, { "epoch": 3.092476489028213, "grad_norm": 10.59632396697998, "learning_rate": 4.932500000000001e-06, "loss": 2.3807, "step": 9865 }, { "epoch": 3.0927899686520375, "grad_norm": 9.522377967834473, "learning_rate": 4.933000000000001e-06, "loss": 2.2056, "step": 9866 }, { "epoch": 3.093103448275862, "grad_norm": 9.066191673278809, "learning_rate": 4.9335e-06, "loss": 2.3421, "step": 9867 }, { "epoch": 3.0934169278996864, "grad_norm": 9.087063789367676, "learning_rate": 4.9340000000000005e-06, "loss": 2.0849, "step": 9868 }, { "epoch": 3.093730407523511, "grad_norm": 11.694283485412598, "learning_rate": 4.934500000000001e-06, "loss": 2.6071, "step": 9869 }, { "epoch": 3.0940438871473352, "grad_norm": 14.85091781616211, "learning_rate": 4.935e-06, "loss": 2.1884, "step": 9870 }, { "epoch": 3.0943573667711597, "grad_norm": 12.765669822692871, "learning_rate": 4.9355000000000004e-06, "loss": 2.103, "step": 9871 }, { "epoch": 3.094670846394984, "grad_norm": 7.0769829750061035, "learning_rate": 4.936e-06, "loss": 1.9826, "step": 9872 }, { "epoch": 3.0949843260188086, "grad_norm": 11.035401344299316, "learning_rate": 4.9365e-06, "loss": 2.2463, "step": 9873 }, { "epoch": 3.095297805642633, "grad_norm": 10.488544464111328, "learning_rate": 4.937e-06, "loss": 1.9462, "step": 9874 }, { "epoch": 3.0956112852664575, "grad_norm": 9.020281791687012, "learning_rate": 4.937500000000001e-06, "loss": 2.1404, "step": 9875 }, { "epoch": 3.0959247648902823, "grad_norm": 12.644916534423828, "learning_rate": 4.938000000000001e-06, "loss": 2.1545, "step": 9876 }, { "epoch": 3.096238244514107, "grad_norm": 6.636775493621826, "learning_rate": 4.9385e-06, "loss": 2.089, "step": 9877 }, { "epoch": 3.0965517241379312, "grad_norm": 9.068178176879883, "learning_rate": 4.9390000000000005e-06, "loss": 2.2043, "step": 9878 }, { "epoch": 3.0968652037617557, "grad_norm": 10.24074649810791, "learning_rate": 4.939500000000001e-06, "loss": 2.0813, "step": 9879 }, { "epoch": 3.09717868338558, "grad_norm": 7.001710414886475, "learning_rate": 4.94e-06, "loss": 2.019, "step": 9880 }, { "epoch": 3.0974921630094046, "grad_norm": 28.04688835144043, "learning_rate": 4.9405e-06, "loss": 2.1886, "step": 9881 }, { "epoch": 3.097805642633229, "grad_norm": 10.570974349975586, "learning_rate": 4.941000000000001e-06, "loss": 1.8986, "step": 9882 }, { "epoch": 3.0981191222570534, "grad_norm": 7.896426200866699, "learning_rate": 4.9415e-06, "loss": 2.0892, "step": 9883 }, { "epoch": 3.098432601880878, "grad_norm": 10.25847053527832, "learning_rate": 4.942e-06, "loss": 2.1231, "step": 9884 }, { "epoch": 3.0987460815047023, "grad_norm": 11.887818336486816, "learning_rate": 4.9425000000000005e-06, "loss": 2.1828, "step": 9885 }, { "epoch": 3.0990595611285268, "grad_norm": 75.09355163574219, "learning_rate": 4.943000000000001e-06, "loss": 2.2029, "step": 9886 }, { "epoch": 3.099373040752351, "grad_norm": 9.202438354492188, "learning_rate": 4.9435e-06, "loss": 1.9979, "step": 9887 }, { "epoch": 3.0996865203761756, "grad_norm": 81.01618194580078, "learning_rate": 4.9440000000000004e-06, "loss": 2.1951, "step": 9888 }, { "epoch": 3.1, "grad_norm": 7.649423122406006, "learning_rate": 4.944500000000001e-06, "loss": 2.3626, "step": 9889 }, { "epoch": 3.1003134796238245, "grad_norm": 8.637659072875977, "learning_rate": 4.945e-06, "loss": 2.0925, "step": 9890 }, { "epoch": 3.100626959247649, "grad_norm": 7.462778091430664, "learning_rate": 4.9455e-06, "loss": 2.0404, "step": 9891 }, { "epoch": 3.1009404388714734, "grad_norm": 5.910890102386475, "learning_rate": 4.946000000000001e-06, "loss": 2.2218, "step": 9892 }, { "epoch": 3.101253918495298, "grad_norm": 242.232666015625, "learning_rate": 4.9465e-06, "loss": 2.1452, "step": 9893 }, { "epoch": 3.1015673981191223, "grad_norm": 125.6268310546875, "learning_rate": 4.947e-06, "loss": 2.2952, "step": 9894 }, { "epoch": 3.1018808777429467, "grad_norm": 6.68086576461792, "learning_rate": 4.9475000000000005e-06, "loss": 1.9514, "step": 9895 }, { "epoch": 3.102194357366771, "grad_norm": 242.34361267089844, "learning_rate": 4.948000000000001e-06, "loss": 2.5085, "step": 9896 }, { "epoch": 3.1025078369905956, "grad_norm": 11.006869316101074, "learning_rate": 4.9485e-06, "loss": 2.1704, "step": 9897 }, { "epoch": 3.10282131661442, "grad_norm": 4.901587963104248, "learning_rate": 4.949e-06, "loss": 2.0534, "step": 9898 }, { "epoch": 3.1031347962382445, "grad_norm": 14.442913055419922, "learning_rate": 4.949500000000001e-06, "loss": 2.0771, "step": 9899 }, { "epoch": 3.103448275862069, "grad_norm": 13.190855979919434, "learning_rate": 4.95e-06, "loss": 2.2036, "step": 9900 }, { "epoch": 3.1037617554858934, "grad_norm": 7.623360633850098, "learning_rate": 4.9505e-06, "loss": 1.8148, "step": 9901 }, { "epoch": 3.104075235109718, "grad_norm": 10.675677299499512, "learning_rate": 4.9510000000000005e-06, "loss": 1.8716, "step": 9902 }, { "epoch": 3.1043887147335423, "grad_norm": 12.488614082336426, "learning_rate": 4.9515e-06, "loss": 2.1638, "step": 9903 }, { "epoch": 3.1047021943573667, "grad_norm": 20.32207489013672, "learning_rate": 4.952e-06, "loss": 2.8587, "step": 9904 }, { "epoch": 3.105015673981191, "grad_norm": 6.231513023376465, "learning_rate": 4.9525000000000004e-06, "loss": 1.8882, "step": 9905 }, { "epoch": 3.1053291536050156, "grad_norm": 14.91689682006836, "learning_rate": 4.953000000000001e-06, "loss": 1.9912, "step": 9906 }, { "epoch": 3.10564263322884, "grad_norm": 7.764899253845215, "learning_rate": 4.953500000000001e-06, "loss": 2.1565, "step": 9907 }, { "epoch": 3.1059561128526645, "grad_norm": 8.366259574890137, "learning_rate": 4.954e-06, "loss": 2.2286, "step": 9908 }, { "epoch": 3.106269592476489, "grad_norm": 8.350461959838867, "learning_rate": 4.954500000000001e-06, "loss": 2.1515, "step": 9909 }, { "epoch": 3.1065830721003134, "grad_norm": 8.524900436401367, "learning_rate": 4.955e-06, "loss": 2.0292, "step": 9910 }, { "epoch": 3.106896551724138, "grad_norm": 17.34325408935547, "learning_rate": 4.9555e-06, "loss": 1.9264, "step": 9911 }, { "epoch": 3.1072100313479623, "grad_norm": 7.175546169281006, "learning_rate": 4.9560000000000005e-06, "loss": 2.2046, "step": 9912 }, { "epoch": 3.1075235109717867, "grad_norm": 64.89925384521484, "learning_rate": 4.9565e-06, "loss": 2.5289, "step": 9913 }, { "epoch": 3.107836990595611, "grad_norm": 12.05226993560791, "learning_rate": 4.957e-06, "loss": 2.1439, "step": 9914 }, { "epoch": 3.1081504702194356, "grad_norm": 11.295367240905762, "learning_rate": 4.9575e-06, "loss": 2.0417, "step": 9915 }, { "epoch": 3.10846394984326, "grad_norm": 4.492518901824951, "learning_rate": 4.958000000000001e-06, "loss": 2.1992, "step": 9916 }, { "epoch": 3.1087774294670845, "grad_norm": 14.11316204071045, "learning_rate": 4.958500000000001e-06, "loss": 2.0748, "step": 9917 }, { "epoch": 3.109090909090909, "grad_norm": 29.50138282775879, "learning_rate": 4.959e-06, "loss": 2.1293, "step": 9918 }, { "epoch": 3.1094043887147333, "grad_norm": 54.70607376098633, "learning_rate": 4.9595000000000005e-06, "loss": 2.0965, "step": 9919 }, { "epoch": 3.109717868338558, "grad_norm": 8.138368606567383, "learning_rate": 4.960000000000001e-06, "loss": 2.0705, "step": 9920 }, { "epoch": 3.1100313479623822, "grad_norm": 9.819841384887695, "learning_rate": 4.9605e-06, "loss": 2.0208, "step": 9921 }, { "epoch": 3.110344827586207, "grad_norm": 8.867249488830566, "learning_rate": 4.9610000000000004e-06, "loss": 2.0535, "step": 9922 }, { "epoch": 3.1106583072100316, "grad_norm": 43.44057846069336, "learning_rate": 4.9615e-06, "loss": 2.1667, "step": 9923 }, { "epoch": 3.110971786833856, "grad_norm": 10.550017356872559, "learning_rate": 4.962e-06, "loss": 2.1053, "step": 9924 }, { "epoch": 3.1112852664576804, "grad_norm": 8.557412147521973, "learning_rate": 4.9625e-06, "loss": 1.9494, "step": 9925 }, { "epoch": 3.111598746081505, "grad_norm": 31.773256301879883, "learning_rate": 4.963000000000001e-06, "loss": 1.8103, "step": 9926 }, { "epoch": 3.1119122257053293, "grad_norm": 13.767291069030762, "learning_rate": 4.963500000000001e-06, "loss": 2.1346, "step": 9927 }, { "epoch": 3.1122257053291538, "grad_norm": 28.15506362915039, "learning_rate": 4.964e-06, "loss": 2.1307, "step": 9928 }, { "epoch": 3.112539184952978, "grad_norm": 14.563071250915527, "learning_rate": 4.9645000000000005e-06, "loss": 2.3161, "step": 9929 }, { "epoch": 3.1128526645768027, "grad_norm": 7.094024181365967, "learning_rate": 4.965000000000001e-06, "loss": 2.016, "step": 9930 }, { "epoch": 3.113166144200627, "grad_norm": 6.884676933288574, "learning_rate": 4.9655e-06, "loss": 2.1129, "step": 9931 }, { "epoch": 3.1134796238244515, "grad_norm": 6.481064319610596, "learning_rate": 4.966e-06, "loss": 2.1619, "step": 9932 }, { "epoch": 3.113793103448276, "grad_norm": 8.408164978027344, "learning_rate": 4.9665e-06, "loss": 1.8933, "step": 9933 }, { "epoch": 3.1141065830721004, "grad_norm": 8.44009017944336, "learning_rate": 4.967e-06, "loss": 2.0744, "step": 9934 }, { "epoch": 3.114420062695925, "grad_norm": 7.526889801025391, "learning_rate": 4.9675e-06, "loss": 2.0943, "step": 9935 }, { "epoch": 3.1147335423197493, "grad_norm": 8.354063987731934, "learning_rate": 4.9680000000000005e-06, "loss": 2.006, "step": 9936 }, { "epoch": 3.1150470219435737, "grad_norm": 10.797410011291504, "learning_rate": 4.968500000000001e-06, "loss": 1.9995, "step": 9937 }, { "epoch": 3.115360501567398, "grad_norm": 37.709468841552734, "learning_rate": 4.969e-06, "loss": 2.1249, "step": 9938 }, { "epoch": 3.1156739811912226, "grad_norm": 13.43321418762207, "learning_rate": 4.9695000000000004e-06, "loss": 1.8984, "step": 9939 }, { "epoch": 3.115987460815047, "grad_norm": 50.003883361816406, "learning_rate": 4.970000000000001e-06, "loss": 2.3696, "step": 9940 }, { "epoch": 3.1163009404388715, "grad_norm": 6.840349197387695, "learning_rate": 4.9705e-06, "loss": 2.0144, "step": 9941 }, { "epoch": 3.116614420062696, "grad_norm": 9.299572944641113, "learning_rate": 4.971e-06, "loss": 2.3369, "step": 9942 }, { "epoch": 3.1169278996865204, "grad_norm": 11.999553680419922, "learning_rate": 4.971500000000001e-06, "loss": 2.0707, "step": 9943 }, { "epoch": 3.117241379310345, "grad_norm": 11.424012184143066, "learning_rate": 4.972e-06, "loss": 2.0969, "step": 9944 }, { "epoch": 3.1175548589341693, "grad_norm": 103.85692596435547, "learning_rate": 4.9725e-06, "loss": 2.4661, "step": 9945 }, { "epoch": 3.1178683385579937, "grad_norm": 10.616049766540527, "learning_rate": 4.9730000000000005e-06, "loss": 1.9998, "step": 9946 }, { "epoch": 3.118181818181818, "grad_norm": 26.419404983520508, "learning_rate": 4.973500000000001e-06, "loss": 2.6809, "step": 9947 }, { "epoch": 3.1184952978056426, "grad_norm": 11.97441577911377, "learning_rate": 4.974e-06, "loss": 2.0419, "step": 9948 }, { "epoch": 3.118808777429467, "grad_norm": 7.687306880950928, "learning_rate": 4.9745e-06, "loss": 1.8797, "step": 9949 }, { "epoch": 3.1191222570532915, "grad_norm": 9.139175415039062, "learning_rate": 4.975000000000001e-06, "loss": 2.0928, "step": 9950 }, { "epoch": 3.119435736677116, "grad_norm": 12.448719024658203, "learning_rate": 4.9755e-06, "loss": 1.9472, "step": 9951 }, { "epoch": 3.1197492163009404, "grad_norm": 35.87437057495117, "learning_rate": 4.976e-06, "loss": 2.5284, "step": 9952 }, { "epoch": 3.120062695924765, "grad_norm": 9.699675559997559, "learning_rate": 4.9765000000000005e-06, "loss": 1.9687, "step": 9953 }, { "epoch": 3.1203761755485893, "grad_norm": 5.603664875030518, "learning_rate": 4.977e-06, "loss": 2.1327, "step": 9954 }, { "epoch": 3.1206896551724137, "grad_norm": 5.152953624725342, "learning_rate": 4.977500000000001e-06, "loss": 2.0747, "step": 9955 }, { "epoch": 3.121003134796238, "grad_norm": 12.962015151977539, "learning_rate": 4.9780000000000005e-06, "loss": 2.0748, "step": 9956 }, { "epoch": 3.1213166144200626, "grad_norm": 10.14980697631836, "learning_rate": 4.978500000000001e-06, "loss": 2.2362, "step": 9957 }, { "epoch": 3.121630094043887, "grad_norm": 84.37296295166016, "learning_rate": 4.979e-06, "loss": 2.2453, "step": 9958 }, { "epoch": 3.1219435736677115, "grad_norm": 104.06275939941406, "learning_rate": 4.9795e-06, "loss": 2.9201, "step": 9959 }, { "epoch": 3.122257053291536, "grad_norm": 6.740980625152588, "learning_rate": 4.980000000000001e-06, "loss": 2.1949, "step": 9960 }, { "epoch": 3.1225705329153604, "grad_norm": 11.904256820678711, "learning_rate": 4.9805e-06, "loss": 2.0978, "step": 9961 }, { "epoch": 3.122884012539185, "grad_norm": 14.267485618591309, "learning_rate": 4.981e-06, "loss": 2.2807, "step": 9962 }, { "epoch": 3.1231974921630092, "grad_norm": 9.122551918029785, "learning_rate": 4.9815000000000005e-06, "loss": 2.1435, "step": 9963 }, { "epoch": 3.1235109717868337, "grad_norm": 6.975362777709961, "learning_rate": 4.982e-06, "loss": 2.0626, "step": 9964 }, { "epoch": 3.123824451410658, "grad_norm": 18.900487899780273, "learning_rate": 4.982500000000001e-06, "loss": 2.6774, "step": 9965 }, { "epoch": 3.1241379310344826, "grad_norm": 14.826176643371582, "learning_rate": 4.983e-06, "loss": 2.3087, "step": 9966 }, { "epoch": 3.124451410658307, "grad_norm": 10.319009780883789, "learning_rate": 4.983500000000001e-06, "loss": 1.798, "step": 9967 }, { "epoch": 3.1247648902821314, "grad_norm": 28.876853942871094, "learning_rate": 4.984000000000001e-06, "loss": 2.0562, "step": 9968 }, { "epoch": 3.125078369905956, "grad_norm": 5.572143077850342, "learning_rate": 4.9845e-06, "loss": 2.2628, "step": 9969 }, { "epoch": 3.1253918495297803, "grad_norm": 16.015830993652344, "learning_rate": 4.9850000000000006e-06, "loss": 2.2064, "step": 9970 }, { "epoch": 3.125705329153605, "grad_norm": 10.716500282287598, "learning_rate": 4.9855e-06, "loss": 2.15, "step": 9971 }, { "epoch": 3.1260188087774297, "grad_norm": 9.304824829101562, "learning_rate": 4.986e-06, "loss": 2.0146, "step": 9972 }, { "epoch": 3.126332288401254, "grad_norm": 6.45346212387085, "learning_rate": 4.9865000000000005e-06, "loss": 2.0679, "step": 9973 }, { "epoch": 3.1266457680250785, "grad_norm": 12.23961353302002, "learning_rate": 4.987e-06, "loss": 2.2256, "step": 9974 }, { "epoch": 3.126959247648903, "grad_norm": 7.018451690673828, "learning_rate": 4.987500000000001e-06, "loss": 1.8775, "step": 9975 }, { "epoch": 3.1272727272727274, "grad_norm": 21.118349075317383, "learning_rate": 4.988e-06, "loss": 2.5563, "step": 9976 }, { "epoch": 3.127586206896552, "grad_norm": 132.11463928222656, "learning_rate": 4.988500000000001e-06, "loss": 2.5035, "step": 9977 }, { "epoch": 3.1278996865203763, "grad_norm": 6.647562503814697, "learning_rate": 4.989000000000001e-06, "loss": 1.9522, "step": 9978 }, { "epoch": 3.1282131661442008, "grad_norm": 15.163290023803711, "learning_rate": 4.9895e-06, "loss": 2.0591, "step": 9979 }, { "epoch": 3.128526645768025, "grad_norm": 12.324861526489258, "learning_rate": 4.9900000000000005e-06, "loss": 1.9417, "step": 9980 }, { "epoch": 3.1288401253918496, "grad_norm": 27.122323989868164, "learning_rate": 4.990500000000001e-06, "loss": 2.1848, "step": 9981 }, { "epoch": 3.129153605015674, "grad_norm": 5.48691463470459, "learning_rate": 4.991e-06, "loss": 2.0786, "step": 9982 }, { "epoch": 3.1294670846394985, "grad_norm": 13.538262367248535, "learning_rate": 4.9915e-06, "loss": 2.0161, "step": 9983 }, { "epoch": 3.129780564263323, "grad_norm": 12.583460807800293, "learning_rate": 4.992e-06, "loss": 2.1737, "step": 9984 }, { "epoch": 3.1300940438871474, "grad_norm": 9.065089225769043, "learning_rate": 4.992500000000001e-06, "loss": 2.1601, "step": 9985 }, { "epoch": 3.130407523510972, "grad_norm": 5.959442615509033, "learning_rate": 4.993e-06, "loss": 2.1382, "step": 9986 }, { "epoch": 3.1307210031347963, "grad_norm": 99.51443481445312, "learning_rate": 4.9935000000000006e-06, "loss": 2.5984, "step": 9987 }, { "epoch": 3.1310344827586207, "grad_norm": 5.98027229309082, "learning_rate": 4.994000000000001e-06, "loss": 2.0561, "step": 9988 }, { "epoch": 3.131347962382445, "grad_norm": 30.842748641967773, "learning_rate": 4.9945e-06, "loss": 2.0603, "step": 9989 }, { "epoch": 3.1316614420062696, "grad_norm": 27.914281845092773, "learning_rate": 4.9950000000000005e-06, "loss": 2.0877, "step": 9990 }, { "epoch": 3.131974921630094, "grad_norm": 9.986546516418457, "learning_rate": 4.995500000000001e-06, "loss": 2.2571, "step": 9991 }, { "epoch": 3.1322884012539185, "grad_norm": 8.913918495178223, "learning_rate": 4.996e-06, "loss": 1.8525, "step": 9992 }, { "epoch": 3.132601880877743, "grad_norm": 7.258603572845459, "learning_rate": 4.9965e-06, "loss": 2.0643, "step": 9993 }, { "epoch": 3.1329153605015674, "grad_norm": 10.037948608398438, "learning_rate": 4.997000000000001e-06, "loss": 2.1347, "step": 9994 }, { "epoch": 3.133228840125392, "grad_norm": 44.09088897705078, "learning_rate": 4.997500000000001e-06, "loss": 2.3415, "step": 9995 }, { "epoch": 3.1335423197492163, "grad_norm": 17.717144012451172, "learning_rate": 4.998e-06, "loss": 2.194, "step": 9996 }, { "epoch": 3.1338557993730407, "grad_norm": 61.563323974609375, "learning_rate": 4.9985000000000005e-06, "loss": 2.4128, "step": 9997 }, { "epoch": 3.134169278996865, "grad_norm": 10.381364822387695, "learning_rate": 4.999000000000001e-06, "loss": 2.2399, "step": 9998 }, { "epoch": 3.1344827586206896, "grad_norm": 7.467041969299316, "learning_rate": 4.9995e-06, "loss": 2.3753, "step": 9999 }, { "epoch": 3.134796238244514, "grad_norm": 29.334957122802734, "learning_rate": 5e-06, "loss": 2.0503, "step": 10000 }, { "epoch": 3.1351097178683385, "grad_norm": 14.755563735961914, "learning_rate": 4.99999838046049e-06, "loss": 2.2375, "step": 10001 }, { "epoch": 3.135423197492163, "grad_norm": 10.341009140014648, "learning_rate": 4.99999352184406e-06, "loss": 2.15, "step": 10002 }, { "epoch": 3.1357366771159874, "grad_norm": 8.07086181640625, "learning_rate": 4.999985424157003e-06, "loss": 2.1543, "step": 10003 }, { "epoch": 3.136050156739812, "grad_norm": 83.61296081542969, "learning_rate": 4.999974087409812e-06, "loss": 2.409, "step": 10004 }, { "epoch": 3.1363636363636362, "grad_norm": 6.419376850128174, "learning_rate": 4.999959511617173e-06, "loss": 1.9869, "step": 10005 }, { "epoch": 3.1366771159874607, "grad_norm": 10.515169143676758, "learning_rate": 4.999941696797974e-06, "loss": 2.3019, "step": 10006 }, { "epoch": 3.136990595611285, "grad_norm": 11.460936546325684, "learning_rate": 4.999920642975294e-06, "loss": 2.0551, "step": 10007 }, { "epoch": 3.1373040752351096, "grad_norm": 19.244932174682617, "learning_rate": 4.999896350176413e-06, "loss": 2.0006, "step": 10008 }, { "epoch": 3.137617554858934, "grad_norm": 8.476037979125977, "learning_rate": 4.9998688184328046e-06, "loss": 2.1243, "step": 10009 }, { "epoch": 3.1379310344827585, "grad_norm": 19.45748519897461, "learning_rate": 4.999838047780139e-06, "loss": 2.3975, "step": 10010 }, { "epoch": 3.138244514106583, "grad_norm": 46.30350112915039, "learning_rate": 4.9998040382582845e-06, "loss": 2.2889, "step": 10011 }, { "epoch": 3.1385579937304073, "grad_norm": 7.7385640144348145, "learning_rate": 4.9997667899113055e-06, "loss": 2.1901, "step": 10012 }, { "epoch": 3.138871473354232, "grad_norm": 25.109907150268555, "learning_rate": 4.999726302787461e-06, "loss": 2.1211, "step": 10013 }, { "epoch": 3.139184952978056, "grad_norm": 5.80121374130249, "learning_rate": 4.999682576939208e-06, "loss": 2.0747, "step": 10014 }, { "epoch": 3.139498432601881, "grad_norm": 13.7593355178833, "learning_rate": 4.999635612423198e-06, "loss": 2.0536, "step": 10015 }, { "epoch": 3.1398119122257055, "grad_norm": 14.891043663024902, "learning_rate": 4.999585409300281e-06, "loss": 2.1567, "step": 10016 }, { "epoch": 3.14012539184953, "grad_norm": 12.14987850189209, "learning_rate": 4.9995319676355015e-06, "loss": 2.2793, "step": 10017 }, { "epoch": 3.1404388714733544, "grad_norm": 9.014960289001465, "learning_rate": 4.9994752874981e-06, "loss": 2.1613, "step": 10018 }, { "epoch": 3.140752351097179, "grad_norm": 6.561460494995117, "learning_rate": 4.9994153689615135e-06, "loss": 2.1155, "step": 10019 }, { "epoch": 3.1410658307210033, "grad_norm": 74.59242248535156, "learning_rate": 4.999352212103373e-06, "loss": 2.2201, "step": 10020 }, { "epoch": 3.1413793103448278, "grad_norm": 11.191134452819824, "learning_rate": 4.999285817005508e-06, "loss": 2.0924, "step": 10021 }, { "epoch": 3.141692789968652, "grad_norm": 10.761069297790527, "learning_rate": 4.999216183753942e-06, "loss": 2.0584, "step": 10022 }, { "epoch": 3.1420062695924766, "grad_norm": 9.08517074584961, "learning_rate": 4.999143312438893e-06, "loss": 2.1136, "step": 10023 }, { "epoch": 3.142319749216301, "grad_norm": 6.125099182128906, "learning_rate": 4.999067203154777e-06, "loss": 2.1047, "step": 10024 }, { "epoch": 3.1426332288401255, "grad_norm": 7.822919845581055, "learning_rate": 4.998987856000202e-06, "loss": 2.0325, "step": 10025 }, { "epoch": 3.14294670846395, "grad_norm": 7.793244361877441, "learning_rate": 4.9989052710779735e-06, "loss": 2.2022, "step": 10026 }, { "epoch": 3.1432601880877744, "grad_norm": 24.539392471313477, "learning_rate": 4.9988194484950905e-06, "loss": 2.4543, "step": 10027 }, { "epoch": 3.143573667711599, "grad_norm": 10.012564659118652, "learning_rate": 4.9987303883627484e-06, "loss": 2.2015, "step": 10028 }, { "epoch": 3.1438871473354233, "grad_norm": 4.804534435272217, "learning_rate": 4.998638090796335e-06, "loss": 1.8954, "step": 10029 }, { "epoch": 3.1442006269592477, "grad_norm": 7.293290138244629, "learning_rate": 4.998542555915435e-06, "loss": 2.0457, "step": 10030 }, { "epoch": 3.144514106583072, "grad_norm": 8.025651931762695, "learning_rate": 4.998443783843827e-06, "loss": 1.6756, "step": 10031 }, { "epoch": 3.1448275862068966, "grad_norm": 8.2313871383667, "learning_rate": 4.998341774709482e-06, "loss": 2.1731, "step": 10032 }, { "epoch": 3.145141065830721, "grad_norm": 5.046492099761963, "learning_rate": 4.998236528644566e-06, "loss": 1.9873, "step": 10033 }, { "epoch": 3.1454545454545455, "grad_norm": 61.49712371826172, "learning_rate": 4.9981280457854406e-06, "loss": 2.3331, "step": 10034 }, { "epoch": 3.14576802507837, "grad_norm": 6.287315368652344, "learning_rate": 4.998016326272658e-06, "loss": 2.1089, "step": 10035 }, { "epoch": 3.1460815047021944, "grad_norm": 10.96717357635498, "learning_rate": 4.997901370250966e-06, "loss": 2.2947, "step": 10036 }, { "epoch": 3.146394984326019, "grad_norm": 13.235350608825684, "learning_rate": 4.997783177869307e-06, "loss": 2.1083, "step": 10037 }, { "epoch": 3.1467084639498433, "grad_norm": 8.936031341552734, "learning_rate": 4.997661749280812e-06, "loss": 2.0805, "step": 10038 }, { "epoch": 3.1470219435736677, "grad_norm": 10.347251892089844, "learning_rate": 4.99753708464281e-06, "loss": 1.8743, "step": 10039 }, { "epoch": 3.147335423197492, "grad_norm": 7.304470062255859, "learning_rate": 4.9974091841168195e-06, "loss": 2.1653, "step": 10040 }, { "epoch": 3.1476489028213166, "grad_norm": 8.110036849975586, "learning_rate": 4.997278047868552e-06, "loss": 2.1342, "step": 10041 }, { "epoch": 3.147962382445141, "grad_norm": 11.166316032409668, "learning_rate": 4.997143676067913e-06, "loss": 1.7876, "step": 10042 }, { "epoch": 3.1482758620689655, "grad_norm": 8.775270462036133, "learning_rate": 4.997006068888997e-06, "loss": 1.9446, "step": 10043 }, { "epoch": 3.14858934169279, "grad_norm": 99.71057891845703, "learning_rate": 4.996865226510094e-06, "loss": 2.2479, "step": 10044 }, { "epoch": 3.1489028213166144, "grad_norm": 15.969762802124023, "learning_rate": 4.996721149113682e-06, "loss": 2.1067, "step": 10045 }, { "epoch": 3.149216300940439, "grad_norm": 16.003082275390625, "learning_rate": 4.9965738368864345e-06, "loss": 1.9282, "step": 10046 }, { "epoch": 3.1495297805642632, "grad_norm": 18.941774368286133, "learning_rate": 4.996423290019213e-06, "loss": 1.8664, "step": 10047 }, { "epoch": 3.1498432601880877, "grad_norm": 6.078216075897217, "learning_rate": 4.99626950870707e-06, "loss": 2.4318, "step": 10048 }, { "epoch": 3.150156739811912, "grad_norm": 8.250240325927734, "learning_rate": 4.99611249314925e-06, "loss": 2.2097, "step": 10049 }, { "epoch": 3.1504702194357366, "grad_norm": 9.107946395874023, "learning_rate": 4.995952243549188e-06, "loss": 2.107, "step": 10050 }, { "epoch": 3.150783699059561, "grad_norm": 67.38969421386719, "learning_rate": 4.995788760114507e-06, "loss": 2.0807, "step": 10051 }, { "epoch": 3.1510971786833855, "grad_norm": 10.320808410644531, "learning_rate": 4.995622043057023e-06, "loss": 1.9398, "step": 10052 }, { "epoch": 3.15141065830721, "grad_norm": 9.035309791564941, "learning_rate": 4.995452092592738e-06, "loss": 1.9624, "step": 10053 }, { "epoch": 3.1517241379310343, "grad_norm": 7.165299892425537, "learning_rate": 4.995278908941845e-06, "loss": 2.0117, "step": 10054 }, { "epoch": 3.152037617554859, "grad_norm": 20.587200164794922, "learning_rate": 4.995102492328729e-06, "loss": 2.2992, "step": 10055 }, { "epoch": 3.1523510971786832, "grad_norm": 13.033390045166016, "learning_rate": 4.994922842981958e-06, "loss": 1.9895, "step": 10056 }, { "epoch": 3.1526645768025077, "grad_norm": 10.80650806427002, "learning_rate": 4.994739961134294e-06, "loss": 2.4796, "step": 10057 }, { "epoch": 3.152978056426332, "grad_norm": 6.395313739776611, "learning_rate": 4.994553847022683e-06, "loss": 2.0733, "step": 10058 }, { "epoch": 3.1532915360501566, "grad_norm": 7.472653865814209, "learning_rate": 4.99436450088826e-06, "loss": 2.0833, "step": 10059 }, { "epoch": 3.153605015673981, "grad_norm": 7.753222942352295, "learning_rate": 4.994171922976349e-06, "loss": 2.0887, "step": 10060 }, { "epoch": 3.1539184952978054, "grad_norm": 7.5873703956604, "learning_rate": 4.993976113536458e-06, "loss": 2.1164, "step": 10061 }, { "epoch": 3.15423197492163, "grad_norm": 10.077157974243164, "learning_rate": 4.993777072822286e-06, "loss": 2.0718, "step": 10062 }, { "epoch": 3.1545454545454543, "grad_norm": 9.079642295837402, "learning_rate": 4.993574801091715e-06, "loss": 2.0094, "step": 10063 }, { "epoch": 3.1548589341692788, "grad_norm": 10.427095413208008, "learning_rate": 4.993369298606817e-06, "loss": 2.0306, "step": 10064 }, { "epoch": 3.1551724137931036, "grad_norm": 6.993827819824219, "learning_rate": 4.993160565633845e-06, "loss": 2.1231, "step": 10065 }, { "epoch": 3.155485893416928, "grad_norm": 23.842500686645508, "learning_rate": 4.9929486024432405e-06, "loss": 2.1972, "step": 10066 }, { "epoch": 3.1557993730407525, "grad_norm": 21.030187606811523, "learning_rate": 4.99273340930963e-06, "loss": 1.973, "step": 10067 }, { "epoch": 3.156112852664577, "grad_norm": 29.955055236816406, "learning_rate": 4.992514986511825e-06, "loss": 2.1152, "step": 10068 }, { "epoch": 3.1564263322884014, "grad_norm": 27.82219696044922, "learning_rate": 4.992293334332821e-06, "loss": 2.637, "step": 10069 }, { "epoch": 3.156739811912226, "grad_norm": 13.661827087402344, "learning_rate": 4.992068453059796e-06, "loss": 2.098, "step": 10070 }, { "epoch": 3.1570532915360503, "grad_norm": 7.089137554168701, "learning_rate": 4.991840342984115e-06, "loss": 2.3151, "step": 10071 }, { "epoch": 3.1573667711598747, "grad_norm": 13.956977844238281, "learning_rate": 4.991609004401324e-06, "loss": 2.0955, "step": 10072 }, { "epoch": 3.157680250783699, "grad_norm": 10.268534660339355, "learning_rate": 4.991374437611153e-06, "loss": 2.2153, "step": 10073 }, { "epoch": 3.1579937304075236, "grad_norm": 9.822632789611816, "learning_rate": 4.991136642917514e-06, "loss": 2.0314, "step": 10074 }, { "epoch": 3.158307210031348, "grad_norm": 8.51823616027832, "learning_rate": 4.990895620628501e-06, "loss": 2.2489, "step": 10075 }, { "epoch": 3.1586206896551725, "grad_norm": 10.767548561096191, "learning_rate": 4.99065137105639e-06, "loss": 2.1945, "step": 10076 }, { "epoch": 3.158934169278997, "grad_norm": 39.25959396362305, "learning_rate": 4.990403894517638e-06, "loss": 2.1128, "step": 10077 }, { "epoch": 3.1592476489028214, "grad_norm": 10.357343673706055, "learning_rate": 4.990153191332885e-06, "loss": 2.1305, "step": 10078 }, { "epoch": 3.159561128526646, "grad_norm": 11.499703407287598, "learning_rate": 4.989899261826948e-06, "loss": 2.1603, "step": 10079 }, { "epoch": 3.1598746081504703, "grad_norm": 11.136299133300781, "learning_rate": 4.989642106328829e-06, "loss": 2.1496, "step": 10080 }, { "epoch": 3.1601880877742947, "grad_norm": 7.036195278167725, "learning_rate": 4.9893817251717034e-06, "loss": 2.1253, "step": 10081 }, { "epoch": 3.160501567398119, "grad_norm": 58.19233703613281, "learning_rate": 4.9891181186929315e-06, "loss": 2.3361, "step": 10082 }, { "epoch": 3.1608150470219436, "grad_norm": 6.382940292358398, "learning_rate": 4.988851287234049e-06, "loss": 2.041, "step": 10083 }, { "epoch": 3.161128526645768, "grad_norm": 9.516130447387695, "learning_rate": 4.988581231140772e-06, "loss": 2.054, "step": 10084 }, { "epoch": 3.1614420062695925, "grad_norm": 13.64483642578125, "learning_rate": 4.988307950762994e-06, "loss": 2.0752, "step": 10085 }, { "epoch": 3.161755485893417, "grad_norm": 9.723690032958984, "learning_rate": 4.988031446454784e-06, "loss": 2.154, "step": 10086 }, { "epoch": 3.1620689655172414, "grad_norm": 7.472517967224121, "learning_rate": 4.987751718574392e-06, "loss": 2.2456, "step": 10087 }, { "epoch": 3.162382445141066, "grad_norm": 77.54561614990234, "learning_rate": 4.98746876748424e-06, "loss": 2.3638, "step": 10088 }, { "epoch": 3.1626959247648903, "grad_norm": 14.868156433105469, "learning_rate": 4.98718259355093e-06, "loss": 2.3435, "step": 10089 }, { "epoch": 3.1630094043887147, "grad_norm": 18.473392486572266, "learning_rate": 4.986893197145238e-06, "loss": 2.1099, "step": 10090 }, { "epoch": 3.163322884012539, "grad_norm": 8.461474418640137, "learning_rate": 4.986600578642114e-06, "loss": 2.031, "step": 10091 }, { "epoch": 3.1636363636363636, "grad_norm": 180.326904296875, "learning_rate": 4.986304738420684e-06, "loss": 3.0385, "step": 10092 }, { "epoch": 3.163949843260188, "grad_norm": 10.383044242858887, "learning_rate": 4.986005676864248e-06, "loss": 1.8876, "step": 10093 }, { "epoch": 3.1642633228840125, "grad_norm": 12.548264503479004, "learning_rate": 4.985703394360281e-06, "loss": 2.143, "step": 10094 }, { "epoch": 3.164576802507837, "grad_norm": 9.232612609863281, "learning_rate": 4.985397891300428e-06, "loss": 2.0059, "step": 10095 }, { "epoch": 3.1648902821316613, "grad_norm": 16.89084243774414, "learning_rate": 4.985089168080509e-06, "loss": 2.391, "step": 10096 }, { "epoch": 3.165203761755486, "grad_norm": 10.708553314208984, "learning_rate": 4.9847772251005165e-06, "loss": 1.9874, "step": 10097 }, { "epoch": 3.1655172413793102, "grad_norm": 8.376543045043945, "learning_rate": 4.9844620627646125e-06, "loss": 2.2165, "step": 10098 }, { "epoch": 3.1658307210031347, "grad_norm": 51.55671310424805, "learning_rate": 4.9841436814811315e-06, "loss": 2.594, "step": 10099 }, { "epoch": 3.166144200626959, "grad_norm": 15.830016136169434, "learning_rate": 4.983822081662578e-06, "loss": 2.3068, "step": 10100 }, { "epoch": 3.1664576802507836, "grad_norm": 7.676496505737305, "learning_rate": 4.983497263725628e-06, "loss": 2.2785, "step": 10101 }, { "epoch": 3.166771159874608, "grad_norm": 58.433284759521484, "learning_rate": 4.983169228091125e-06, "loss": 2.2606, "step": 10102 }, { "epoch": 3.1670846394984324, "grad_norm": 7.262620449066162, "learning_rate": 4.9828379751840825e-06, "loss": 1.9238, "step": 10103 }, { "epoch": 3.167398119122257, "grad_norm": 8.287198066711426, "learning_rate": 4.982503505433683e-06, "loss": 2.1793, "step": 10104 }, { "epoch": 3.1677115987460813, "grad_norm": 56.644466400146484, "learning_rate": 4.982165819273275e-06, "loss": 2.4948, "step": 10105 }, { "epoch": 3.1680250783699058, "grad_norm": 7.558988571166992, "learning_rate": 4.981824917140376e-06, "loss": 2.1551, "step": 10106 }, { "epoch": 3.16833855799373, "grad_norm": 8.287984848022461, "learning_rate": 4.981480799476669e-06, "loss": 2.1425, "step": 10107 }, { "epoch": 3.1686520376175547, "grad_norm": 50.695518493652344, "learning_rate": 4.981133466728004e-06, "loss": 2.838, "step": 10108 }, { "epoch": 3.1689655172413795, "grad_norm": 6.426566123962402, "learning_rate": 4.980782919344397e-06, "loss": 1.9652, "step": 10109 }, { "epoch": 3.169278996865204, "grad_norm": 22.17452621459961, "learning_rate": 4.980429157780027e-06, "loss": 2.4055, "step": 10110 }, { "epoch": 3.1695924764890284, "grad_norm": 7.178967475891113, "learning_rate": 4.98007218249324e-06, "loss": 2.0155, "step": 10111 }, { "epoch": 3.169905956112853, "grad_norm": 26.847415924072266, "learning_rate": 4.979711993946543e-06, "loss": 2.0844, "step": 10112 }, { "epoch": 3.1702194357366773, "grad_norm": 9.120711326599121, "learning_rate": 4.9793485926066086e-06, "loss": 1.9308, "step": 10113 }, { "epoch": 3.1705329153605017, "grad_norm": 9.894770622253418, "learning_rate": 4.978981978944271e-06, "loss": 2.3352, "step": 10114 }, { "epoch": 3.170846394984326, "grad_norm": 12.968624114990234, "learning_rate": 4.978612153434527e-06, "loss": 2.1535, "step": 10115 }, { "epoch": 3.1711598746081506, "grad_norm": 7.2515387535095215, "learning_rate": 4.978239116556533e-06, "loss": 2.0999, "step": 10116 }, { "epoch": 3.171473354231975, "grad_norm": 8.109625816345215, "learning_rate": 4.977862868793607e-06, "loss": 2.1223, "step": 10117 }, { "epoch": 3.1717868338557995, "grad_norm": 71.0289306640625, "learning_rate": 4.977483410633229e-06, "loss": 2.4858, "step": 10118 }, { "epoch": 3.172100313479624, "grad_norm": 15.042811393737793, "learning_rate": 4.977100742567037e-06, "loss": 1.8713, "step": 10119 }, { "epoch": 3.1724137931034484, "grad_norm": 7.42080545425415, "learning_rate": 4.976714865090827e-06, "loss": 2.1731, "step": 10120 }, { "epoch": 3.172727272727273, "grad_norm": 11.799166679382324, "learning_rate": 4.976325778704555e-06, "loss": 2.0544, "step": 10121 }, { "epoch": 3.1730407523510973, "grad_norm": 8.801580429077148, "learning_rate": 4.975933483912332e-06, "loss": 2.1248, "step": 10122 }, { "epoch": 3.1733542319749217, "grad_norm": 9.492209434509277, "learning_rate": 4.975537981222429e-06, "loss": 2.031, "step": 10123 }, { "epoch": 3.173667711598746, "grad_norm": 37.21613693237305, "learning_rate": 4.97513927114727e-06, "loss": 2.3914, "step": 10124 }, { "epoch": 3.1739811912225706, "grad_norm": 9.097406387329102, "learning_rate": 4.974737354203439e-06, "loss": 2.4783, "step": 10125 }, { "epoch": 3.174294670846395, "grad_norm": 7.6060051918029785, "learning_rate": 4.97433223091167e-06, "loss": 2.0825, "step": 10126 }, { "epoch": 3.1746081504702195, "grad_norm": 7.525833606719971, "learning_rate": 4.973923901796856e-06, "loss": 2.21, "step": 10127 }, { "epoch": 3.174921630094044, "grad_norm": 6.649914264678955, "learning_rate": 4.973512367388038e-06, "loss": 2.344, "step": 10128 }, { "epoch": 3.1752351097178684, "grad_norm": 7.706682205200195, "learning_rate": 4.973097628218415e-06, "loss": 2.0059, "step": 10129 }, { "epoch": 3.175548589341693, "grad_norm": 14.445734024047852, "learning_rate": 4.9726796848253365e-06, "loss": 2.1198, "step": 10130 }, { "epoch": 3.1758620689655173, "grad_norm": 14.331486701965332, "learning_rate": 4.9722585377503014e-06, "loss": 2.0519, "step": 10131 }, { "epoch": 3.1761755485893417, "grad_norm": 7.217585563659668, "learning_rate": 4.971834187538963e-06, "loss": 2.1761, "step": 10132 }, { "epoch": 3.176489028213166, "grad_norm": 8.836959838867188, "learning_rate": 4.971406634741121e-06, "loss": 2.112, "step": 10133 }, { "epoch": 3.1768025078369906, "grad_norm": 8.91513729095459, "learning_rate": 4.970975879910728e-06, "loss": 2.2292, "step": 10134 }, { "epoch": 3.177115987460815, "grad_norm": 9.225442886352539, "learning_rate": 4.9705419236058825e-06, "loss": 2.2962, "step": 10135 }, { "epoch": 3.1774294670846395, "grad_norm": 9.509936332702637, "learning_rate": 4.970104766388833e-06, "loss": 2.3073, "step": 10136 }, { "epoch": 3.177742946708464, "grad_norm": 60.8763313293457, "learning_rate": 4.969664408825973e-06, "loss": 2.1062, "step": 10137 }, { "epoch": 3.1780564263322884, "grad_norm": 10.467870712280273, "learning_rate": 4.9692208514878445e-06, "loss": 2.1352, "step": 10138 }, { "epoch": 3.178369905956113, "grad_norm": 26.236255645751953, "learning_rate": 4.968774094949135e-06, "loss": 2.1424, "step": 10139 }, { "epoch": 3.1786833855799372, "grad_norm": 7.876876354217529, "learning_rate": 4.968324139788675e-06, "loss": 2.1388, "step": 10140 }, { "epoch": 3.1789968652037617, "grad_norm": 9.857775688171387, "learning_rate": 4.96787098658944e-06, "loss": 1.9912, "step": 10141 }, { "epoch": 3.179310344827586, "grad_norm": 16.92973518371582, "learning_rate": 4.967414635938552e-06, "loss": 2.2557, "step": 10142 }, { "epoch": 3.1796238244514106, "grad_norm": 25.699729919433594, "learning_rate": 4.966955088427273e-06, "loss": 2.0162, "step": 10143 }, { "epoch": 3.179937304075235, "grad_norm": 9.593525886535645, "learning_rate": 4.966492344651006e-06, "loss": 1.8685, "step": 10144 }, { "epoch": 3.1802507836990594, "grad_norm": 35.92695617675781, "learning_rate": 4.9660264052092954e-06, "loss": 2.0443, "step": 10145 }, { "epoch": 3.180564263322884, "grad_norm": 8.509702682495117, "learning_rate": 4.965557270705831e-06, "loss": 1.9615, "step": 10146 }, { "epoch": 3.1808777429467083, "grad_norm": 40.05824279785156, "learning_rate": 4.965084941748434e-06, "loss": 2.1523, "step": 10147 }, { "epoch": 3.1811912225705328, "grad_norm": 8.13050365447998, "learning_rate": 4.9646094189490716e-06, "loss": 2.1289, "step": 10148 }, { "epoch": 3.181504702194357, "grad_norm": 6.308853626251221, "learning_rate": 4.964130702923844e-06, "loss": 1.9261, "step": 10149 }, { "epoch": 3.1818181818181817, "grad_norm": 6.656688213348389, "learning_rate": 4.963648794292992e-06, "loss": 2.1389, "step": 10150 }, { "epoch": 3.182131661442006, "grad_norm": 15.271326065063477, "learning_rate": 4.963163693680893e-06, "loss": 2.0596, "step": 10151 }, { "epoch": 3.1824451410658305, "grad_norm": 5.64087438583374, "learning_rate": 4.962675401716056e-06, "loss": 2.0276, "step": 10152 }, { "epoch": 3.182758620689655, "grad_norm": 9.619490623474121, "learning_rate": 4.96218391903113e-06, "loss": 1.9465, "step": 10153 }, { "epoch": 3.1830721003134794, "grad_norm": 6.722213268280029, "learning_rate": 4.961689246262894e-06, "loss": 2.1282, "step": 10154 }, { "epoch": 3.183385579937304, "grad_norm": 11.597489356994629, "learning_rate": 4.961191384052261e-06, "loss": 2.2211, "step": 10155 }, { "epoch": 3.1836990595611283, "grad_norm": 5.950822353363037, "learning_rate": 4.960690333044279e-06, "loss": 1.9535, "step": 10156 }, { "epoch": 3.1840125391849528, "grad_norm": 8.629708290100098, "learning_rate": 4.960186093888124e-06, "loss": 2.3181, "step": 10157 }, { "epoch": 3.1843260188087776, "grad_norm": 10.962076187133789, "learning_rate": 4.959678667237106e-06, "loss": 2.5209, "step": 10158 }, { "epoch": 3.184639498432602, "grad_norm": 5.0298943519592285, "learning_rate": 4.95916805374866e-06, "loss": 2.2074, "step": 10159 }, { "epoch": 3.1849529780564265, "grad_norm": 6.174869060516357, "learning_rate": 4.958654254084356e-06, "loss": 2.1809, "step": 10160 }, { "epoch": 3.185266457680251, "grad_norm": 20.36189079284668, "learning_rate": 4.958137268909887e-06, "loss": 2.346, "step": 10161 }, { "epoch": 3.1855799373040754, "grad_norm": 14.660477638244629, "learning_rate": 4.957617098895076e-06, "loss": 1.9395, "step": 10162 }, { "epoch": 3.1858934169279, "grad_norm": 33.34031295776367, "learning_rate": 4.957093744713872e-06, "loss": 2.0744, "step": 10163 }, { "epoch": 3.1862068965517243, "grad_norm": 59.17599105834961, "learning_rate": 4.956567207044349e-06, "loss": 2.1416, "step": 10164 }, { "epoch": 3.1865203761755487, "grad_norm": 7.621762275695801, "learning_rate": 4.956037486568707e-06, "loss": 1.9883, "step": 10165 }, { "epoch": 3.186833855799373, "grad_norm": 7.222348690032959, "learning_rate": 4.955504583973266e-06, "loss": 2.0853, "step": 10166 }, { "epoch": 3.1871473354231976, "grad_norm": 7.455623626708984, "learning_rate": 4.954968499948474e-06, "loss": 1.9018, "step": 10167 }, { "epoch": 3.187460815047022, "grad_norm": 9.011062622070312, "learning_rate": 4.954429235188897e-06, "loss": 2.4855, "step": 10168 }, { "epoch": 3.1877742946708465, "grad_norm": 8.30917739868164, "learning_rate": 4.953886790393224e-06, "loss": 1.7544, "step": 10169 }, { "epoch": 3.188087774294671, "grad_norm": 83.94273376464844, "learning_rate": 4.9533411662642625e-06, "loss": 2.3086, "step": 10170 }, { "epoch": 3.1884012539184954, "grad_norm": 34.87735366821289, "learning_rate": 4.952792363508943e-06, "loss": 2.431, "step": 10171 }, { "epoch": 3.18871473354232, "grad_norm": 14.460996627807617, "learning_rate": 4.9522403828383085e-06, "loss": 2.0841, "step": 10172 }, { "epoch": 3.1890282131661443, "grad_norm": 84.48480987548828, "learning_rate": 4.951685224967524e-06, "loss": 2.1964, "step": 10173 }, { "epoch": 3.1893416927899687, "grad_norm": 39.105796813964844, "learning_rate": 4.951126890615871e-06, "loss": 2.0742, "step": 10174 }, { "epoch": 3.189655172413793, "grad_norm": 9.493748664855957, "learning_rate": 4.950565380506742e-06, "loss": 2.2183, "step": 10175 }, { "epoch": 3.1899686520376176, "grad_norm": 7.700687885284424, "learning_rate": 4.95000069536765e-06, "loss": 2.2104, "step": 10176 }, { "epoch": 3.190282131661442, "grad_norm": 12.290258407592773, "learning_rate": 4.949432835930219e-06, "loss": 2.4045, "step": 10177 }, { "epoch": 3.1905956112852665, "grad_norm": 7.603089809417725, "learning_rate": 4.948861802930184e-06, "loss": 1.8463, "step": 10178 }, { "epoch": 3.190909090909091, "grad_norm": 8.37252140045166, "learning_rate": 4.948287597107393e-06, "loss": 2.38, "step": 10179 }, { "epoch": 3.1912225705329154, "grad_norm": 8.213652610778809, "learning_rate": 4.947710219205808e-06, "loss": 1.7764, "step": 10180 }, { "epoch": 3.19153605015674, "grad_norm": 11.823604583740234, "learning_rate": 4.947129669973495e-06, "loss": 2.2695, "step": 10181 }, { "epoch": 3.1918495297805642, "grad_norm": 7.708943843841553, "learning_rate": 4.946545950162634e-06, "loss": 1.9848, "step": 10182 }, { "epoch": 3.1921630094043887, "grad_norm": 10.535895347595215, "learning_rate": 4.945959060529509e-06, "loss": 2.1492, "step": 10183 }, { "epoch": 3.192476489028213, "grad_norm": 6.885127067565918, "learning_rate": 4.9453690018345144e-06, "loss": 1.975, "step": 10184 }, { "epoch": 3.1927899686520376, "grad_norm": 14.497472763061523, "learning_rate": 4.944775774842149e-06, "loss": 2.2, "step": 10185 }, { "epoch": 3.193103448275862, "grad_norm": 9.273137092590332, "learning_rate": 4.944179380321015e-06, "loss": 1.9535, "step": 10186 }, { "epoch": 3.1934169278996865, "grad_norm": 6.478431701660156, "learning_rate": 4.9435798190438206e-06, "loss": 2.0358, "step": 10187 }, { "epoch": 3.193730407523511, "grad_norm": 6.529944896697998, "learning_rate": 4.942977091787376e-06, "loss": 2.3338, "step": 10188 }, { "epoch": 3.1940438871473353, "grad_norm": 8.661742210388184, "learning_rate": 4.942371199332596e-06, "loss": 2.325, "step": 10189 }, { "epoch": 3.19435736677116, "grad_norm": 7.856261730194092, "learning_rate": 4.941762142464491e-06, "loss": 2.0684, "step": 10190 }, { "epoch": 3.194670846394984, "grad_norm": 11.17569637298584, "learning_rate": 4.9411499219721754e-06, "loss": 2.233, "step": 10191 }, { "epoch": 3.1949843260188087, "grad_norm": 18.203805923461914, "learning_rate": 4.940534538648862e-06, "loss": 2.1027, "step": 10192 }, { "epoch": 3.195297805642633, "grad_norm": 70.25038146972656, "learning_rate": 4.93991599329186e-06, "loss": 2.1583, "step": 10193 }, { "epoch": 3.1956112852664575, "grad_norm": 6.400642395019531, "learning_rate": 4.9392942867025775e-06, "loss": 2.0302, "step": 10194 }, { "epoch": 3.195924764890282, "grad_norm": 9.489896774291992, "learning_rate": 4.938669419686516e-06, "loss": 2.0612, "step": 10195 }, { "epoch": 3.1962382445141064, "grad_norm": 7.376988410949707, "learning_rate": 4.938041393053273e-06, "loss": 2.2046, "step": 10196 }, { "epoch": 3.196551724137931, "grad_norm": 6.3598127365112305, "learning_rate": 4.937410207616541e-06, "loss": 2.2869, "step": 10197 }, { "epoch": 3.1968652037617553, "grad_norm": 12.441479682922363, "learning_rate": 4.936775864194101e-06, "loss": 2.379, "step": 10198 }, { "epoch": 3.1971786833855798, "grad_norm": 10.65619945526123, "learning_rate": 4.936138363607832e-06, "loss": 2.0554, "step": 10199 }, { "epoch": 3.197492163009404, "grad_norm": 6.433727741241455, "learning_rate": 4.935497706683698e-06, "loss": 2.2514, "step": 10200 }, { "epoch": 3.1978056426332286, "grad_norm": 19.275232315063477, "learning_rate": 4.934853894251754e-06, "loss": 2.1702, "step": 10201 }, { "epoch": 3.1981191222570535, "grad_norm": 11.46334171295166, "learning_rate": 4.934206927146145e-06, "loss": 2.0721, "step": 10202 }, { "epoch": 3.198432601880878, "grad_norm": 10.792070388793945, "learning_rate": 4.933556806205101e-06, "loss": 2.1566, "step": 10203 }, { "epoch": 3.1987460815047024, "grad_norm": 5.8852667808532715, "learning_rate": 4.932903532270939e-06, "loss": 2.133, "step": 10204 }, { "epoch": 3.199059561128527, "grad_norm": 6.422491073608398, "learning_rate": 4.932247106190063e-06, "loss": 2.129, "step": 10205 }, { "epoch": 3.1993730407523513, "grad_norm": 8.684357643127441, "learning_rate": 4.931587528812957e-06, "loss": 2.0384, "step": 10206 }, { "epoch": 3.1996865203761757, "grad_norm": 25.378263473510742, "learning_rate": 4.930924800994192e-06, "loss": 2.5752, "step": 10207 }, { "epoch": 3.2, "grad_norm": 8.196176528930664, "learning_rate": 4.9302589235924185e-06, "loss": 2.0487, "step": 10208 }, { "epoch": 3.2003134796238246, "grad_norm": 11.427424430847168, "learning_rate": 4.9295898974703695e-06, "loss": 2.0834, "step": 10209 }, { "epoch": 3.200626959247649, "grad_norm": 7.6105475425720215, "learning_rate": 4.928917723494854e-06, "loss": 2.1894, "step": 10210 }, { "epoch": 3.2009404388714735, "grad_norm": 5.3308258056640625, "learning_rate": 4.928242402536764e-06, "loss": 2.1444, "step": 10211 }, { "epoch": 3.201253918495298, "grad_norm": 45.00204086303711, "learning_rate": 4.927563935471066e-06, "loss": 2.2768, "step": 10212 }, { "epoch": 3.2015673981191224, "grad_norm": 10.13443374633789, "learning_rate": 4.926882323176803e-06, "loss": 2.1276, "step": 10213 }, { "epoch": 3.201880877742947, "grad_norm": 44.9454460144043, "learning_rate": 4.926197566537094e-06, "loss": 2.1207, "step": 10214 }, { "epoch": 3.2021943573667713, "grad_norm": 5.935682773590088, "learning_rate": 4.9255096664391305e-06, "loss": 2.0419, "step": 10215 }, { "epoch": 3.2025078369905957, "grad_norm": 6.412038803100586, "learning_rate": 4.924818623774178e-06, "loss": 2.6359, "step": 10216 }, { "epoch": 3.20282131661442, "grad_norm": 6.320741176605225, "learning_rate": 4.924124439437574e-06, "loss": 2.0084, "step": 10217 }, { "epoch": 3.2031347962382446, "grad_norm": 11.658798217773438, "learning_rate": 4.923427114328725e-06, "loss": 2.2649, "step": 10218 }, { "epoch": 3.203448275862069, "grad_norm": 55.57625198364258, "learning_rate": 4.922726649351108e-06, "loss": 2.3795, "step": 10219 }, { "epoch": 3.2037617554858935, "grad_norm": 10.661152839660645, "learning_rate": 4.922023045412266e-06, "loss": 2.0168, "step": 10220 }, { "epoch": 3.204075235109718, "grad_norm": 10.782994270324707, "learning_rate": 4.921316303423812e-06, "loss": 2.4755, "step": 10221 }, { "epoch": 3.2043887147335424, "grad_norm": 7.619351863861084, "learning_rate": 4.920606424301424e-06, "loss": 1.9396, "step": 10222 }, { "epoch": 3.204702194357367, "grad_norm": 20.153839111328125, "learning_rate": 4.919893408964841e-06, "loss": 2.4844, "step": 10223 }, { "epoch": 3.2050156739811912, "grad_norm": 11.450020790100098, "learning_rate": 4.91917725833787e-06, "loss": 2.068, "step": 10224 }, { "epoch": 3.2053291536050157, "grad_norm": 6.951805591583252, "learning_rate": 4.91845797334838e-06, "loss": 2.006, "step": 10225 }, { "epoch": 3.20564263322884, "grad_norm": 7.529156684875488, "learning_rate": 4.917735554928296e-06, "loss": 2.1315, "step": 10226 }, { "epoch": 3.2059561128526646, "grad_norm": 21.872188568115234, "learning_rate": 4.9170100040136076e-06, "loss": 1.8463, "step": 10227 }, { "epoch": 3.206269592476489, "grad_norm": 7.5848164558410645, "learning_rate": 4.916281321544362e-06, "loss": 2.0687, "step": 10228 }, { "epoch": 3.2065830721003135, "grad_norm": 11.074767112731934, "learning_rate": 4.915549508464663e-06, "loss": 1.9767, "step": 10229 }, { "epoch": 3.206896551724138, "grad_norm": 16.35045051574707, "learning_rate": 4.914814565722671e-06, "loss": 1.9851, "step": 10230 }, { "epoch": 3.2072100313479623, "grad_norm": 9.384275436401367, "learning_rate": 4.914076494270601e-06, "loss": 2.3271, "step": 10231 }, { "epoch": 3.207523510971787, "grad_norm": 21.516138076782227, "learning_rate": 4.913335295064721e-06, "loss": 2.2024, "step": 10232 }, { "epoch": 3.2078369905956112, "grad_norm": 7.182125568389893, "learning_rate": 4.912590969065352e-06, "loss": 1.9425, "step": 10233 }, { "epoch": 3.2081504702194357, "grad_norm": 7.3275146484375, "learning_rate": 4.911843517236867e-06, "loss": 1.8435, "step": 10234 }, { "epoch": 3.20846394984326, "grad_norm": 166.70196533203125, "learning_rate": 4.911092940547688e-06, "loss": 2.6872, "step": 10235 }, { "epoch": 3.2087774294670846, "grad_norm": 5.064785480499268, "learning_rate": 4.910339239970286e-06, "loss": 2.0313, "step": 10236 }, { "epoch": 3.209090909090909, "grad_norm": 8.595926284790039, "learning_rate": 4.90958241648118e-06, "loss": 2.0378, "step": 10237 }, { "epoch": 3.2094043887147334, "grad_norm": 9.702130317687988, "learning_rate": 4.908822471060932e-06, "loss": 2.3029, "step": 10238 }, { "epoch": 3.209717868338558, "grad_norm": 16.16521644592285, "learning_rate": 4.908059404694155e-06, "loss": 2.4376, "step": 10239 }, { "epoch": 3.2100313479623823, "grad_norm": 24.82144546508789, "learning_rate": 4.907293218369499e-06, "loss": 2.277, "step": 10240 }, { "epoch": 3.2103448275862068, "grad_norm": 8.200942993164062, "learning_rate": 4.906523913079659e-06, "loss": 1.9526, "step": 10241 }, { "epoch": 3.210658307210031, "grad_norm": 111.89910888671875, "learning_rate": 4.905751489821374e-06, "loss": 2.4181, "step": 10242 }, { "epoch": 3.2109717868338556, "grad_norm": 24.466325759887695, "learning_rate": 4.904975949595417e-06, "loss": 2.055, "step": 10243 }, { "epoch": 3.21128526645768, "grad_norm": 157.03518676757812, "learning_rate": 4.904197293406604e-06, "loss": 2.6675, "step": 10244 }, { "epoch": 3.2115987460815045, "grad_norm": 5.92139196395874, "learning_rate": 4.903415522263787e-06, "loss": 1.9266, "step": 10245 }, { "epoch": 3.211912225705329, "grad_norm": 4.725953102111816, "learning_rate": 4.9026306371798526e-06, "loss": 2.1913, "step": 10246 }, { "epoch": 3.2122257053291534, "grad_norm": 8.969285011291504, "learning_rate": 4.901842639171722e-06, "loss": 2.3202, "step": 10247 }, { "epoch": 3.212539184952978, "grad_norm": 9.10229778289795, "learning_rate": 4.901051529260352e-06, "loss": 1.9313, "step": 10248 }, { "epoch": 3.2128526645768023, "grad_norm": 15.433198928833008, "learning_rate": 4.9002573084707284e-06, "loss": 1.9121, "step": 10249 }, { "epoch": 3.2131661442006267, "grad_norm": 18.19792938232422, "learning_rate": 4.899459977831869e-06, "loss": 1.9703, "step": 10250 }, { "epoch": 3.213479623824451, "grad_norm": 7.2077741622924805, "learning_rate": 4.898659538376821e-06, "loss": 2.0011, "step": 10251 }, { "epoch": 3.213793103448276, "grad_norm": 30.596385955810547, "learning_rate": 4.897855991142658e-06, "loss": 2.052, "step": 10252 }, { "epoch": 3.2141065830721005, "grad_norm": 16.962915420532227, "learning_rate": 4.897049337170483e-06, "loss": 2.0775, "step": 10253 }, { "epoch": 3.214420062695925, "grad_norm": 8.502519607543945, "learning_rate": 4.896239577505421e-06, "loss": 2.0855, "step": 10254 }, { "epoch": 3.2147335423197494, "grad_norm": 39.96928787231445, "learning_rate": 4.895426713196623e-06, "loss": 2.2766, "step": 10255 }, { "epoch": 3.215047021943574, "grad_norm": 27.422832489013672, "learning_rate": 4.89461074529726e-06, "loss": 2.0565, "step": 10256 }, { "epoch": 3.2153605015673983, "grad_norm": 14.682738304138184, "learning_rate": 4.893791674864527e-06, "loss": 2.2054, "step": 10257 }, { "epoch": 3.2156739811912227, "grad_norm": 59.32605743408203, "learning_rate": 4.892969502959639e-06, "loss": 2.1576, "step": 10258 }, { "epoch": 3.215987460815047, "grad_norm": 52.33721160888672, "learning_rate": 4.892144230647826e-06, "loss": 2.2773, "step": 10259 }, { "epoch": 3.2163009404388716, "grad_norm": 9.785981178283691, "learning_rate": 4.891315858998337e-06, "loss": 2.1005, "step": 10260 }, { "epoch": 3.216614420062696, "grad_norm": 13.950303077697754, "learning_rate": 4.890484389084437e-06, "loss": 2.2541, "step": 10261 }, { "epoch": 3.2169278996865205, "grad_norm": 15.94414234161377, "learning_rate": 4.889649821983405e-06, "loss": 1.8839, "step": 10262 }, { "epoch": 3.217241379310345, "grad_norm": 36.551116943359375, "learning_rate": 4.888812158776532e-06, "loss": 3.1031, "step": 10263 }, { "epoch": 3.2175548589341694, "grad_norm": 13.444183349609375, "learning_rate": 4.8879714005491205e-06, "loss": 2.3097, "step": 10264 }, { "epoch": 3.217868338557994, "grad_norm": 21.944334030151367, "learning_rate": 4.8871275483904845e-06, "loss": 1.9407, "step": 10265 }, { "epoch": 3.2181818181818183, "grad_norm": 8.637795448303223, "learning_rate": 4.886280603393945e-06, "loss": 2.0959, "step": 10266 }, { "epoch": 3.2184952978056427, "grad_norm": 11.036513328552246, "learning_rate": 4.88543056665683e-06, "loss": 2.4281, "step": 10267 }, { "epoch": 3.218808777429467, "grad_norm": 119.82063293457031, "learning_rate": 4.884577439280475e-06, "loss": 2.0519, "step": 10268 }, { "epoch": 3.2191222570532916, "grad_norm": 6.599746227264404, "learning_rate": 4.883721222370219e-06, "loss": 2.21, "step": 10269 }, { "epoch": 3.219435736677116, "grad_norm": 71.69346618652344, "learning_rate": 4.882861917035403e-06, "loss": 2.9776, "step": 10270 }, { "epoch": 3.2197492163009405, "grad_norm": 6.783605575561523, "learning_rate": 4.88199952438937e-06, "loss": 2.0966, "step": 10271 }, { "epoch": 3.220062695924765, "grad_norm": 4.352674961090088, "learning_rate": 4.881134045549463e-06, "loss": 2.1331, "step": 10272 }, { "epoch": 3.2203761755485893, "grad_norm": 12.64493465423584, "learning_rate": 4.880265481637024e-06, "loss": 2.3506, "step": 10273 }, { "epoch": 3.220689655172414, "grad_norm": 6.9362473487854, "learning_rate": 4.8793938337773935e-06, "loss": 2.003, "step": 10274 }, { "epoch": 3.2210031347962382, "grad_norm": 20.729984283447266, "learning_rate": 4.878519103099904e-06, "loss": 1.9904, "step": 10275 }, { "epoch": 3.2213166144200627, "grad_norm": 7.76502799987793, "learning_rate": 4.8776412907378845e-06, "loss": 2.4409, "step": 10276 }, { "epoch": 3.221630094043887, "grad_norm": 10.143837928771973, "learning_rate": 4.876760397828657e-06, "loss": 2.04, "step": 10277 }, { "epoch": 3.2219435736677116, "grad_norm": 7.845150947570801, "learning_rate": 4.875876425513534e-06, "loss": 2.151, "step": 10278 }, { "epoch": 3.222257053291536, "grad_norm": 9.973098754882812, "learning_rate": 4.874989374937818e-06, "loss": 1.9528, "step": 10279 }, { "epoch": 3.2225705329153604, "grad_norm": 8.456871032714844, "learning_rate": 4.874099247250799e-06, "loss": 2.0931, "step": 10280 }, { "epoch": 3.222884012539185, "grad_norm": 17.216747283935547, "learning_rate": 4.873206043605755e-06, "loss": 2.2131, "step": 10281 }, { "epoch": 3.2231974921630093, "grad_norm": 7.438287734985352, "learning_rate": 4.87230976515995e-06, "loss": 1.8673, "step": 10282 }, { "epoch": 3.2235109717868338, "grad_norm": 14.359661102294922, "learning_rate": 4.8714104130746296e-06, "loss": 2.1498, "step": 10283 }, { "epoch": 3.223824451410658, "grad_norm": 5.519262790679932, "learning_rate": 4.870507988515023e-06, "loss": 2.0173, "step": 10284 }, { "epoch": 3.2241379310344827, "grad_norm": 10.237730026245117, "learning_rate": 4.86960249265034e-06, "loss": 2.1715, "step": 10285 }, { "epoch": 3.224451410658307, "grad_norm": 10.111181259155273, "learning_rate": 4.8686939266537695e-06, "loss": 2.2568, "step": 10286 }, { "epoch": 3.2247648902821315, "grad_norm": 11.922489166259766, "learning_rate": 4.867782291702479e-06, "loss": 1.9438, "step": 10287 }, { "epoch": 3.225078369905956, "grad_norm": 6.986643314361572, "learning_rate": 4.8668675889776095e-06, "loss": 2.2104, "step": 10288 }, { "epoch": 3.2253918495297804, "grad_norm": 224.27944946289062, "learning_rate": 4.865949819664282e-06, "loss": 2.5096, "step": 10289 }, { "epoch": 3.225705329153605, "grad_norm": 9.93991470336914, "learning_rate": 4.865028984951585e-06, "loss": 1.8592, "step": 10290 }, { "epoch": 3.2260188087774293, "grad_norm": 4.558305740356445, "learning_rate": 4.864105086032581e-06, "loss": 2.1298, "step": 10291 }, { "epoch": 3.2263322884012537, "grad_norm": 6.327045440673828, "learning_rate": 4.863178124104305e-06, "loss": 2.012, "step": 10292 }, { "epoch": 3.226645768025078, "grad_norm": 6.6480207443237305, "learning_rate": 4.862248100367755e-06, "loss": 1.8633, "step": 10293 }, { "epoch": 3.2269592476489026, "grad_norm": 8.046250343322754, "learning_rate": 4.861315016027902e-06, "loss": 1.9742, "step": 10294 }, { "epoch": 3.227272727272727, "grad_norm": 10.25992202758789, "learning_rate": 4.860378872293677e-06, "loss": 2.157, "step": 10295 }, { "epoch": 3.227586206896552, "grad_norm": 10.590046882629395, "learning_rate": 4.85943967037798e-06, "loss": 2.7937, "step": 10296 }, { "epoch": 3.2278996865203764, "grad_norm": 10.363182067871094, "learning_rate": 4.858497411497669e-06, "loss": 1.9519, "step": 10297 }, { "epoch": 3.228213166144201, "grad_norm": 9.303900718688965, "learning_rate": 4.857552096873564e-06, "loss": 1.9508, "step": 10298 }, { "epoch": 3.2285266457680253, "grad_norm": 8.812919616699219, "learning_rate": 4.856603727730446e-06, "loss": 2.2269, "step": 10299 }, { "epoch": 3.2288401253918497, "grad_norm": 8.864347457885742, "learning_rate": 4.855652305297052e-06, "loss": 2.0194, "step": 10300 }, { "epoch": 3.229153605015674, "grad_norm": 8.419734954833984, "learning_rate": 4.854697830806073e-06, "loss": 2.0467, "step": 10301 }, { "epoch": 3.2294670846394986, "grad_norm": 9.969587326049805, "learning_rate": 4.853740305494159e-06, "loss": 2.2287, "step": 10302 }, { "epoch": 3.229780564263323, "grad_norm": 52.11187744140625, "learning_rate": 4.852779730601908e-06, "loss": 1.9668, "step": 10303 }, { "epoch": 3.2300940438871475, "grad_norm": 6.189080715179443, "learning_rate": 4.851816107373871e-06, "loss": 2.1122, "step": 10304 }, { "epoch": 3.230407523510972, "grad_norm": 46.5469970703125, "learning_rate": 4.85084943705855e-06, "loss": 1.7906, "step": 10305 }, { "epoch": 3.2307210031347964, "grad_norm": 10.677568435668945, "learning_rate": 4.849879720908394e-06, "loss": 1.9082, "step": 10306 }, { "epoch": 3.231034482758621, "grad_norm": 8.884016036987305, "learning_rate": 4.848906960179796e-06, "loss": 2.3364, "step": 10307 }, { "epoch": 3.2313479623824453, "grad_norm": 14.040743827819824, "learning_rate": 4.847931156133097e-06, "loss": 2.2402, "step": 10308 }, { "epoch": 3.2316614420062697, "grad_norm": 6.579909801483154, "learning_rate": 4.84695231003258e-06, "loss": 1.9163, "step": 10309 }, { "epoch": 3.231974921630094, "grad_norm": 10.11977481842041, "learning_rate": 4.845970423146468e-06, "loss": 2.3195, "step": 10310 }, { "epoch": 3.2322884012539186, "grad_norm": 5.71937370300293, "learning_rate": 4.844985496746925e-06, "loss": 2.0537, "step": 10311 }, { "epoch": 3.232601880877743, "grad_norm": 9.752167701721191, "learning_rate": 4.843997532110051e-06, "loss": 2.0321, "step": 10312 }, { "epoch": 3.2329153605015675, "grad_norm": 12.409279823303223, "learning_rate": 4.843006530515887e-06, "loss": 2.131, "step": 10313 }, { "epoch": 3.233228840125392, "grad_norm": 35.08982467651367, "learning_rate": 4.842012493248405e-06, "loss": 2.1811, "step": 10314 }, { "epoch": 3.2335423197492164, "grad_norm": 6.847688674926758, "learning_rate": 4.841015421595511e-06, "loss": 2.3358, "step": 10315 }, { "epoch": 3.233855799373041, "grad_norm": 9.216371536254883, "learning_rate": 4.840015316849042e-06, "loss": 2.1147, "step": 10316 }, { "epoch": 3.2341692789968652, "grad_norm": 6.77595853805542, "learning_rate": 4.839012180304765e-06, "loss": 2.1066, "step": 10317 }, { "epoch": 3.2344827586206897, "grad_norm": 7.259424686431885, "learning_rate": 4.8380060132623776e-06, "loss": 2.0784, "step": 10318 }, { "epoch": 3.234796238244514, "grad_norm": 7.5641770362854, "learning_rate": 4.836996817025499e-06, "loss": 2.0171, "step": 10319 }, { "epoch": 3.2351097178683386, "grad_norm": 12.968236923217773, "learning_rate": 4.835984592901678e-06, "loss": 2.1611, "step": 10320 }, { "epoch": 3.235423197492163, "grad_norm": 6.371007442474365, "learning_rate": 4.834969342202383e-06, "loss": 2.0084, "step": 10321 }, { "epoch": 3.2357366771159874, "grad_norm": 8.4412202835083, "learning_rate": 4.833951066243004e-06, "loss": 2.3583, "step": 10322 }, { "epoch": 3.236050156739812, "grad_norm": 8.975794792175293, "learning_rate": 4.832929766342854e-06, "loss": 2.0482, "step": 10323 }, { "epoch": 3.2363636363636363, "grad_norm": 23.33732795715332, "learning_rate": 4.83190544382516e-06, "loss": 2.516, "step": 10324 }, { "epoch": 3.2366771159874608, "grad_norm": 9.635971069335938, "learning_rate": 4.8308781000170656e-06, "loss": 2.0449, "step": 10325 }, { "epoch": 3.236990595611285, "grad_norm": 8.985801696777344, "learning_rate": 4.829847736249631e-06, "loss": 2.1092, "step": 10326 }, { "epoch": 3.2373040752351097, "grad_norm": 8.741048812866211, "learning_rate": 4.828814353857828e-06, "loss": 2.1617, "step": 10327 }, { "epoch": 3.237617554858934, "grad_norm": 9.730113983154297, "learning_rate": 4.82777795418054e-06, "loss": 2.0301, "step": 10328 }, { "epoch": 3.2379310344827585, "grad_norm": 9.582262992858887, "learning_rate": 4.8267385385605585e-06, "loss": 2.0888, "step": 10329 }, { "epoch": 3.238244514106583, "grad_norm": 25.030933380126953, "learning_rate": 4.825696108344583e-06, "loss": 2.2963, "step": 10330 }, { "epoch": 3.2385579937304074, "grad_norm": 9.378925323486328, "learning_rate": 4.82465066488322e-06, "loss": 2.2511, "step": 10331 }, { "epoch": 3.238871473354232, "grad_norm": 8.080329895019531, "learning_rate": 4.823602209530978e-06, "loss": 2.0876, "step": 10332 }, { "epoch": 3.2391849529780563, "grad_norm": 54.74850082397461, "learning_rate": 4.82255074364627e-06, "loss": 2.3705, "step": 10333 }, { "epoch": 3.2394984326018808, "grad_norm": 11.48755168914795, "learning_rate": 4.821496268591408e-06, "loss": 1.6742, "step": 10334 }, { "epoch": 3.239811912225705, "grad_norm": 8.086645126342773, "learning_rate": 4.820438785732602e-06, "loss": 2.1766, "step": 10335 }, { "epoch": 3.2401253918495296, "grad_norm": 43.881805419921875, "learning_rate": 4.819378296439962e-06, "loss": 2.3673, "step": 10336 }, { "epoch": 3.240438871473354, "grad_norm": 15.614360809326172, "learning_rate": 4.81831480208749e-06, "loss": 2.0429, "step": 10337 }, { "epoch": 3.2407523510971785, "grad_norm": 7.359541416168213, "learning_rate": 4.8172483040530845e-06, "loss": 2.2163, "step": 10338 }, { "epoch": 3.241065830721003, "grad_norm": 8.983024597167969, "learning_rate": 4.816178803718532e-06, "loss": 1.6976, "step": 10339 }, { "epoch": 3.2413793103448274, "grad_norm": 73.78611755371094, "learning_rate": 4.815106302469513e-06, "loss": 2.6532, "step": 10340 }, { "epoch": 3.241692789968652, "grad_norm": 10.156828880310059, "learning_rate": 4.814030801695594e-06, "loss": 1.9626, "step": 10341 }, { "epoch": 3.2420062695924763, "grad_norm": 17.165592193603516, "learning_rate": 4.812952302790226e-06, "loss": 2.1766, "step": 10342 }, { "epoch": 3.2423197492163007, "grad_norm": 7.247964382171631, "learning_rate": 4.811870807150747e-06, "loss": 2.2771, "step": 10343 }, { "epoch": 3.242633228840125, "grad_norm": 12.131818771362305, "learning_rate": 4.810786316178377e-06, "loss": 2.0874, "step": 10344 }, { "epoch": 3.2429467084639496, "grad_norm": 7.1404290199279785, "learning_rate": 4.809698831278217e-06, "loss": 2.145, "step": 10345 }, { "epoch": 3.2432601880877745, "grad_norm": 9.405982971191406, "learning_rate": 4.808608353859247e-06, "loss": 2.279, "step": 10346 }, { "epoch": 3.243573667711599, "grad_norm": 9.214646339416504, "learning_rate": 4.807514885334324e-06, "loss": 2.018, "step": 10347 }, { "epoch": 3.2438871473354234, "grad_norm": 15.025487899780273, "learning_rate": 4.80641842712018e-06, "loss": 1.9574, "step": 10348 }, { "epoch": 3.244200626959248, "grad_norm": 12.505887031555176, "learning_rate": 4.8053189806374205e-06, "loss": 2.1147, "step": 10349 }, { "epoch": 3.2445141065830723, "grad_norm": 30.858016967773438, "learning_rate": 4.804216547310524e-06, "loss": 2.2292, "step": 10350 }, { "epoch": 3.2448275862068967, "grad_norm": 145.07565307617188, "learning_rate": 4.803111128567838e-06, "loss": 2.4291, "step": 10351 }, { "epoch": 3.245141065830721, "grad_norm": 30.84320831298828, "learning_rate": 4.802002725841577e-06, "loss": 2.3916, "step": 10352 }, { "epoch": 3.2454545454545456, "grad_norm": 10.323012351989746, "learning_rate": 4.800891340567823e-06, "loss": 2.2977, "step": 10353 }, { "epoch": 3.24576802507837, "grad_norm": 8.595511436462402, "learning_rate": 4.799776974186523e-06, "loss": 2.0791, "step": 10354 }, { "epoch": 3.2460815047021945, "grad_norm": 9.700237274169922, "learning_rate": 4.798659628141484e-06, "loss": 2.0393, "step": 10355 }, { "epoch": 3.246394984326019, "grad_norm": 9.655471801757812, "learning_rate": 4.797539303880375e-06, "loss": 2.2179, "step": 10356 }, { "epoch": 3.2467084639498434, "grad_norm": 25.284435272216797, "learning_rate": 4.796416002854724e-06, "loss": 2.1148, "step": 10357 }, { "epoch": 3.247021943573668, "grad_norm": 15.506875991821289, "learning_rate": 4.795289726519915e-06, "loss": 2.1117, "step": 10358 }, { "epoch": 3.2473354231974922, "grad_norm": 44.546043395996094, "learning_rate": 4.794160476335187e-06, "loss": 2.2422, "step": 10359 }, { "epoch": 3.2476489028213167, "grad_norm": 16.44324493408203, "learning_rate": 4.793028253763633e-06, "loss": 2.0253, "step": 10360 }, { "epoch": 3.247962382445141, "grad_norm": 41.03581237792969, "learning_rate": 4.791893060272196e-06, "loss": 2.2656, "step": 10361 }, { "epoch": 3.2482758620689656, "grad_norm": 7.839076042175293, "learning_rate": 4.790754897331668e-06, "loss": 2.0344, "step": 10362 }, { "epoch": 3.24858934169279, "grad_norm": 17.740598678588867, "learning_rate": 4.789613766416689e-06, "loss": 2.3889, "step": 10363 }, { "epoch": 3.2489028213166145, "grad_norm": 9.250628471374512, "learning_rate": 4.788469669005745e-06, "loss": 2.1377, "step": 10364 }, { "epoch": 3.249216300940439, "grad_norm": 25.41570281982422, "learning_rate": 4.7873226065811645e-06, "loss": 2.4065, "step": 10365 }, { "epoch": 3.2495297805642633, "grad_norm": 9.12908935546875, "learning_rate": 4.786172580629118e-06, "loss": 2.1678, "step": 10366 }, { "epoch": 3.249843260188088, "grad_norm": 5.376280307769775, "learning_rate": 4.785019592639615e-06, "loss": 2.1066, "step": 10367 }, { "epoch": 3.250156739811912, "grad_norm": 48.19829559326172, "learning_rate": 4.783863644106502e-06, "loss": 2.322, "step": 10368 }, { "epoch": 3.2504702194357367, "grad_norm": 6.740322113037109, "learning_rate": 4.782704736527466e-06, "loss": 2.0831, "step": 10369 }, { "epoch": 3.250783699059561, "grad_norm": 11.225165367126465, "learning_rate": 4.78154287140402e-06, "loss": 2.4518, "step": 10370 }, { "epoch": 3.2510971786833855, "grad_norm": 9.698965072631836, "learning_rate": 4.780378050241517e-06, "loss": 2.0971, "step": 10371 }, { "epoch": 3.25141065830721, "grad_norm": 126.59722137451172, "learning_rate": 4.7792102745491345e-06, "loss": 2.3834, "step": 10372 }, { "epoch": 3.2517241379310344, "grad_norm": 9.660218238830566, "learning_rate": 4.7780395458398785e-06, "loss": 2.0487, "step": 10373 }, { "epoch": 3.252037617554859, "grad_norm": 13.880728721618652, "learning_rate": 4.776865865630583e-06, "loss": 2.2354, "step": 10374 }, { "epoch": 3.252037617554859, "eval_loss": 2.2600181102752686, "eval_runtime": 21.1969, "eval_samples_per_second": 126.764, "eval_steps_per_second": 7.926, "step": 10374 }, { "epoch": 3.2523510971786833, "grad_norm": 46.527835845947266, "learning_rate": 4.775689235441906e-06, "loss": 2.85, "step": 10375 }, { "epoch": 3.2526645768025078, "grad_norm": 16.594409942626953, "learning_rate": 4.774509656798326e-06, "loss": 2.1882, "step": 10376 }, { "epoch": 3.252978056426332, "grad_norm": 10.438210487365723, "learning_rate": 4.7733271312281415e-06, "loss": 2.5737, "step": 10377 }, { "epoch": 3.2532915360501566, "grad_norm": 10.818836212158203, "learning_rate": 4.772141660263472e-06, "loss": 1.9554, "step": 10378 }, { "epoch": 3.253605015673981, "grad_norm": 12.170820236206055, "learning_rate": 4.77095324544025e-06, "loss": 2.1241, "step": 10379 }, { "epoch": 3.2539184952978055, "grad_norm": 7.167964935302734, "learning_rate": 4.769761888298223e-06, "loss": 2.2016, "step": 10380 }, { "epoch": 3.25423197492163, "grad_norm": 12.129745483398438, "learning_rate": 4.768567590380951e-06, "loss": 1.9695, "step": 10381 }, { "epoch": 3.2545454545454544, "grad_norm": 9.69056510925293, "learning_rate": 4.767370353235805e-06, "loss": 2.2543, "step": 10382 }, { "epoch": 3.254858934169279, "grad_norm": 7.305293560028076, "learning_rate": 4.766170178413964e-06, "loss": 1.9935, "step": 10383 }, { "epoch": 3.2551724137931033, "grad_norm": 9.180840492248535, "learning_rate": 4.764967067470409e-06, "loss": 2.2853, "step": 10384 }, { "epoch": 3.2554858934169277, "grad_norm": 11.487865447998047, "learning_rate": 4.763761021963932e-06, "loss": 2.127, "step": 10385 }, { "epoch": 3.255799373040752, "grad_norm": 7.05090856552124, "learning_rate": 4.762552043457122e-06, "loss": 2.1581, "step": 10386 }, { "epoch": 3.2561128526645766, "grad_norm": 13.31796646118164, "learning_rate": 4.761340133516371e-06, "loss": 2.2403, "step": 10387 }, { "epoch": 3.256426332288401, "grad_norm": 4.173501491546631, "learning_rate": 4.7601252937118665e-06, "loss": 2.0715, "step": 10388 }, { "epoch": 3.256739811912226, "grad_norm": 11.309921264648438, "learning_rate": 4.758907525617594e-06, "loss": 2.0216, "step": 10389 }, { "epoch": 3.2570532915360504, "grad_norm": 29.2979793548584, "learning_rate": 4.757686830811332e-06, "loss": 2.072, "step": 10390 }, { "epoch": 3.257366771159875, "grad_norm": 7.402588367462158, "learning_rate": 4.7564632108746524e-06, "loss": 2.132, "step": 10391 }, { "epoch": 3.2576802507836993, "grad_norm": 17.37984275817871, "learning_rate": 4.755236667392914e-06, "loss": 1.8682, "step": 10392 }, { "epoch": 3.2579937304075237, "grad_norm": 13.152661323547363, "learning_rate": 4.754007201955267e-06, "loss": 2.4216, "step": 10393 }, { "epoch": 3.258307210031348, "grad_norm": 8.95935344696045, "learning_rate": 4.752774816154644e-06, "loss": 1.8313, "step": 10394 }, { "epoch": 3.2586206896551726, "grad_norm": 45.11777114868164, "learning_rate": 4.751539511587765e-06, "loss": 2.3189, "step": 10395 }, { "epoch": 3.258934169278997, "grad_norm": 10.797253608703613, "learning_rate": 4.750301289855128e-06, "loss": 1.9961, "step": 10396 }, { "epoch": 3.2592476489028215, "grad_norm": 126.47699737548828, "learning_rate": 4.749060152561012e-06, "loss": 2.3851, "step": 10397 }, { "epoch": 3.259561128526646, "grad_norm": 16.552989959716797, "learning_rate": 4.747816101313476e-06, "loss": 2.0118, "step": 10398 }, { "epoch": 3.2598746081504704, "grad_norm": 10.174836158752441, "learning_rate": 4.746569137724351e-06, "loss": 2.0644, "step": 10399 }, { "epoch": 3.260188087774295, "grad_norm": 16.841550827026367, "learning_rate": 4.745319263409241e-06, "loss": 2.221, "step": 10400 }, { "epoch": 3.2605015673981192, "grad_norm": 9.622954368591309, "learning_rate": 4.744066479987524e-06, "loss": 2.1373, "step": 10401 }, { "epoch": 3.2608150470219437, "grad_norm": 10.109556198120117, "learning_rate": 4.742810789082345e-06, "loss": 2.054, "step": 10402 }, { "epoch": 3.261128526645768, "grad_norm": 6.1477813720703125, "learning_rate": 4.741552192320618e-06, "loss": 2.0426, "step": 10403 }, { "epoch": 3.2614420062695926, "grad_norm": 8.723674774169922, "learning_rate": 4.740290691333021e-06, "loss": 2.5009, "step": 10404 }, { "epoch": 3.261755485893417, "grad_norm": 24.61112403869629, "learning_rate": 4.7390262877539915e-06, "loss": 2.6446, "step": 10405 }, { "epoch": 3.2620689655172415, "grad_norm": 12.557799339294434, "learning_rate": 4.737758983221734e-06, "loss": 2.2973, "step": 10406 }, { "epoch": 3.262382445141066, "grad_norm": 9.406915664672852, "learning_rate": 4.736488779378207e-06, "loss": 1.919, "step": 10407 }, { "epoch": 3.2626959247648903, "grad_norm": 8.748431205749512, "learning_rate": 4.735215677869129e-06, "loss": 2.0645, "step": 10408 }, { "epoch": 3.263009404388715, "grad_norm": 17.627410888671875, "learning_rate": 4.733939680343966e-06, "loss": 2.1003, "step": 10409 }, { "epoch": 3.2633228840125392, "grad_norm": 13.824841499328613, "learning_rate": 4.732660788455944e-06, "loss": 1.9944, "step": 10410 }, { "epoch": 3.2636363636363637, "grad_norm": 6.7091498374938965, "learning_rate": 4.731379003862036e-06, "loss": 1.912, "step": 10411 }, { "epoch": 3.263949843260188, "grad_norm": 16.23147201538086, "learning_rate": 4.730094328222961e-06, "loss": 2.2401, "step": 10412 }, { "epoch": 3.2642633228840126, "grad_norm": 6.213670253753662, "learning_rate": 4.728806763203185e-06, "loss": 2.1351, "step": 10413 }, { "epoch": 3.264576802507837, "grad_norm": 13.011263847351074, "learning_rate": 4.72751631047092e-06, "loss": 2.1091, "step": 10414 }, { "epoch": 3.2648902821316614, "grad_norm": 19.286300659179688, "learning_rate": 4.726222971698116e-06, "loss": 2.1509, "step": 10415 }, { "epoch": 3.265203761755486, "grad_norm": 10.077512741088867, "learning_rate": 4.724926748560464e-06, "loss": 2.3614, "step": 10416 }, { "epoch": 3.2655172413793103, "grad_norm": 10.3480863571167, "learning_rate": 4.723627642737391e-06, "loss": 1.9134, "step": 10417 }, { "epoch": 3.2658307210031348, "grad_norm": 7.155325412750244, "learning_rate": 4.7223256559120614e-06, "loss": 2.19, "step": 10418 }, { "epoch": 3.266144200626959, "grad_norm": 10.499210357666016, "learning_rate": 4.721020789771368e-06, "loss": 2.2161, "step": 10419 }, { "epoch": 3.2664576802507836, "grad_norm": 99.06182861328125, "learning_rate": 4.7197130460059385e-06, "loss": 2.3174, "step": 10420 }, { "epoch": 3.266771159874608, "grad_norm": 21.754396438598633, "learning_rate": 4.7184024263101256e-06, "loss": 2.0864, "step": 10421 }, { "epoch": 3.2670846394984325, "grad_norm": 43.53794479370117, "learning_rate": 4.717088932382011e-06, "loss": 2.2504, "step": 10422 }, { "epoch": 3.267398119122257, "grad_norm": 8.198076248168945, "learning_rate": 4.715772565923399e-06, "loss": 2.1976, "step": 10423 }, { "epoch": 3.2677115987460814, "grad_norm": 9.555412292480469, "learning_rate": 4.714453328639814e-06, "loss": 1.9331, "step": 10424 }, { "epoch": 3.268025078369906, "grad_norm": 8.589456558227539, "learning_rate": 4.713131222240502e-06, "loss": 1.9377, "step": 10425 }, { "epoch": 3.2683385579937303, "grad_norm": 7.327493667602539, "learning_rate": 4.711806248438428e-06, "loss": 2.1299, "step": 10426 }, { "epoch": 3.2686520376175547, "grad_norm": 13.145010948181152, "learning_rate": 4.7104784089502664e-06, "loss": 2.1513, "step": 10427 }, { "epoch": 3.268965517241379, "grad_norm": 30.21534538269043, "learning_rate": 4.7091477054964105e-06, "loss": 3.0554, "step": 10428 }, { "epoch": 3.2692789968652036, "grad_norm": 5.527064323425293, "learning_rate": 4.707814139800962e-06, "loss": 2.0126, "step": 10429 }, { "epoch": 3.269592476489028, "grad_norm": 138.116943359375, "learning_rate": 4.7064777135917284e-06, "loss": 2.4851, "step": 10430 }, { "epoch": 3.2699059561128525, "grad_norm": 9.801795959472656, "learning_rate": 4.705138428600228e-06, "loss": 2.3413, "step": 10431 }, { "epoch": 3.270219435736677, "grad_norm": 8.715144157409668, "learning_rate": 4.7037962865616795e-06, "loss": 2.1686, "step": 10432 }, { "epoch": 3.2705329153605014, "grad_norm": 10.247964859008789, "learning_rate": 4.702451289215005e-06, "loss": 1.9807, "step": 10433 }, { "epoch": 3.270846394984326, "grad_norm": 8.9537353515625, "learning_rate": 4.701103438302827e-06, "loss": 2.594, "step": 10434 }, { "epoch": 3.2711598746081503, "grad_norm": 25.924991607666016, "learning_rate": 4.69975273557146e-06, "loss": 2.2794, "step": 10435 }, { "epoch": 3.2714733542319747, "grad_norm": 7.841407299041748, "learning_rate": 4.698399182770921e-06, "loss": 2.038, "step": 10436 }, { "epoch": 3.271786833855799, "grad_norm": 90.70893859863281, "learning_rate": 4.697042781654913e-06, "loss": 2.4831, "step": 10437 }, { "epoch": 3.2721003134796236, "grad_norm": 7.357234001159668, "learning_rate": 4.695683533980835e-06, "loss": 2.4035, "step": 10438 }, { "epoch": 3.272413793103448, "grad_norm": 11.27476692199707, "learning_rate": 4.694321441509769e-06, "loss": 2.2272, "step": 10439 }, { "epoch": 3.2727272727272725, "grad_norm": 7.490302085876465, "learning_rate": 4.692956506006486e-06, "loss": 2.251, "step": 10440 }, { "epoch": 3.2730407523510974, "grad_norm": 7.2692341804504395, "learning_rate": 4.6915887292394395e-06, "loss": 1.9701, "step": 10441 }, { "epoch": 3.273354231974922, "grad_norm": 22.438905715942383, "learning_rate": 4.690218112980763e-06, "loss": 1.9807, "step": 10442 }, { "epoch": 3.2736677115987463, "grad_norm": 5.56432580947876, "learning_rate": 4.688844659006272e-06, "loss": 1.951, "step": 10443 }, { "epoch": 3.2739811912225707, "grad_norm": 6.554599761962891, "learning_rate": 4.687468369095457e-06, "loss": 2.0409, "step": 10444 }, { "epoch": 3.274294670846395, "grad_norm": 10.092485427856445, "learning_rate": 4.686089245031481e-06, "loss": 2.0089, "step": 10445 }, { "epoch": 3.2746081504702196, "grad_norm": 7.641839981079102, "learning_rate": 4.684707288601182e-06, "loss": 1.9374, "step": 10446 }, { "epoch": 3.274921630094044, "grad_norm": 7.833014488220215, "learning_rate": 4.683322501595066e-06, "loss": 1.9843, "step": 10447 }, { "epoch": 3.2752351097178685, "grad_norm": 17.84506607055664, "learning_rate": 4.681934885807307e-06, "loss": 2.1336, "step": 10448 }, { "epoch": 3.275548589341693, "grad_norm": 82.46455383300781, "learning_rate": 4.680544443035744e-06, "loss": 2.4479, "step": 10449 }, { "epoch": 3.2758620689655173, "grad_norm": 10.83971881866455, "learning_rate": 4.679151175081879e-06, "loss": 2.1703, "step": 10450 }, { "epoch": 3.276175548589342, "grad_norm": 8.121395111083984, "learning_rate": 4.677755083750872e-06, "loss": 2.0442, "step": 10451 }, { "epoch": 3.2764890282131662, "grad_norm": 11.723223686218262, "learning_rate": 4.676356170851545e-06, "loss": 1.9114, "step": 10452 }, { "epoch": 3.2768025078369907, "grad_norm": 8.587835311889648, "learning_rate": 4.674954438196374e-06, "loss": 2.3379, "step": 10453 }, { "epoch": 3.277115987460815, "grad_norm": 33.29867172241211, "learning_rate": 4.673549887601486e-06, "loss": 2.576, "step": 10454 }, { "epoch": 3.2774294670846396, "grad_norm": 6.669020175933838, "learning_rate": 4.6721425208866646e-06, "loss": 2.2817, "step": 10455 }, { "epoch": 3.277742946708464, "grad_norm": 284.3402099609375, "learning_rate": 4.6707323398753346e-06, "loss": 2.5671, "step": 10456 }, { "epoch": 3.2780564263322884, "grad_norm": 9.107257843017578, "learning_rate": 4.669319346394574e-06, "loss": 2.5478, "step": 10457 }, { "epoch": 3.278369905956113, "grad_norm": 19.172096252441406, "learning_rate": 4.6679035422751e-06, "loss": 2.2971, "step": 10458 }, { "epoch": 3.2786833855799373, "grad_norm": 11.62237548828125, "learning_rate": 4.666484929351275e-06, "loss": 2.2724, "step": 10459 }, { "epoch": 3.2789968652037618, "grad_norm": 10.682868957519531, "learning_rate": 4.665063509461098e-06, "loss": 2.085, "step": 10460 }, { "epoch": 3.279310344827586, "grad_norm": 8.384634971618652, "learning_rate": 4.663639284446204e-06, "loss": 2.137, "step": 10461 }, { "epoch": 3.2796238244514107, "grad_norm": 11.892141342163086, "learning_rate": 4.662212256151865e-06, "loss": 2.3115, "step": 10462 }, { "epoch": 3.279937304075235, "grad_norm": 6.1182379722595215, "learning_rate": 4.660782426426985e-06, "loss": 2.0586, "step": 10463 }, { "epoch": 3.2802507836990595, "grad_norm": 10.7253999710083, "learning_rate": 4.659349797124096e-06, "loss": 2.1253, "step": 10464 }, { "epoch": 3.280564263322884, "grad_norm": 11.572819709777832, "learning_rate": 4.657914370099357e-06, "loss": 1.8696, "step": 10465 }, { "epoch": 3.2808777429467084, "grad_norm": 12.076330184936523, "learning_rate": 4.656476147212554e-06, "loss": 2.1268, "step": 10466 }, { "epoch": 3.281191222570533, "grad_norm": 29.247495651245117, "learning_rate": 4.6550351303270924e-06, "loss": 2.0302, "step": 10467 }, { "epoch": 3.2815047021943573, "grad_norm": 9.934165000915527, "learning_rate": 4.6535913213100005e-06, "loss": 2.2102, "step": 10468 }, { "epoch": 3.2818181818181817, "grad_norm": 15.2339506149292, "learning_rate": 4.652144722031922e-06, "loss": 2.1644, "step": 10469 }, { "epoch": 3.282131661442006, "grad_norm": 9.378424644470215, "learning_rate": 4.650695334367118e-06, "loss": 2.2531, "step": 10470 }, { "epoch": 3.2824451410658306, "grad_norm": 9.435429573059082, "learning_rate": 4.64924316019346e-06, "loss": 1.9331, "step": 10471 }, { "epoch": 3.282758620689655, "grad_norm": 9.564239501953125, "learning_rate": 4.647788201392429e-06, "loss": 2.1116, "step": 10472 }, { "epoch": 3.2830721003134795, "grad_norm": 7.870461463928223, "learning_rate": 4.6463304598491196e-06, "loss": 2.1544, "step": 10473 }, { "epoch": 3.283385579937304, "grad_norm": 49.229209899902344, "learning_rate": 4.644869937452224e-06, "loss": 2.0037, "step": 10474 }, { "epoch": 3.2836990595611284, "grad_norm": 8.002914428710938, "learning_rate": 4.643406636094045e-06, "loss": 1.9227, "step": 10475 }, { "epoch": 3.284012539184953, "grad_norm": 10.762214660644531, "learning_rate": 4.641940557670478e-06, "loss": 1.7552, "step": 10476 }, { "epoch": 3.2843260188087773, "grad_norm": 8.29768180847168, "learning_rate": 4.6404717040810235e-06, "loss": 2.2675, "step": 10477 }, { "epoch": 3.2846394984326017, "grad_norm": 10.743518829345703, "learning_rate": 4.639000077228773e-06, "loss": 1.9389, "step": 10478 }, { "epoch": 3.284952978056426, "grad_norm": 7.485596179962158, "learning_rate": 4.637525679020414e-06, "loss": 2.2195, "step": 10479 }, { "epoch": 3.2852664576802506, "grad_norm": 7.215515613555908, "learning_rate": 4.636048511366222e-06, "loss": 2.0942, "step": 10480 }, { "epoch": 3.285579937304075, "grad_norm": 6.592324733734131, "learning_rate": 4.634568576180063e-06, "loss": 1.9999, "step": 10481 }, { "epoch": 3.2858934169279, "grad_norm": 8.913631439208984, "learning_rate": 4.633085875379388e-06, "loss": 2.0415, "step": 10482 }, { "epoch": 3.2862068965517244, "grad_norm": 30.650266647338867, "learning_rate": 4.631600410885231e-06, "loss": 2.2775, "step": 10483 }, { "epoch": 3.286520376175549, "grad_norm": 11.043688774108887, "learning_rate": 4.630112184622207e-06, "loss": 1.7416, "step": 10484 }, { "epoch": 3.2868338557993733, "grad_norm": 62.258541107177734, "learning_rate": 4.628621198518507e-06, "loss": 2.4139, "step": 10485 }, { "epoch": 3.2871473354231977, "grad_norm": 7.590291500091553, "learning_rate": 4.627127454505902e-06, "loss": 2.0398, "step": 10486 }, { "epoch": 3.287460815047022, "grad_norm": 8.70261287689209, "learning_rate": 4.625630954519733e-06, "loss": 1.9384, "step": 10487 }, { "epoch": 3.2877742946708466, "grad_norm": 68.99636840820312, "learning_rate": 4.624131700498913e-06, "loss": 1.8825, "step": 10488 }, { "epoch": 3.288087774294671, "grad_norm": 54.55644607543945, "learning_rate": 4.6226296943859225e-06, "loss": 2.3678, "step": 10489 }, { "epoch": 3.2884012539184955, "grad_norm": 6.302337646484375, "learning_rate": 4.621124938126809e-06, "loss": 2.0009, "step": 10490 }, { "epoch": 3.28871473354232, "grad_norm": 7.0113444328308105, "learning_rate": 4.619617433671181e-06, "loss": 2.3015, "step": 10491 }, { "epoch": 3.2890282131661444, "grad_norm": 7.7312469482421875, "learning_rate": 4.618107182972209e-06, "loss": 2.2354, "step": 10492 }, { "epoch": 3.289341692789969, "grad_norm": 11.6555814743042, "learning_rate": 4.6165941879866225e-06, "loss": 1.9589, "step": 10493 }, { "epoch": 3.2896551724137932, "grad_norm": 14.91659164428711, "learning_rate": 4.615078450674706e-06, "loss": 2.4757, "step": 10494 }, { "epoch": 3.2899686520376177, "grad_norm": 8.412989616394043, "learning_rate": 4.613559973000295e-06, "loss": 2.045, "step": 10495 }, { "epoch": 3.290282131661442, "grad_norm": 7.8893656730651855, "learning_rate": 4.612038756930778e-06, "loss": 1.9222, "step": 10496 }, { "epoch": 3.2905956112852666, "grad_norm": 7.117119312286377, "learning_rate": 4.610514804437091e-06, "loss": 1.9772, "step": 10497 }, { "epoch": 3.290909090909091, "grad_norm": 120.8690185546875, "learning_rate": 4.6089881174937146e-06, "loss": 2.5146, "step": 10498 }, { "epoch": 3.2912225705329154, "grad_norm": 7.320493221282959, "learning_rate": 4.607458698078673e-06, "loss": 2.0375, "step": 10499 }, { "epoch": 3.29153605015674, "grad_norm": 8.508760452270508, "learning_rate": 4.6059265481735295e-06, "loss": 2.1363, "step": 10500 }, { "epoch": 3.2918495297805643, "grad_norm": 19.052244186401367, "learning_rate": 4.604391669763386e-06, "loss": 2.4773, "step": 10501 }, { "epoch": 3.2921630094043888, "grad_norm": 9.841686248779297, "learning_rate": 4.602854064836881e-06, "loss": 2.2128, "step": 10502 }, { "epoch": 3.292476489028213, "grad_norm": 12.013845443725586, "learning_rate": 4.601313735386183e-06, "loss": 2.0414, "step": 10503 }, { "epoch": 3.2927899686520377, "grad_norm": 25.531024932861328, "learning_rate": 4.599770683406992e-06, "loss": 2.1491, "step": 10504 }, { "epoch": 3.293103448275862, "grad_norm": 6.9195990562438965, "learning_rate": 4.598224910898534e-06, "loss": 2.1143, "step": 10505 }, { "epoch": 3.2934169278996865, "grad_norm": 28.196706771850586, "learning_rate": 4.596676419863561e-06, "loss": 2.3084, "step": 10506 }, { "epoch": 3.293730407523511, "grad_norm": 14.49538803100586, "learning_rate": 4.595125212308347e-06, "loss": 2.3327, "step": 10507 }, { "epoch": 3.2940438871473354, "grad_norm": 53.91624450683594, "learning_rate": 4.593571290242685e-06, "loss": 2.2059, "step": 10508 }, { "epoch": 3.29435736677116, "grad_norm": 52.86344528198242, "learning_rate": 4.592014655679888e-06, "loss": 2.1242, "step": 10509 }, { "epoch": 3.2946708463949843, "grad_norm": 41.343894958496094, "learning_rate": 4.590455310636778e-06, "loss": 2.4776, "step": 10510 }, { "epoch": 3.2949843260188088, "grad_norm": 14.026541709899902, "learning_rate": 4.588893257133692e-06, "loss": 2.2336, "step": 10511 }, { "epoch": 3.295297805642633, "grad_norm": 8.391745567321777, "learning_rate": 4.587328497194478e-06, "loss": 2.1112, "step": 10512 }, { "epoch": 3.2956112852664576, "grad_norm": 8.39424991607666, "learning_rate": 4.585761032846488e-06, "loss": 1.8825, "step": 10513 }, { "epoch": 3.295924764890282, "grad_norm": 6.620738983154297, "learning_rate": 4.584190866120576e-06, "loss": 1.7994, "step": 10514 }, { "epoch": 3.2962382445141065, "grad_norm": 11.981104850769043, "learning_rate": 4.582617999051102e-06, "loss": 2.1997, "step": 10515 }, { "epoch": 3.296551724137931, "grad_norm": 7.295839786529541, "learning_rate": 4.58104243367592e-06, "loss": 2.0043, "step": 10516 }, { "epoch": 3.2968652037617554, "grad_norm": 9.454835891723633, "learning_rate": 4.579464172036386e-06, "loss": 2.618, "step": 10517 }, { "epoch": 3.29717868338558, "grad_norm": 12.250143051147461, "learning_rate": 4.577883216177342e-06, "loss": 2.3298, "step": 10518 }, { "epoch": 3.2974921630094043, "grad_norm": 7.611923694610596, "learning_rate": 4.576299568147127e-06, "loss": 2.3694, "step": 10519 }, { "epoch": 3.2978056426332287, "grad_norm": 12.556456565856934, "learning_rate": 4.5747132299975634e-06, "loss": 1.6992, "step": 10520 }, { "epoch": 3.298119122257053, "grad_norm": 12.933649063110352, "learning_rate": 4.573124203783962e-06, "loss": 1.8866, "step": 10521 }, { "epoch": 3.2984326018808776, "grad_norm": 8.405144691467285, "learning_rate": 4.571532491565115e-06, "loss": 2.1254, "step": 10522 }, { "epoch": 3.298746081504702, "grad_norm": 6.481987476348877, "learning_rate": 4.569938095403296e-06, "loss": 2.1173, "step": 10523 }, { "epoch": 3.2990595611285265, "grad_norm": 7.640778541564941, "learning_rate": 4.568341017364254e-06, "loss": 2.0806, "step": 10524 }, { "epoch": 3.299373040752351, "grad_norm": 9.014044761657715, "learning_rate": 4.566741259517214e-06, "loss": 2.2309, "step": 10525 }, { "epoch": 3.2996865203761754, "grad_norm": 5.837136268615723, "learning_rate": 4.565138823934874e-06, "loss": 2.037, "step": 10526 }, { "epoch": 3.3, "grad_norm": 101.85710906982422, "learning_rate": 4.563533712693399e-06, "loss": 2.0999, "step": 10527 }, { "epoch": 3.3003134796238243, "grad_norm": 8.360709190368652, "learning_rate": 4.561925927872421e-06, "loss": 2.0076, "step": 10528 }, { "epoch": 3.3006269592476487, "grad_norm": 7.278694152832031, "learning_rate": 4.560315471555039e-06, "loss": 2.0726, "step": 10529 }, { "epoch": 3.300940438871473, "grad_norm": 6.338781833648682, "learning_rate": 4.55870234582781e-06, "loss": 2.1914, "step": 10530 }, { "epoch": 3.3012539184952976, "grad_norm": 7.050608158111572, "learning_rate": 4.55708655278075e-06, "loss": 2.1935, "step": 10531 }, { "epoch": 3.301567398119122, "grad_norm": 6.537271976470947, "learning_rate": 4.555468094507334e-06, "loss": 2.0212, "step": 10532 }, { "epoch": 3.3018808777429465, "grad_norm": 9.044611930847168, "learning_rate": 4.5538469731044844e-06, "loss": 2.3177, "step": 10533 }, { "epoch": 3.302194357366771, "grad_norm": 15.778304100036621, "learning_rate": 4.55222319067258e-06, "loss": 2.1738, "step": 10534 }, { "epoch": 3.302507836990596, "grad_norm": 8.753870010375977, "learning_rate": 4.550596749315443e-06, "loss": 2.0819, "step": 10535 }, { "epoch": 3.3028213166144202, "grad_norm": 60.38290023803711, "learning_rate": 4.548967651140341e-06, "loss": 2.0147, "step": 10536 }, { "epoch": 3.3031347962382447, "grad_norm": 11.745752334594727, "learning_rate": 4.547335898257989e-06, "loss": 1.9596, "step": 10537 }, { "epoch": 3.303448275862069, "grad_norm": 9.124713897705078, "learning_rate": 4.545701492782535e-06, "loss": 2.1182, "step": 10538 }, { "epoch": 3.3037617554858936, "grad_norm": 10.425442695617676, "learning_rate": 4.5440644368315665e-06, "loss": 2.2141, "step": 10539 }, { "epoch": 3.304075235109718, "grad_norm": 9.204667091369629, "learning_rate": 4.542424732526105e-06, "loss": 1.8912, "step": 10540 }, { "epoch": 3.3043887147335425, "grad_norm": 29.369857788085938, "learning_rate": 4.540782381990604e-06, "loss": 2.4693, "step": 10541 }, { "epoch": 3.304702194357367, "grad_norm": 6.793385028839111, "learning_rate": 4.539137387352945e-06, "loss": 2.0111, "step": 10542 }, { "epoch": 3.3050156739811913, "grad_norm": 21.462987899780273, "learning_rate": 4.537489750744434e-06, "loss": 2.57, "step": 10543 }, { "epoch": 3.305329153605016, "grad_norm": 6.233617305755615, "learning_rate": 4.5358394742998e-06, "loss": 2.1908, "step": 10544 }, { "epoch": 3.30564263322884, "grad_norm": 5.586977005004883, "learning_rate": 4.534186560157196e-06, "loss": 2.3305, "step": 10545 }, { "epoch": 3.3059561128526647, "grad_norm": 7.0360612869262695, "learning_rate": 4.532531010458188e-06, "loss": 2.0999, "step": 10546 }, { "epoch": 3.306269592476489, "grad_norm": 9.117305755615234, "learning_rate": 4.5308728273477594e-06, "loss": 1.9703, "step": 10547 }, { "epoch": 3.3065830721003135, "grad_norm": 19.407020568847656, "learning_rate": 4.5292120129743044e-06, "loss": 2.7353, "step": 10548 }, { "epoch": 3.306896551724138, "grad_norm": 6.835867881774902, "learning_rate": 4.527548569489626e-06, "loss": 2.0551, "step": 10549 }, { "epoch": 3.3072100313479624, "grad_norm": 14.972426414489746, "learning_rate": 4.525882499048935e-06, "loss": 2.8677, "step": 10550 }, { "epoch": 3.307523510971787, "grad_norm": 7.417217254638672, "learning_rate": 4.524213803810844e-06, "loss": 1.8852, "step": 10551 }, { "epoch": 3.3078369905956113, "grad_norm": 16.260683059692383, "learning_rate": 4.522542485937369e-06, "loss": 2.608, "step": 10552 }, { "epoch": 3.3081504702194358, "grad_norm": 21.031753540039062, "learning_rate": 4.520868547593921e-06, "loss": 1.6914, "step": 10553 }, { "epoch": 3.30846394984326, "grad_norm": 25.315378189086914, "learning_rate": 4.519191990949307e-06, "loss": 2.0309, "step": 10554 }, { "epoch": 3.3087774294670846, "grad_norm": 13.460307121276855, "learning_rate": 4.517512818175726e-06, "loss": 2.2108, "step": 10555 }, { "epoch": 3.309090909090909, "grad_norm": 7.529034614562988, "learning_rate": 4.51583103144877e-06, "loss": 2.0819, "step": 10556 }, { "epoch": 3.3094043887147335, "grad_norm": 31.585485458374023, "learning_rate": 4.514146632947415e-06, "loss": 1.9471, "step": 10557 }, { "epoch": 3.309717868338558, "grad_norm": 6.0264892578125, "learning_rate": 4.512459624854017e-06, "loss": 2.062, "step": 10558 }, { "epoch": 3.3100313479623824, "grad_norm": 6.8912553787231445, "learning_rate": 4.5107700093543215e-06, "loss": 1.91, "step": 10559 }, { "epoch": 3.310344827586207, "grad_norm": 10.873221397399902, "learning_rate": 4.509077788637446e-06, "loss": 2.2393, "step": 10560 }, { "epoch": 3.3106583072100313, "grad_norm": 28.232507705688477, "learning_rate": 4.507382964895885e-06, "loss": 2.0571, "step": 10561 }, { "epoch": 3.3109717868338557, "grad_norm": 10.064257621765137, "learning_rate": 4.505685540325504e-06, "loss": 2.0624, "step": 10562 }, { "epoch": 3.31128526645768, "grad_norm": 22.21202278137207, "learning_rate": 4.503985517125543e-06, "loss": 2.144, "step": 10563 }, { "epoch": 3.3115987460815046, "grad_norm": 52.697509765625, "learning_rate": 4.5022828974986044e-06, "loss": 1.9421, "step": 10564 }, { "epoch": 3.311912225705329, "grad_norm": 14.799664497375488, "learning_rate": 4.500577683650656e-06, "loss": 2.3043, "step": 10565 }, { "epoch": 3.3122257053291535, "grad_norm": 13.567487716674805, "learning_rate": 4.498869877791026e-06, "loss": 2.1563, "step": 10566 }, { "epoch": 3.312539184952978, "grad_norm": 44.73188018798828, "learning_rate": 4.497159482132404e-06, "loss": 2.2046, "step": 10567 }, { "epoch": 3.3128526645768024, "grad_norm": 41.122501373291016, "learning_rate": 4.4954464988908306e-06, "loss": 2.7747, "step": 10568 }, { "epoch": 3.313166144200627, "grad_norm": 67.87079620361328, "learning_rate": 4.493730930285702e-06, "loss": 2.3926, "step": 10569 }, { "epoch": 3.3134796238244513, "grad_norm": 7.390591621398926, "learning_rate": 4.4920127785397615e-06, "loss": 1.9369, "step": 10570 }, { "epoch": 3.3137931034482757, "grad_norm": 6.170122146606445, "learning_rate": 4.490292045879103e-06, "loss": 1.938, "step": 10571 }, { "epoch": 3.3141065830721, "grad_norm": 6.6823272705078125, "learning_rate": 4.488568734533161e-06, "loss": 2.2302, "step": 10572 }, { "epoch": 3.3144200626959246, "grad_norm": 10.845130920410156, "learning_rate": 4.486842846734712e-06, "loss": 2.0129, "step": 10573 }, { "epoch": 3.314733542319749, "grad_norm": 6.4328227043151855, "learning_rate": 4.4851143847198706e-06, "loss": 2.2318, "step": 10574 }, { "epoch": 3.3150470219435735, "grad_norm": 9.398799896240234, "learning_rate": 4.4833833507280884e-06, "loss": 2.2984, "step": 10575 }, { "epoch": 3.3153605015673984, "grad_norm": 9.751736640930176, "learning_rate": 4.481649747002146e-06, "loss": 2.0736, "step": 10576 }, { "epoch": 3.315673981191223, "grad_norm": 6.572675704956055, "learning_rate": 4.479913575788156e-06, "loss": 1.9561, "step": 10577 }, { "epoch": 3.3159874608150472, "grad_norm": 11.857147216796875, "learning_rate": 4.478174839335556e-06, "loss": 2.257, "step": 10578 }, { "epoch": 3.3163009404388717, "grad_norm": 7.344451904296875, "learning_rate": 4.476433539897109e-06, "loss": 2.2431, "step": 10579 }, { "epoch": 3.316614420062696, "grad_norm": 5.111598968505859, "learning_rate": 4.474689679728897e-06, "loss": 2.172, "step": 10580 }, { "epoch": 3.3169278996865206, "grad_norm": 6.215763568878174, "learning_rate": 4.47294326109032e-06, "loss": 2.074, "step": 10581 }, { "epoch": 3.317241379310345, "grad_norm": 10.93439769744873, "learning_rate": 4.471194286244094e-06, "loss": 1.8825, "step": 10582 }, { "epoch": 3.3175548589341695, "grad_norm": 5.86237096786499, "learning_rate": 4.469442757456246e-06, "loss": 2.1207, "step": 10583 }, { "epoch": 3.317868338557994, "grad_norm": 19.579931259155273, "learning_rate": 4.467688676996111e-06, "loss": 2.1798, "step": 10584 }, { "epoch": 3.3181818181818183, "grad_norm": 7.737973690032959, "learning_rate": 4.465932047136331e-06, "loss": 2.0573, "step": 10585 }, { "epoch": 3.318495297805643, "grad_norm": 6.775116920471191, "learning_rate": 4.4641728701528535e-06, "loss": 2.0408, "step": 10586 }, { "epoch": 3.3188087774294672, "grad_norm": 9.981524467468262, "learning_rate": 4.462411148324922e-06, "loss": 1.9763, "step": 10587 }, { "epoch": 3.3191222570532917, "grad_norm": 37.12565612792969, "learning_rate": 4.460646883935079e-06, "loss": 1.6032, "step": 10588 }, { "epoch": 3.319435736677116, "grad_norm": 6.987195014953613, "learning_rate": 4.458880079269161e-06, "loss": 1.996, "step": 10589 }, { "epoch": 3.3197492163009406, "grad_norm": 11.206425666809082, "learning_rate": 4.457110736616297e-06, "loss": 1.9071, "step": 10590 }, { "epoch": 3.320062695924765, "grad_norm": 8.5078706741333, "learning_rate": 4.455338858268903e-06, "loss": 2.2492, "step": 10591 }, { "epoch": 3.3203761755485894, "grad_norm": 17.317359924316406, "learning_rate": 4.4535644465226795e-06, "loss": 1.8946, "step": 10592 }, { "epoch": 3.320689655172414, "grad_norm": 10.194339752197266, "learning_rate": 4.451787503676612e-06, "loss": 2.0866, "step": 10593 }, { "epoch": 3.3210031347962383, "grad_norm": 11.17991828918457, "learning_rate": 4.4500080320329615e-06, "loss": 2.1482, "step": 10594 }, { "epoch": 3.3213166144200628, "grad_norm": 157.06407165527344, "learning_rate": 4.448226033897271e-06, "loss": 2.2378, "step": 10595 }, { "epoch": 3.321630094043887, "grad_norm": 7.081401348114014, "learning_rate": 4.446441511578351e-06, "loss": 2.072, "step": 10596 }, { "epoch": 3.3219435736677116, "grad_norm": 6.328521251678467, "learning_rate": 4.444654467388286e-06, "loss": 1.9127, "step": 10597 }, { "epoch": 3.322257053291536, "grad_norm": 8.459552764892578, "learning_rate": 4.442864903642428e-06, "loss": 2.2318, "step": 10598 }, { "epoch": 3.3225705329153605, "grad_norm": 7.550581455230713, "learning_rate": 4.441072822659389e-06, "loss": 2.2025, "step": 10599 }, { "epoch": 3.322884012539185, "grad_norm": 12.635766983032227, "learning_rate": 4.43927822676105e-06, "loss": 2.0703, "step": 10600 }, { "epoch": 3.3231974921630094, "grad_norm": 14.8805513381958, "learning_rate": 4.437481118272543e-06, "loss": 2.0603, "step": 10601 }, { "epoch": 3.323510971786834, "grad_norm": 28.298828125, "learning_rate": 4.43568149952226e-06, "loss": 1.9196, "step": 10602 }, { "epoch": 3.3238244514106583, "grad_norm": 18.404848098754883, "learning_rate": 4.433879372841844e-06, "loss": 2.1083, "step": 10603 }, { "epoch": 3.3241379310344827, "grad_norm": 17.702035903930664, "learning_rate": 4.432074740566185e-06, "loss": 2.0459, "step": 10604 }, { "epoch": 3.324451410658307, "grad_norm": 9.670708656311035, "learning_rate": 4.4302676050334255e-06, "loss": 2.2026, "step": 10605 }, { "epoch": 3.3247648902821316, "grad_norm": 9.294676780700684, "learning_rate": 4.428457968584945e-06, "loss": 1.8637, "step": 10606 }, { "epoch": 3.325078369905956, "grad_norm": 11.56635570526123, "learning_rate": 4.426645833565366e-06, "loss": 2.0998, "step": 10607 }, { "epoch": 3.3253918495297805, "grad_norm": 10.114940643310547, "learning_rate": 4.424831202322548e-06, "loss": 1.9711, "step": 10608 }, { "epoch": 3.325705329153605, "grad_norm": 8.873268127441406, "learning_rate": 4.423014077207585e-06, "loss": 3.512, "step": 10609 }, { "epoch": 3.3260188087774294, "grad_norm": 16.395692825317383, "learning_rate": 4.4211944605748016e-06, "loss": 2.0916, "step": 10610 }, { "epoch": 3.326332288401254, "grad_norm": 103.45442962646484, "learning_rate": 4.4193723547817494e-06, "loss": 2.3516, "step": 10611 }, { "epoch": 3.3266457680250783, "grad_norm": 14.908660888671875, "learning_rate": 4.417547762189207e-06, "loss": 2.196, "step": 10612 }, { "epoch": 3.3269592476489027, "grad_norm": 191.55343627929688, "learning_rate": 4.4157206851611754e-06, "loss": 2.3475, "step": 10613 }, { "epoch": 3.327272727272727, "grad_norm": 7.88029670715332, "learning_rate": 4.413891126064872e-06, "loss": 2.3723, "step": 10614 }, { "epoch": 3.3275862068965516, "grad_norm": 23.775056838989258, "learning_rate": 4.412059087270732e-06, "loss": 2.1727, "step": 10615 }, { "epoch": 3.327899686520376, "grad_norm": 26.0772762298584, "learning_rate": 4.410224571152402e-06, "loss": 2.2823, "step": 10616 }, { "epoch": 3.3282131661442005, "grad_norm": 9.880824089050293, "learning_rate": 4.408387580086741e-06, "loss": 1.7552, "step": 10617 }, { "epoch": 3.328526645768025, "grad_norm": 6.011771202087402, "learning_rate": 4.40654811645381e-06, "loss": 2.1222, "step": 10618 }, { "epoch": 3.3288401253918494, "grad_norm": 9.072969436645508, "learning_rate": 4.404706182636878e-06, "loss": 2.0812, "step": 10619 }, { "epoch": 3.329153605015674, "grad_norm": 12.775172233581543, "learning_rate": 4.4028617810224125e-06, "loss": 1.9771, "step": 10620 }, { "epoch": 3.3294670846394983, "grad_norm": 28.611703872680664, "learning_rate": 4.401014914000078e-06, "loss": 2.2296, "step": 10621 }, { "epoch": 3.3297805642633227, "grad_norm": 18.85086441040039, "learning_rate": 4.399165583962734e-06, "loss": 2.314, "step": 10622 }, { "epoch": 3.330094043887147, "grad_norm": 7.410982608795166, "learning_rate": 4.397313793306431e-06, "loss": 2.2559, "step": 10623 }, { "epoch": 3.3304075235109716, "grad_norm": 9.493101119995117, "learning_rate": 4.395459544430407e-06, "loss": 2.1583, "step": 10624 }, { "epoch": 3.330721003134796, "grad_norm": 6.791162967681885, "learning_rate": 4.393602839737087e-06, "loss": 2.1051, "step": 10625 }, { "epoch": 3.3310344827586205, "grad_norm": 166.51715087890625, "learning_rate": 4.391743681632075e-06, "loss": 2.2164, "step": 10626 }, { "epoch": 3.331347962382445, "grad_norm": 8.590611457824707, "learning_rate": 4.3898820725241544e-06, "loss": 1.9744, "step": 10627 }, { "epoch": 3.3316614420062693, "grad_norm": 5.063475608825684, "learning_rate": 4.388018014825287e-06, "loss": 2.2486, "step": 10628 }, { "epoch": 3.3319749216300942, "grad_norm": 136.18661499023438, "learning_rate": 4.386151510950602e-06, "loss": 1.8571, "step": 10629 }, { "epoch": 3.3322884012539187, "grad_norm": 46.020137786865234, "learning_rate": 4.384282563318403e-06, "loss": 2.4314, "step": 10630 }, { "epoch": 3.332601880877743, "grad_norm": 8.841217994689941, "learning_rate": 4.382411174350157e-06, "loss": 2.0914, "step": 10631 }, { "epoch": 3.3329153605015676, "grad_norm": 7.53122615814209, "learning_rate": 4.380537346470495e-06, "loss": 2.1038, "step": 10632 }, { "epoch": 3.333228840125392, "grad_norm": 5.751541614532471, "learning_rate": 4.378661082107207e-06, "loss": 2.5677, "step": 10633 }, { "epoch": 3.3335423197492164, "grad_norm": 12.781482696533203, "learning_rate": 4.37678238369124e-06, "loss": 2.164, "step": 10634 }, { "epoch": 3.333855799373041, "grad_norm": 7.5511980056762695, "learning_rate": 4.374901253656697e-06, "loss": 2.1939, "step": 10635 }, { "epoch": 3.3341692789968653, "grad_norm": 52.25824737548828, "learning_rate": 4.373017694440828e-06, "loss": 2.741, "step": 10636 }, { "epoch": 3.3344827586206898, "grad_norm": 8.277274131774902, "learning_rate": 4.3711317084840325e-06, "loss": 2.0904, "step": 10637 }, { "epoch": 3.334796238244514, "grad_norm": 120.23431396484375, "learning_rate": 4.3692432982298515e-06, "loss": 2.6316, "step": 10638 }, { "epoch": 3.3351097178683387, "grad_norm": 8.727564811706543, "learning_rate": 4.367352466124972e-06, "loss": 2.045, "step": 10639 }, { "epoch": 3.335423197492163, "grad_norm": 8.940752983093262, "learning_rate": 4.3654592146192146e-06, "loss": 2.1773, "step": 10640 }, { "epoch": 3.3357366771159875, "grad_norm": 22.79124641418457, "learning_rate": 4.3635635461655345e-06, "loss": 2.1215, "step": 10641 }, { "epoch": 3.336050156739812, "grad_norm": 8.699186325073242, "learning_rate": 4.361665463220023e-06, "loss": 2.0663, "step": 10642 }, { "epoch": 3.3363636363636364, "grad_norm": 7.220518112182617, "learning_rate": 4.359764968241892e-06, "loss": 2.2074, "step": 10643 }, { "epoch": 3.336677115987461, "grad_norm": 6.068549156188965, "learning_rate": 4.357862063693486e-06, "loss": 2.0819, "step": 10644 }, { "epoch": 3.3369905956112853, "grad_norm": 13.745779991149902, "learning_rate": 4.355956752040267e-06, "loss": 2.4479, "step": 10645 }, { "epoch": 3.3373040752351097, "grad_norm": 9.521445274353027, "learning_rate": 4.354049035750818e-06, "loss": 2.7339, "step": 10646 }, { "epoch": 3.337617554858934, "grad_norm": 7.462592124938965, "learning_rate": 4.352138917296836e-06, "loss": 2.1377, "step": 10647 }, { "epoch": 3.3379310344827586, "grad_norm": 13.738550186157227, "learning_rate": 4.35022639915313e-06, "loss": 1.9165, "step": 10648 }, { "epoch": 3.338244514106583, "grad_norm": 12.895854949951172, "learning_rate": 4.34831148379762e-06, "loss": 2.1331, "step": 10649 }, { "epoch": 3.3385579937304075, "grad_norm": 24.07604217529297, "learning_rate": 4.346394173711331e-06, "loss": 2.2519, "step": 10650 }, { "epoch": 3.338871473354232, "grad_norm": 8.710412979125977, "learning_rate": 4.344474471378389e-06, "loss": 2.2082, "step": 10651 }, { "epoch": 3.3391849529780564, "grad_norm": 66.51384735107422, "learning_rate": 4.3425523792860234e-06, "loss": 2.3036, "step": 10652 }, { "epoch": 3.339498432601881, "grad_norm": 10.355724334716797, "learning_rate": 4.340627899924555e-06, "loss": 2.1987, "step": 10653 }, { "epoch": 3.3398119122257053, "grad_norm": 9.321538925170898, "learning_rate": 4.338701035787403e-06, "loss": 2.4221, "step": 10654 }, { "epoch": 3.3401253918495297, "grad_norm": 8.261653900146484, "learning_rate": 4.3367717893710705e-06, "loss": 2.1657, "step": 10655 }, { "epoch": 3.340438871473354, "grad_norm": 8.702301025390625, "learning_rate": 4.334840163175152e-06, "loss": 2.0899, "step": 10656 }, { "epoch": 3.3407523510971786, "grad_norm": 6.991822719573975, "learning_rate": 4.332906159702322e-06, "loss": 2.6259, "step": 10657 }, { "epoch": 3.341065830721003, "grad_norm": 7.080436706542969, "learning_rate": 4.330969781458338e-06, "loss": 1.9751, "step": 10658 }, { "epoch": 3.3413793103448275, "grad_norm": 17.487451553344727, "learning_rate": 4.3290310309520325e-06, "loss": 2.0957, "step": 10659 }, { "epoch": 3.341692789968652, "grad_norm": 10.384638786315918, "learning_rate": 4.32708991069531e-06, "loss": 2.1484, "step": 10660 }, { "epoch": 3.3420062695924764, "grad_norm": 6.204391002655029, "learning_rate": 4.3251464232031505e-06, "loss": 2.1604, "step": 10661 }, { "epoch": 3.342319749216301, "grad_norm": 24.644617080688477, "learning_rate": 4.3232005709935965e-06, "loss": 2.4392, "step": 10662 }, { "epoch": 3.3426332288401253, "grad_norm": 15.438302993774414, "learning_rate": 4.3212523565877545e-06, "loss": 2.1004, "step": 10663 }, { "epoch": 3.3429467084639497, "grad_norm": 10.112593650817871, "learning_rate": 4.319301782509794e-06, "loss": 2.1487, "step": 10664 }, { "epoch": 3.343260188087774, "grad_norm": 13.612513542175293, "learning_rate": 4.31734885128694e-06, "loss": 2.145, "step": 10665 }, { "epoch": 3.3435736677115986, "grad_norm": 20.36163902282715, "learning_rate": 4.315393565449472e-06, "loss": 1.9855, "step": 10666 }, { "epoch": 3.343887147335423, "grad_norm": 8.70656967163086, "learning_rate": 4.313435927530719e-06, "loss": 2.1275, "step": 10667 }, { "epoch": 3.3442006269592475, "grad_norm": 28.868633270263672, "learning_rate": 4.311475940067061e-06, "loss": 2.2513, "step": 10668 }, { "epoch": 3.344514106583072, "grad_norm": 9.321325302124023, "learning_rate": 4.309513605597918e-06, "loss": 2.2205, "step": 10669 }, { "epoch": 3.344827586206897, "grad_norm": 6.292479515075684, "learning_rate": 4.307548926665752e-06, "loss": 2.0109, "step": 10670 }, { "epoch": 3.3451410658307212, "grad_norm": 4.511789321899414, "learning_rate": 4.305581905816064e-06, "loss": 2.0825, "step": 10671 }, { "epoch": 3.3454545454545457, "grad_norm": 8.6427640914917, "learning_rate": 4.30361254559739e-06, "loss": 1.8514, "step": 10672 }, { "epoch": 3.34576802507837, "grad_norm": 11.082159996032715, "learning_rate": 4.301640848561293e-06, "loss": 2.4138, "step": 10673 }, { "epoch": 3.3460815047021946, "grad_norm": 13.721964836120605, "learning_rate": 4.299666817262366e-06, "loss": 1.8782, "step": 10674 }, { "epoch": 3.346394984326019, "grad_norm": 11.668535232543945, "learning_rate": 4.297690454258227e-06, "loss": 2.3069, "step": 10675 }, { "epoch": 3.3467084639498434, "grad_norm": 12.756232261657715, "learning_rate": 4.295711762109515e-06, "loss": 2.1148, "step": 10676 }, { "epoch": 3.347021943573668, "grad_norm": 13.293801307678223, "learning_rate": 4.293730743379886e-06, "loss": 2.3005, "step": 10677 }, { "epoch": 3.3473354231974923, "grad_norm": 6.3511834144592285, "learning_rate": 4.291747400636009e-06, "loss": 1.9325, "step": 10678 }, { "epoch": 3.3476489028213168, "grad_norm": 7.686220169067383, "learning_rate": 4.289761736447567e-06, "loss": 2.0488, "step": 10679 }, { "epoch": 3.347962382445141, "grad_norm": 5.471682548522949, "learning_rate": 4.287773753387249e-06, "loss": 2.0447, "step": 10680 }, { "epoch": 3.3482758620689657, "grad_norm": 22.380220413208008, "learning_rate": 4.285783454030748e-06, "loss": 2.4845, "step": 10681 }, { "epoch": 3.34858934169279, "grad_norm": 6.578351020812988, "learning_rate": 4.2837908409567595e-06, "loss": 2.031, "step": 10682 }, { "epoch": 3.3489028213166145, "grad_norm": 6.1908111572265625, "learning_rate": 4.281795916746977e-06, "loss": 2.1546, "step": 10683 }, { "epoch": 3.349216300940439, "grad_norm": 30.757667541503906, "learning_rate": 4.279798683986084e-06, "loss": 2.0838, "step": 10684 }, { "epoch": 3.3495297805642634, "grad_norm": 11.816805839538574, "learning_rate": 4.2777991452617625e-06, "loss": 1.921, "step": 10685 }, { "epoch": 3.349843260188088, "grad_norm": 6.553493022918701, "learning_rate": 4.275797303164675e-06, "loss": 2.1386, "step": 10686 }, { "epoch": 3.3501567398119123, "grad_norm": 6.928806781768799, "learning_rate": 4.273793160288473e-06, "loss": 1.8635, "step": 10687 }, { "epoch": 3.3504702194357368, "grad_norm": 4.811969757080078, "learning_rate": 4.271786719229787e-06, "loss": 2.1152, "step": 10688 }, { "epoch": 3.350783699059561, "grad_norm": 102.9781494140625, "learning_rate": 4.269777982588225e-06, "loss": 2.1162, "step": 10689 }, { "epoch": 3.3510971786833856, "grad_norm": 6.197340488433838, "learning_rate": 4.267766952966369e-06, "loss": 2.0346, "step": 10690 }, { "epoch": 3.35141065830721, "grad_norm": 13.40040111541748, "learning_rate": 4.265753632969775e-06, "loss": 1.9045, "step": 10691 }, { "epoch": 3.3517241379310345, "grad_norm": 85.40364837646484, "learning_rate": 4.263738025206961e-06, "loss": 2.2269, "step": 10692 }, { "epoch": 3.352037617554859, "grad_norm": 28.283105850219727, "learning_rate": 4.261720132289415e-06, "loss": 1.9304, "step": 10693 }, { "epoch": 3.3523510971786834, "grad_norm": 21.724851608276367, "learning_rate": 4.259699956831582e-06, "loss": 2.0131, "step": 10694 }, { "epoch": 3.352664576802508, "grad_norm": 5.477083683013916, "learning_rate": 4.257677501450863e-06, "loss": 2.0347, "step": 10695 }, { "epoch": 3.3529780564263323, "grad_norm": 9.372947692871094, "learning_rate": 4.255652768767619e-06, "loss": 1.8661, "step": 10696 }, { "epoch": 3.3532915360501567, "grad_norm": 11.529804229736328, "learning_rate": 4.253625761405154e-06, "loss": 2.023, "step": 10697 }, { "epoch": 3.353605015673981, "grad_norm": 15.674795150756836, "learning_rate": 4.251596481989724e-06, "loss": 2.1426, "step": 10698 }, { "epoch": 3.3539184952978056, "grad_norm": 10.012611389160156, "learning_rate": 4.2495649331505284e-06, "loss": 1.9288, "step": 10699 }, { "epoch": 3.35423197492163, "grad_norm": 13.128251075744629, "learning_rate": 4.247531117519705e-06, "loss": 2.07, "step": 10700 }, { "epoch": 3.3545454545454545, "grad_norm": 6.1495280265808105, "learning_rate": 4.245495037732331e-06, "loss": 2.1774, "step": 10701 }, { "epoch": 3.354858934169279, "grad_norm": 7.6022162437438965, "learning_rate": 4.243456696426415e-06, "loss": 2.251, "step": 10702 }, { "epoch": 3.3551724137931034, "grad_norm": 14.009737968444824, "learning_rate": 4.241416096242895e-06, "loss": 2.7565, "step": 10703 }, { "epoch": 3.355485893416928, "grad_norm": 7.093069553375244, "learning_rate": 4.23937323982564e-06, "loss": 2.1497, "step": 10704 }, { "epoch": 3.3557993730407523, "grad_norm": 6.690266132354736, "learning_rate": 4.237328129821437e-06, "loss": 1.9347, "step": 10705 }, { "epoch": 3.3561128526645767, "grad_norm": 23.861154556274414, "learning_rate": 4.235280768879996e-06, "loss": 2.2263, "step": 10706 }, { "epoch": 3.356426332288401, "grad_norm": 15.750188827514648, "learning_rate": 4.233231159653942e-06, "loss": 2.0124, "step": 10707 }, { "epoch": 3.3567398119122256, "grad_norm": 8.348553657531738, "learning_rate": 4.2311793047988145e-06, "loss": 1.9755, "step": 10708 }, { "epoch": 3.35705329153605, "grad_norm": 5.684324741363525, "learning_rate": 4.229125206973061e-06, "loss": 1.9629, "step": 10709 }, { "epoch": 3.3573667711598745, "grad_norm": 8.771926879882812, "learning_rate": 4.227068868838035e-06, "loss": 2.0577, "step": 10710 }, { "epoch": 3.357680250783699, "grad_norm": 7.476025104522705, "learning_rate": 4.225010293057994e-06, "loss": 1.945, "step": 10711 }, { "epoch": 3.3579937304075234, "grad_norm": 14.13808536529541, "learning_rate": 4.222949482300094e-06, "loss": 2.211, "step": 10712 }, { "epoch": 3.358307210031348, "grad_norm": 6.991470813751221, "learning_rate": 4.220886439234385e-06, "loss": 2.0398, "step": 10713 }, { "epoch": 3.3586206896551722, "grad_norm": 6.682497978210449, "learning_rate": 4.218821166533813e-06, "loss": 2.2521, "step": 10714 }, { "epoch": 3.3589341692789967, "grad_norm": 5.606316089630127, "learning_rate": 4.2167536668742094e-06, "loss": 2.0789, "step": 10715 }, { "epoch": 3.359247648902821, "grad_norm": 7.374332427978516, "learning_rate": 4.214683942934291e-06, "loss": 2.066, "step": 10716 }, { "epoch": 3.3595611285266456, "grad_norm": 9.77920913696289, "learning_rate": 4.2126119973956604e-06, "loss": 1.8608, "step": 10717 }, { "epoch": 3.35987460815047, "grad_norm": 8.570023536682129, "learning_rate": 4.210537832942794e-06, "loss": 1.9333, "step": 10718 }, { "epoch": 3.3601880877742945, "grad_norm": 9.03072738647461, "learning_rate": 4.2084614522630435e-06, "loss": 2.5508, "step": 10719 }, { "epoch": 3.360501567398119, "grad_norm": 8.618317604064941, "learning_rate": 4.206382858046636e-06, "loss": 1.9743, "step": 10720 }, { "epoch": 3.3608150470219433, "grad_norm": 7.675114154815674, "learning_rate": 4.204302052986662e-06, "loss": 2.2861, "step": 10721 }, { "epoch": 3.3611285266457678, "grad_norm": 70.33665466308594, "learning_rate": 4.202219039779078e-06, "loss": 1.8926, "step": 10722 }, { "epoch": 3.3614420062695927, "grad_norm": 22.482452392578125, "learning_rate": 4.200133821122705e-06, "loss": 2.1991, "step": 10723 }, { "epoch": 3.361755485893417, "grad_norm": 11.290307998657227, "learning_rate": 4.1980463997192146e-06, "loss": 2.054, "step": 10724 }, { "epoch": 3.3620689655172415, "grad_norm": 86.1988296508789, "learning_rate": 4.195956778273137e-06, "loss": 2.1612, "step": 10725 }, { "epoch": 3.362382445141066, "grad_norm": 8.064861297607422, "learning_rate": 4.193864959491853e-06, "loss": 1.9708, "step": 10726 }, { "epoch": 3.3626959247648904, "grad_norm": 37.787147521972656, "learning_rate": 4.191770946085587e-06, "loss": 2.1491, "step": 10727 }, { "epoch": 3.363009404388715, "grad_norm": 79.85929870605469, "learning_rate": 4.189674740767411e-06, "loss": 2.3765, "step": 10728 }, { "epoch": 3.3633228840125393, "grad_norm": 4.968204021453857, "learning_rate": 4.187576346253234e-06, "loss": 2.0274, "step": 10729 }, { "epoch": 3.3636363636363638, "grad_norm": 13.553655624389648, "learning_rate": 4.185475765261801e-06, "loss": 1.9646, "step": 10730 }, { "epoch": 3.363949843260188, "grad_norm": 25.97132110595703, "learning_rate": 4.1833730005146936e-06, "loss": 2.3007, "step": 10731 }, { "epoch": 3.3642633228840126, "grad_norm": 67.50879669189453, "learning_rate": 4.181268054736319e-06, "loss": 2.2488, "step": 10732 }, { "epoch": 3.364576802507837, "grad_norm": 7.950389862060547, "learning_rate": 4.17916093065391e-06, "loss": 2.0386, "step": 10733 }, { "epoch": 3.3648902821316615, "grad_norm": 10.269017219543457, "learning_rate": 4.1770516309975264e-06, "loss": 2.0447, "step": 10734 }, { "epoch": 3.365203761755486, "grad_norm": 6.556334972381592, "learning_rate": 4.1749401585000415e-06, "loss": 1.8913, "step": 10735 }, { "epoch": 3.3655172413793104, "grad_norm": 13.808581352233887, "learning_rate": 4.172826515897146e-06, "loss": 1.9711, "step": 10736 }, { "epoch": 3.365830721003135, "grad_norm": 13.410728454589844, "learning_rate": 4.170710705927343e-06, "loss": 1.9945, "step": 10737 }, { "epoch": 3.3661442006269593, "grad_norm": 4.52083158493042, "learning_rate": 4.16859273133194e-06, "loss": 2.3791, "step": 10738 }, { "epoch": 3.3664576802507837, "grad_norm": 13.265490531921387, "learning_rate": 4.1664725948550545e-06, "loss": 2.1748, "step": 10739 }, { "epoch": 3.366771159874608, "grad_norm": 9.368539810180664, "learning_rate": 4.164350299243601e-06, "loss": 1.9677, "step": 10740 }, { "epoch": 3.3670846394984326, "grad_norm": 11.863995552062988, "learning_rate": 4.1622258472472955e-06, "loss": 2.0721, "step": 10741 }, { "epoch": 3.367398119122257, "grad_norm": 11.4646577835083, "learning_rate": 4.160099241618642e-06, "loss": 2.1004, "step": 10742 }, { "epoch": 3.3677115987460815, "grad_norm": 8.982125282287598, "learning_rate": 4.1579704851129385e-06, "loss": 2.3009, "step": 10743 }, { "epoch": 3.368025078369906, "grad_norm": 85.43389129638672, "learning_rate": 4.15583958048827e-06, "loss": 2.6656, "step": 10744 }, { "epoch": 3.3683385579937304, "grad_norm": 9.407835006713867, "learning_rate": 4.153706530505504e-06, "loss": 2.1429, "step": 10745 }, { "epoch": 3.368652037617555, "grad_norm": 7.429203987121582, "learning_rate": 4.151571337928285e-06, "loss": 1.9336, "step": 10746 }, { "epoch": 3.3689655172413793, "grad_norm": 21.516958236694336, "learning_rate": 4.14943400552304e-06, "loss": 2.0077, "step": 10747 }, { "epoch": 3.3692789968652037, "grad_norm": 17.183666229248047, "learning_rate": 4.1472945360589626e-06, "loss": 2.1651, "step": 10748 }, { "epoch": 3.369592476489028, "grad_norm": 7.366785049438477, "learning_rate": 4.145152932308016e-06, "loss": 2.7777, "step": 10749 }, { "epoch": 3.3699059561128526, "grad_norm": 7.48547887802124, "learning_rate": 4.143009197044932e-06, "loss": 2.2144, "step": 10750 }, { "epoch": 3.370219435736677, "grad_norm": 8.660316467285156, "learning_rate": 4.1408633330471995e-06, "loss": 2.0482, "step": 10751 }, { "epoch": 3.3705329153605015, "grad_norm": 6.705793380737305, "learning_rate": 4.138715343095069e-06, "loss": 2.074, "step": 10752 }, { "epoch": 3.370846394984326, "grad_norm": 8.455639839172363, "learning_rate": 4.136565229971543e-06, "loss": 2.2797, "step": 10753 }, { "epoch": 3.3711598746081504, "grad_norm": 11.431133270263672, "learning_rate": 4.1344129964623765e-06, "loss": 1.9381, "step": 10754 }, { "epoch": 3.371473354231975, "grad_norm": 10.318950653076172, "learning_rate": 4.132258645356073e-06, "loss": 2.3564, "step": 10755 }, { "epoch": 3.3717868338557992, "grad_norm": 4.6567792892456055, "learning_rate": 4.130102179443877e-06, "loss": 1.9943, "step": 10756 }, { "epoch": 3.3721003134796237, "grad_norm": 5.861505508422852, "learning_rate": 4.127943601519772e-06, "loss": 2.1839, "step": 10757 }, { "epoch": 3.372413793103448, "grad_norm": 11.272713661193848, "learning_rate": 4.125782914380482e-06, "loss": 2.1853, "step": 10758 }, { "epoch": 3.3727272727272726, "grad_norm": 8.851056098937988, "learning_rate": 4.123620120825459e-06, "loss": 2.0857, "step": 10759 }, { "epoch": 3.373040752351097, "grad_norm": 42.94769287109375, "learning_rate": 4.12145522365689e-06, "loss": 1.9958, "step": 10760 }, { "epoch": 3.3733542319749215, "grad_norm": 8.313215255737305, "learning_rate": 4.119288225679683e-06, "loss": 1.8239, "step": 10761 }, { "epoch": 3.373667711598746, "grad_norm": 6.260233402252197, "learning_rate": 4.117119129701468e-06, "loss": 1.8021, "step": 10762 }, { "epoch": 3.3739811912225703, "grad_norm": 7.642189979553223, "learning_rate": 4.114947938532595e-06, "loss": 2.0327, "step": 10763 }, { "epoch": 3.3742946708463952, "grad_norm": 7.243236541748047, "learning_rate": 4.112774654986128e-06, "loss": 2.0629, "step": 10764 }, { "epoch": 3.3746081504702197, "grad_norm": 6.622078895568848, "learning_rate": 4.1105992818778416e-06, "loss": 2.0556, "step": 10765 }, { "epoch": 3.374921630094044, "grad_norm": 13.255139350891113, "learning_rate": 4.1084218220262175e-06, "loss": 2.0276, "step": 10766 }, { "epoch": 3.3752351097178686, "grad_norm": 8.4006986618042, "learning_rate": 4.106242278252443e-06, "loss": 2.033, "step": 10767 }, { "epoch": 3.375548589341693, "grad_norm": 226.27035522460938, "learning_rate": 4.104060653380403e-06, "loss": 2.4989, "step": 10768 }, { "epoch": 3.3758620689655174, "grad_norm": 6.4070000648498535, "learning_rate": 4.10187695023668e-06, "loss": 2.1556, "step": 10769 }, { "epoch": 3.376175548589342, "grad_norm": 19.5567569732666, "learning_rate": 4.099691171650547e-06, "loss": 2.3102, "step": 10770 }, { "epoch": 3.3764890282131663, "grad_norm": 9.81298828125, "learning_rate": 4.097503320453971e-06, "loss": 2.1019, "step": 10771 }, { "epoch": 3.3768025078369908, "grad_norm": 15.00754165649414, "learning_rate": 4.095313399481599e-06, "loss": 2.2757, "step": 10772 }, { "epoch": 3.377115987460815, "grad_norm": 8.540647506713867, "learning_rate": 4.093121411570762e-06, "loss": 2.1027, "step": 10773 }, { "epoch": 3.3774294670846396, "grad_norm": 7.158480167388916, "learning_rate": 4.090927359561469e-06, "loss": 1.9182, "step": 10774 }, { "epoch": 3.377742946708464, "grad_norm": 8.468045234680176, "learning_rate": 4.0887312462964035e-06, "loss": 1.8684, "step": 10775 }, { "epoch": 3.3780564263322885, "grad_norm": 6.360187530517578, "learning_rate": 4.086533074620919e-06, "loss": 2.3909, "step": 10776 }, { "epoch": 3.378369905956113, "grad_norm": 9.827460289001465, "learning_rate": 4.084332847383037e-06, "loss": 2.386, "step": 10777 }, { "epoch": 3.3786833855799374, "grad_norm": 6.558335304260254, "learning_rate": 4.082130567433439e-06, "loss": 2.3486, "step": 10778 }, { "epoch": 3.378996865203762, "grad_norm": 7.773528099060059, "learning_rate": 4.079926237625472e-06, "loss": 1.9192, "step": 10779 }, { "epoch": 3.3793103448275863, "grad_norm": 10.564056396484375, "learning_rate": 4.077719860815132e-06, "loss": 2.2389, "step": 10780 }, { "epoch": 3.3796238244514107, "grad_norm": 13.442465782165527, "learning_rate": 4.075511439861073e-06, "loss": 2.2275, "step": 10781 }, { "epoch": 3.379937304075235, "grad_norm": 8.406206130981445, "learning_rate": 4.073300977624594e-06, "loss": 2.5843, "step": 10782 }, { "epoch": 3.3802507836990596, "grad_norm": 9.475699424743652, "learning_rate": 4.071088476969639e-06, "loss": 1.9934, "step": 10783 }, { "epoch": 3.380564263322884, "grad_norm": 113.31139373779297, "learning_rate": 4.068873940762796e-06, "loss": 2.4407, "step": 10784 }, { "epoch": 3.3808777429467085, "grad_norm": 13.223235130310059, "learning_rate": 4.066657371873286e-06, "loss": 1.9374, "step": 10785 }, { "epoch": 3.381191222570533, "grad_norm": 9.68596363067627, "learning_rate": 4.064438773172966e-06, "loss": 2.736, "step": 10786 }, { "epoch": 3.3815047021943574, "grad_norm": 41.88019561767578, "learning_rate": 4.062218147536324e-06, "loss": 2.0198, "step": 10787 }, { "epoch": 3.381818181818182, "grad_norm": 7.205523490905762, "learning_rate": 4.059995497840471e-06, "loss": 2.0766, "step": 10788 }, { "epoch": 3.3821316614420063, "grad_norm": 10.764029502868652, "learning_rate": 4.057770826965143e-06, "loss": 1.9265, "step": 10789 }, { "epoch": 3.3824451410658307, "grad_norm": 5.847966194152832, "learning_rate": 4.055544137792695e-06, "loss": 2.0429, "step": 10790 }, { "epoch": 3.382758620689655, "grad_norm": 28.853614807128906, "learning_rate": 4.053315433208093e-06, "loss": 2.2192, "step": 10791 }, { "epoch": 3.3830721003134796, "grad_norm": 9.981918334960938, "learning_rate": 4.051084716098921e-06, "loss": 1.9155, "step": 10792 }, { "epoch": 3.383385579937304, "grad_norm": 8.752816200256348, "learning_rate": 4.048851989355363e-06, "loss": 2.0176, "step": 10793 }, { "epoch": 3.3836990595611285, "grad_norm": 6.9544525146484375, "learning_rate": 4.046617255870212e-06, "loss": 2.0317, "step": 10794 }, { "epoch": 3.384012539184953, "grad_norm": 9.405781745910645, "learning_rate": 4.044380518538859e-06, "loss": 1.8862, "step": 10795 }, { "epoch": 3.3843260188087774, "grad_norm": 12.022509574890137, "learning_rate": 4.042141780259292e-06, "loss": 1.97, "step": 10796 }, { "epoch": 3.384639498432602, "grad_norm": 12.161920547485352, "learning_rate": 4.039901043932091e-06, "loss": 2.1778, "step": 10797 }, { "epoch": 3.3849529780564263, "grad_norm": 17.332122802734375, "learning_rate": 4.037658312460424e-06, "loss": 2.297, "step": 10798 }, { "epoch": 3.3852664576802507, "grad_norm": 7.26715612411499, "learning_rate": 4.035413588750046e-06, "loss": 1.9227, "step": 10799 }, { "epoch": 3.385579937304075, "grad_norm": 5.877050399780273, "learning_rate": 4.033166875709291e-06, "loss": 2.1393, "step": 10800 }, { "epoch": 3.3858934169278996, "grad_norm": 64.21208953857422, "learning_rate": 4.030918176249072e-06, "loss": 1.9164, "step": 10801 }, { "epoch": 3.386206896551724, "grad_norm": 21.667987823486328, "learning_rate": 4.028667493282875e-06, "loss": 2.0301, "step": 10802 }, { "epoch": 3.3865203761755485, "grad_norm": 21.85112762451172, "learning_rate": 4.0264148297267555e-06, "loss": 2.2854, "step": 10803 }, { "epoch": 3.386833855799373, "grad_norm": 8.487278938293457, "learning_rate": 4.024160188499337e-06, "loss": 2.2863, "step": 10804 }, { "epoch": 3.3871473354231973, "grad_norm": 20.624835968017578, "learning_rate": 4.021903572521802e-06, "loss": 1.9501, "step": 10805 }, { "epoch": 3.387460815047022, "grad_norm": 8.263264656066895, "learning_rate": 4.0196449847178945e-06, "loss": 2.0953, "step": 10806 }, { "epoch": 3.3877742946708462, "grad_norm": 11.19937515258789, "learning_rate": 4.017384428013913e-06, "loss": 2.4941, "step": 10807 }, { "epoch": 3.3880877742946707, "grad_norm": 11.611288070678711, "learning_rate": 4.015121905338704e-06, "loss": 1.9391, "step": 10808 }, { "epoch": 3.388401253918495, "grad_norm": 6.225773811340332, "learning_rate": 4.012857419623666e-06, "loss": 2.0682, "step": 10809 }, { "epoch": 3.3887147335423196, "grad_norm": 7.784508228302002, "learning_rate": 4.010590973802737e-06, "loss": 2.1068, "step": 10810 }, { "epoch": 3.389028213166144, "grad_norm": 77.31217193603516, "learning_rate": 4.008322570812395e-06, "loss": 2.4672, "step": 10811 }, { "epoch": 3.3893416927899684, "grad_norm": 71.845458984375, "learning_rate": 4.006052213591657e-06, "loss": 2.5533, "step": 10812 }, { "epoch": 3.389655172413793, "grad_norm": 6.865128040313721, "learning_rate": 4.0037799050820664e-06, "loss": 2.1013, "step": 10813 }, { "epoch": 3.3899686520376173, "grad_norm": 12.165755271911621, "learning_rate": 4.0015056482277e-06, "loss": 2.2856, "step": 10814 }, { "epoch": 3.3902821316614418, "grad_norm": 10.216269493103027, "learning_rate": 3.999229445975158e-06, "loss": 2.1956, "step": 10815 }, { "epoch": 3.390595611285266, "grad_norm": 112.8543930053711, "learning_rate": 3.996951301273556e-06, "loss": 2.3692, "step": 10816 }, { "epoch": 3.390909090909091, "grad_norm": 8.711501121520996, "learning_rate": 3.994671217074535e-06, "loss": 2.1524, "step": 10817 }, { "epoch": 3.3912225705329155, "grad_norm": 7.944595813751221, "learning_rate": 3.992389196332241e-06, "loss": 2.0668, "step": 10818 }, { "epoch": 3.39153605015674, "grad_norm": 16.247802734375, "learning_rate": 3.990105242003333e-06, "loss": 2.3624, "step": 10819 }, { "epoch": 3.3918495297805644, "grad_norm": 8.16081428527832, "learning_rate": 3.987819357046975e-06, "loss": 2.3707, "step": 10820 }, { "epoch": 3.392163009404389, "grad_norm": 18.1226863861084, "learning_rate": 3.9855315444248305e-06, "loss": 2.0996, "step": 10821 }, { "epoch": 3.3924764890282133, "grad_norm": 10.490715980529785, "learning_rate": 3.983241807101064e-06, "loss": 2.3358, "step": 10822 }, { "epoch": 3.3927899686520377, "grad_norm": 97.77823638916016, "learning_rate": 3.980950148042329e-06, "loss": 2.0417, "step": 10823 }, { "epoch": 3.393103448275862, "grad_norm": 7.553221702575684, "learning_rate": 3.9786565702177725e-06, "loss": 2.0297, "step": 10824 }, { "epoch": 3.3934169278996866, "grad_norm": 7.390120029449463, "learning_rate": 3.976361076599027e-06, "loss": 1.9088, "step": 10825 }, { "epoch": 3.393730407523511, "grad_norm": 9.604377746582031, "learning_rate": 3.9740636701602065e-06, "loss": 2.1685, "step": 10826 }, { "epoch": 3.3940438871473355, "grad_norm": 7.918846607208252, "learning_rate": 3.971764353877903e-06, "loss": 2.0245, "step": 10827 }, { "epoch": 3.39435736677116, "grad_norm": 7.60604190826416, "learning_rate": 3.969463130731183e-06, "loss": 2.0085, "step": 10828 }, { "epoch": 3.3946708463949844, "grad_norm": 38.89109420776367, "learning_rate": 3.9671600037015844e-06, "loss": 2.027, "step": 10829 }, { "epoch": 3.394984326018809, "grad_norm": 10.890227317810059, "learning_rate": 3.964854975773112e-06, "loss": 2.0307, "step": 10830 }, { "epoch": 3.3952978056426333, "grad_norm": 7.403882026672363, "learning_rate": 3.962548049932232e-06, "loss": 2.3909, "step": 10831 }, { "epoch": 3.3956112852664577, "grad_norm": 22.334733963012695, "learning_rate": 3.960239229167869e-06, "loss": 2.1785, "step": 10832 }, { "epoch": 3.395924764890282, "grad_norm": 10.155850410461426, "learning_rate": 3.957928516471407e-06, "loss": 2.0727, "step": 10833 }, { "epoch": 3.3962382445141066, "grad_norm": 36.20935821533203, "learning_rate": 3.955615914836678e-06, "loss": 2.5631, "step": 10834 }, { "epoch": 3.396551724137931, "grad_norm": 8.983194351196289, "learning_rate": 3.9533014272599605e-06, "loss": 2.3022, "step": 10835 }, { "epoch": 3.3968652037617555, "grad_norm": 5.979011535644531, "learning_rate": 3.950985056739978e-06, "loss": 2.0025, "step": 10836 }, { "epoch": 3.39717868338558, "grad_norm": 11.568971633911133, "learning_rate": 3.948666806277893e-06, "loss": 2.2836, "step": 10837 }, { "epoch": 3.3974921630094044, "grad_norm": 6.572995662689209, "learning_rate": 3.946346678877305e-06, "loss": 2.2522, "step": 10838 }, { "epoch": 3.397805642633229, "grad_norm": 9.133193016052246, "learning_rate": 3.9440246775442436e-06, "loss": 2.1148, "step": 10839 }, { "epoch": 3.3981191222570533, "grad_norm": 4.974786281585693, "learning_rate": 3.941700805287169e-06, "loss": 2.1095, "step": 10840 }, { "epoch": 3.3984326018808777, "grad_norm": 9.506574630737305, "learning_rate": 3.9393750651169604e-06, "loss": 1.9472, "step": 10841 }, { "epoch": 3.398746081504702, "grad_norm": 8.275567054748535, "learning_rate": 3.937047460046923e-06, "loss": 1.9508, "step": 10842 }, { "epoch": 3.3990595611285266, "grad_norm": 21.16493034362793, "learning_rate": 3.934717993092774e-06, "loss": 1.9425, "step": 10843 }, { "epoch": 3.399373040752351, "grad_norm": 9.492948532104492, "learning_rate": 3.932386667272645e-06, "loss": 2.2176, "step": 10844 }, { "epoch": 3.3996865203761755, "grad_norm": 10.626721382141113, "learning_rate": 3.930053485607075e-06, "loss": 2.139, "step": 10845 }, { "epoch": 3.4, "grad_norm": 5.863310813903809, "learning_rate": 3.927718451119009e-06, "loss": 2.0718, "step": 10846 }, { "epoch": 3.4003134796238244, "grad_norm": 6.200711250305176, "learning_rate": 3.92538156683379e-06, "loss": 2.1365, "step": 10847 }, { "epoch": 3.400626959247649, "grad_norm": 15.739635467529297, "learning_rate": 3.92304283577916e-06, "loss": 2.1471, "step": 10848 }, { "epoch": 3.4009404388714732, "grad_norm": 7.017481327056885, "learning_rate": 3.920702260985253e-06, "loss": 2.1232, "step": 10849 }, { "epoch": 3.4012539184952977, "grad_norm": 7.369931221008301, "learning_rate": 3.918359845484591e-06, "loss": 2.0687, "step": 10850 }, { "epoch": 3.401567398119122, "grad_norm": 4.725981712341309, "learning_rate": 3.916015592312083e-06, "loss": 1.95, "step": 10851 }, { "epoch": 3.4018808777429466, "grad_norm": 19.353862762451172, "learning_rate": 3.913669504505015e-06, "loss": 2.1949, "step": 10852 }, { "epoch": 3.402194357366771, "grad_norm": 4.883718490600586, "learning_rate": 3.911321585103055e-06, "loss": 2.0482, "step": 10853 }, { "epoch": 3.4025078369905954, "grad_norm": 5.468510627746582, "learning_rate": 3.90897183714824e-06, "loss": 2.1924, "step": 10854 }, { "epoch": 3.40282131661442, "grad_norm": 8.344846725463867, "learning_rate": 3.906620263684979e-06, "loss": 2.3124, "step": 10855 }, { "epoch": 3.4031347962382443, "grad_norm": 10.216719627380371, "learning_rate": 3.904266867760044e-06, "loss": 1.9382, "step": 10856 }, { "epoch": 3.4034482758620688, "grad_norm": 10.284916877746582, "learning_rate": 3.901911652422569e-06, "loss": 2.0044, "step": 10857 }, { "epoch": 3.4037617554858937, "grad_norm": 10.76353931427002, "learning_rate": 3.8995546207240455e-06, "loss": 2.0272, "step": 10858 }, { "epoch": 3.404075235109718, "grad_norm": 16.008403778076172, "learning_rate": 3.89719577571832e-06, "loss": 2.0551, "step": 10859 }, { "epoch": 3.4043887147335425, "grad_norm": 8.031129837036133, "learning_rate": 3.8948351204615846e-06, "loss": 2.2074, "step": 10860 }, { "epoch": 3.404702194357367, "grad_norm": 55.48567199707031, "learning_rate": 3.892472658012379e-06, "loss": 2.2865, "step": 10861 }, { "epoch": 3.4050156739811914, "grad_norm": 29.972814559936523, "learning_rate": 3.890108391431584e-06, "loss": 2.089, "step": 10862 }, { "epoch": 3.405329153605016, "grad_norm": 52.226375579833984, "learning_rate": 3.88774232378242e-06, "loss": 2.4591, "step": 10863 }, { "epoch": 3.4056426332288403, "grad_norm": 4.53100061416626, "learning_rate": 3.8853744581304376e-06, "loss": 2.1062, "step": 10864 }, { "epoch": 3.4059561128526648, "grad_norm": 8.754895210266113, "learning_rate": 3.8830047975435184e-06, "loss": 2.1172, "step": 10865 }, { "epoch": 3.406269592476489, "grad_norm": 65.5130615234375, "learning_rate": 3.88063334509187e-06, "loss": 2.2382, "step": 10866 }, { "epoch": 3.4065830721003136, "grad_norm": 8.199430465698242, "learning_rate": 3.87826010384802e-06, "loss": 2.0718, "step": 10867 }, { "epoch": 3.406896551724138, "grad_norm": 9.792732238769531, "learning_rate": 3.875885076886817e-06, "loss": 2.0834, "step": 10868 }, { "epoch": 3.4072100313479625, "grad_norm": 72.27893829345703, "learning_rate": 3.8735082672854195e-06, "loss": 2.1093, "step": 10869 }, { "epoch": 3.407523510971787, "grad_norm": 5.609694957733154, "learning_rate": 3.871129678123297e-06, "loss": 2.4369, "step": 10870 }, { "epoch": 3.4078369905956114, "grad_norm": 6.411273002624512, "learning_rate": 3.868749312482225e-06, "loss": 1.8181, "step": 10871 }, { "epoch": 3.408150470219436, "grad_norm": 6.639737129211426, "learning_rate": 3.866367173446281e-06, "loss": 1.9926, "step": 10872 }, { "epoch": 3.4084639498432603, "grad_norm": 7.9556074142456055, "learning_rate": 3.86398326410184e-06, "loss": 2.0639, "step": 10873 }, { "epoch": 3.4087774294670847, "grad_norm": 5.1251220703125, "learning_rate": 3.861597587537568e-06, "loss": 2.0506, "step": 10874 }, { "epoch": 3.409090909090909, "grad_norm": 26.09616470336914, "learning_rate": 3.859210146844425e-06, "loss": 1.9024, "step": 10875 }, { "epoch": 3.4094043887147336, "grad_norm": 19.23230743408203, "learning_rate": 3.856820945115655e-06, "loss": 1.9895, "step": 10876 }, { "epoch": 3.409717868338558, "grad_norm": 59.43765640258789, "learning_rate": 3.854429985446782e-06, "loss": 2.6159, "step": 10877 }, { "epoch": 3.4100313479623825, "grad_norm": 35.685237884521484, "learning_rate": 3.852037270935608e-06, "loss": 2.0801, "step": 10878 }, { "epoch": 3.410344827586207, "grad_norm": 8.916625022888184, "learning_rate": 3.849642804682212e-06, "loss": 2.1222, "step": 10879 }, { "epoch": 3.4106583072100314, "grad_norm": 18.509033203125, "learning_rate": 3.84724658978894e-06, "loss": 1.9819, "step": 10880 }, { "epoch": 3.410971786833856, "grad_norm": 7.983786582946777, "learning_rate": 3.844848629360401e-06, "loss": 1.8548, "step": 10881 }, { "epoch": 3.4112852664576803, "grad_norm": 11.210798263549805, "learning_rate": 3.84244892650347e-06, "loss": 2.0754, "step": 10882 }, { "epoch": 3.4115987460815047, "grad_norm": 19.546371459960938, "learning_rate": 3.8400474843272795e-06, "loss": 2.2166, "step": 10883 }, { "epoch": 3.411912225705329, "grad_norm": 6.267324447631836, "learning_rate": 3.837644305943211e-06, "loss": 1.9143, "step": 10884 }, { "epoch": 3.4122257053291536, "grad_norm": 8.867936134338379, "learning_rate": 3.835239394464901e-06, "loss": 2.3302, "step": 10885 }, { "epoch": 3.412539184952978, "grad_norm": 10.167280197143555, "learning_rate": 3.832832753008227e-06, "loss": 1.9591, "step": 10886 }, { "epoch": 3.4128526645768025, "grad_norm": 8.547554016113281, "learning_rate": 3.83042438469131e-06, "loss": 1.8901, "step": 10887 }, { "epoch": 3.413166144200627, "grad_norm": 7.002828121185303, "learning_rate": 3.828014292634508e-06, "loss": 1.998, "step": 10888 }, { "epoch": 3.4134796238244514, "grad_norm": 5.7365593910217285, "learning_rate": 3.825602479960414e-06, "loss": 2.6167, "step": 10889 }, { "epoch": 3.413793103448276, "grad_norm": 4.981120586395264, "learning_rate": 3.8231889497938475e-06, "loss": 2.0426, "step": 10890 }, { "epoch": 3.4141065830721002, "grad_norm": 8.140067100524902, "learning_rate": 3.820773705261854e-06, "loss": 2.2638, "step": 10891 }, { "epoch": 3.4144200626959247, "grad_norm": 8.550292015075684, "learning_rate": 3.818356749493703e-06, "loss": 1.9444, "step": 10892 }, { "epoch": 3.414733542319749, "grad_norm": 9.156805992126465, "learning_rate": 3.815938085620875e-06, "loss": 2.0203, "step": 10893 }, { "epoch": 3.4150470219435736, "grad_norm": 6.535094738006592, "learning_rate": 3.813517716777069e-06, "loss": 1.9239, "step": 10894 }, { "epoch": 3.415360501567398, "grad_norm": 6.792361736297607, "learning_rate": 3.811095646098193e-06, "loss": 2.013, "step": 10895 }, { "epoch": 3.4156739811912225, "grad_norm": 5.072399616241455, "learning_rate": 3.808671876722357e-06, "loss": 2.1825, "step": 10896 }, { "epoch": 3.415987460815047, "grad_norm": 41.467891693115234, "learning_rate": 3.806246411789872e-06, "loss": 2.3175, "step": 10897 }, { "epoch": 3.4163009404388713, "grad_norm": 8.823486328125, "learning_rate": 3.80381925444325e-06, "loss": 2.0604, "step": 10898 }, { "epoch": 3.416614420062696, "grad_norm": 4.983666896820068, "learning_rate": 3.80139040782719e-06, "loss": 2.0654, "step": 10899 }, { "epoch": 3.41692789968652, "grad_norm": 4.745028972625732, "learning_rate": 3.798959875088584e-06, "loss": 1.8773, "step": 10900 }, { "epoch": 3.4172413793103447, "grad_norm": 6.674015522003174, "learning_rate": 3.796527659376507e-06, "loss": 1.9681, "step": 10901 }, { "epoch": 3.417554858934169, "grad_norm": 11.620408058166504, "learning_rate": 3.794093763842214e-06, "loss": 2.0158, "step": 10902 }, { "epoch": 3.4178683385579935, "grad_norm": 10.81863784790039, "learning_rate": 3.7916581916391364e-06, "loss": 2.0491, "step": 10903 }, { "epoch": 3.418181818181818, "grad_norm": 7.805420398712158, "learning_rate": 3.7892209459228802e-06, "loss": 2.2106, "step": 10904 }, { "epoch": 3.4184952978056424, "grad_norm": 20.89044761657715, "learning_rate": 3.786782029851216e-06, "loss": 2.7557, "step": 10905 }, { "epoch": 3.418808777429467, "grad_norm": 18.082212448120117, "learning_rate": 3.7843414465840823e-06, "loss": 2.0073, "step": 10906 }, { "epoch": 3.4191222570532913, "grad_norm": 6.865312576293945, "learning_rate": 3.7818991992835747e-06, "loss": 2.1888, "step": 10907 }, { "epoch": 3.4194357366771158, "grad_norm": 8.140865325927734, "learning_rate": 3.7794552911139472e-06, "loss": 1.9934, "step": 10908 }, { "epoch": 3.41974921630094, "grad_norm": 125.36479949951172, "learning_rate": 3.777009725241604e-06, "loss": 2.1633, "step": 10909 }, { "epoch": 3.420062695924765, "grad_norm": 8.449714660644531, "learning_rate": 3.7745625048350963e-06, "loss": 2.2191, "step": 10910 }, { "epoch": 3.4203761755485895, "grad_norm": 5.4379072189331055, "learning_rate": 3.772113633065122e-06, "loss": 1.9709, "step": 10911 }, { "epoch": 3.420689655172414, "grad_norm": 7.302900314331055, "learning_rate": 3.769663113104516e-06, "loss": 2.3516, "step": 10912 }, { "epoch": 3.4210031347962384, "grad_norm": 16.038848876953125, "learning_rate": 3.767210948128248e-06, "loss": 1.9691, "step": 10913 }, { "epoch": 3.421316614420063, "grad_norm": 11.58913516998291, "learning_rate": 3.7647571413134236e-06, "loss": 2.0511, "step": 10914 }, { "epoch": 3.4216300940438873, "grad_norm": 9.429380416870117, "learning_rate": 3.7623016958392706e-06, "loss": 2.1825, "step": 10915 }, { "epoch": 3.4219435736677117, "grad_norm": 5.387939929962158, "learning_rate": 3.759844614887141e-06, "loss": 2.0963, "step": 10916 }, { "epoch": 3.422257053291536, "grad_norm": 7.6942548751831055, "learning_rate": 3.757385901640508e-06, "loss": 2.0004, "step": 10917 }, { "epoch": 3.4225705329153606, "grad_norm": 26.28522300720215, "learning_rate": 3.7549255592849575e-06, "loss": 2.4933, "step": 10918 }, { "epoch": 3.422884012539185, "grad_norm": 9.656407356262207, "learning_rate": 3.752463591008187e-06, "loss": 2.2848, "step": 10919 }, { "epoch": 3.4231974921630095, "grad_norm": 9.803751945495605, "learning_rate": 3.7500000000000005e-06, "loss": 2.0816, "step": 10920 }, { "epoch": 3.423510971786834, "grad_norm": 5.5497589111328125, "learning_rate": 3.747534789452304e-06, "loss": 1.9784, "step": 10921 }, { "epoch": 3.4238244514106584, "grad_norm": 63.77427673339844, "learning_rate": 3.7450679625591023e-06, "loss": 2.399, "step": 10922 }, { "epoch": 3.424137931034483, "grad_norm": 40.35515594482422, "learning_rate": 3.742599522516496e-06, "loss": 2.1435, "step": 10923 }, { "epoch": 3.4244514106583073, "grad_norm": 8.390459060668945, "learning_rate": 3.7401294725226707e-06, "loss": 2.2187, "step": 10924 }, { "epoch": 3.4247648902821317, "grad_norm": 18.5019474029541, "learning_rate": 3.737657815777904e-06, "loss": 1.9731, "step": 10925 }, { "epoch": 3.425078369905956, "grad_norm": 7.686516761779785, "learning_rate": 3.7351845554845505e-06, "loss": 2.0409, "step": 10926 }, { "epoch": 3.4253918495297806, "grad_norm": 8.184749603271484, "learning_rate": 3.7327096948470466e-06, "loss": 2.0209, "step": 10927 }, { "epoch": 3.425705329153605, "grad_norm": 11.67505168914795, "learning_rate": 3.7302332370718988e-06, "loss": 2.0202, "step": 10928 }, { "epoch": 3.4260188087774295, "grad_norm": 6.302692890167236, "learning_rate": 3.7277551853676835e-06, "loss": 2.1096, "step": 10929 }, { "epoch": 3.426332288401254, "grad_norm": 24.631319046020508, "learning_rate": 3.7252755429450437e-06, "loss": 2.9299, "step": 10930 }, { "epoch": 3.4266457680250784, "grad_norm": 48.342559814453125, "learning_rate": 3.722794313016682e-06, "loss": 2.374, "step": 10931 }, { "epoch": 3.426959247648903, "grad_norm": 7.080175399780273, "learning_rate": 3.720311498797359e-06, "loss": 1.9744, "step": 10932 }, { "epoch": 3.4272727272727272, "grad_norm": 28.517501831054688, "learning_rate": 3.7178271035038867e-06, "loss": 1.8996, "step": 10933 }, { "epoch": 3.4275862068965517, "grad_norm": 13.506369590759277, "learning_rate": 3.7153411303551255e-06, "loss": 2.1799, "step": 10934 }, { "epoch": 3.427899686520376, "grad_norm": 5.940223693847656, "learning_rate": 3.7128535825719823e-06, "loss": 2.0632, "step": 10935 }, { "epoch": 3.4282131661442006, "grad_norm": 10.841436386108398, "learning_rate": 3.7103644633774015e-06, "loss": 2.7999, "step": 10936 }, { "epoch": 3.428526645768025, "grad_norm": 81.27136993408203, "learning_rate": 3.7078737759963652e-06, "loss": 2.4626, "step": 10937 }, { "epoch": 3.4288401253918495, "grad_norm": 28.791250228881836, "learning_rate": 3.7053815236558865e-06, "loss": 2.3022, "step": 10938 }, { "epoch": 3.429153605015674, "grad_norm": 12.518988609313965, "learning_rate": 3.702887709585007e-06, "loss": 2.1541, "step": 10939 }, { "epoch": 3.4294670846394983, "grad_norm": 12.396989822387695, "learning_rate": 3.7003923370147894e-06, "loss": 2.0618, "step": 10940 }, { "epoch": 3.429780564263323, "grad_norm": 14.469995498657227, "learning_rate": 3.6978954091783186e-06, "loss": 2.076, "step": 10941 }, { "epoch": 3.4300940438871472, "grad_norm": 13.837172508239746, "learning_rate": 3.695396929310693e-06, "loss": 1.9561, "step": 10942 }, { "epoch": 3.4304075235109717, "grad_norm": 17.387622833251953, "learning_rate": 3.6928969006490212e-06, "loss": 1.678, "step": 10943 }, { "epoch": 3.430721003134796, "grad_norm": 9.320486068725586, "learning_rate": 3.690395326432421e-06, "loss": 2.7374, "step": 10944 }, { "epoch": 3.4310344827586206, "grad_norm": 143.2814483642578, "learning_rate": 3.687892209902009e-06, "loss": 2.0133, "step": 10945 }, { "epoch": 3.431347962382445, "grad_norm": 64.74868774414062, "learning_rate": 3.6853875543009045e-06, "loss": 2.179, "step": 10946 }, { "epoch": 3.4316614420062694, "grad_norm": 7.276220798492432, "learning_rate": 3.6828813628742166e-06, "loss": 1.9521, "step": 10947 }, { "epoch": 3.431974921630094, "grad_norm": 6.149008274078369, "learning_rate": 3.680373638869047e-06, "loss": 2.3269, "step": 10948 }, { "epoch": 3.4322884012539183, "grad_norm": 11.50536823272705, "learning_rate": 3.677864385534481e-06, "loss": 2.1655, "step": 10949 }, { "epoch": 3.4326018808777428, "grad_norm": 7.519523620605469, "learning_rate": 3.6753536061215882e-06, "loss": 1.9256, "step": 10950 }, { "epoch": 3.4329153605015676, "grad_norm": 26.182857513427734, "learning_rate": 3.6728413038834132e-06, "loss": 2.4951, "step": 10951 }, { "epoch": 3.433228840125392, "grad_norm": 10.153072357177734, "learning_rate": 3.6703274820749736e-06, "loss": 2.066, "step": 10952 }, { "epoch": 3.4335423197492165, "grad_norm": 6.846949100494385, "learning_rate": 3.6678121439532568e-06, "loss": 2.224, "step": 10953 }, { "epoch": 3.433855799373041, "grad_norm": 8.070419311523438, "learning_rate": 3.665295292777214e-06, "loss": 1.941, "step": 10954 }, { "epoch": 3.4341692789968654, "grad_norm": 66.25064849853516, "learning_rate": 3.662776931807758e-06, "loss": 2.259, "step": 10955 }, { "epoch": 3.43448275862069, "grad_norm": 6.157934188842773, "learning_rate": 3.6602570643077555e-06, "loss": 2.1849, "step": 10956 }, { "epoch": 3.4347962382445143, "grad_norm": 21.516454696655273, "learning_rate": 3.657735693542028e-06, "loss": 2.3162, "step": 10957 }, { "epoch": 3.4351097178683387, "grad_norm": 10.787699699401855, "learning_rate": 3.6552128227773428e-06, "loss": 1.5886, "step": 10958 }, { "epoch": 3.435423197492163, "grad_norm": 21.704341888427734, "learning_rate": 3.65268845528241e-06, "loss": 2.303, "step": 10959 }, { "epoch": 3.4357366771159876, "grad_norm": 30.892200469970703, "learning_rate": 3.650162594327881e-06, "loss": 2.2283, "step": 10960 }, { "epoch": 3.436050156739812, "grad_norm": 10.457781791687012, "learning_rate": 3.6476352431863397e-06, "loss": 1.9565, "step": 10961 }, { "epoch": 3.4363636363636365, "grad_norm": 5.369532108306885, "learning_rate": 3.6451064051323043e-06, "loss": 2.077, "step": 10962 }, { "epoch": 3.436677115987461, "grad_norm": 4.225282669067383, "learning_rate": 3.6425760834422162e-06, "loss": 1.9213, "step": 10963 }, { "epoch": 3.4369905956112854, "grad_norm": 7.098785400390625, "learning_rate": 3.6400442813944394e-06, "loss": 1.8193, "step": 10964 }, { "epoch": 3.43730407523511, "grad_norm": 7.5919904708862305, "learning_rate": 3.6375110022692576e-06, "loss": 2.666, "step": 10965 }, { "epoch": 3.4376175548589343, "grad_norm": 6.028403282165527, "learning_rate": 3.634976249348867e-06, "loss": 2.0945, "step": 10966 }, { "epoch": 3.4379310344827587, "grad_norm": 15.112696647644043, "learning_rate": 3.6324400259173743e-06, "loss": 1.9864, "step": 10967 }, { "epoch": 3.438244514106583, "grad_norm": 7.946347236633301, "learning_rate": 3.6299023352607894e-06, "loss": 2.375, "step": 10968 }, { "epoch": 3.4385579937304076, "grad_norm": 14.49914836883545, "learning_rate": 3.627363180667025e-06, "loss": 3.0762, "step": 10969 }, { "epoch": 3.438871473354232, "grad_norm": 9.440372467041016, "learning_rate": 3.624822565425891e-06, "loss": 1.9018, "step": 10970 }, { "epoch": 3.4391849529780565, "grad_norm": 7.301163196563721, "learning_rate": 3.6222804928290877e-06, "loss": 1.9935, "step": 10971 }, { "epoch": 3.439498432601881, "grad_norm": 7.819459915161133, "learning_rate": 3.6197369661702052e-06, "loss": 1.8863, "step": 10972 }, { "epoch": 3.4398119122257054, "grad_norm": 6.456066608428955, "learning_rate": 3.6171919887447173e-06, "loss": 1.9046, "step": 10973 }, { "epoch": 3.44012539184953, "grad_norm": 5.784896373748779, "learning_rate": 3.6146455638499763e-06, "loss": 1.9515, "step": 10974 }, { "epoch": 3.4404388714733543, "grad_norm": 10.429377555847168, "learning_rate": 3.612097694785211e-06, "loss": 2.2119, "step": 10975 }, { "epoch": 3.4407523510971787, "grad_norm": 9.02249526977539, "learning_rate": 3.6095483848515223e-06, "loss": 1.8855, "step": 10976 }, { "epoch": 3.441065830721003, "grad_norm": 97.81578826904297, "learning_rate": 3.606997637351875e-06, "loss": 2.3632, "step": 10977 }, { "epoch": 3.4413793103448276, "grad_norm": 18.098007202148438, "learning_rate": 3.604445455591099e-06, "loss": 1.9074, "step": 10978 }, { "epoch": 3.441692789968652, "grad_norm": 10.020901679992676, "learning_rate": 3.601891842875882e-06, "loss": 2.1503, "step": 10979 }, { "epoch": 3.4420062695924765, "grad_norm": 6.823008060455322, "learning_rate": 3.5993368025147647e-06, "loss": 2.1735, "step": 10980 }, { "epoch": 3.442319749216301, "grad_norm": 28.041271209716797, "learning_rate": 3.5967803378181387e-06, "loss": 2.6673, "step": 10981 }, { "epoch": 3.4426332288401253, "grad_norm": 206.98020935058594, "learning_rate": 3.5942224520982405e-06, "loss": 2.5859, "step": 10982 }, { "epoch": 3.44294670846395, "grad_norm": 9.063553810119629, "learning_rate": 3.5916631486691467e-06, "loss": 2.0452, "step": 10983 }, { "epoch": 3.4432601880877742, "grad_norm": 11.605732917785645, "learning_rate": 3.589102430846773e-06, "loss": 1.9399, "step": 10984 }, { "epoch": 3.4435736677115987, "grad_norm": 8.095850944519043, "learning_rate": 3.586540301948866e-06, "loss": 2.1125, "step": 10985 }, { "epoch": 3.443887147335423, "grad_norm": 5.688508987426758, "learning_rate": 3.5839767652949998e-06, "loss": 2.195, "step": 10986 }, { "epoch": 3.4442006269592476, "grad_norm": 6.556223392486572, "learning_rate": 3.5814118242065755e-06, "loss": 1.9742, "step": 10987 }, { "epoch": 3.444514106583072, "grad_norm": 6.326581001281738, "learning_rate": 3.57884548200681e-06, "loss": 2.1886, "step": 10988 }, { "epoch": 3.4448275862068964, "grad_norm": 118.64533233642578, "learning_rate": 3.5762777420207382e-06, "loss": 2.7056, "step": 10989 }, { "epoch": 3.445141065830721, "grad_norm": 8.37398910522461, "learning_rate": 3.5737086075752054e-06, "loss": 1.9705, "step": 10990 }, { "epoch": 3.4454545454545453, "grad_norm": 178.28038024902344, "learning_rate": 3.5711380819988627e-06, "loss": 2.2533, "step": 10991 }, { "epoch": 3.4457680250783698, "grad_norm": 6.293056011199951, "learning_rate": 3.5685661686221644e-06, "loss": 2.087, "step": 10992 }, { "epoch": 3.446081504702194, "grad_norm": 20.558147430419922, "learning_rate": 3.565992870777364e-06, "loss": 1.6956, "step": 10993 }, { "epoch": 3.4463949843260187, "grad_norm": 17.61791229248047, "learning_rate": 3.5634181917985057e-06, "loss": 2.1633, "step": 10994 }, { "epoch": 3.446708463949843, "grad_norm": 5.013230323791504, "learning_rate": 3.5608421350214256e-06, "loss": 2.0434, "step": 10995 }, { "epoch": 3.4470219435736675, "grad_norm": 7.378649711608887, "learning_rate": 3.5582647037837446e-06, "loss": 2.0821, "step": 10996 }, { "epoch": 3.447335423197492, "grad_norm": 6.336434841156006, "learning_rate": 3.5556859014248646e-06, "loss": 2.268, "step": 10997 }, { "epoch": 3.4476489028213164, "grad_norm": 68.13605499267578, "learning_rate": 3.553105731285963e-06, "loss": 2.4221, "step": 10998 }, { "epoch": 3.447962382445141, "grad_norm": 7.13245964050293, "learning_rate": 3.5505241967099883e-06, "loss": 2.3866, "step": 10999 }, { "epoch": 3.4482758620689653, "grad_norm": 9.987151145935059, "learning_rate": 3.5479413010416606e-06, "loss": 2.223, "step": 11000 }, { "epoch": 3.4485893416927897, "grad_norm": 8.28502368927002, "learning_rate": 3.54535704762746e-06, "loss": 2.0354, "step": 11001 }, { "epoch": 3.448902821316614, "grad_norm": 21.46816635131836, "learning_rate": 3.5427714398156267e-06, "loss": 1.9602, "step": 11002 }, { "epoch": 3.4492163009404386, "grad_norm": 8.501631736755371, "learning_rate": 3.540184480956157e-06, "loss": 2.0336, "step": 11003 }, { "epoch": 3.4495297805642635, "grad_norm": 7.380056858062744, "learning_rate": 3.5375961744007954e-06, "loss": 2.1308, "step": 11004 }, { "epoch": 3.449843260188088, "grad_norm": 7.649080276489258, "learning_rate": 3.535006523503034e-06, "loss": 2.034, "step": 11005 }, { "epoch": 3.4501567398119124, "grad_norm": 9.880663871765137, "learning_rate": 3.532415531618107e-06, "loss": 1.8521, "step": 11006 }, { "epoch": 3.450470219435737, "grad_norm": 6.39438533782959, "learning_rate": 3.5298232021029845e-06, "loss": 2.0519, "step": 11007 }, { "epoch": 3.4507836990595613, "grad_norm": 6.558131694793701, "learning_rate": 3.527229538316371e-06, "loss": 1.9733, "step": 11008 }, { "epoch": 3.4510971786833857, "grad_norm": 6.167349815368652, "learning_rate": 3.524634543618699e-06, "loss": 2.1121, "step": 11009 }, { "epoch": 3.45141065830721, "grad_norm": 5.200660705566406, "learning_rate": 3.522038221372126e-06, "loss": 2.2715, "step": 11010 }, { "epoch": 3.4517241379310346, "grad_norm": 11.069522857666016, "learning_rate": 3.519440574940529e-06, "loss": 2.2117, "step": 11011 }, { "epoch": 3.452037617554859, "grad_norm": 13.359758377075195, "learning_rate": 3.516841607689501e-06, "loss": 1.9694, "step": 11012 }, { "epoch": 3.4523510971786835, "grad_norm": 12.218761444091797, "learning_rate": 3.514241322986346e-06, "loss": 2.0473, "step": 11013 }, { "epoch": 3.452664576802508, "grad_norm": 8.149621963500977, "learning_rate": 3.5116397242000748e-06, "loss": 2.1054, "step": 11014 }, { "epoch": 3.4529780564263324, "grad_norm": 13.959660530090332, "learning_rate": 3.509036814701401e-06, "loss": 1.9517, "step": 11015 }, { "epoch": 3.453291536050157, "grad_norm": 5.786566734313965, "learning_rate": 3.5064325978627365e-06, "loss": 2.076, "step": 11016 }, { "epoch": 3.4536050156739813, "grad_norm": 13.605454444885254, "learning_rate": 3.5038270770581883e-06, "loss": 2.0511, "step": 11017 }, { "epoch": 3.4539184952978057, "grad_norm": 7.2624897956848145, "learning_rate": 3.501220255663549e-06, "loss": 2.0421, "step": 11018 }, { "epoch": 3.45423197492163, "grad_norm": 16.76471710205078, "learning_rate": 3.4986121370563007e-06, "loss": 2.1701, "step": 11019 }, { "epoch": 3.4545454545454546, "grad_norm": 5.968671798706055, "learning_rate": 3.4960027246156043e-06, "loss": 1.8678, "step": 11020 }, { "epoch": 3.454858934169279, "grad_norm": 9.906926155090332, "learning_rate": 3.4933920217222955e-06, "loss": 2.145, "step": 11021 }, { "epoch": 3.4551724137931035, "grad_norm": 61.301612854003906, "learning_rate": 3.4907800317588845e-06, "loss": 2.0987, "step": 11022 }, { "epoch": 3.455485893416928, "grad_norm": 6.117892265319824, "learning_rate": 3.488166758109548e-06, "loss": 2.107, "step": 11023 }, { "epoch": 3.4557993730407524, "grad_norm": 6.5693559646606445, "learning_rate": 3.4855522041601265e-06, "loss": 1.9699, "step": 11024 }, { "epoch": 3.456112852664577, "grad_norm": 9.39967155456543, "learning_rate": 3.482936373298118e-06, "loss": 2.2463, "step": 11025 }, { "epoch": 3.4564263322884012, "grad_norm": 19.323734283447266, "learning_rate": 3.480319268912676e-06, "loss": 2.6698, "step": 11026 }, { "epoch": 3.4567398119122257, "grad_norm": 9.556253433227539, "learning_rate": 3.4777008943946032e-06, "loss": 2.2039, "step": 11027 }, { "epoch": 3.45705329153605, "grad_norm": 55.92209243774414, "learning_rate": 3.4750812531363486e-06, "loss": 1.527, "step": 11028 }, { "epoch": 3.4573667711598746, "grad_norm": 8.4572172164917, "learning_rate": 3.472460348532002e-06, "loss": 2.1097, "step": 11029 }, { "epoch": 3.457680250783699, "grad_norm": 14.268648147583008, "learning_rate": 3.469838183977291e-06, "loss": 2.183, "step": 11030 }, { "epoch": 3.4579937304075234, "grad_norm": 5.959068775177002, "learning_rate": 3.467214762869574e-06, "loss": 2.0911, "step": 11031 }, { "epoch": 3.458307210031348, "grad_norm": 24.565610885620117, "learning_rate": 3.4645900886078388e-06, "loss": 2.0553, "step": 11032 }, { "epoch": 3.4586206896551723, "grad_norm": 9.25796127319336, "learning_rate": 3.4619641645926966e-06, "loss": 1.9901, "step": 11033 }, { "epoch": 3.4589341692789968, "grad_norm": 14.883484840393066, "learning_rate": 3.4593369942263766e-06, "loss": 2.0997, "step": 11034 }, { "epoch": 3.459247648902821, "grad_norm": 21.498735427856445, "learning_rate": 3.4567085809127247e-06, "loss": 2.103, "step": 11035 }, { "epoch": 3.4595611285266457, "grad_norm": 12.249140739440918, "learning_rate": 3.4540789280571962e-06, "loss": 2.1483, "step": 11036 }, { "epoch": 3.45987460815047, "grad_norm": 7.707129955291748, "learning_rate": 3.4514480390668516e-06, "loss": 1.9802, "step": 11037 }, { "epoch": 3.4601880877742945, "grad_norm": 10.302163124084473, "learning_rate": 3.448815917350355e-06, "loss": 2.0533, "step": 11038 }, { "epoch": 3.460501567398119, "grad_norm": 8.79355525970459, "learning_rate": 3.446182566317966e-06, "loss": 2.3678, "step": 11039 }, { "epoch": 3.4608150470219434, "grad_norm": 4.989346027374268, "learning_rate": 3.443547989381536e-06, "loss": 1.8659, "step": 11040 }, { "epoch": 3.461128526645768, "grad_norm": 7.2120256423950195, "learning_rate": 3.4409121899545087e-06, "loss": 2.5581, "step": 11041 }, { "epoch": 3.4614420062695923, "grad_norm": 5.775761604309082, "learning_rate": 3.4382751714519073e-06, "loss": 2.1519, "step": 11042 }, { "epoch": 3.4617554858934168, "grad_norm": 7.934549808502197, "learning_rate": 3.4356369372903377e-06, "loss": 2.1329, "step": 11043 }, { "epoch": 3.462068965517241, "grad_norm": 7.046741485595703, "learning_rate": 3.432997490887979e-06, "loss": 2.0619, "step": 11044 }, { "epoch": 3.462382445141066, "grad_norm": 24.24644660949707, "learning_rate": 3.4303568356645804e-06, "loss": 2.2077, "step": 11045 }, { "epoch": 3.4626959247648905, "grad_norm": 10.682208061218262, "learning_rate": 3.4277149750414597e-06, "loss": 2.12, "step": 11046 }, { "epoch": 3.463009404388715, "grad_norm": 12.537620544433594, "learning_rate": 3.4250719124414933e-06, "loss": 2.2576, "step": 11047 }, { "epoch": 3.4633228840125394, "grad_norm": 7.227124214172363, "learning_rate": 3.422427651289118e-06, "loss": 2.0312, "step": 11048 }, { "epoch": 3.463636363636364, "grad_norm": 8.081364631652832, "learning_rate": 3.419782195010322e-06, "loss": 1.9108, "step": 11049 }, { "epoch": 3.4639498432601883, "grad_norm": 10.107020378112793, "learning_rate": 3.417135547032642e-06, "loss": 2.2735, "step": 11050 }, { "epoch": 3.4642633228840127, "grad_norm": 9.05185604095459, "learning_rate": 3.4144877107851582e-06, "loss": 1.9523, "step": 11051 }, { "epoch": 3.464576802507837, "grad_norm": 16.421157836914062, "learning_rate": 3.411838689698492e-06, "loss": 1.9878, "step": 11052 }, { "epoch": 3.4648902821316616, "grad_norm": 6.11166524887085, "learning_rate": 3.409188487204797e-06, "loss": 1.9571, "step": 11053 }, { "epoch": 3.465203761755486, "grad_norm": 9.85101318359375, "learning_rate": 3.406537106737762e-06, "loss": 2.0877, "step": 11054 }, { "epoch": 3.4655172413793105, "grad_norm": 6.166840076446533, "learning_rate": 3.4038845517325968e-06, "loss": 2.1842, "step": 11055 }, { "epoch": 3.465830721003135, "grad_norm": 17.8258056640625, "learning_rate": 3.4012308256260366e-06, "loss": 2.0533, "step": 11056 }, { "epoch": 3.4661442006269594, "grad_norm": 6.9129180908203125, "learning_rate": 3.3985759318563338e-06, "loss": 2.0014, "step": 11057 }, { "epoch": 3.466457680250784, "grad_norm": 6.286008358001709, "learning_rate": 3.39591987386325e-06, "loss": 2.3088, "step": 11058 }, { "epoch": 3.4667711598746083, "grad_norm": 6.414161682128906, "learning_rate": 3.3932626550880617e-06, "loss": 2.0587, "step": 11059 }, { "epoch": 3.4670846394984327, "grad_norm": 26.817668914794922, "learning_rate": 3.3906042789735427e-06, "loss": 2.6531, "step": 11060 }, { "epoch": 3.467398119122257, "grad_norm": 9.486750602722168, "learning_rate": 3.3879447489639687e-06, "loss": 2.1643, "step": 11061 }, { "epoch": 3.4677115987460816, "grad_norm": 8.974705696105957, "learning_rate": 3.385284068505113e-06, "loss": 2.2145, "step": 11062 }, { "epoch": 3.468025078369906, "grad_norm": 9.027792930603027, "learning_rate": 3.382622241044237e-06, "loss": 1.923, "step": 11063 }, { "epoch": 3.4683385579937305, "grad_norm": 70.34439849853516, "learning_rate": 3.3799592700300867e-06, "loss": 2.4458, "step": 11064 }, { "epoch": 3.468652037617555, "grad_norm": 6.379955768585205, "learning_rate": 3.3772951589128926e-06, "loss": 2.1882, "step": 11065 }, { "epoch": 3.4689655172413794, "grad_norm": 20.917125701904297, "learning_rate": 3.3746299111443616e-06, "loss": 2.2328, "step": 11066 }, { "epoch": 3.469278996865204, "grad_norm": 9.714945793151855, "learning_rate": 3.371963530177672e-06, "loss": 2.0691, "step": 11067 }, { "epoch": 3.4695924764890282, "grad_norm": 10.9085111618042, "learning_rate": 3.369296019467473e-06, "loss": 2.0091, "step": 11068 }, { "epoch": 3.4699059561128527, "grad_norm": 9.702077865600586, "learning_rate": 3.366627382469873e-06, "loss": 2.0083, "step": 11069 }, { "epoch": 3.470219435736677, "grad_norm": 6.689896106719971, "learning_rate": 3.3639576226424454e-06, "loss": 1.7491, "step": 11070 }, { "epoch": 3.4705329153605016, "grad_norm": 8.191032409667969, "learning_rate": 3.3612867434442135e-06, "loss": 2.1535, "step": 11071 }, { "epoch": 3.470846394984326, "grad_norm": 9.611598014831543, "learning_rate": 3.3586147483356534e-06, "loss": 2.0944, "step": 11072 }, { "epoch": 3.4711598746081505, "grad_norm": 5.798028469085693, "learning_rate": 3.355941640778687e-06, "loss": 1.9113, "step": 11073 }, { "epoch": 3.471473354231975, "grad_norm": 9.092846870422363, "learning_rate": 3.3532674242366764e-06, "loss": 1.9596, "step": 11074 }, { "epoch": 3.4717868338557993, "grad_norm": 10.695840835571289, "learning_rate": 3.3505921021744203e-06, "loss": 2.0528, "step": 11075 }, { "epoch": 3.472100313479624, "grad_norm": 14.2547025680542, "learning_rate": 3.3479156780581523e-06, "loss": 2.2038, "step": 11076 }, { "epoch": 3.472413793103448, "grad_norm": 22.972145080566406, "learning_rate": 3.34523815535553e-06, "loss": 1.9594, "step": 11077 }, { "epoch": 3.4727272727272727, "grad_norm": 5.744856834411621, "learning_rate": 3.342559537535638e-06, "loss": 1.8504, "step": 11078 }, { "epoch": 3.473040752351097, "grad_norm": 14.52711009979248, "learning_rate": 3.3398798280689782e-06, "loss": 2.5, "step": 11079 }, { "epoch": 3.4733542319749215, "grad_norm": 5.648051738739014, "learning_rate": 3.3371990304274654e-06, "loss": 2.1141, "step": 11080 }, { "epoch": 3.473667711598746, "grad_norm": 7.034724712371826, "learning_rate": 3.3345171480844275e-06, "loss": 1.919, "step": 11081 }, { "epoch": 3.4739811912225704, "grad_norm": 12.782983779907227, "learning_rate": 3.3318341845145953e-06, "loss": 2.3516, "step": 11082 }, { "epoch": 3.474294670846395, "grad_norm": 6.823465347290039, "learning_rate": 3.3291501431941013e-06, "loss": 1.9575, "step": 11083 }, { "epoch": 3.4746081504702193, "grad_norm": 6.222175121307373, "learning_rate": 3.326465027600474e-06, "loss": 2.1627, "step": 11084 }, { "epoch": 3.4749216300940438, "grad_norm": 5.017480373382568, "learning_rate": 3.3237788412126337e-06, "loss": 2.0625, "step": 11085 }, { "epoch": 3.475235109717868, "grad_norm": 7.903756141662598, "learning_rate": 3.3210915875108895e-06, "loss": 2.224, "step": 11086 }, { "epoch": 3.4755485893416926, "grad_norm": 13.201666831970215, "learning_rate": 3.318403269976932e-06, "loss": 2.0254, "step": 11087 }, { "epoch": 3.475862068965517, "grad_norm": 183.51268005371094, "learning_rate": 3.315713892093829e-06, "loss": 2.3963, "step": 11088 }, { "epoch": 3.4761755485893415, "grad_norm": 6.881157398223877, "learning_rate": 3.313023457346025e-06, "loss": 1.912, "step": 11089 }, { "epoch": 3.476489028213166, "grad_norm": 8.789355278015137, "learning_rate": 3.310331969219332e-06, "loss": 2.1456, "step": 11090 }, { "epoch": 3.4768025078369904, "grad_norm": 13.569602012634277, "learning_rate": 3.3076394312009263e-06, "loss": 3.0151, "step": 11091 }, { "epoch": 3.477115987460815, "grad_norm": 6.707271575927734, "learning_rate": 3.304945846779346e-06, "loss": 1.9398, "step": 11092 }, { "epoch": 3.4774294670846393, "grad_norm": 126.74052429199219, "learning_rate": 3.3022512194444847e-06, "loss": 2.0631, "step": 11093 }, { "epoch": 3.4777429467084637, "grad_norm": 7.4158172607421875, "learning_rate": 3.2995555526875856e-06, "loss": 2.77, "step": 11094 }, { "epoch": 3.478056426332288, "grad_norm": 6.635753154754639, "learning_rate": 3.29685885000124e-06, "loss": 1.8408, "step": 11095 }, { "epoch": 3.4783699059561126, "grad_norm": 74.57502746582031, "learning_rate": 3.294161114879382e-06, "loss": 2.0175, "step": 11096 }, { "epoch": 3.478683385579937, "grad_norm": 21.890514373779297, "learning_rate": 3.291462350817281e-06, "loss": 2.0819, "step": 11097 }, { "epoch": 3.478996865203762, "grad_norm": 12.054509162902832, "learning_rate": 3.2887625613115427e-06, "loss": 2.0422, "step": 11098 }, { "epoch": 3.4793103448275864, "grad_norm": 9.005194664001465, "learning_rate": 3.2860617498600982e-06, "loss": 2.6503, "step": 11099 }, { "epoch": 3.479623824451411, "grad_norm": 7.964498519897461, "learning_rate": 3.283359919962206e-06, "loss": 2.1657, "step": 11100 }, { "epoch": 3.4799373040752353, "grad_norm": 7.380356788635254, "learning_rate": 3.2806570751184406e-06, "loss": 2.1884, "step": 11101 }, { "epoch": 3.4802507836990597, "grad_norm": 6.798099517822266, "learning_rate": 3.277953218830694e-06, "loss": 1.9055, "step": 11102 }, { "epoch": 3.480564263322884, "grad_norm": 6.1082634925842285, "learning_rate": 3.2752483546021674e-06, "loss": 2.206, "step": 11103 }, { "epoch": 3.4808777429467086, "grad_norm": 6.170208930969238, "learning_rate": 3.272542485937369e-06, "loss": 2.1506, "step": 11104 }, { "epoch": 3.481191222570533, "grad_norm": 6.914505481719971, "learning_rate": 3.2698356163421073e-06, "loss": 2.062, "step": 11105 }, { "epoch": 3.4815047021943575, "grad_norm": 9.373636245727539, "learning_rate": 3.267127749323489e-06, "loss": 1.9532, "step": 11106 }, { "epoch": 3.481818181818182, "grad_norm": 8.881158828735352, "learning_rate": 3.2644188883899095e-06, "loss": 2.7879, "step": 11107 }, { "epoch": 3.4821316614420064, "grad_norm": 6.513095378875732, "learning_rate": 3.2617090370510584e-06, "loss": 2.0543, "step": 11108 }, { "epoch": 3.482445141065831, "grad_norm": 8.898777961730957, "learning_rate": 3.2589981988179027e-06, "loss": 2.1269, "step": 11109 }, { "epoch": 3.4827586206896552, "grad_norm": 6.822129249572754, "learning_rate": 3.25628637720269e-06, "loss": 2.1374, "step": 11110 }, { "epoch": 3.4830721003134797, "grad_norm": 12.747669219970703, "learning_rate": 3.2535735757189425e-06, "loss": 2.1019, "step": 11111 }, { "epoch": 3.483385579937304, "grad_norm": 6.935823440551758, "learning_rate": 3.2508597978814515e-06, "loss": 2.0127, "step": 11112 }, { "epoch": 3.4836990595611286, "grad_norm": 12.284368515014648, "learning_rate": 3.248145047206273e-06, "loss": 2.049, "step": 11113 }, { "epoch": 3.484012539184953, "grad_norm": 9.033076286315918, "learning_rate": 3.245429327210725e-06, "loss": 2.2939, "step": 11114 }, { "epoch": 3.4843260188087775, "grad_norm": 6.336668491363525, "learning_rate": 3.2427126414133787e-06, "loss": 2.0198, "step": 11115 }, { "epoch": 3.484639498432602, "grad_norm": 42.827552795410156, "learning_rate": 3.239994993334059e-06, "loss": 2.1234, "step": 11116 }, { "epoch": 3.4849529780564263, "grad_norm": 11.656134605407715, "learning_rate": 3.2372763864938363e-06, "loss": 2.2056, "step": 11117 }, { "epoch": 3.485266457680251, "grad_norm": 18.528879165649414, "learning_rate": 3.234556824415023e-06, "loss": 2.4466, "step": 11118 }, { "epoch": 3.4855799373040752, "grad_norm": 7.781280517578125, "learning_rate": 3.2318363106211707e-06, "loss": 1.9909, "step": 11119 }, { "epoch": 3.4858934169278997, "grad_norm": 10.028493881225586, "learning_rate": 3.2291148486370626e-06, "loss": 2.0446, "step": 11120 }, { "epoch": 3.486206896551724, "grad_norm": 8.279256820678711, "learning_rate": 3.2263924419887116e-06, "loss": 2.0269, "step": 11121 }, { "epoch": 3.4865203761755486, "grad_norm": 5.392518520355225, "learning_rate": 3.2236690942033523e-06, "loss": 2.4413, "step": 11122 }, { "epoch": 3.486833855799373, "grad_norm": 5.363646030426025, "learning_rate": 3.2209448088094406e-06, "loss": 1.9882, "step": 11123 }, { "epoch": 3.4871473354231974, "grad_norm": 9.451504707336426, "learning_rate": 3.218219589336648e-06, "loss": 2.0262, "step": 11124 }, { "epoch": 3.487460815047022, "grad_norm": 7.592034816741943, "learning_rate": 3.215493439315854e-06, "loss": 2.1545, "step": 11125 }, { "epoch": 3.4877742946708463, "grad_norm": 14.205779075622559, "learning_rate": 3.2127663622791445e-06, "loss": 2.2117, "step": 11126 }, { "epoch": 3.4880877742946708, "grad_norm": 11.300971984863281, "learning_rate": 3.2100383617598075e-06, "loss": 2.1665, "step": 11127 }, { "epoch": 3.488401253918495, "grad_norm": 10.335318565368652, "learning_rate": 3.207309441292325e-06, "loss": 2.1383, "step": 11128 }, { "epoch": 3.4887147335423196, "grad_norm": 10.229572296142578, "learning_rate": 3.204579604412375e-06, "loss": 2.125, "step": 11129 }, { "epoch": 3.489028213166144, "grad_norm": 53.29576110839844, "learning_rate": 3.2018488546568194e-06, "loss": 1.9363, "step": 11130 }, { "epoch": 3.4893416927899685, "grad_norm": 15.112397193908691, "learning_rate": 3.1991171955637036e-06, "loss": 2.1269, "step": 11131 }, { "epoch": 3.489655172413793, "grad_norm": 36.49161148071289, "learning_rate": 3.196384630672251e-06, "loss": 2.1813, "step": 11132 }, { "epoch": 3.4899686520376174, "grad_norm": 176.57106018066406, "learning_rate": 3.193651163522861e-06, "loss": 2.3964, "step": 11133 }, { "epoch": 3.490282131661442, "grad_norm": 5.329248428344727, "learning_rate": 3.1909167976570977e-06, "loss": 2.0269, "step": 11134 }, { "epoch": 3.4905956112852663, "grad_norm": 7.735179901123047, "learning_rate": 3.188181536617694e-06, "loss": 2.1236, "step": 11135 }, { "epoch": 3.4909090909090907, "grad_norm": 5.927901268005371, "learning_rate": 3.185445383948539e-06, "loss": 2.0804, "step": 11136 }, { "epoch": 3.491222570532915, "grad_norm": 9.126822471618652, "learning_rate": 3.1827083431946786e-06, "loss": 2.444, "step": 11137 }, { "epoch": 3.4915360501567396, "grad_norm": 7.816540241241455, "learning_rate": 3.1799704179023105e-06, "loss": 2.0691, "step": 11138 }, { "epoch": 3.4918495297805645, "grad_norm": 7.356496810913086, "learning_rate": 3.177231611618777e-06, "loss": 2.1215, "step": 11139 }, { "epoch": 3.492163009404389, "grad_norm": 6.415617942810059, "learning_rate": 3.174491927892561e-06, "loss": 2.169, "step": 11140 }, { "epoch": 3.4924764890282134, "grad_norm": 7.045026779174805, "learning_rate": 3.1717513702732844e-06, "loss": 2.155, "step": 11141 }, { "epoch": 3.492789968652038, "grad_norm": 7.240963459014893, "learning_rate": 3.1690099423117e-06, "loss": 2.1786, "step": 11142 }, { "epoch": 3.4931034482758623, "grad_norm": 7.880476474761963, "learning_rate": 3.166267647559688e-06, "loss": 2.178, "step": 11143 }, { "epoch": 3.4934169278996867, "grad_norm": 10.45797061920166, "learning_rate": 3.1635244895702527e-06, "loss": 2.1378, "step": 11144 }, { "epoch": 3.493730407523511, "grad_norm": 94.32008361816406, "learning_rate": 3.160780471897516e-06, "loss": 2.2031, "step": 11145 }, { "epoch": 3.4940438871473356, "grad_norm": 10.3287935256958, "learning_rate": 3.158035598096715e-06, "loss": 1.9397, "step": 11146 }, { "epoch": 3.49435736677116, "grad_norm": 12.260371208190918, "learning_rate": 3.1552898717241927e-06, "loss": 2.2229, "step": 11147 }, { "epoch": 3.4946708463949845, "grad_norm": 8.14364242553711, "learning_rate": 3.152543296337401e-06, "loss": 1.8871, "step": 11148 }, { "epoch": 3.494984326018809, "grad_norm": 12.951444625854492, "learning_rate": 3.1497958754948894e-06, "loss": 2.4241, "step": 11149 }, { "epoch": 3.4952978056426334, "grad_norm": 7.937882900238037, "learning_rate": 3.147047612756302e-06, "loss": 2.0508, "step": 11150 }, { "epoch": 3.495611285266458, "grad_norm": 5.501919746398926, "learning_rate": 3.1442985116823764e-06, "loss": 2.038, "step": 11151 }, { "epoch": 3.4959247648902823, "grad_norm": 16.341705322265625, "learning_rate": 3.1415485758349344e-06, "loss": 2.102, "step": 11152 }, { "epoch": 3.4962382445141067, "grad_norm": 8.208110809326172, "learning_rate": 3.1387978087768796e-06, "loss": 2.2208, "step": 11153 }, { "epoch": 3.496551724137931, "grad_norm": 185.7803192138672, "learning_rate": 3.1360462140721925e-06, "loss": 2.2489, "step": 11154 }, { "epoch": 3.4968652037617556, "grad_norm": 16.41879653930664, "learning_rate": 3.133293795285926e-06, "loss": 2.4954, "step": 11155 }, { "epoch": 3.49717868338558, "grad_norm": 28.10996437072754, "learning_rate": 3.1305405559842016e-06, "loss": 2.5217, "step": 11156 }, { "epoch": 3.4974921630094045, "grad_norm": 289.1875915527344, "learning_rate": 3.1277864997342037e-06, "loss": 2.2584, "step": 11157 }, { "epoch": 3.497805642633229, "grad_norm": 6.011585712432861, "learning_rate": 3.1250316301041727e-06, "loss": 2.1543, "step": 11158 }, { "epoch": 3.4981191222570533, "grad_norm": 9.811304092407227, "learning_rate": 3.1222759506634063e-06, "loss": 2.1709, "step": 11159 }, { "epoch": 3.498432601880878, "grad_norm": 10.861211776733398, "learning_rate": 3.11951946498225e-06, "loss": 2.0472, "step": 11160 }, { "epoch": 3.4987460815047022, "grad_norm": 5.319398880004883, "learning_rate": 3.1167621766320932e-06, "loss": 1.9691, "step": 11161 }, { "epoch": 3.4990595611285267, "grad_norm": 15.501347541809082, "learning_rate": 3.114004089185365e-06, "loss": 2.0216, "step": 11162 }, { "epoch": 3.499373040752351, "grad_norm": 9.702702522277832, "learning_rate": 3.1112452062155324e-06, "loss": 2.2803, "step": 11163 }, { "epoch": 3.4996865203761756, "grad_norm": 6.2375664710998535, "learning_rate": 3.1084855312970897e-06, "loss": 2.063, "step": 11164 }, { "epoch": 3.5, "grad_norm": 8.431381225585938, "learning_rate": 3.1057250680055604e-06, "loss": 2.4914, "step": 11165 }, { "epoch": 3.5003134796238244, "grad_norm": 7.694077968597412, "learning_rate": 3.1029638199174865e-06, "loss": 2.5003, "step": 11166 }, { "epoch": 3.500626959247649, "grad_norm": 59.8624267578125, "learning_rate": 3.1002017906104303e-06, "loss": 2.0974, "step": 11167 }, { "epoch": 3.5009404388714733, "grad_norm": 6.842008113861084, "learning_rate": 3.0974389836629628e-06, "loss": 1.9321, "step": 11168 }, { "epoch": 3.5012539184952978, "grad_norm": 5.644147872924805, "learning_rate": 3.094675402654664e-06, "loss": 2.0064, "step": 11169 }, { "epoch": 3.501567398119122, "grad_norm": 15.47774600982666, "learning_rate": 3.091911051166117e-06, "loss": 2.268, "step": 11170 }, { "epoch": 3.5018808777429467, "grad_norm": 4.801971435546875, "learning_rate": 3.0891459327789035e-06, "loss": 2.1988, "step": 11171 }, { "epoch": 3.502194357366771, "grad_norm": 9.197260856628418, "learning_rate": 3.0863800510755976e-06, "loss": 2.0795, "step": 11172 }, { "epoch": 3.502194357366771, "eval_loss": 2.2918038368225098, "eval_runtime": 21.0847, "eval_samples_per_second": 127.438, "eval_steps_per_second": 7.968, "step": 11172 }, { "epoch": 3.5025078369905955, "grad_norm": 26.205333709716797, "learning_rate": 3.0836134096397642e-06, "loss": 2.1207, "step": 11173 }, { "epoch": 3.50282131661442, "grad_norm": 59.40201187133789, "learning_rate": 3.0808460120559493e-06, "loss": 2.2126, "step": 11174 }, { "epoch": 3.5031347962382444, "grad_norm": 76.23147583007812, "learning_rate": 3.0780778619096827e-06, "loss": 2.3692, "step": 11175 }, { "epoch": 3.503448275862069, "grad_norm": 31.889497756958008, "learning_rate": 3.0753089627874668e-06, "loss": 2.1223, "step": 11176 }, { "epoch": 3.5037617554858933, "grad_norm": 66.89705657958984, "learning_rate": 3.072539318276773e-06, "loss": 2.1611, "step": 11177 }, { "epoch": 3.5040752351097177, "grad_norm": 9.225419044494629, "learning_rate": 3.069768931966043e-06, "loss": 2.0245, "step": 11178 }, { "epoch": 3.504388714733542, "grad_norm": 24.017433166503906, "learning_rate": 3.0669978074446756e-06, "loss": 2.3171, "step": 11179 }, { "epoch": 3.5047021943573666, "grad_norm": 7.461001396179199, "learning_rate": 3.064225948303027e-06, "loss": 2.129, "step": 11180 }, { "epoch": 3.505015673981191, "grad_norm": 6.006809711456299, "learning_rate": 3.0614533581324063e-06, "loss": 2.1501, "step": 11181 }, { "epoch": 3.5053291536050155, "grad_norm": 16.280696868896484, "learning_rate": 3.0586800405250677e-06, "loss": 2.1298, "step": 11182 }, { "epoch": 3.50564263322884, "grad_norm": 18.322460174560547, "learning_rate": 3.0559059990742103e-06, "loss": 2.0537, "step": 11183 }, { "epoch": 3.5059561128526644, "grad_norm": 18.12174415588379, "learning_rate": 3.0531312373739695e-06, "loss": 2.1752, "step": 11184 }, { "epoch": 3.506269592476489, "grad_norm": 25.697303771972656, "learning_rate": 3.050355759019414e-06, "loss": 2.1339, "step": 11185 }, { "epoch": 3.5065830721003133, "grad_norm": 7.857831001281738, "learning_rate": 3.0475795676065424e-06, "loss": 1.8981, "step": 11186 }, { "epoch": 3.5068965517241377, "grad_norm": 6.3715739250183105, "learning_rate": 3.044802666732275e-06, "loss": 2.1942, "step": 11187 }, { "epoch": 3.507210031347962, "grad_norm": 7.807705402374268, "learning_rate": 3.0420250599944525e-06, "loss": 2.1148, "step": 11188 }, { "epoch": 3.5075235109717866, "grad_norm": 7.151328086853027, "learning_rate": 3.0392467509918294e-06, "loss": 2.1886, "step": 11189 }, { "epoch": 3.507836990595611, "grad_norm": 8.021864891052246, "learning_rate": 3.036467743324072e-06, "loss": 2.8278, "step": 11190 }, { "epoch": 3.5081504702194355, "grad_norm": 6.848351955413818, "learning_rate": 3.0336880405917496e-06, "loss": 1.8056, "step": 11191 }, { "epoch": 3.50846394984326, "grad_norm": 9.438998222351074, "learning_rate": 3.030907646396333e-06, "loss": 2.0343, "step": 11192 }, { "epoch": 3.5087774294670844, "grad_norm": 6.154539585113525, "learning_rate": 3.0281265643401893e-06, "loss": 1.9385, "step": 11193 }, { "epoch": 3.509090909090909, "grad_norm": 10.282724380493164, "learning_rate": 3.0253447980265754e-06, "loss": 2.2504, "step": 11194 }, { "epoch": 3.5094043887147337, "grad_norm": 8.483940124511719, "learning_rate": 3.022562351059637e-06, "loss": 1.7973, "step": 11195 }, { "epoch": 3.509717868338558, "grad_norm": 8.859427452087402, "learning_rate": 3.019779227044398e-06, "loss": 2.1112, "step": 11196 }, { "epoch": 3.5100313479623826, "grad_norm": 18.39027976989746, "learning_rate": 3.016995429586765e-06, "loss": 2.1993, "step": 11197 }, { "epoch": 3.510344827586207, "grad_norm": 8.796746253967285, "learning_rate": 3.0142109622935118e-06, "loss": 1.82, "step": 11198 }, { "epoch": 3.5106583072100315, "grad_norm": 10.031035423278809, "learning_rate": 3.011425828772283e-06, "loss": 1.9269, "step": 11199 }, { "epoch": 3.510971786833856, "grad_norm": 6.190243244171143, "learning_rate": 3.0086400326315853e-06, "loss": 2.338, "step": 11200 }, { "epoch": 3.5112852664576804, "grad_norm": 7.417943000793457, "learning_rate": 3.0058535774807835e-06, "loss": 2.0706, "step": 11201 }, { "epoch": 3.511598746081505, "grad_norm": 5.152398586273193, "learning_rate": 3.003066466930099e-06, "loss": 2.0832, "step": 11202 }, { "epoch": 3.5119122257053292, "grad_norm": 10.092495918273926, "learning_rate": 3.0002787045905985e-06, "loss": 2.3026, "step": 11203 }, { "epoch": 3.5122257053291537, "grad_norm": 9.795414924621582, "learning_rate": 2.9974902940741957e-06, "loss": 1.982, "step": 11204 }, { "epoch": 3.512539184952978, "grad_norm": 5.26927375793457, "learning_rate": 2.9947012389936433e-06, "loss": 1.9348, "step": 11205 }, { "epoch": 3.5128526645768026, "grad_norm": 6.439382553100586, "learning_rate": 2.9919115429625295e-06, "loss": 2.2652, "step": 11206 }, { "epoch": 3.513166144200627, "grad_norm": 90.46839141845703, "learning_rate": 2.9891212095952715e-06, "loss": 2.1576, "step": 11207 }, { "epoch": 3.5134796238244514, "grad_norm": 6.638188362121582, "learning_rate": 2.9863302425071156e-06, "loss": 2.3311, "step": 11208 }, { "epoch": 3.513793103448276, "grad_norm": 9.768985748291016, "learning_rate": 2.9835386453141245e-06, "loss": 2.1908, "step": 11209 }, { "epoch": 3.5141065830721003, "grad_norm": 8.835453033447266, "learning_rate": 2.9807464216331815e-06, "loss": 2.0614, "step": 11210 }, { "epoch": 3.5144200626959248, "grad_norm": 26.101186752319336, "learning_rate": 2.97795357508198e-06, "loss": 2.3591, "step": 11211 }, { "epoch": 3.514733542319749, "grad_norm": 6.017193794250488, "learning_rate": 2.9751601092790185e-06, "loss": 2.0781, "step": 11212 }, { "epoch": 3.5150470219435737, "grad_norm": 6.449441909790039, "learning_rate": 2.972366027843601e-06, "loss": 2.1275, "step": 11213 }, { "epoch": 3.515360501567398, "grad_norm": 11.049010276794434, "learning_rate": 2.9695713343958277e-06, "loss": 2.5652, "step": 11214 }, { "epoch": 3.5156739811912225, "grad_norm": 6.207465648651123, "learning_rate": 2.966776032556591e-06, "loss": 2.0708, "step": 11215 }, { "epoch": 3.515987460815047, "grad_norm": 21.507081985473633, "learning_rate": 2.963980125947573e-06, "loss": 2.1759, "step": 11216 }, { "epoch": 3.5163009404388714, "grad_norm": 6.145802021026611, "learning_rate": 2.9611836181912386e-06, "loss": 2.0316, "step": 11217 }, { "epoch": 3.516614420062696, "grad_norm": 8.326297760009766, "learning_rate": 2.958386512910831e-06, "loss": 2.324, "step": 11218 }, { "epoch": 3.5169278996865203, "grad_norm": 13.854616165161133, "learning_rate": 2.9555888137303695e-06, "loss": 2.2443, "step": 11219 }, { "epoch": 3.5172413793103448, "grad_norm": 139.50689697265625, "learning_rate": 2.9527905242746397e-06, "loss": 2.5295, "step": 11220 }, { "epoch": 3.517554858934169, "grad_norm": 6.109231472015381, "learning_rate": 2.949991648169196e-06, "loss": 2.1979, "step": 11221 }, { "epoch": 3.5178683385579936, "grad_norm": 23.111881256103516, "learning_rate": 2.9471921890403487e-06, "loss": 2.2106, "step": 11222 }, { "epoch": 3.518181818181818, "grad_norm": 10.992758750915527, "learning_rate": 2.944392150515167e-06, "loss": 2.1079, "step": 11223 }, { "epoch": 3.5184952978056425, "grad_norm": 13.174660682678223, "learning_rate": 2.941591536221469e-06, "loss": 2.1143, "step": 11224 }, { "epoch": 3.518808777429467, "grad_norm": 23.607070922851562, "learning_rate": 2.938790349787819e-06, "loss": 2.0059, "step": 11225 }, { "epoch": 3.5191222570532914, "grad_norm": 7.399129390716553, "learning_rate": 2.935988594843522e-06, "loss": 2.0848, "step": 11226 }, { "epoch": 3.519435736677116, "grad_norm": 25.538963317871094, "learning_rate": 2.933186275018621e-06, "loss": 2.8202, "step": 11227 }, { "epoch": 3.5197492163009403, "grad_norm": 5.3722310066223145, "learning_rate": 2.9303833939438905e-06, "loss": 1.7063, "step": 11228 }, { "epoch": 3.5200626959247647, "grad_norm": 8.189468383789062, "learning_rate": 2.927579955250831e-06, "loss": 2.0312, "step": 11229 }, { "epoch": 3.5203761755485896, "grad_norm": 6.609261512756348, "learning_rate": 2.924775962571667e-06, "loss": 2.3004, "step": 11230 }, { "epoch": 3.520689655172414, "grad_norm": 23.234697341918945, "learning_rate": 2.921971419539339e-06, "loss": 2.2396, "step": 11231 }, { "epoch": 3.5210031347962385, "grad_norm": 5.9522504806518555, "learning_rate": 2.9191663297875027e-06, "loss": 2.1394, "step": 11232 }, { "epoch": 3.521316614420063, "grad_norm": 7.370587348937988, "learning_rate": 2.9163606969505208e-06, "loss": 2.3857, "step": 11233 }, { "epoch": 3.5216300940438874, "grad_norm": 7.891048908233643, "learning_rate": 2.9135545246634595e-06, "loss": 2.1126, "step": 11234 }, { "epoch": 3.521943573667712, "grad_norm": 5.744268894195557, "learning_rate": 2.9107478165620845e-06, "loss": 2.1908, "step": 11235 }, { "epoch": 3.5222570532915363, "grad_norm": 9.51618480682373, "learning_rate": 2.907940576282856e-06, "loss": 2.0337, "step": 11236 }, { "epoch": 3.5225705329153607, "grad_norm": 105.05807495117188, "learning_rate": 2.9051328074629226e-06, "loss": 2.3332, "step": 11237 }, { "epoch": 3.522884012539185, "grad_norm": 5.896337032318115, "learning_rate": 2.902324513740118e-06, "loss": 2.0951, "step": 11238 }, { "epoch": 3.5231974921630096, "grad_norm": 7.1210856437683105, "learning_rate": 2.8995156987529567e-06, "loss": 1.8643, "step": 11239 }, { "epoch": 3.523510971786834, "grad_norm": 6.486800193786621, "learning_rate": 2.896706366140629e-06, "loss": 2.0583, "step": 11240 }, { "epoch": 3.5238244514106585, "grad_norm": 9.412314414978027, "learning_rate": 2.893896519542994e-06, "loss": 2.1112, "step": 11241 }, { "epoch": 3.524137931034483, "grad_norm": 7.889289379119873, "learning_rate": 2.8910861626005774e-06, "loss": 2.3149, "step": 11242 }, { "epoch": 3.5244514106583074, "grad_norm": 5.926351547241211, "learning_rate": 2.8882752989545683e-06, "loss": 2.0865, "step": 11243 }, { "epoch": 3.524764890282132, "grad_norm": 8.007075309753418, "learning_rate": 2.8854639322468082e-06, "loss": 2.1207, "step": 11244 }, { "epoch": 3.5250783699059562, "grad_norm": 8.206079483032227, "learning_rate": 2.882652066119795e-06, "loss": 2.0902, "step": 11245 }, { "epoch": 3.5253918495297807, "grad_norm": 33.69438934326172, "learning_rate": 2.8798397042166693e-06, "loss": 2.3478, "step": 11246 }, { "epoch": 3.525705329153605, "grad_norm": 6.085033416748047, "learning_rate": 2.877026850181217e-06, "loss": 2.1627, "step": 11247 }, { "epoch": 3.5260188087774296, "grad_norm": 9.154694557189941, "learning_rate": 2.8742135076578608e-06, "loss": 2.1616, "step": 11248 }, { "epoch": 3.526332288401254, "grad_norm": 8.268950462341309, "learning_rate": 2.8713996802916566e-06, "loss": 2.1676, "step": 11249 }, { "epoch": 3.5266457680250785, "grad_norm": 9.457715034484863, "learning_rate": 2.8685853717282865e-06, "loss": 2.1167, "step": 11250 }, { "epoch": 3.526959247648903, "grad_norm": 6.487046718597412, "learning_rate": 2.8657705856140596e-06, "loss": 1.8855, "step": 11251 }, { "epoch": 3.5272727272727273, "grad_norm": 7.380247116088867, "learning_rate": 2.862955325595899e-06, "loss": 2.1347, "step": 11252 }, { "epoch": 3.527586206896552, "grad_norm": 6.130248546600342, "learning_rate": 2.8601395953213483e-06, "loss": 2.0218, "step": 11253 }, { "epoch": 3.527899686520376, "grad_norm": 15.305278778076172, "learning_rate": 2.857323398438554e-06, "loss": 2.2471, "step": 11254 }, { "epoch": 3.5282131661442007, "grad_norm": 9.61702823638916, "learning_rate": 2.8545067385962706e-06, "loss": 2.1287, "step": 11255 }, { "epoch": 3.528526645768025, "grad_norm": 9.96308422088623, "learning_rate": 2.8516896194438515e-06, "loss": 2.6845, "step": 11256 }, { "epoch": 3.5288401253918495, "grad_norm": 15.594147682189941, "learning_rate": 2.8488720446312456e-06, "loss": 2.7972, "step": 11257 }, { "epoch": 3.529153605015674, "grad_norm": 53.376739501953125, "learning_rate": 2.8460540178089907e-06, "loss": 2.3955, "step": 11258 }, { "epoch": 3.5294670846394984, "grad_norm": 23.199403762817383, "learning_rate": 2.843235542628213e-06, "loss": 2.1178, "step": 11259 }, { "epoch": 3.529780564263323, "grad_norm": 6.430031776428223, "learning_rate": 2.840416622740617e-06, "loss": 2.0224, "step": 11260 }, { "epoch": 3.5300940438871473, "grad_norm": 8.830323219299316, "learning_rate": 2.8375972617984847e-06, "loss": 2.0036, "step": 11261 }, { "epoch": 3.5304075235109718, "grad_norm": 7.854339599609375, "learning_rate": 2.83477746345467e-06, "loss": 2.1283, "step": 11262 }, { "epoch": 3.530721003134796, "grad_norm": 6.260110378265381, "learning_rate": 2.831957231362591e-06, "loss": 2.1894, "step": 11263 }, { "epoch": 3.5310344827586206, "grad_norm": 5.980480670928955, "learning_rate": 2.8291365691762313e-06, "loss": 2.2322, "step": 11264 }, { "epoch": 3.531347962382445, "grad_norm": 8.269782066345215, "learning_rate": 2.82631548055013e-06, "loss": 2.4492, "step": 11265 }, { "epoch": 3.5316614420062695, "grad_norm": 6.556046485900879, "learning_rate": 2.8234939691393765e-06, "loss": 1.8258, "step": 11266 }, { "epoch": 3.531974921630094, "grad_norm": 10.521160125732422, "learning_rate": 2.8206720385996125e-06, "loss": 2.0899, "step": 11267 }, { "epoch": 3.5322884012539184, "grad_norm": 106.22428131103516, "learning_rate": 2.81784969258702e-06, "loss": 2.1916, "step": 11268 }, { "epoch": 3.532601880877743, "grad_norm": 8.112627983093262, "learning_rate": 2.8150269347583175e-06, "loss": 2.1528, "step": 11269 }, { "epoch": 3.5329153605015673, "grad_norm": 8.793839454650879, "learning_rate": 2.8122037687707628e-06, "loss": 2.0673, "step": 11270 }, { "epoch": 3.5332288401253917, "grad_norm": 17.790088653564453, "learning_rate": 2.809380198282136e-06, "loss": 1.8837, "step": 11271 }, { "epoch": 3.533542319749216, "grad_norm": 57.2846794128418, "learning_rate": 2.8065562269507464e-06, "loss": 2.2485, "step": 11272 }, { "epoch": 3.5338557993730406, "grad_norm": 8.697456359863281, "learning_rate": 2.8037318584354195e-06, "loss": 2.406, "step": 11273 }, { "epoch": 3.534169278996865, "grad_norm": 5.884538650512695, "learning_rate": 2.8009070963954967e-06, "loss": 2.0238, "step": 11274 }, { "epoch": 3.5344827586206895, "grad_norm": 25.83014488220215, "learning_rate": 2.798081944490828e-06, "loss": 2.4418, "step": 11275 }, { "epoch": 3.534796238244514, "grad_norm": 6.689133167266846, "learning_rate": 2.7952564063817707e-06, "loss": 2.2064, "step": 11276 }, { "epoch": 3.5351097178683384, "grad_norm": 8.424864768981934, "learning_rate": 2.792430485729181e-06, "loss": 2.3145, "step": 11277 }, { "epoch": 3.535423197492163, "grad_norm": 7.159041404724121, "learning_rate": 2.789604186194411e-06, "loss": 2.0479, "step": 11278 }, { "epoch": 3.5357366771159873, "grad_norm": 7.937317371368408, "learning_rate": 2.786777511439303e-06, "loss": 2.1056, "step": 11279 }, { "epoch": 3.5360501567398117, "grad_norm": 5.449398040771484, "learning_rate": 2.7839504651261873e-06, "loss": 2.2146, "step": 11280 }, { "epoch": 3.536363636363636, "grad_norm": 83.26216125488281, "learning_rate": 2.7811230509178745e-06, "loss": 2.4459, "step": 11281 }, { "epoch": 3.5366771159874606, "grad_norm": 8.541850090026855, "learning_rate": 2.7782952724776503e-06, "loss": 2.0049, "step": 11282 }, { "epoch": 3.536990595611285, "grad_norm": 9.53398323059082, "learning_rate": 2.775467133469275e-06, "loss": 2.225, "step": 11283 }, { "epoch": 3.5373040752351095, "grad_norm": 9.536582946777344, "learning_rate": 2.7726386375569748e-06, "loss": 2.2078, "step": 11284 }, { "epoch": 3.537617554858934, "grad_norm": 5.440367221832275, "learning_rate": 2.7698097884054383e-06, "loss": 2.0122, "step": 11285 }, { "epoch": 3.5379310344827584, "grad_norm": 11.26754093170166, "learning_rate": 2.766980589679812e-06, "loss": 1.9332, "step": 11286 }, { "epoch": 3.538244514106583, "grad_norm": 8.057870864868164, "learning_rate": 2.7641510450456944e-06, "loss": 2.0218, "step": 11287 }, { "epoch": 3.5385579937304072, "grad_norm": 4.983579635620117, "learning_rate": 2.761321158169134e-06, "loss": 2.0229, "step": 11288 }, { "epoch": 3.538871473354232, "grad_norm": 47.393795013427734, "learning_rate": 2.7584909327166213e-06, "loss": 2.2207, "step": 11289 }, { "epoch": 3.5391849529780566, "grad_norm": 15.97496223449707, "learning_rate": 2.7556603723550855e-06, "loss": 2.246, "step": 11290 }, { "epoch": 3.539498432601881, "grad_norm": 9.167826652526855, "learning_rate": 2.752829480751891e-06, "loss": 2.0821, "step": 11291 }, { "epoch": 3.5398119122257055, "grad_norm": 31.31332778930664, "learning_rate": 2.74999826157483e-06, "loss": 2.1968, "step": 11292 }, { "epoch": 3.54012539184953, "grad_norm": 8.228386878967285, "learning_rate": 2.7471667184921193e-06, "loss": 2.2066, "step": 11293 }, { "epoch": 3.5404388714733543, "grad_norm": 5.404594898223877, "learning_rate": 2.7443348551723945e-06, "loss": 2.1219, "step": 11294 }, { "epoch": 3.540752351097179, "grad_norm": 9.407115936279297, "learning_rate": 2.7415026752847102e-06, "loss": 2.3975, "step": 11295 }, { "epoch": 3.5410658307210032, "grad_norm": 6.41239595413208, "learning_rate": 2.7386701824985257e-06, "loss": 2.1406, "step": 11296 }, { "epoch": 3.5413793103448277, "grad_norm": 13.184613227844238, "learning_rate": 2.735837380483709e-06, "loss": 2.4022, "step": 11297 }, { "epoch": 3.541692789968652, "grad_norm": 7.3516058921813965, "learning_rate": 2.7330042729105276e-06, "loss": 1.99, "step": 11298 }, { "epoch": 3.5420062695924766, "grad_norm": 6.975371360778809, "learning_rate": 2.7301708634496453e-06, "loss": 2.3033, "step": 11299 }, { "epoch": 3.542319749216301, "grad_norm": 7.538755893707275, "learning_rate": 2.727337155772118e-06, "loss": 2.1094, "step": 11300 }, { "epoch": 3.5426332288401254, "grad_norm": 7.60914421081543, "learning_rate": 2.724503153549385e-06, "loss": 2.0052, "step": 11301 }, { "epoch": 3.54294670846395, "grad_norm": 5.245004653930664, "learning_rate": 2.721668860453271e-06, "loss": 2.1849, "step": 11302 }, { "epoch": 3.5432601880877743, "grad_norm": 11.255414009094238, "learning_rate": 2.7188342801559747e-06, "loss": 1.9329, "step": 11303 }, { "epoch": 3.5435736677115988, "grad_norm": 7.622250556945801, "learning_rate": 2.715999416330068e-06, "loss": 1.9396, "step": 11304 }, { "epoch": 3.543887147335423, "grad_norm": 5.935868740081787, "learning_rate": 2.7131642726484902e-06, "loss": 2.0393, "step": 11305 }, { "epoch": 3.5442006269592476, "grad_norm": 5.176035404205322, "learning_rate": 2.710328852784543e-06, "loss": 2.0357, "step": 11306 }, { "epoch": 3.544514106583072, "grad_norm": 7.796402931213379, "learning_rate": 2.7074931604118864e-06, "loss": 2.1374, "step": 11307 }, { "epoch": 3.5448275862068965, "grad_norm": 11.635645866394043, "learning_rate": 2.7046571992045334e-06, "loss": 2.098, "step": 11308 }, { "epoch": 3.545141065830721, "grad_norm": 7.9825029373168945, "learning_rate": 2.701820972836843e-06, "loss": 2.0334, "step": 11309 }, { "epoch": 3.5454545454545454, "grad_norm": 23.519559860229492, "learning_rate": 2.698984484983522e-06, "loss": 2.1177, "step": 11310 }, { "epoch": 3.54576802507837, "grad_norm": 9.359636306762695, "learning_rate": 2.696147739319613e-06, "loss": 1.9709, "step": 11311 }, { "epoch": 3.5460815047021943, "grad_norm": 102.8886489868164, "learning_rate": 2.6933107395204926e-06, "loss": 2.2416, "step": 11312 }, { "epoch": 3.5463949843260187, "grad_norm": 34.96095657348633, "learning_rate": 2.6904734892618676e-06, "loss": 1.9439, "step": 11313 }, { "epoch": 3.546708463949843, "grad_norm": 7.254970073699951, "learning_rate": 2.6876359922197703e-06, "loss": 2.2337, "step": 11314 }, { "epoch": 3.5470219435736676, "grad_norm": 4.984001159667969, "learning_rate": 2.6847982520705508e-06, "loss": 2.1214, "step": 11315 }, { "epoch": 3.547335423197492, "grad_norm": 7.9591288566589355, "learning_rate": 2.6819602724908744e-06, "loss": 2.0914, "step": 11316 }, { "epoch": 3.5476489028213165, "grad_norm": 8.138660430908203, "learning_rate": 2.679122057157717e-06, "loss": 2.0538, "step": 11317 }, { "epoch": 3.547962382445141, "grad_norm": 33.66203689575195, "learning_rate": 2.6762836097483615e-06, "loss": 2.2114, "step": 11318 }, { "epoch": 3.5482758620689654, "grad_norm": 28.185697555541992, "learning_rate": 2.673444933940389e-06, "loss": 2.2431, "step": 11319 }, { "epoch": 3.54858934169279, "grad_norm": 101.81204223632812, "learning_rate": 2.670606033411678e-06, "loss": 2.1485, "step": 11320 }, { "epoch": 3.5489028213166143, "grad_norm": 8.00770092010498, "learning_rate": 2.6677669118403978e-06, "loss": 2.2293, "step": 11321 }, { "epoch": 3.5492163009404387, "grad_norm": 10.383228302001953, "learning_rate": 2.664927572905003e-06, "loss": 2.2112, "step": 11322 }, { "epoch": 3.549529780564263, "grad_norm": 11.358299255371094, "learning_rate": 2.6620880202842325e-06, "loss": 2.0361, "step": 11323 }, { "epoch": 3.549843260188088, "grad_norm": 60.09614944458008, "learning_rate": 2.6592482576571e-06, "loss": 2.2672, "step": 11324 }, { "epoch": 3.5501567398119125, "grad_norm": 10.911402702331543, "learning_rate": 2.65640828870289e-06, "loss": 2.2167, "step": 11325 }, { "epoch": 3.550470219435737, "grad_norm": 24.003320693969727, "learning_rate": 2.653568117101159e-06, "loss": 1.9568, "step": 11326 }, { "epoch": 3.5507836990595614, "grad_norm": 9.297477722167969, "learning_rate": 2.650727746531721e-06, "loss": 2.1705, "step": 11327 }, { "epoch": 3.551097178683386, "grad_norm": 39.57312774658203, "learning_rate": 2.6478871806746496e-06, "loss": 2.0682, "step": 11328 }, { "epoch": 3.5514106583072103, "grad_norm": 77.75819396972656, "learning_rate": 2.645046423210274e-06, "loss": 2.1825, "step": 11329 }, { "epoch": 3.5517241379310347, "grad_norm": 10.202300071716309, "learning_rate": 2.6422054778191674e-06, "loss": 2.2805, "step": 11330 }, { "epoch": 3.552037617554859, "grad_norm": 6.1370530128479, "learning_rate": 2.6393643481821484e-06, "loss": 2.1085, "step": 11331 }, { "epoch": 3.5523510971786836, "grad_norm": 6.16062593460083, "learning_rate": 2.636523037980275e-06, "loss": 2.0951, "step": 11332 }, { "epoch": 3.552664576802508, "grad_norm": 8.823986053466797, "learning_rate": 2.6336815508948387e-06, "loss": 1.9941, "step": 11333 }, { "epoch": 3.5529780564263325, "grad_norm": 5.1614155769348145, "learning_rate": 2.6308398906073603e-06, "loss": 1.9209, "step": 11334 }, { "epoch": 3.553291536050157, "grad_norm": 17.318532943725586, "learning_rate": 2.6279980607995837e-06, "loss": 2.1861, "step": 11335 }, { "epoch": 3.5536050156739813, "grad_norm": 26.256093978881836, "learning_rate": 2.625156065153473e-06, "loss": 2.6122, "step": 11336 }, { "epoch": 3.553918495297806, "grad_norm": 6.146217346191406, "learning_rate": 2.6223139073512087e-06, "loss": 2.1, "step": 11337 }, { "epoch": 3.5542319749216302, "grad_norm": 7.3534064292907715, "learning_rate": 2.6194715910751806e-06, "loss": 2.4837, "step": 11338 }, { "epoch": 3.5545454545454547, "grad_norm": 9.928845405578613, "learning_rate": 2.616629120007982e-06, "loss": 2.0748, "step": 11339 }, { "epoch": 3.554858934169279, "grad_norm": 5.978449821472168, "learning_rate": 2.6137864978324097e-06, "loss": 1.9598, "step": 11340 }, { "epoch": 3.5551724137931036, "grad_norm": 5.722709655761719, "learning_rate": 2.6109437282314535e-06, "loss": 2.0207, "step": 11341 }, { "epoch": 3.555485893416928, "grad_norm": 11.040054321289062, "learning_rate": 2.608100814888297e-06, "loss": 2.0746, "step": 11342 }, { "epoch": 3.5557993730407524, "grad_norm": 10.1940336227417, "learning_rate": 2.6052577614863077e-06, "loss": 1.978, "step": 11343 }, { "epoch": 3.556112852664577, "grad_norm": 4.017271995544434, "learning_rate": 2.602414571709036e-06, "loss": 2.0924, "step": 11344 }, { "epoch": 3.5564263322884013, "grad_norm": 8.851007461547852, "learning_rate": 2.5995712492402074e-06, "loss": 2.1488, "step": 11345 }, { "epoch": 3.5567398119122258, "grad_norm": 6.489303112030029, "learning_rate": 2.596727797763722e-06, "loss": 2.038, "step": 11346 }, { "epoch": 3.55705329153605, "grad_norm": 5.884336948394775, "learning_rate": 2.5938842209636438e-06, "loss": 2.0518, "step": 11347 }, { "epoch": 3.5573667711598747, "grad_norm": 8.861178398132324, "learning_rate": 2.591040522524202e-06, "loss": 2.0556, "step": 11348 }, { "epoch": 3.557680250783699, "grad_norm": 7.840882301330566, "learning_rate": 2.588196706129781e-06, "loss": 1.7638, "step": 11349 }, { "epoch": 3.5579937304075235, "grad_norm": 5.565367221832275, "learning_rate": 2.58535277546492e-06, "loss": 2.1887, "step": 11350 }, { "epoch": 3.558307210031348, "grad_norm": 8.32569408416748, "learning_rate": 2.5825087342143052e-06, "loss": 2.0645, "step": 11351 }, { "epoch": 3.5586206896551724, "grad_norm": 5.718750476837158, "learning_rate": 2.5796645860627665e-06, "loss": 2.081, "step": 11352 }, { "epoch": 3.558934169278997, "grad_norm": 7.714791297912598, "learning_rate": 2.576820334695273e-06, "loss": 2.0606, "step": 11353 }, { "epoch": 3.5592476489028213, "grad_norm": 8.946866989135742, "learning_rate": 2.5739759837969247e-06, "loss": 2.2403, "step": 11354 }, { "epoch": 3.5595611285266457, "grad_norm": 5.838191986083984, "learning_rate": 2.5711315370529537e-06, "loss": 2.1072, "step": 11355 }, { "epoch": 3.55987460815047, "grad_norm": 9.702314376831055, "learning_rate": 2.5682869981487154e-06, "loss": 2.0346, "step": 11356 }, { "epoch": 3.5601880877742946, "grad_norm": 6.4387526512146, "learning_rate": 2.5654423707696834e-06, "loss": 2.0189, "step": 11357 }, { "epoch": 3.560501567398119, "grad_norm": 14.80083179473877, "learning_rate": 2.562597658601447e-06, "loss": 2.2894, "step": 11358 }, { "epoch": 3.5608150470219435, "grad_norm": 7.054652690887451, "learning_rate": 2.5597528653297068e-06, "loss": 1.9877, "step": 11359 }, { "epoch": 3.561128526645768, "grad_norm": 28.741910934448242, "learning_rate": 2.556907994640264e-06, "loss": 2.2427, "step": 11360 }, { "epoch": 3.5614420062695924, "grad_norm": 6.6990885734558105, "learning_rate": 2.554063050219026e-06, "loss": 2.0917, "step": 11361 }, { "epoch": 3.561755485893417, "grad_norm": 10.130668640136719, "learning_rate": 2.5512180357519913e-06, "loss": 2.16, "step": 11362 }, { "epoch": 3.5620689655172413, "grad_norm": 6.595481872558594, "learning_rate": 2.5483729549252496e-06, "loss": 1.9888, "step": 11363 }, { "epoch": 3.5623824451410657, "grad_norm": 4.856290340423584, "learning_rate": 2.545527811424979e-06, "loss": 2.0912, "step": 11364 }, { "epoch": 3.56269592476489, "grad_norm": 7.8082661628723145, "learning_rate": 2.5426826089374374e-06, "loss": 2.1337, "step": 11365 }, { "epoch": 3.5630094043887146, "grad_norm": 4.415545463562012, "learning_rate": 2.539837351148958e-06, "loss": 2.332, "step": 11366 }, { "epoch": 3.563322884012539, "grad_norm": 9.791839599609375, "learning_rate": 2.536992041745947e-06, "loss": 1.9178, "step": 11367 }, { "epoch": 3.5636363636363635, "grad_norm": 36.28117752075195, "learning_rate": 2.5341466844148775e-06, "loss": 2.1547, "step": 11368 }, { "epoch": 3.563949843260188, "grad_norm": 7.272012233734131, "learning_rate": 2.531301282842285e-06, "loss": 1.9427, "step": 11369 }, { "epoch": 3.5642633228840124, "grad_norm": 5.187603950500488, "learning_rate": 2.5284558407147606e-06, "loss": 1.8336, "step": 11370 }, { "epoch": 3.564576802507837, "grad_norm": 5.636175155639648, "learning_rate": 2.5256103617189504e-06, "loss": 1.9866, "step": 11371 }, { "epoch": 3.5648902821316613, "grad_norm": 10.430776596069336, "learning_rate": 2.522764849541546e-06, "loss": 2.0755, "step": 11372 }, { "epoch": 3.5652037617554857, "grad_norm": 9.759641647338867, "learning_rate": 2.5199193078692824e-06, "loss": 1.9218, "step": 11373 }, { "epoch": 3.56551724137931, "grad_norm": 5.468115329742432, "learning_rate": 2.5170737403889334e-06, "loss": 2.0608, "step": 11374 }, { "epoch": 3.5658307210031346, "grad_norm": 9.022682189941406, "learning_rate": 2.5142281507873072e-06, "loss": 2.0349, "step": 11375 }, { "epoch": 3.566144200626959, "grad_norm": 26.72218132019043, "learning_rate": 2.511382542751239e-06, "loss": 2.0127, "step": 11376 }, { "epoch": 3.5664576802507835, "grad_norm": 6.210046291351318, "learning_rate": 2.508536919967589e-06, "loss": 2.1675, "step": 11377 }, { "epoch": 3.566771159874608, "grad_norm": 73.68148040771484, "learning_rate": 2.5056912861232356e-06, "loss": 2.3682, "step": 11378 }, { "epoch": 3.5670846394984324, "grad_norm": 11.815535545349121, "learning_rate": 2.5028456449050716e-06, "loss": 2.077, "step": 11379 }, { "epoch": 3.567398119122257, "grad_norm": 6.094146728515625, "learning_rate": 2.5e-06, "loss": 2.1895, "step": 11380 }, { "epoch": 3.5677115987460812, "grad_norm": 8.78769302368164, "learning_rate": 2.4971543550949292e-06, "loss": 2.0935, "step": 11381 }, { "epoch": 3.568025078369906, "grad_norm": 9.703916549682617, "learning_rate": 2.4943087138767656e-06, "loss": 2.0373, "step": 11382 }, { "epoch": 3.5683385579937306, "grad_norm": 7.823214530944824, "learning_rate": 2.4914630800324114e-06, "loss": 2.1369, "step": 11383 }, { "epoch": 3.568652037617555, "grad_norm": 13.259880065917969, "learning_rate": 2.488617457248761e-06, "loss": 2.0231, "step": 11384 }, { "epoch": 3.5689655172413794, "grad_norm": 6.548025131225586, "learning_rate": 2.485771849212693e-06, "loss": 2.1289, "step": 11385 }, { "epoch": 3.569278996865204, "grad_norm": 5.996400356292725, "learning_rate": 2.4829262596110674e-06, "loss": 2.1658, "step": 11386 }, { "epoch": 3.5695924764890283, "grad_norm": 6.920334339141846, "learning_rate": 2.4800806921307192e-06, "loss": 2.1263, "step": 11387 }, { "epoch": 3.5699059561128528, "grad_norm": 6.4435930252075195, "learning_rate": 2.4772351504584557e-06, "loss": 2.1135, "step": 11388 }, { "epoch": 3.570219435736677, "grad_norm": 5.440578937530518, "learning_rate": 2.4743896382810505e-06, "loss": 2.1195, "step": 11389 }, { "epoch": 3.5705329153605017, "grad_norm": 12.536064147949219, "learning_rate": 2.4715441592852398e-06, "loss": 2.0436, "step": 11390 }, { "epoch": 3.570846394984326, "grad_norm": 49.37378692626953, "learning_rate": 2.4686987171577155e-06, "loss": 2.2896, "step": 11391 }, { "epoch": 3.5711598746081505, "grad_norm": 21.68727684020996, "learning_rate": 2.465853315585123e-06, "loss": 3.0053, "step": 11392 }, { "epoch": 3.571473354231975, "grad_norm": 11.05690860748291, "learning_rate": 2.4630079582540542e-06, "loss": 2.0219, "step": 11393 }, { "epoch": 3.5717868338557994, "grad_norm": 24.286428451538086, "learning_rate": 2.4601626488510426e-06, "loss": 2.2427, "step": 11394 }, { "epoch": 3.572100313479624, "grad_norm": 11.52802848815918, "learning_rate": 2.4573173910625634e-06, "loss": 2.2583, "step": 11395 }, { "epoch": 3.5724137931034483, "grad_norm": 160.24774169921875, "learning_rate": 2.4544721885750217e-06, "loss": 2.4812, "step": 11396 }, { "epoch": 3.5727272727272728, "grad_norm": 7.380281448364258, "learning_rate": 2.451627045074751e-06, "loss": 2.2264, "step": 11397 }, { "epoch": 3.573040752351097, "grad_norm": 13.409160614013672, "learning_rate": 2.44878196424801e-06, "loss": 2.1865, "step": 11398 }, { "epoch": 3.5733542319749216, "grad_norm": 9.340126991271973, "learning_rate": 2.445936949780974e-06, "loss": 1.8651, "step": 11399 }, { "epoch": 3.573667711598746, "grad_norm": 5.682707786560059, "learning_rate": 2.443092005359736e-06, "loss": 2.1842, "step": 11400 }, { "epoch": 3.5739811912225705, "grad_norm": 37.51350784301758, "learning_rate": 2.440247134670294e-06, "loss": 2.1214, "step": 11401 }, { "epoch": 3.574294670846395, "grad_norm": 6.767791271209717, "learning_rate": 2.4374023413985532e-06, "loss": 2.1171, "step": 11402 }, { "epoch": 3.5746081504702194, "grad_norm": 4.187166213989258, "learning_rate": 2.434557629230318e-06, "loss": 2.112, "step": 11403 }, { "epoch": 3.574921630094044, "grad_norm": 19.817096710205078, "learning_rate": 2.431713001851286e-06, "loss": 2.0536, "step": 11404 }, { "epoch": 3.5752351097178683, "grad_norm": 6.55342960357666, "learning_rate": 2.4288684629470467e-06, "loss": 1.9664, "step": 11405 }, { "epoch": 3.5755485893416927, "grad_norm": 67.46814727783203, "learning_rate": 2.4260240162030757e-06, "loss": 2.1739, "step": 11406 }, { "epoch": 3.575862068965517, "grad_norm": 14.932708740234375, "learning_rate": 2.423179665304728e-06, "loss": 2.2223, "step": 11407 }, { "epoch": 3.5761755485893416, "grad_norm": 222.7375030517578, "learning_rate": 2.420335413937234e-06, "loss": 2.0937, "step": 11408 }, { "epoch": 3.576489028213166, "grad_norm": 6.960488796234131, "learning_rate": 2.417491265785695e-06, "loss": 2.138, "step": 11409 }, { "epoch": 3.5768025078369905, "grad_norm": 10.27044677734375, "learning_rate": 2.4146472245350804e-06, "loss": 2.2022, "step": 11410 }, { "epoch": 3.577115987460815, "grad_norm": 6.446353912353516, "learning_rate": 2.4118032938702195e-06, "loss": 1.9951, "step": 11411 }, { "epoch": 3.5774294670846394, "grad_norm": 10.347004890441895, "learning_rate": 2.408959477475799e-06, "loss": 1.9168, "step": 11412 }, { "epoch": 3.577742946708464, "grad_norm": 16.020021438598633, "learning_rate": 2.406115779036357e-06, "loss": 2.2801, "step": 11413 }, { "epoch": 3.5780564263322883, "grad_norm": 40.483489990234375, "learning_rate": 2.403272202236279e-06, "loss": 2.0946, "step": 11414 }, { "epoch": 3.5783699059561127, "grad_norm": 14.644736289978027, "learning_rate": 2.4004287507597935e-06, "loss": 2.2263, "step": 11415 }, { "epoch": 3.578683385579937, "grad_norm": 12.376179695129395, "learning_rate": 2.3975854282909645e-06, "loss": 2.1238, "step": 11416 }, { "epoch": 3.5789968652037616, "grad_norm": 8.800227165222168, "learning_rate": 2.3947422385136927e-06, "loss": 2.2004, "step": 11417 }, { "epoch": 3.5793103448275865, "grad_norm": 8.045197486877441, "learning_rate": 2.3918991851117036e-06, "loss": 2.2735, "step": 11418 }, { "epoch": 3.579623824451411, "grad_norm": 10.062227249145508, "learning_rate": 2.3890562717685473e-06, "loss": 1.9123, "step": 11419 }, { "epoch": 3.5799373040752354, "grad_norm": 16.313602447509766, "learning_rate": 2.3862135021675916e-06, "loss": 2.3234, "step": 11420 }, { "epoch": 3.58025078369906, "grad_norm": 14.958221435546875, "learning_rate": 2.383370879992018e-06, "loss": 2.1919, "step": 11421 }, { "epoch": 3.5805642633228842, "grad_norm": 17.250741958618164, "learning_rate": 2.3805284089248203e-06, "loss": 2.5871, "step": 11422 }, { "epoch": 3.5808777429467087, "grad_norm": 7.560488700866699, "learning_rate": 2.3776860926487917e-06, "loss": 2.0392, "step": 11423 }, { "epoch": 3.581191222570533, "grad_norm": 9.702984809875488, "learning_rate": 2.374843934846528e-06, "loss": 2.2733, "step": 11424 }, { "epoch": 3.5815047021943576, "grad_norm": 15.56987476348877, "learning_rate": 2.372001939200418e-06, "loss": 2.0927, "step": 11425 }, { "epoch": 3.581818181818182, "grad_norm": 7.438675880432129, "learning_rate": 2.3691601093926406e-06, "loss": 2.1643, "step": 11426 }, { "epoch": 3.5821316614420065, "grad_norm": 8.263379096984863, "learning_rate": 2.3663184491051617e-06, "loss": 1.9417, "step": 11427 }, { "epoch": 3.582445141065831, "grad_norm": 7.893310546875, "learning_rate": 2.3634769620197253e-06, "loss": 2.1408, "step": 11428 }, { "epoch": 3.5827586206896553, "grad_norm": 10.996240615844727, "learning_rate": 2.3606356518178524e-06, "loss": 2.119, "step": 11429 }, { "epoch": 3.58307210031348, "grad_norm": 7.955410957336426, "learning_rate": 2.3577945221808334e-06, "loss": 1.6572, "step": 11430 }, { "epoch": 3.583385579937304, "grad_norm": 7.34541654586792, "learning_rate": 2.354953576789727e-06, "loss": 1.92, "step": 11431 }, { "epoch": 3.5836990595611287, "grad_norm": 12.074618339538574, "learning_rate": 2.35211281932535e-06, "loss": 2.1691, "step": 11432 }, { "epoch": 3.584012539184953, "grad_norm": 6.624088287353516, "learning_rate": 2.3492722534682798e-06, "loss": 2.218, "step": 11433 }, { "epoch": 3.5843260188087775, "grad_norm": 7.03781795501709, "learning_rate": 2.3464318828988416e-06, "loss": 1.8448, "step": 11434 }, { "epoch": 3.584639498432602, "grad_norm": 8.769946098327637, "learning_rate": 2.3435917112971104e-06, "loss": 2.0611, "step": 11435 }, { "epoch": 3.5849529780564264, "grad_norm": 7.288632869720459, "learning_rate": 2.3407517423429016e-06, "loss": 2.1449, "step": 11436 }, { "epoch": 3.585266457680251, "grad_norm": 15.279923439025879, "learning_rate": 2.3379119797157674e-06, "loss": 3.0568, "step": 11437 }, { "epoch": 3.5855799373040753, "grad_norm": 10.475753784179688, "learning_rate": 2.335072427094997e-06, "loss": 1.7407, "step": 11438 }, { "epoch": 3.5858934169278998, "grad_norm": 7.617859363555908, "learning_rate": 2.3322330881596035e-06, "loss": 2.1112, "step": 11439 }, { "epoch": 3.586206896551724, "grad_norm": 6.927984714508057, "learning_rate": 2.3293939665883233e-06, "loss": 2.039, "step": 11440 }, { "epoch": 3.5865203761755486, "grad_norm": 5.035408020019531, "learning_rate": 2.326555066059612e-06, "loss": 2.1287, "step": 11441 }, { "epoch": 3.586833855799373, "grad_norm": 10.87588119506836, "learning_rate": 2.323716390251639e-06, "loss": 2.0546, "step": 11442 }, { "epoch": 3.5871473354231975, "grad_norm": 5.73219108581543, "learning_rate": 2.320877942842283e-06, "loss": 2.0922, "step": 11443 }, { "epoch": 3.587460815047022, "grad_norm": 5.34437894821167, "learning_rate": 2.3180397275091264e-06, "loss": 2.1125, "step": 11444 }, { "epoch": 3.5877742946708464, "grad_norm": 6.99710750579834, "learning_rate": 2.3152017479294505e-06, "loss": 2.0793, "step": 11445 }, { "epoch": 3.588087774294671, "grad_norm": 4.825108528137207, "learning_rate": 2.3123640077802305e-06, "loss": 2.1218, "step": 11446 }, { "epoch": 3.5884012539184953, "grad_norm": 7.8436665534973145, "learning_rate": 2.309526510738133e-06, "loss": 2.0224, "step": 11447 }, { "epoch": 3.5887147335423197, "grad_norm": 7.289486408233643, "learning_rate": 2.306689260479508e-06, "loss": 2.3405, "step": 11448 }, { "epoch": 3.589028213166144, "grad_norm": 6.176151752471924, "learning_rate": 2.3038522606803882e-06, "loss": 2.3286, "step": 11449 }, { "epoch": 3.5893416927899686, "grad_norm": 4.993758678436279, "learning_rate": 2.3010155150164787e-06, "loss": 2.0874, "step": 11450 }, { "epoch": 3.589655172413793, "grad_norm": 30.139158248901367, "learning_rate": 2.2981790271631577e-06, "loss": 2.3078, "step": 11451 }, { "epoch": 3.5899686520376175, "grad_norm": 5.149630069732666, "learning_rate": 2.2953428007954682e-06, "loss": 2.5379, "step": 11452 }, { "epoch": 3.590282131661442, "grad_norm": 9.684494018554688, "learning_rate": 2.292506839588113e-06, "loss": 2.1881, "step": 11453 }, { "epoch": 3.5905956112852664, "grad_norm": 7.063571929931641, "learning_rate": 2.289671147215457e-06, "loss": 2.0869, "step": 11454 }, { "epoch": 3.590909090909091, "grad_norm": 5.938957691192627, "learning_rate": 2.28683572735151e-06, "loss": 2.3778, "step": 11455 }, { "epoch": 3.5912225705329153, "grad_norm": 8.369854927062988, "learning_rate": 2.284000583669933e-06, "loss": 2.0947, "step": 11456 }, { "epoch": 3.5915360501567397, "grad_norm": 7.429238319396973, "learning_rate": 2.2811657198440265e-06, "loss": 2.1125, "step": 11457 }, { "epoch": 3.591849529780564, "grad_norm": 219.13316345214844, "learning_rate": 2.2783311395467304e-06, "loss": 2.3188, "step": 11458 }, { "epoch": 3.5921630094043886, "grad_norm": 6.726437091827393, "learning_rate": 2.275496846450615e-06, "loss": 1.9073, "step": 11459 }, { "epoch": 3.592476489028213, "grad_norm": 5.0409955978393555, "learning_rate": 2.272662844227883e-06, "loss": 2.0989, "step": 11460 }, { "epoch": 3.5927899686520375, "grad_norm": 5.161118030548096, "learning_rate": 2.269829136550355e-06, "loss": 2.0011, "step": 11461 }, { "epoch": 3.593103448275862, "grad_norm": 4.813777923583984, "learning_rate": 2.2669957270894733e-06, "loss": 2.0337, "step": 11462 }, { "epoch": 3.5934169278996864, "grad_norm": 5.5670166015625, "learning_rate": 2.2641626195162923e-06, "loss": 2.2346, "step": 11463 }, { "epoch": 3.593730407523511, "grad_norm": 10.863298416137695, "learning_rate": 2.261329817501475e-06, "loss": 2.056, "step": 11464 }, { "epoch": 3.5940438871473352, "grad_norm": 6.841751575469971, "learning_rate": 2.2584973247152906e-06, "loss": 1.8757, "step": 11465 }, { "epoch": 3.5943573667711597, "grad_norm": 18.133041381835938, "learning_rate": 2.255665144827606e-06, "loss": 2.2915, "step": 11466 }, { "epoch": 3.594670846394984, "grad_norm": 10.801653861999512, "learning_rate": 2.2528332815078816e-06, "loss": 1.899, "step": 11467 }, { "epoch": 3.5949843260188086, "grad_norm": 8.806497573852539, "learning_rate": 2.2500017384251705e-06, "loss": 2.1798, "step": 11468 }, { "epoch": 3.595297805642633, "grad_norm": 5.047201156616211, "learning_rate": 2.2471705192481092e-06, "loss": 2.3134, "step": 11469 }, { "epoch": 3.5956112852664575, "grad_norm": 12.538493156433105, "learning_rate": 2.2443396276449145e-06, "loss": 1.8421, "step": 11470 }, { "epoch": 3.595924764890282, "grad_norm": 25.76827049255371, "learning_rate": 2.2415090672833795e-06, "loss": 1.7987, "step": 11471 }, { "epoch": 3.5962382445141063, "grad_norm": 10.981163024902344, "learning_rate": 2.238678841830867e-06, "loss": 2.0433, "step": 11472 }, { "epoch": 3.596551724137931, "grad_norm": 7.071147441864014, "learning_rate": 2.2358489549543065e-06, "loss": 2.2629, "step": 11473 }, { "epoch": 3.5968652037617552, "grad_norm": 8.050207138061523, "learning_rate": 2.23301941032019e-06, "loss": 2.2669, "step": 11474 }, { "epoch": 3.5971786833855797, "grad_norm": 9.670609474182129, "learning_rate": 2.230190211594562e-06, "loss": 2.1855, "step": 11475 }, { "epoch": 3.5974921630094046, "grad_norm": 10.03724479675293, "learning_rate": 2.2273613624430256e-06, "loss": 1.6016, "step": 11476 }, { "epoch": 3.597805642633229, "grad_norm": 46.72837829589844, "learning_rate": 2.224532866530726e-06, "loss": 2.2845, "step": 11477 }, { "epoch": 3.5981191222570534, "grad_norm": 185.36083984375, "learning_rate": 2.221704727522351e-06, "loss": 2.7322, "step": 11478 }, { "epoch": 3.598432601880878, "grad_norm": 14.758145332336426, "learning_rate": 2.2188769490821272e-06, "loss": 1.9735, "step": 11479 }, { "epoch": 3.5987460815047023, "grad_norm": 117.49163818359375, "learning_rate": 2.2160495348738127e-06, "loss": 2.46, "step": 11480 }, { "epoch": 3.5990595611285268, "grad_norm": 32.57691192626953, "learning_rate": 2.213222488560697e-06, "loss": 1.9569, "step": 11481 }, { "epoch": 3.599373040752351, "grad_norm": 6.699265480041504, "learning_rate": 2.2103958138055897e-06, "loss": 2.1181, "step": 11482 }, { "epoch": 3.5996865203761756, "grad_norm": 8.136775970458984, "learning_rate": 2.2075695142708197e-06, "loss": 2.7382, "step": 11483 }, { "epoch": 3.6, "grad_norm": 7.226282119750977, "learning_rate": 2.20474359361823e-06, "loss": 2.2792, "step": 11484 }, { "epoch": 3.6003134796238245, "grad_norm": 5.95620584487915, "learning_rate": 2.201918055509173e-06, "loss": 2.0035, "step": 11485 }, { "epoch": 3.600626959247649, "grad_norm": 6.660520076751709, "learning_rate": 2.199092903604504e-06, "loss": 2.1824, "step": 11486 }, { "epoch": 3.6009404388714734, "grad_norm": 10.2006254196167, "learning_rate": 2.1962681415645813e-06, "loss": 2.2021, "step": 11487 }, { "epoch": 3.601253918495298, "grad_norm": 7.2167887687683105, "learning_rate": 2.1934437730492544e-06, "loss": 1.9338, "step": 11488 }, { "epoch": 3.6015673981191223, "grad_norm": 13.12779426574707, "learning_rate": 2.1906198017178646e-06, "loss": 2.3841, "step": 11489 }, { "epoch": 3.6018808777429467, "grad_norm": 7.888892650604248, "learning_rate": 2.1877962312292385e-06, "loss": 1.878, "step": 11490 }, { "epoch": 3.602194357366771, "grad_norm": 6.468718528747559, "learning_rate": 2.1849730652416825e-06, "loss": 2.1029, "step": 11491 }, { "epoch": 3.6025078369905956, "grad_norm": 9.453620910644531, "learning_rate": 2.1821503074129814e-06, "loss": 2.3225, "step": 11492 }, { "epoch": 3.60282131661442, "grad_norm": 5.000191688537598, "learning_rate": 2.179327961400388e-06, "loss": 1.9624, "step": 11493 }, { "epoch": 3.6031347962382445, "grad_norm": 8.068221092224121, "learning_rate": 2.1765060308606243e-06, "loss": 2.1989, "step": 11494 }, { "epoch": 3.603448275862069, "grad_norm": 12.782742500305176, "learning_rate": 2.173684519449872e-06, "loss": 1.7962, "step": 11495 }, { "epoch": 3.6037617554858934, "grad_norm": 35.254112243652344, "learning_rate": 2.1708634308237687e-06, "loss": 2.095, "step": 11496 }, { "epoch": 3.604075235109718, "grad_norm": 14.080343246459961, "learning_rate": 2.168042768637409e-06, "loss": 2.2697, "step": 11497 }, { "epoch": 3.6043887147335423, "grad_norm": 7.393898963928223, "learning_rate": 2.165222536545331e-06, "loss": 1.8314, "step": 11498 }, { "epoch": 3.6047021943573667, "grad_norm": 8.744771003723145, "learning_rate": 2.1624027382015157e-06, "loss": 2.0956, "step": 11499 }, { "epoch": 3.605015673981191, "grad_norm": 10.059080123901367, "learning_rate": 2.159583377259384e-06, "loss": 1.9923, "step": 11500 }, { "epoch": 3.6053291536050156, "grad_norm": 11.946632385253906, "learning_rate": 2.156764457371788e-06, "loss": 2.248, "step": 11501 }, { "epoch": 3.60564263322884, "grad_norm": 8.335786819458008, "learning_rate": 2.15394598219101e-06, "loss": 1.9286, "step": 11502 }, { "epoch": 3.6059561128526645, "grad_norm": 7.8888444900512695, "learning_rate": 2.1511279553687553e-06, "loss": 2.3373, "step": 11503 }, { "epoch": 3.606269592476489, "grad_norm": 5.340022563934326, "learning_rate": 2.1483103805561493e-06, "loss": 1.9347, "step": 11504 }, { "epoch": 3.6065830721003134, "grad_norm": 8.023561477661133, "learning_rate": 2.14549326140373e-06, "loss": 2.1525, "step": 11505 }, { "epoch": 3.606896551724138, "grad_norm": 16.372819900512695, "learning_rate": 2.142676601561447e-06, "loss": 2.081, "step": 11506 }, { "epoch": 3.6072100313479623, "grad_norm": 7.255988597869873, "learning_rate": 2.139860404678652e-06, "loss": 2.114, "step": 11507 }, { "epoch": 3.6075235109717867, "grad_norm": 7.121533393859863, "learning_rate": 2.137044674404101e-06, "loss": 2.2709, "step": 11508 }, { "epoch": 3.607836990595611, "grad_norm": 15.037862777709961, "learning_rate": 2.1342294143859417e-06, "loss": 2.0303, "step": 11509 }, { "epoch": 3.6081504702194356, "grad_norm": 6.914831638336182, "learning_rate": 2.1314146282717144e-06, "loss": 2.0052, "step": 11510 }, { "epoch": 3.6084639498432605, "grad_norm": 6.5555219650268555, "learning_rate": 2.128600319708345e-06, "loss": 2.1746, "step": 11511 }, { "epoch": 3.608777429467085, "grad_norm": 11.545136451721191, "learning_rate": 2.1257864923421405e-06, "loss": 2.1874, "step": 11512 }, { "epoch": 3.6090909090909093, "grad_norm": 7.069450855255127, "learning_rate": 2.1229731498187833e-06, "loss": 2.1048, "step": 11513 }, { "epoch": 3.609404388714734, "grad_norm": 8.055485725402832, "learning_rate": 2.120160295783331e-06, "loss": 2.2146, "step": 11514 }, { "epoch": 3.6097178683385582, "grad_norm": 11.921868324279785, "learning_rate": 2.117347933880206e-06, "loss": 1.754, "step": 11515 }, { "epoch": 3.6100313479623827, "grad_norm": 181.00352478027344, "learning_rate": 2.114536067753192e-06, "loss": 2.2671, "step": 11516 }, { "epoch": 3.610344827586207, "grad_norm": 5.72701358795166, "learning_rate": 2.111724701045433e-06, "loss": 1.8761, "step": 11517 }, { "epoch": 3.6106583072100316, "grad_norm": 5.233709335327148, "learning_rate": 2.1089138373994226e-06, "loss": 2.222, "step": 11518 }, { "epoch": 3.610971786833856, "grad_norm": 5.8093647956848145, "learning_rate": 2.106103480457007e-06, "loss": 2.0934, "step": 11519 }, { "epoch": 3.6112852664576804, "grad_norm": 24.724773406982422, "learning_rate": 2.1032936338593716e-06, "loss": 2.1791, "step": 11520 }, { "epoch": 3.611598746081505, "grad_norm": 14.597319602966309, "learning_rate": 2.1004843012470437e-06, "loss": 2.0683, "step": 11521 }, { "epoch": 3.6119122257053293, "grad_norm": 24.620615005493164, "learning_rate": 2.097675486259883e-06, "loss": 2.1527, "step": 11522 }, { "epoch": 3.6122257053291538, "grad_norm": 7.181385040283203, "learning_rate": 2.0948671925370782e-06, "loss": 2.0619, "step": 11523 }, { "epoch": 3.612539184952978, "grad_norm": 7.788346767425537, "learning_rate": 2.092059423717145e-06, "loss": 1.994, "step": 11524 }, { "epoch": 3.6128526645768027, "grad_norm": 8.456772804260254, "learning_rate": 2.089252183437916e-06, "loss": 2.1114, "step": 11525 }, { "epoch": 3.613166144200627, "grad_norm": 9.400517463684082, "learning_rate": 2.0864454753365414e-06, "loss": 2.1438, "step": 11526 }, { "epoch": 3.6134796238244515, "grad_norm": 6.079730987548828, "learning_rate": 2.08363930304948e-06, "loss": 1.905, "step": 11527 }, { "epoch": 3.613793103448276, "grad_norm": 11.112736701965332, "learning_rate": 2.080833670212498e-06, "loss": 2.4792, "step": 11528 }, { "epoch": 3.6141065830721004, "grad_norm": 10.673262596130371, "learning_rate": 2.0780285804606607e-06, "loss": 1.849, "step": 11529 }, { "epoch": 3.614420062695925, "grad_norm": 17.951900482177734, "learning_rate": 2.0752240374283334e-06, "loss": 2.3818, "step": 11530 }, { "epoch": 3.6147335423197493, "grad_norm": 9.670760154724121, "learning_rate": 2.0724200447491694e-06, "loss": 2.2815, "step": 11531 }, { "epoch": 3.6150470219435737, "grad_norm": 7.939383029937744, "learning_rate": 2.0696166060561104e-06, "loss": 2.2158, "step": 11532 }, { "epoch": 3.615360501567398, "grad_norm": 12.452754020690918, "learning_rate": 2.06681372498138e-06, "loss": 1.9761, "step": 11533 }, { "epoch": 3.6156739811912226, "grad_norm": 9.472728729248047, "learning_rate": 2.064011405156478e-06, "loss": 1.6042, "step": 11534 }, { "epoch": 3.615987460815047, "grad_norm": 10.215024948120117, "learning_rate": 2.0612096502121817e-06, "loss": 2.5567, "step": 11535 }, { "epoch": 3.6163009404388715, "grad_norm": 5.554502964019775, "learning_rate": 2.0584084637785316e-06, "loss": 1.9986, "step": 11536 }, { "epoch": 3.616614420062696, "grad_norm": 8.699304580688477, "learning_rate": 2.0556078494848335e-06, "loss": 2.2273, "step": 11537 }, { "epoch": 3.6169278996865204, "grad_norm": 62.734901428222656, "learning_rate": 2.052807810959652e-06, "loss": 1.9697, "step": 11538 }, { "epoch": 3.617241379310345, "grad_norm": 7.919193744659424, "learning_rate": 2.0500083518308057e-06, "loss": 2.1921, "step": 11539 }, { "epoch": 3.6175548589341693, "grad_norm": 8.559066772460938, "learning_rate": 2.0472094757253603e-06, "loss": 1.9436, "step": 11540 }, { "epoch": 3.6178683385579937, "grad_norm": 6.509829998016357, "learning_rate": 2.0444111862696313e-06, "loss": 2.0487, "step": 11541 }, { "epoch": 3.618181818181818, "grad_norm": 80.18795013427734, "learning_rate": 2.0416134870891697e-06, "loss": 2.5565, "step": 11542 }, { "epoch": 3.6184952978056426, "grad_norm": 7.359169006347656, "learning_rate": 2.038816381808762e-06, "loss": 1.8321, "step": 11543 }, { "epoch": 3.618808777429467, "grad_norm": 4.356690883636475, "learning_rate": 2.036019874052428e-06, "loss": 2.1353, "step": 11544 }, { "epoch": 3.6191222570532915, "grad_norm": 7.738428115844727, "learning_rate": 2.0332239674434096e-06, "loss": 2.018, "step": 11545 }, { "epoch": 3.619435736677116, "grad_norm": 13.654322624206543, "learning_rate": 2.030428665604173e-06, "loss": 2.2243, "step": 11546 }, { "epoch": 3.6197492163009404, "grad_norm": 25.214168548583984, "learning_rate": 2.0276339721564e-06, "loss": 1.8978, "step": 11547 }, { "epoch": 3.620062695924765, "grad_norm": 17.90543556213379, "learning_rate": 2.0248398907209827e-06, "loss": 2.5814, "step": 11548 }, { "epoch": 3.6203761755485893, "grad_norm": 7.338598728179932, "learning_rate": 2.0220464249180216e-06, "loss": 1.9968, "step": 11549 }, { "epoch": 3.6206896551724137, "grad_norm": 9.468785285949707, "learning_rate": 2.0192535783668185e-06, "loss": 1.9747, "step": 11550 }, { "epoch": 3.621003134796238, "grad_norm": 4.4217658042907715, "learning_rate": 2.016461354685876e-06, "loss": 2.0273, "step": 11551 }, { "epoch": 3.6213166144200626, "grad_norm": 8.932258605957031, "learning_rate": 2.0136697574928853e-06, "loss": 2.0588, "step": 11552 }, { "epoch": 3.621630094043887, "grad_norm": 5.496156692504883, "learning_rate": 2.010878790404729e-06, "loss": 1.8538, "step": 11553 }, { "epoch": 3.6219435736677115, "grad_norm": 8.168444633483887, "learning_rate": 2.008088457037472e-06, "loss": 2.0138, "step": 11554 }, { "epoch": 3.622257053291536, "grad_norm": 9.743471145629883, "learning_rate": 2.005298761006358e-06, "loss": 2.0585, "step": 11555 }, { "epoch": 3.6225705329153604, "grad_norm": 9.404460906982422, "learning_rate": 2.0025097059258047e-06, "loss": 1.9657, "step": 11556 }, { "epoch": 3.622884012539185, "grad_norm": 9.091063499450684, "learning_rate": 1.9997212954094023e-06, "loss": 2.2233, "step": 11557 }, { "epoch": 3.6231974921630092, "grad_norm": 11.956096649169922, "learning_rate": 1.9969335330699017e-06, "loss": 2.37, "step": 11558 }, { "epoch": 3.6235109717868337, "grad_norm": 11.498491287231445, "learning_rate": 1.994146422519217e-06, "loss": 2.1698, "step": 11559 }, { "epoch": 3.623824451410658, "grad_norm": 7.5298309326171875, "learning_rate": 1.991359967368416e-06, "loss": 2.2464, "step": 11560 }, { "epoch": 3.6241379310344826, "grad_norm": 7.275935173034668, "learning_rate": 1.988574171227718e-06, "loss": 2.2367, "step": 11561 }, { "epoch": 3.624451410658307, "grad_norm": 21.69171905517578, "learning_rate": 1.985789037706489e-06, "loss": 2.0912, "step": 11562 }, { "epoch": 3.6247648902821314, "grad_norm": 10.1316499710083, "learning_rate": 1.9830045704132358e-06, "loss": 2.2498, "step": 11563 }, { "epoch": 3.625078369905956, "grad_norm": 8.579560279846191, "learning_rate": 1.9802207729556023e-06, "loss": 2.0326, "step": 11564 }, { "epoch": 3.6253918495297803, "grad_norm": 8.67944049835205, "learning_rate": 1.9774376489403646e-06, "loss": 2.0028, "step": 11565 }, { "epoch": 3.6257053291536048, "grad_norm": 12.613636016845703, "learning_rate": 1.9746552019734246e-06, "loss": 2.0085, "step": 11566 }, { "epoch": 3.626018808777429, "grad_norm": 7.667855262756348, "learning_rate": 1.971873435659811e-06, "loss": 1.9712, "step": 11567 }, { "epoch": 3.6263322884012537, "grad_norm": 6.6727190017700195, "learning_rate": 1.9690923536036673e-06, "loss": 2.1172, "step": 11568 }, { "epoch": 3.626645768025078, "grad_norm": 22.403911590576172, "learning_rate": 1.9663119594082512e-06, "loss": 2.3045, "step": 11569 }, { "epoch": 3.626959247648903, "grad_norm": 12.713897705078125, "learning_rate": 1.963532256675929e-06, "loss": 1.8989, "step": 11570 }, { "epoch": 3.6272727272727274, "grad_norm": 7.370038986206055, "learning_rate": 1.9607532490081714e-06, "loss": 2.3136, "step": 11571 }, { "epoch": 3.627586206896552, "grad_norm": 13.558568954467773, "learning_rate": 1.957974940005548e-06, "loss": 2.3994, "step": 11572 }, { "epoch": 3.6278996865203763, "grad_norm": 38.79823684692383, "learning_rate": 1.955197333267726e-06, "loss": 2.5457, "step": 11573 }, { "epoch": 3.6282131661442008, "grad_norm": 12.509878158569336, "learning_rate": 1.952420432393458e-06, "loss": 2.4358, "step": 11574 }, { "epoch": 3.628526645768025, "grad_norm": 6.929110527038574, "learning_rate": 1.9496442409805867e-06, "loss": 1.8663, "step": 11575 }, { "epoch": 3.6288401253918496, "grad_norm": 41.035457611083984, "learning_rate": 1.9468687626260314e-06, "loss": 2.3242, "step": 11576 }, { "epoch": 3.629153605015674, "grad_norm": 73.693603515625, "learning_rate": 1.9440940009257896e-06, "loss": 2.0783, "step": 11577 }, { "epoch": 3.6294670846394985, "grad_norm": 13.503495216369629, "learning_rate": 1.9413199594749327e-06, "loss": 2.0865, "step": 11578 }, { "epoch": 3.629780564263323, "grad_norm": 5.166501045227051, "learning_rate": 1.9385466418675945e-06, "loss": 2.1312, "step": 11579 }, { "epoch": 3.6300940438871474, "grad_norm": 4.969310283660889, "learning_rate": 1.9357740516969736e-06, "loss": 2.3799, "step": 11580 }, { "epoch": 3.630407523510972, "grad_norm": 38.956607818603516, "learning_rate": 1.9330021925553253e-06, "loss": 2.8809, "step": 11581 }, { "epoch": 3.6307210031347963, "grad_norm": 21.057231903076172, "learning_rate": 1.9302310680339573e-06, "loss": 1.8827, "step": 11582 }, { "epoch": 3.6310344827586207, "grad_norm": 7.386157512664795, "learning_rate": 1.927460681723227e-06, "loss": 1.839, "step": 11583 }, { "epoch": 3.631347962382445, "grad_norm": 6.302616119384766, "learning_rate": 1.9246910372125345e-06, "loss": 2.3603, "step": 11584 }, { "epoch": 3.6316614420062696, "grad_norm": 7.82151985168457, "learning_rate": 1.921922138090318e-06, "loss": 1.9822, "step": 11585 }, { "epoch": 3.631974921630094, "grad_norm": 8.439537048339844, "learning_rate": 1.9191539879440515e-06, "loss": 2.2721, "step": 11586 }, { "epoch": 3.6322884012539185, "grad_norm": 5.153085231781006, "learning_rate": 1.9163865903602374e-06, "loss": 1.7537, "step": 11587 }, { "epoch": 3.632601880877743, "grad_norm": 7.642799377441406, "learning_rate": 1.9136199489244023e-06, "loss": 2.1697, "step": 11588 }, { "epoch": 3.6329153605015674, "grad_norm": 6.38833475112915, "learning_rate": 1.910854067221097e-06, "loss": 2.2607, "step": 11589 }, { "epoch": 3.633228840125392, "grad_norm": 7.445428848266602, "learning_rate": 1.9080889488338833e-06, "loss": 2.0384, "step": 11590 }, { "epoch": 3.6335423197492163, "grad_norm": 77.0081558227539, "learning_rate": 1.9053245973453368e-06, "loss": 2.2948, "step": 11591 }, { "epoch": 3.6338557993730407, "grad_norm": 5.65863037109375, "learning_rate": 1.9025610163370385e-06, "loss": 2.0959, "step": 11592 }, { "epoch": 3.634169278996865, "grad_norm": 5.821320533752441, "learning_rate": 1.8997982093895701e-06, "loss": 2.2153, "step": 11593 }, { "epoch": 3.6344827586206896, "grad_norm": 6.529261589050293, "learning_rate": 1.8970361800825132e-06, "loss": 1.9161, "step": 11594 }, { "epoch": 3.634796238244514, "grad_norm": 9.945510864257812, "learning_rate": 1.8942749319944404e-06, "loss": 1.9981, "step": 11595 }, { "epoch": 3.6351097178683385, "grad_norm": 31.246259689331055, "learning_rate": 1.8915144687029107e-06, "loss": 2.059, "step": 11596 }, { "epoch": 3.635423197492163, "grad_norm": 181.88890075683594, "learning_rate": 1.8887547937844686e-06, "loss": 2.0355, "step": 11597 }, { "epoch": 3.6357366771159874, "grad_norm": 14.962053298950195, "learning_rate": 1.8859959108146359e-06, "loss": 2.9362, "step": 11598 }, { "epoch": 3.636050156739812, "grad_norm": 8.645916938781738, "learning_rate": 1.8832378233679078e-06, "loss": 2.0397, "step": 11599 }, { "epoch": 3.6363636363636362, "grad_norm": 9.704648971557617, "learning_rate": 1.8804805350177507e-06, "loss": 2.3762, "step": 11600 }, { "epoch": 3.6366771159874607, "grad_norm": 38.042903900146484, "learning_rate": 1.8777240493365939e-06, "loss": 2.1962, "step": 11601 }, { "epoch": 3.636990595611285, "grad_norm": 26.133310317993164, "learning_rate": 1.874968369895828e-06, "loss": 2.553, "step": 11602 }, { "epoch": 3.6373040752351096, "grad_norm": 37.567840576171875, "learning_rate": 1.8722135002657971e-06, "loss": 2.2558, "step": 11603 }, { "epoch": 3.637617554858934, "grad_norm": 8.035416603088379, "learning_rate": 1.8694594440157981e-06, "loss": 2.2243, "step": 11604 }, { "epoch": 3.637931034482759, "grad_norm": 17.579418182373047, "learning_rate": 1.866706204714074e-06, "loss": 2.6775, "step": 11605 }, { "epoch": 3.6382445141065833, "grad_norm": 9.68446159362793, "learning_rate": 1.8639537859278084e-06, "loss": 2.6876, "step": 11606 }, { "epoch": 3.638557993730408, "grad_norm": 7.799139022827148, "learning_rate": 1.8612021912231215e-06, "loss": 2.1402, "step": 11607 }, { "epoch": 3.6388714733542322, "grad_norm": 7.3362717628479, "learning_rate": 1.8584514241650667e-06, "loss": 2.0991, "step": 11608 }, { "epoch": 3.6391849529780567, "grad_norm": 211.7080078125, "learning_rate": 1.8557014883176244e-06, "loss": 2.2165, "step": 11609 }, { "epoch": 3.639498432601881, "grad_norm": 10.073141098022461, "learning_rate": 1.852952387243698e-06, "loss": 1.9793, "step": 11610 }, { "epoch": 3.6398119122257055, "grad_norm": 5.143453121185303, "learning_rate": 1.8502041245051114e-06, "loss": 2.016, "step": 11611 }, { "epoch": 3.64012539184953, "grad_norm": 5.499497413635254, "learning_rate": 1.8474567036625996e-06, "loss": 2.1273, "step": 11612 }, { "epoch": 3.6404388714733544, "grad_norm": 14.047724723815918, "learning_rate": 1.8447101282758081e-06, "loss": 2.1642, "step": 11613 }, { "epoch": 3.640752351097179, "grad_norm": 5.549081802368164, "learning_rate": 1.8419644019032868e-06, "loss": 2.0693, "step": 11614 }, { "epoch": 3.6410658307210033, "grad_norm": 23.11883544921875, "learning_rate": 1.839219528102484e-06, "loss": 2.1899, "step": 11615 }, { "epoch": 3.6413793103448278, "grad_norm": 7.786057949066162, "learning_rate": 1.8364755104297477e-06, "loss": 2.1467, "step": 11616 }, { "epoch": 3.641692789968652, "grad_norm": 9.604482650756836, "learning_rate": 1.8337323524403129e-06, "loss": 1.9266, "step": 11617 }, { "epoch": 3.6420062695924766, "grad_norm": 8.403105735778809, "learning_rate": 1.830990057688301e-06, "loss": 2.2699, "step": 11618 }, { "epoch": 3.642319749216301, "grad_norm": 8.628679275512695, "learning_rate": 1.8282486297267166e-06, "loss": 2.2615, "step": 11619 }, { "epoch": 3.6426332288401255, "grad_norm": 9.474409103393555, "learning_rate": 1.8255080721074391e-06, "loss": 1.802, "step": 11620 }, { "epoch": 3.64294670846395, "grad_norm": 11.634696960449219, "learning_rate": 1.8227683883812235e-06, "loss": 2.1411, "step": 11621 }, { "epoch": 3.6432601880877744, "grad_norm": 4.764641284942627, "learning_rate": 1.8200295820976899e-06, "loss": 1.892, "step": 11622 }, { "epoch": 3.643573667711599, "grad_norm": 7.15770959854126, "learning_rate": 1.817291656805322e-06, "loss": 1.9543, "step": 11623 }, { "epoch": 3.6438871473354233, "grad_norm": 8.03456974029541, "learning_rate": 1.8145546160514622e-06, "loss": 2.1107, "step": 11624 }, { "epoch": 3.6442006269592477, "grad_norm": 6.211708068847656, "learning_rate": 1.8118184633823074e-06, "loss": 2.0309, "step": 11625 }, { "epoch": 3.644514106583072, "grad_norm": 97.80658721923828, "learning_rate": 1.8090832023429022e-06, "loss": 2.1978, "step": 11626 }, { "epoch": 3.6448275862068966, "grad_norm": 13.504722595214844, "learning_rate": 1.8063488364771397e-06, "loss": 2.1552, "step": 11627 }, { "epoch": 3.645141065830721, "grad_norm": 12.71108627319336, "learning_rate": 1.8036153693277492e-06, "loss": 1.9503, "step": 11628 }, { "epoch": 3.6454545454545455, "grad_norm": 8.857804298400879, "learning_rate": 1.8008828044362974e-06, "loss": 2.2143, "step": 11629 }, { "epoch": 3.64576802507837, "grad_norm": 64.8749771118164, "learning_rate": 1.7981511453431817e-06, "loss": 2.0784, "step": 11630 }, { "epoch": 3.6460815047021944, "grad_norm": 7.438033103942871, "learning_rate": 1.7954203955876248e-06, "loss": 2.0539, "step": 11631 }, { "epoch": 3.646394984326019, "grad_norm": 6.119851112365723, "learning_rate": 1.792690558707675e-06, "loss": 2.0479, "step": 11632 }, { "epoch": 3.6467084639498433, "grad_norm": 6.727479457855225, "learning_rate": 1.7899616382401935e-06, "loss": 2.1778, "step": 11633 }, { "epoch": 3.6470219435736677, "grad_norm": 7.472078323364258, "learning_rate": 1.7872336377208565e-06, "loss": 2.1575, "step": 11634 }, { "epoch": 3.647335423197492, "grad_norm": 6.349815845489502, "learning_rate": 1.7845065606841471e-06, "loss": 2.1054, "step": 11635 }, { "epoch": 3.6476489028213166, "grad_norm": 8.153975486755371, "learning_rate": 1.7817804106633531e-06, "loss": 2.1381, "step": 11636 }, { "epoch": 3.647962382445141, "grad_norm": 18.622947692871094, "learning_rate": 1.7790551911905598e-06, "loss": 2.2291, "step": 11637 }, { "epoch": 3.6482758620689655, "grad_norm": 10.861016273498535, "learning_rate": 1.7763309057966487e-06, "loss": 2.1563, "step": 11638 }, { "epoch": 3.64858934169279, "grad_norm": 5.634410858154297, "learning_rate": 1.7736075580112894e-06, "loss": 2.0826, "step": 11639 }, { "epoch": 3.6489028213166144, "grad_norm": 12.840315818786621, "learning_rate": 1.7708851513629376e-06, "loss": 1.7845, "step": 11640 }, { "epoch": 3.649216300940439, "grad_norm": 6.831881523132324, "learning_rate": 1.7681636893788302e-06, "loss": 2.1144, "step": 11641 }, { "epoch": 3.6495297805642632, "grad_norm": 8.859375, "learning_rate": 1.7654431755849771e-06, "loss": 2.1295, "step": 11642 }, { "epoch": 3.6498432601880877, "grad_norm": 6.944246292114258, "learning_rate": 1.7627236135061645e-06, "loss": 2.4352, "step": 11643 }, { "epoch": 3.650156739811912, "grad_norm": 28.120595932006836, "learning_rate": 1.7600050066659418e-06, "loss": 2.2125, "step": 11644 }, { "epoch": 3.6504702194357366, "grad_norm": 5.566560745239258, "learning_rate": 1.7572873585866224e-06, "loss": 2.0414, "step": 11645 }, { "epoch": 3.650783699059561, "grad_norm": 12.085586547851562, "learning_rate": 1.7545706727892763e-06, "loss": 2.1475, "step": 11646 }, { "epoch": 3.6510971786833855, "grad_norm": 28.920869827270508, "learning_rate": 1.7518549527937268e-06, "loss": 2.3151, "step": 11647 }, { "epoch": 3.65141065830721, "grad_norm": 19.62807273864746, "learning_rate": 1.7491402021185489e-06, "loss": 1.9256, "step": 11648 }, { "epoch": 3.6517241379310343, "grad_norm": 46.08668899536133, "learning_rate": 1.746426424281058e-06, "loss": 1.9962, "step": 11649 }, { "epoch": 3.652037617554859, "grad_norm": 9.752584457397461, "learning_rate": 1.7437136227973108e-06, "loss": 2.5038, "step": 11650 }, { "epoch": 3.6523510971786832, "grad_norm": 9.515190124511719, "learning_rate": 1.7410018011820983e-06, "loss": 2.0274, "step": 11651 }, { "epoch": 3.6526645768025077, "grad_norm": 12.174535751342773, "learning_rate": 1.7382909629489424e-06, "loss": 2.1838, "step": 11652 }, { "epoch": 3.652978056426332, "grad_norm": 18.119808197021484, "learning_rate": 1.73558111161009e-06, "loss": 2.1419, "step": 11653 }, { "epoch": 3.6532915360501566, "grad_norm": 7.058817386627197, "learning_rate": 1.7328722506765122e-06, "loss": 2.3553, "step": 11654 }, { "epoch": 3.653605015673981, "grad_norm": 52.2440071105957, "learning_rate": 1.7301643836578933e-06, "loss": 2.5367, "step": 11655 }, { "epoch": 3.6539184952978054, "grad_norm": 10.301461219787598, "learning_rate": 1.7274575140626318e-06, "loss": 2.1568, "step": 11656 }, { "epoch": 3.65423197492163, "grad_norm": 8.351689338684082, "learning_rate": 1.7247516453978336e-06, "loss": 2.0751, "step": 11657 }, { "epoch": 3.6545454545454543, "grad_norm": 7.104780673980713, "learning_rate": 1.7220467811693064e-06, "loss": 1.7757, "step": 11658 }, { "epoch": 3.6548589341692788, "grad_norm": 16.77413558959961, "learning_rate": 1.7193429248815602e-06, "loss": 2.1466, "step": 11659 }, { "epoch": 3.655172413793103, "grad_norm": 9.08117961883545, "learning_rate": 1.7166400800377947e-06, "loss": 2.3186, "step": 11660 }, { "epoch": 3.6554858934169276, "grad_norm": 6.697193622589111, "learning_rate": 1.7139382501399022e-06, "loss": 2.0031, "step": 11661 }, { "epoch": 3.655799373040752, "grad_norm": 6.478462219238281, "learning_rate": 1.7112374386884583e-06, "loss": 2.0971, "step": 11662 }, { "epoch": 3.6561128526645765, "grad_norm": 8.004417419433594, "learning_rate": 1.70853764918272e-06, "loss": 1.937, "step": 11663 }, { "epoch": 3.6564263322884014, "grad_norm": 137.7964324951172, "learning_rate": 1.7058388851206187e-06, "loss": 2.4803, "step": 11664 }, { "epoch": 3.656739811912226, "grad_norm": 28.591154098510742, "learning_rate": 1.7031411499987604e-06, "loss": 1.996, "step": 11665 }, { "epoch": 3.6570532915360503, "grad_norm": 7.494840621948242, "learning_rate": 1.7004444473124154e-06, "loss": 1.9261, "step": 11666 }, { "epoch": 3.6573667711598747, "grad_norm": 8.256072044372559, "learning_rate": 1.6977487805555166e-06, "loss": 1.931, "step": 11667 }, { "epoch": 3.657680250783699, "grad_norm": 9.120476722717285, "learning_rate": 1.695054153220655e-06, "loss": 1.9462, "step": 11668 }, { "epoch": 3.6579937304075236, "grad_norm": 5.052282810211182, "learning_rate": 1.6923605687990741e-06, "loss": 1.9897, "step": 11669 }, { "epoch": 3.658307210031348, "grad_norm": 5.830216407775879, "learning_rate": 1.6896680307806687e-06, "loss": 1.8817, "step": 11670 }, { "epoch": 3.6586206896551725, "grad_norm": 5.462981700897217, "learning_rate": 1.6869765426539759e-06, "loss": 2.1543, "step": 11671 }, { "epoch": 3.658934169278997, "grad_norm": 6.617360591888428, "learning_rate": 1.6842861079061717e-06, "loss": 2.22, "step": 11672 }, { "epoch": 3.6592476489028214, "grad_norm": 5.443572521209717, "learning_rate": 1.6815967300230695e-06, "loss": 2.0429, "step": 11673 }, { "epoch": 3.659561128526646, "grad_norm": 7.331566333770752, "learning_rate": 1.678908412489111e-06, "loss": 2.1688, "step": 11674 }, { "epoch": 3.6598746081504703, "grad_norm": 14.696879386901855, "learning_rate": 1.6762211587873667e-06, "loss": 2.3682, "step": 11675 }, { "epoch": 3.6601880877742947, "grad_norm": 5.999813079833984, "learning_rate": 1.6735349723995263e-06, "loss": 2.0704, "step": 11676 }, { "epoch": 3.660501567398119, "grad_norm": 5.223268032073975, "learning_rate": 1.6708498568058995e-06, "loss": 2.1581, "step": 11677 }, { "epoch": 3.6608150470219436, "grad_norm": 14.84890079498291, "learning_rate": 1.6681658154854053e-06, "loss": 2.194, "step": 11678 }, { "epoch": 3.661128526645768, "grad_norm": 6.616195201873779, "learning_rate": 1.665482851915573e-06, "loss": 1.9965, "step": 11679 }, { "epoch": 3.6614420062695925, "grad_norm": 10.937299728393555, "learning_rate": 1.6628009695725348e-06, "loss": 2.0508, "step": 11680 }, { "epoch": 3.661755485893417, "grad_norm": 62.7127799987793, "learning_rate": 1.6601201719310228e-06, "loss": 2.5188, "step": 11681 }, { "epoch": 3.6620689655172414, "grad_norm": 6.545909404754639, "learning_rate": 1.6574404624643626e-06, "loss": 2.2083, "step": 11682 }, { "epoch": 3.662382445141066, "grad_norm": 7.34820556640625, "learning_rate": 1.6547618446444708e-06, "loss": 2.0596, "step": 11683 }, { "epoch": 3.6626959247648903, "grad_norm": 7.670024871826172, "learning_rate": 1.6520843219418492e-06, "loss": 1.9369, "step": 11684 }, { "epoch": 3.6630094043887147, "grad_norm": 9.41492748260498, "learning_rate": 1.6494078978255797e-06, "loss": 2.152, "step": 11685 }, { "epoch": 3.663322884012539, "grad_norm": 15.69944953918457, "learning_rate": 1.6467325757633242e-06, "loss": 1.7932, "step": 11686 }, { "epoch": 3.6636363636363636, "grad_norm": 5.934609889984131, "learning_rate": 1.6440583592213135e-06, "loss": 2.1927, "step": 11687 }, { "epoch": 3.663949843260188, "grad_norm": 21.97848129272461, "learning_rate": 1.6413852516643468e-06, "loss": 2.2826, "step": 11688 }, { "epoch": 3.6642633228840125, "grad_norm": 250.330810546875, "learning_rate": 1.6387132565557873e-06, "loss": 2.8469, "step": 11689 }, { "epoch": 3.664576802507837, "grad_norm": 5.819884300231934, "learning_rate": 1.6360423773575548e-06, "loss": 1.9373, "step": 11690 }, { "epoch": 3.6648902821316613, "grad_norm": 7.263125896453857, "learning_rate": 1.633372617530127e-06, "loss": 2.0426, "step": 11691 }, { "epoch": 3.665203761755486, "grad_norm": 6.426148891448975, "learning_rate": 1.630703980532528e-06, "loss": 2.2287, "step": 11692 }, { "epoch": 3.6655172413793102, "grad_norm": 10.547592163085938, "learning_rate": 1.6280364698223286e-06, "loss": 2.0351, "step": 11693 }, { "epoch": 3.6658307210031347, "grad_norm": 7.237731456756592, "learning_rate": 1.6253700888556395e-06, "loss": 1.9902, "step": 11694 }, { "epoch": 3.666144200626959, "grad_norm": 6.188724994659424, "learning_rate": 1.6227048410871085e-06, "loss": 2.0927, "step": 11695 }, { "epoch": 3.6664576802507836, "grad_norm": 16.31299591064453, "learning_rate": 1.6200407299699141e-06, "loss": 1.9516, "step": 11696 }, { "epoch": 3.666771159874608, "grad_norm": 14.734925270080566, "learning_rate": 1.6173777589557644e-06, "loss": 2.0276, "step": 11697 }, { "epoch": 3.6670846394984324, "grad_norm": 43.48994064331055, "learning_rate": 1.6147159314948873e-06, "loss": 2.388, "step": 11698 }, { "epoch": 3.6673981191222573, "grad_norm": 10.911104202270508, "learning_rate": 1.6120552510360315e-06, "loss": 2.1792, "step": 11699 }, { "epoch": 3.6677115987460818, "grad_norm": 5.053783416748047, "learning_rate": 1.6093957210264588e-06, "loss": 2.1398, "step": 11700 }, { "epoch": 3.668025078369906, "grad_norm": 18.647415161132812, "learning_rate": 1.6067373449119387e-06, "loss": 2.3261, "step": 11701 }, { "epoch": 3.6683385579937307, "grad_norm": 5.8810272216796875, "learning_rate": 1.6040801261367494e-06, "loss": 2.2722, "step": 11702 }, { "epoch": 3.668652037617555, "grad_norm": 7.026515483856201, "learning_rate": 1.601424068143667e-06, "loss": 2.1877, "step": 11703 }, { "epoch": 3.6689655172413795, "grad_norm": 10.661813735961914, "learning_rate": 1.5987691743739636e-06, "loss": 2.284, "step": 11704 }, { "epoch": 3.669278996865204, "grad_norm": 4.740579128265381, "learning_rate": 1.596115448267404e-06, "loss": 2.1293, "step": 11705 }, { "epoch": 3.6695924764890284, "grad_norm": 9.776100158691406, "learning_rate": 1.5934628932622395e-06, "loss": 1.9303, "step": 11706 }, { "epoch": 3.669905956112853, "grad_norm": 7.999732971191406, "learning_rate": 1.590811512795203e-06, "loss": 2.3003, "step": 11707 }, { "epoch": 3.6702194357366773, "grad_norm": 39.85436248779297, "learning_rate": 1.588161310301509e-06, "loss": 2.109, "step": 11708 }, { "epoch": 3.6705329153605017, "grad_norm": 9.471391677856445, "learning_rate": 1.5855122892148426e-06, "loss": 1.9515, "step": 11709 }, { "epoch": 3.670846394984326, "grad_norm": 23.845766067504883, "learning_rate": 1.5828644529673592e-06, "loss": 2.0153, "step": 11710 }, { "epoch": 3.6711598746081506, "grad_norm": 7.57498025894165, "learning_rate": 1.5802178049896791e-06, "loss": 2.109, "step": 11711 }, { "epoch": 3.671473354231975, "grad_norm": 11.625082969665527, "learning_rate": 1.5775723487108821e-06, "loss": 2.1269, "step": 11712 }, { "epoch": 3.6717868338557995, "grad_norm": 6.491303443908691, "learning_rate": 1.5749280875585071e-06, "loss": 1.9525, "step": 11713 }, { "epoch": 3.672100313479624, "grad_norm": 7.155837059020996, "learning_rate": 1.5722850249585415e-06, "loss": 2.2901, "step": 11714 }, { "epoch": 3.6724137931034484, "grad_norm": 12.836321830749512, "learning_rate": 1.5696431643354204e-06, "loss": 2.3517, "step": 11715 }, { "epoch": 3.672727272727273, "grad_norm": 39.042579650878906, "learning_rate": 1.5670025091120219e-06, "loss": 2.2983, "step": 11716 }, { "epoch": 3.6730407523510973, "grad_norm": 37.21473693847656, "learning_rate": 1.5643630627096623e-06, "loss": 1.9752, "step": 11717 }, { "epoch": 3.6733542319749217, "grad_norm": 5.630346298217773, "learning_rate": 1.5617248285480924e-06, "loss": 2.1339, "step": 11718 }, { "epoch": 3.673667711598746, "grad_norm": 199.8052978515625, "learning_rate": 1.5590878100454917e-06, "loss": 2.3737, "step": 11719 }, { "epoch": 3.6739811912225706, "grad_norm": 11.058389663696289, "learning_rate": 1.5564520106184643e-06, "loss": 2.1772, "step": 11720 }, { "epoch": 3.674294670846395, "grad_norm": 12.803619384765625, "learning_rate": 1.5538174336820355e-06, "loss": 2.1849, "step": 11721 }, { "epoch": 3.6746081504702195, "grad_norm": 11.396381378173828, "learning_rate": 1.5511840826496462e-06, "loss": 2.1207, "step": 11722 }, { "epoch": 3.674921630094044, "grad_norm": 5.611574649810791, "learning_rate": 1.5485519609331483e-06, "loss": 2.2174, "step": 11723 }, { "epoch": 3.6752351097178684, "grad_norm": 17.215118408203125, "learning_rate": 1.5459210719428044e-06, "loss": 1.857, "step": 11724 }, { "epoch": 3.675548589341693, "grad_norm": 4.186439514160156, "learning_rate": 1.5432914190872757e-06, "loss": 1.9571, "step": 11725 }, { "epoch": 3.6758620689655173, "grad_norm": 14.145862579345703, "learning_rate": 1.540663005773624e-06, "loss": 2.3395, "step": 11726 }, { "epoch": 3.6761755485893417, "grad_norm": 16.338916778564453, "learning_rate": 1.5380358354073047e-06, "loss": 2.6062, "step": 11727 }, { "epoch": 3.676489028213166, "grad_norm": 16.0185604095459, "learning_rate": 1.5354099113921614e-06, "loss": 1.9462, "step": 11728 }, { "epoch": 3.6768025078369906, "grad_norm": 8.589717864990234, "learning_rate": 1.5327852371304264e-06, "loss": 2.1104, "step": 11729 }, { "epoch": 3.677115987460815, "grad_norm": 7.032997131347656, "learning_rate": 1.5301618160227098e-06, "loss": 2.0847, "step": 11730 }, { "epoch": 3.6774294670846395, "grad_norm": 8.03189754486084, "learning_rate": 1.5275396514679986e-06, "loss": 2.2512, "step": 11731 }, { "epoch": 3.677742946708464, "grad_norm": 13.380393028259277, "learning_rate": 1.5249187468636526e-06, "loss": 1.6909, "step": 11732 }, { "epoch": 3.6780564263322884, "grad_norm": 18.77620506286621, "learning_rate": 1.5222991056053982e-06, "loss": 2.327, "step": 11733 }, { "epoch": 3.678369905956113, "grad_norm": 5.382106781005859, "learning_rate": 1.519680731087325e-06, "loss": 2.0483, "step": 11734 }, { "epoch": 3.6786833855799372, "grad_norm": 5.012572288513184, "learning_rate": 1.5170636267018829e-06, "loss": 2.1606, "step": 11735 }, { "epoch": 3.6789968652037617, "grad_norm": 5.118175029754639, "learning_rate": 1.514447795839874e-06, "loss": 2.0007, "step": 11736 }, { "epoch": 3.679310344827586, "grad_norm": 6.835746765136719, "learning_rate": 1.5118332418904523e-06, "loss": 2.3279, "step": 11737 }, { "epoch": 3.6796238244514106, "grad_norm": 7.555581092834473, "learning_rate": 1.5092199682411163e-06, "loss": 2.3095, "step": 11738 }, { "epoch": 3.679937304075235, "grad_norm": 5.14214563369751, "learning_rate": 1.5066079782777049e-06, "loss": 1.915, "step": 11739 }, { "epoch": 3.6802507836990594, "grad_norm": 37.92070770263672, "learning_rate": 1.5039972753843966e-06, "loss": 2.0178, "step": 11740 }, { "epoch": 3.680564263322884, "grad_norm": 12.094643592834473, "learning_rate": 1.5013878629436995e-06, "loss": 2.689, "step": 11741 }, { "epoch": 3.6808777429467083, "grad_norm": 13.63865852355957, "learning_rate": 1.4987797443364516e-06, "loss": 2.2014, "step": 11742 }, { "epoch": 3.6811912225705328, "grad_norm": 5.498270511627197, "learning_rate": 1.496172922941813e-06, "loss": 2.1423, "step": 11743 }, { "epoch": 3.681504702194357, "grad_norm": 5.736780643463135, "learning_rate": 1.493567402137263e-06, "loss": 2.0358, "step": 11744 }, { "epoch": 3.6818181818181817, "grad_norm": 8.55668830871582, "learning_rate": 1.4909631852985992e-06, "loss": 1.8452, "step": 11745 }, { "epoch": 3.682131661442006, "grad_norm": 12.387293815612793, "learning_rate": 1.488360275799926e-06, "loss": 2.4505, "step": 11746 }, { "epoch": 3.6824451410658305, "grad_norm": 6.1985650062561035, "learning_rate": 1.485758677013655e-06, "loss": 2.0528, "step": 11747 }, { "epoch": 3.682758620689655, "grad_norm": 4.48468017578125, "learning_rate": 1.4831583923105e-06, "loss": 2.0694, "step": 11748 }, { "epoch": 3.6830721003134794, "grad_norm": 9.11159896850586, "learning_rate": 1.4805594250594718e-06, "loss": 1.9138, "step": 11749 }, { "epoch": 3.683385579937304, "grad_norm": 10.589515686035156, "learning_rate": 1.4779617786278743e-06, "loss": 1.8996, "step": 11750 }, { "epoch": 3.6836990595611283, "grad_norm": 8.000334739685059, "learning_rate": 1.4753654563813013e-06, "loss": 1.9556, "step": 11751 }, { "epoch": 3.6840125391849528, "grad_norm": 10.879490852355957, "learning_rate": 1.4727704616836297e-06, "loss": 2.315, "step": 11752 }, { "epoch": 3.684326018808777, "grad_norm": 6.881276607513428, "learning_rate": 1.4701767978970163e-06, "loss": 2.0328, "step": 11753 }, { "epoch": 3.6846394984326016, "grad_norm": 10.272948265075684, "learning_rate": 1.467584468381894e-06, "loss": 1.9852, "step": 11754 }, { "epoch": 3.684952978056426, "grad_norm": 7.214583873748779, "learning_rate": 1.4649934764969665e-06, "loss": 2.0348, "step": 11755 }, { "epoch": 3.6852664576802505, "grad_norm": 6.5928826332092285, "learning_rate": 1.462403825599205e-06, "loss": 2.0462, "step": 11756 }, { "epoch": 3.685579937304075, "grad_norm": 5.643819808959961, "learning_rate": 1.459815519043844e-06, "loss": 1.8846, "step": 11757 }, { "epoch": 3.6858934169279, "grad_norm": 7.746495723724365, "learning_rate": 1.457228560184374e-06, "loss": 2.1538, "step": 11758 }, { "epoch": 3.6862068965517243, "grad_norm": 198.83099365234375, "learning_rate": 1.4546429523725412e-06, "loss": 2.2578, "step": 11759 }, { "epoch": 3.6865203761755487, "grad_norm": 10.768177032470703, "learning_rate": 1.4520586989583406e-06, "loss": 2.2586, "step": 11760 }, { "epoch": 3.686833855799373, "grad_norm": 150.4090118408203, "learning_rate": 1.4494758032900119e-06, "loss": 2.0744, "step": 11761 }, { "epoch": 3.6871473354231976, "grad_norm": 21.2949161529541, "learning_rate": 1.446894268714038e-06, "loss": 2.5971, "step": 11762 }, { "epoch": 3.687460815047022, "grad_norm": 7.99656343460083, "learning_rate": 1.4443140985751362e-06, "loss": 2.2514, "step": 11763 }, { "epoch": 3.6877742946708465, "grad_norm": 5.955010414123535, "learning_rate": 1.441735296216256e-06, "loss": 1.9313, "step": 11764 }, { "epoch": 3.688087774294671, "grad_norm": 8.720565795898438, "learning_rate": 1.4391578649785754e-06, "loss": 2.1394, "step": 11765 }, { "epoch": 3.6884012539184954, "grad_norm": 48.91127014160156, "learning_rate": 1.4365818082014947e-06, "loss": 2.0028, "step": 11766 }, { "epoch": 3.68871473354232, "grad_norm": 8.661713600158691, "learning_rate": 1.434007129222637e-06, "loss": 1.8853, "step": 11767 }, { "epoch": 3.6890282131661443, "grad_norm": 7.375433921813965, "learning_rate": 1.431433831377836e-06, "loss": 2.0659, "step": 11768 }, { "epoch": 3.6893416927899687, "grad_norm": 6.784815311431885, "learning_rate": 1.4288619180011381e-06, "loss": 2.0757, "step": 11769 }, { "epoch": 3.689655172413793, "grad_norm": 11.209379196166992, "learning_rate": 1.4262913924247956e-06, "loss": 1.9631, "step": 11770 }, { "epoch": 3.6899686520376176, "grad_norm": 30.62173080444336, "learning_rate": 1.4237222579792618e-06, "loss": 1.9274, "step": 11771 }, { "epoch": 3.690282131661442, "grad_norm": 4.857993125915527, "learning_rate": 1.42115451799319e-06, "loss": 1.7847, "step": 11772 }, { "epoch": 3.6905956112852665, "grad_norm": 11.043049812316895, "learning_rate": 1.418588175793425e-06, "loss": 2.0297, "step": 11773 }, { "epoch": 3.690909090909091, "grad_norm": 6.848451614379883, "learning_rate": 1.4160232347050002e-06, "loss": 2.2642, "step": 11774 }, { "epoch": 3.6912225705329154, "grad_norm": 6.972294807434082, "learning_rate": 1.4134596980511346e-06, "loss": 2.2528, "step": 11775 }, { "epoch": 3.69153605015674, "grad_norm": 8.242999076843262, "learning_rate": 1.4108975691532273e-06, "loss": 1.9648, "step": 11776 }, { "epoch": 3.6918495297805642, "grad_norm": 7.3917155265808105, "learning_rate": 1.4083368513308526e-06, "loss": 1.9793, "step": 11777 }, { "epoch": 3.6921630094043887, "grad_norm": 7.806635856628418, "learning_rate": 1.4057775479017597e-06, "loss": 2.1481, "step": 11778 }, { "epoch": 3.692476489028213, "grad_norm": 9.372099876403809, "learning_rate": 1.403219662181861e-06, "loss": 2.0592, "step": 11779 }, { "epoch": 3.6927899686520376, "grad_norm": 9.984753608703613, "learning_rate": 1.4006631974852353e-06, "loss": 2.484, "step": 11780 }, { "epoch": 3.693103448275862, "grad_norm": 7.239848613739014, "learning_rate": 1.3981081571241183e-06, "loss": 2.3147, "step": 11781 }, { "epoch": 3.6934169278996865, "grad_norm": 7.818907737731934, "learning_rate": 1.3955545444089017e-06, "loss": 2.2628, "step": 11782 }, { "epoch": 3.693730407523511, "grad_norm": 6.544126033782959, "learning_rate": 1.393002362648126e-06, "loss": 2.061, "step": 11783 }, { "epoch": 3.6940438871473353, "grad_norm": 5.5599541664123535, "learning_rate": 1.3904516151484794e-06, "loss": 1.9544, "step": 11784 }, { "epoch": 3.69435736677116, "grad_norm": 6.268325328826904, "learning_rate": 1.38790230521479e-06, "loss": 2.1108, "step": 11785 }, { "epoch": 3.694670846394984, "grad_norm": 16.578083038330078, "learning_rate": 1.3853544361500254e-06, "loss": 2.8702, "step": 11786 }, { "epoch": 3.6949843260188087, "grad_norm": 39.09196853637695, "learning_rate": 1.3828080112552835e-06, "loss": 1.9043, "step": 11787 }, { "epoch": 3.695297805642633, "grad_norm": 6.13904333114624, "learning_rate": 1.3802630338297956e-06, "loss": 2.1311, "step": 11788 }, { "epoch": 3.6956112852664575, "grad_norm": 18.247425079345703, "learning_rate": 1.3777195071709131e-06, "loss": 3.0786, "step": 11789 }, { "epoch": 3.695924764890282, "grad_norm": 6.552885055541992, "learning_rate": 1.37517743457411e-06, "loss": 2.112, "step": 11790 }, { "epoch": 3.6962382445141064, "grad_norm": 56.67654800415039, "learning_rate": 1.372636819332976e-06, "loss": 2.1574, "step": 11791 }, { "epoch": 3.696551724137931, "grad_norm": 7.688971042633057, "learning_rate": 1.370097664739212e-06, "loss": 2.147, "step": 11792 }, { "epoch": 3.6968652037617558, "grad_norm": 5.020254135131836, "learning_rate": 1.3675599740826263e-06, "loss": 2.2492, "step": 11793 }, { "epoch": 3.69717868338558, "grad_norm": 14.688599586486816, "learning_rate": 1.3650237506511333e-06, "loss": 2.2149, "step": 11794 }, { "epoch": 3.6974921630094046, "grad_norm": 20.622234344482422, "learning_rate": 1.3624889977307432e-06, "loss": 2.2964, "step": 11795 }, { "epoch": 3.697805642633229, "grad_norm": 13.549787521362305, "learning_rate": 1.3599557186055612e-06, "loss": 2.1162, "step": 11796 }, { "epoch": 3.6981191222570535, "grad_norm": 5.157261848449707, "learning_rate": 1.3574239165577852e-06, "loss": 2.0889, "step": 11797 }, { "epoch": 3.698432601880878, "grad_norm": 13.384552001953125, "learning_rate": 1.3548935948676955e-06, "loss": 2.0487, "step": 11798 }, { "epoch": 3.6987460815047024, "grad_norm": 6.736470699310303, "learning_rate": 1.3523647568136603e-06, "loss": 2.0707, "step": 11799 }, { "epoch": 3.699059561128527, "grad_norm": 8.252944946289062, "learning_rate": 1.3498374056721198e-06, "loss": 2.0475, "step": 11800 }, { "epoch": 3.6993730407523513, "grad_norm": 6.680192947387695, "learning_rate": 1.347311544717591e-06, "loss": 2.1927, "step": 11801 }, { "epoch": 3.6996865203761757, "grad_norm": 9.055883407592773, "learning_rate": 1.3447871772226585e-06, "loss": 2.1622, "step": 11802 }, { "epoch": 3.7, "grad_norm": 5.5665812492370605, "learning_rate": 1.3422643064579733e-06, "loss": 2.0799, "step": 11803 }, { "epoch": 3.7003134796238246, "grad_norm": 11.996819496154785, "learning_rate": 1.3397429356922447e-06, "loss": 2.1095, "step": 11804 }, { "epoch": 3.700626959247649, "grad_norm": 7.040694713592529, "learning_rate": 1.337223068192243e-06, "loss": 2.0694, "step": 11805 }, { "epoch": 3.7009404388714735, "grad_norm": 8.759737014770508, "learning_rate": 1.334704707222787e-06, "loss": 1.8193, "step": 11806 }, { "epoch": 3.701253918495298, "grad_norm": 8.668539047241211, "learning_rate": 1.3321878560467445e-06, "loss": 1.9121, "step": 11807 }, { "epoch": 3.7015673981191224, "grad_norm": 125.59626007080078, "learning_rate": 1.3296725179250274e-06, "loss": 2.5304, "step": 11808 }, { "epoch": 3.701880877742947, "grad_norm": 73.62805938720703, "learning_rate": 1.3271586961165872e-06, "loss": 2.3012, "step": 11809 }, { "epoch": 3.7021943573667713, "grad_norm": 7.111425399780273, "learning_rate": 1.3246463938784122e-06, "loss": 2.0756, "step": 11810 }, { "epoch": 3.7025078369905957, "grad_norm": 7.729673385620117, "learning_rate": 1.322135614465519e-06, "loss": 1.9035, "step": 11811 }, { "epoch": 3.70282131661442, "grad_norm": 7.400961399078369, "learning_rate": 1.3196263611309539e-06, "loss": 1.8918, "step": 11812 }, { "epoch": 3.7031347962382446, "grad_norm": 6.012514591217041, "learning_rate": 1.317118637125784e-06, "loss": 2.2498, "step": 11813 }, { "epoch": 3.703448275862069, "grad_norm": 8.967763900756836, "learning_rate": 1.3146124456990955e-06, "loss": 2.2437, "step": 11814 }, { "epoch": 3.7037617554858935, "grad_norm": 6.538351535797119, "learning_rate": 1.3121077900979907e-06, "loss": 2.0137, "step": 11815 }, { "epoch": 3.704075235109718, "grad_norm": 6.979898929595947, "learning_rate": 1.3096046735675795e-06, "loss": 2.1082, "step": 11816 }, { "epoch": 3.7043887147335424, "grad_norm": 116.82706451416016, "learning_rate": 1.307103099350979e-06, "loss": 2.3042, "step": 11817 }, { "epoch": 3.704702194357367, "grad_norm": 8.139591217041016, "learning_rate": 1.3046030706893079e-06, "loss": 1.9615, "step": 11818 }, { "epoch": 3.7050156739811912, "grad_norm": 14.859047889709473, "learning_rate": 1.3021045908216822e-06, "loss": 1.9726, "step": 11819 }, { "epoch": 3.7053291536050157, "grad_norm": 12.418227195739746, "learning_rate": 1.2996076629852114e-06, "loss": 2.4654, "step": 11820 }, { "epoch": 3.70564263322884, "grad_norm": 6.4466633796691895, "learning_rate": 1.2971122904149944e-06, "loss": 2.3289, "step": 11821 }, { "epoch": 3.7059561128526646, "grad_norm": 75.96076965332031, "learning_rate": 1.2946184763441145e-06, "loss": 2.0216, "step": 11822 }, { "epoch": 3.706269592476489, "grad_norm": 5.0251851081848145, "learning_rate": 1.2921262240036358e-06, "loss": 1.9208, "step": 11823 }, { "epoch": 3.7065830721003135, "grad_norm": 17.49217987060547, "learning_rate": 1.2896355366226e-06, "loss": 2.3403, "step": 11824 }, { "epoch": 3.706896551724138, "grad_norm": 5.339019775390625, "learning_rate": 1.2871464174280185e-06, "loss": 2.0545, "step": 11825 }, { "epoch": 3.7072100313479623, "grad_norm": 101.44503021240234, "learning_rate": 1.2846588696448753e-06, "loss": 2.1268, "step": 11826 }, { "epoch": 3.707523510971787, "grad_norm": 5.224442481994629, "learning_rate": 1.282172896496115e-06, "loss": 2.3213, "step": 11827 }, { "epoch": 3.7078369905956112, "grad_norm": 21.15250015258789, "learning_rate": 1.2796885012026426e-06, "loss": 2.3184, "step": 11828 }, { "epoch": 3.7081504702194357, "grad_norm": 7.682493686676025, "learning_rate": 1.2772056869833194e-06, "loss": 2.0897, "step": 11829 }, { "epoch": 3.70846394984326, "grad_norm": 7.9188995361328125, "learning_rate": 1.2747244570549578e-06, "loss": 2.2244, "step": 11830 }, { "epoch": 3.7087774294670846, "grad_norm": 26.036285400390625, "learning_rate": 1.2722448146323175e-06, "loss": 2.0861, "step": 11831 }, { "epoch": 3.709090909090909, "grad_norm": 7.994349956512451, "learning_rate": 1.2697667629281025e-06, "loss": 1.9414, "step": 11832 }, { "epoch": 3.7094043887147334, "grad_norm": 6.124399662017822, "learning_rate": 1.2672903051529545e-06, "loss": 1.99, "step": 11833 }, { "epoch": 3.709717868338558, "grad_norm": 5.268187522888184, "learning_rate": 1.2648154445154503e-06, "loss": 2.1028, "step": 11834 }, { "epoch": 3.7100313479623823, "grad_norm": 22.05226707458496, "learning_rate": 1.2623421842220976e-06, "loss": 2.1581, "step": 11835 }, { "epoch": 3.7103448275862068, "grad_norm": 8.214548110961914, "learning_rate": 1.2598705274773299e-06, "loss": 1.9614, "step": 11836 }, { "epoch": 3.710658307210031, "grad_norm": 5.832287788391113, "learning_rate": 1.2574004774835052e-06, "loss": 2.0839, "step": 11837 }, { "epoch": 3.7109717868338556, "grad_norm": 7.274767875671387, "learning_rate": 1.254932037440898e-06, "loss": 1.9255, "step": 11838 }, { "epoch": 3.71128526645768, "grad_norm": 8.902657508850098, "learning_rate": 1.2524652105476967e-06, "loss": 2.152, "step": 11839 }, { "epoch": 3.7115987460815045, "grad_norm": 11.36705493927002, "learning_rate": 1.2500000000000007e-06, "loss": 2.1918, "step": 11840 }, { "epoch": 3.711912225705329, "grad_norm": 7.800271511077881, "learning_rate": 1.2475364089918132e-06, "loss": 2.1403, "step": 11841 }, { "epoch": 3.7122257053291534, "grad_norm": 5.661422252655029, "learning_rate": 1.2450744407150427e-06, "loss": 2.0716, "step": 11842 }, { "epoch": 3.712539184952978, "grad_norm": 6.595630168914795, "learning_rate": 1.2426140983594925e-06, "loss": 1.9813, "step": 11843 }, { "epoch": 3.7128526645768023, "grad_norm": 6.165835380554199, "learning_rate": 1.2401553851128598e-06, "loss": 2.0971, "step": 11844 }, { "epoch": 3.7131661442006267, "grad_norm": 73.97936248779297, "learning_rate": 1.2376983041607307e-06, "loss": 2.2902, "step": 11845 }, { "epoch": 3.713479623824451, "grad_norm": 29.427825927734375, "learning_rate": 1.2352428586865775e-06, "loss": 2.2948, "step": 11846 }, { "epoch": 3.7137931034482756, "grad_norm": 50.02643966674805, "learning_rate": 1.2327890518717518e-06, "loss": 2.4208, "step": 11847 }, { "epoch": 3.7141065830721, "grad_norm": 7.618636131286621, "learning_rate": 1.2303368868954848e-06, "loss": 2.1415, "step": 11848 }, { "epoch": 3.7144200626959245, "grad_norm": 6.388270378112793, "learning_rate": 1.2278863669348788e-06, "loss": 1.8524, "step": 11849 }, { "epoch": 3.714733542319749, "grad_norm": 5.000609397888184, "learning_rate": 1.2254374951649043e-06, "loss": 2.2476, "step": 11850 }, { "epoch": 3.7150470219435734, "grad_norm": 148.49742126464844, "learning_rate": 1.2229902747583972e-06, "loss": 2.1885, "step": 11851 }, { "epoch": 3.7153605015673983, "grad_norm": 14.361618041992188, "learning_rate": 1.2205447088860528e-06, "loss": 2.1006, "step": 11852 }, { "epoch": 3.7156739811912227, "grad_norm": 6.653857231140137, "learning_rate": 1.2181008007164249e-06, "loss": 1.8287, "step": 11853 }, { "epoch": 3.715987460815047, "grad_norm": 6.7451491355896, "learning_rate": 1.215658553415918e-06, "loss": 2.0672, "step": 11854 }, { "epoch": 3.7163009404388716, "grad_norm": 7.844499588012695, "learning_rate": 1.2132179701487841e-06, "loss": 2.3321, "step": 11855 }, { "epoch": 3.716614420062696, "grad_norm": 5.583634853363037, "learning_rate": 1.2107790540771208e-06, "loss": 2.1358, "step": 11856 }, { "epoch": 3.7169278996865205, "grad_norm": 4.272064208984375, "learning_rate": 1.2083418083608642e-06, "loss": 2.1212, "step": 11857 }, { "epoch": 3.717241379310345, "grad_norm": 7.327932834625244, "learning_rate": 1.2059062361577871e-06, "loss": 2.2172, "step": 11858 }, { "epoch": 3.7175548589341694, "grad_norm": 5.6187639236450195, "learning_rate": 1.2034723406234942e-06, "loss": 2.0069, "step": 11859 }, { "epoch": 3.717868338557994, "grad_norm": 9.447388648986816, "learning_rate": 1.2010401249114166e-06, "loss": 1.9757, "step": 11860 }, { "epoch": 3.7181818181818183, "grad_norm": 9.47160816192627, "learning_rate": 1.1986095921728099e-06, "loss": 2.1374, "step": 11861 }, { "epoch": 3.7184952978056427, "grad_norm": 223.70350646972656, "learning_rate": 1.1961807455567505e-06, "loss": 2.3751, "step": 11862 }, { "epoch": 3.718808777429467, "grad_norm": 7.952380180358887, "learning_rate": 1.193753588210128e-06, "loss": 1.8767, "step": 11863 }, { "epoch": 3.7191222570532916, "grad_norm": 10.086857795715332, "learning_rate": 1.1913281232776445e-06, "loss": 2.518, "step": 11864 }, { "epoch": 3.719435736677116, "grad_norm": 9.72911548614502, "learning_rate": 1.188904353901808e-06, "loss": 1.8619, "step": 11865 }, { "epoch": 3.7197492163009405, "grad_norm": 20.856670379638672, "learning_rate": 1.1864822832229319e-06, "loss": 2.2472, "step": 11866 }, { "epoch": 3.720062695924765, "grad_norm": 5.762858867645264, "learning_rate": 1.184061914379127e-06, "loss": 2.14, "step": 11867 }, { "epoch": 3.7203761755485893, "grad_norm": 10.269980430603027, "learning_rate": 1.1816432505062985e-06, "loss": 2.0626, "step": 11868 }, { "epoch": 3.720689655172414, "grad_norm": 4.2453765869140625, "learning_rate": 1.1792262947381464e-06, "loss": 2.1691, "step": 11869 }, { "epoch": 3.7210031347962382, "grad_norm": 5.187534332275391, "learning_rate": 1.1768110502061534e-06, "loss": 2.234, "step": 11870 }, { "epoch": 3.7213166144200627, "grad_norm": 6.6286444664001465, "learning_rate": 1.1743975200395865e-06, "loss": 2.1836, "step": 11871 }, { "epoch": 3.721630094043887, "grad_norm": 13.652948379516602, "learning_rate": 1.1719857073654923e-06, "loss": 2.2805, "step": 11872 }, { "epoch": 3.7219435736677116, "grad_norm": 8.836125373840332, "learning_rate": 1.1695756153086911e-06, "loss": 2.0856, "step": 11873 }, { "epoch": 3.722257053291536, "grad_norm": 75.47357940673828, "learning_rate": 1.1671672469917735e-06, "loss": 2.6994, "step": 11874 }, { "epoch": 3.7225705329153604, "grad_norm": 6.333208084106445, "learning_rate": 1.1647606055350997e-06, "loss": 2.1445, "step": 11875 }, { "epoch": 3.722884012539185, "grad_norm": 6.5105414390563965, "learning_rate": 1.1623556940567892e-06, "loss": 1.9541, "step": 11876 }, { "epoch": 3.7231974921630093, "grad_norm": 5.982640266418457, "learning_rate": 1.1599525156727215e-06, "loss": 2.0062, "step": 11877 }, { "epoch": 3.7235109717868338, "grad_norm": 5.674196720123291, "learning_rate": 1.1575510734965305e-06, "loss": 2.036, "step": 11878 }, { "epoch": 3.723824451410658, "grad_norm": 43.82720184326172, "learning_rate": 1.1551513706395994e-06, "loss": 2.1511, "step": 11879 }, { "epoch": 3.7241379310344827, "grad_norm": 5.682617664337158, "learning_rate": 1.1527534102110613e-06, "loss": 2.2001, "step": 11880 }, { "epoch": 3.724451410658307, "grad_norm": 28.62896728515625, "learning_rate": 1.1503571953177884e-06, "loss": 2.0612, "step": 11881 }, { "epoch": 3.7247648902821315, "grad_norm": 7.450265884399414, "learning_rate": 1.1479627290643924e-06, "loss": 2.1884, "step": 11882 }, { "epoch": 3.725078369905956, "grad_norm": 10.637771606445312, "learning_rate": 1.1455700145532198e-06, "loss": 2.3818, "step": 11883 }, { "epoch": 3.7253918495297804, "grad_norm": 23.543020248413086, "learning_rate": 1.1431790548843464e-06, "loss": 2.1492, "step": 11884 }, { "epoch": 3.725705329153605, "grad_norm": 6.182159900665283, "learning_rate": 1.140789853155575e-06, "loss": 2.0998, "step": 11885 }, { "epoch": 3.7260188087774293, "grad_norm": 11.623204231262207, "learning_rate": 1.1384024124624324e-06, "loss": 1.8997, "step": 11886 }, { "epoch": 3.726332288401254, "grad_norm": 16.657573699951172, "learning_rate": 1.1360167358981612e-06, "loss": 1.9186, "step": 11887 }, { "epoch": 3.7266457680250786, "grad_norm": 7.484588623046875, "learning_rate": 1.1336328265537195e-06, "loss": 2.1852, "step": 11888 }, { "epoch": 3.726959247648903, "grad_norm": 11.662534713745117, "learning_rate": 1.1312506875177754e-06, "loss": 2.064, "step": 11889 }, { "epoch": 3.7272727272727275, "grad_norm": 185.8316192626953, "learning_rate": 1.1288703218767027e-06, "loss": 2.6783, "step": 11890 }, { "epoch": 3.727586206896552, "grad_norm": 6.86447811126709, "learning_rate": 1.1264917327145805e-06, "loss": 2.4577, "step": 11891 }, { "epoch": 3.7278996865203764, "grad_norm": 8.464423179626465, "learning_rate": 1.124114923113183e-06, "loss": 2.1735, "step": 11892 }, { "epoch": 3.728213166144201, "grad_norm": 11.48192310333252, "learning_rate": 1.1217398961519798e-06, "loss": 1.6965, "step": 11893 }, { "epoch": 3.7285266457680253, "grad_norm": 8.207538604736328, "learning_rate": 1.119366654908131e-06, "loss": 2.1114, "step": 11894 }, { "epoch": 3.7288401253918497, "grad_norm": 5.960636615753174, "learning_rate": 1.1169952024564824e-06, "loss": 2.0493, "step": 11895 }, { "epoch": 3.729153605015674, "grad_norm": 6.30183219909668, "learning_rate": 1.1146255418695635e-06, "loss": 2.1816, "step": 11896 }, { "epoch": 3.7294670846394986, "grad_norm": 13.153573036193848, "learning_rate": 1.1122576762175813e-06, "loss": 2.4404, "step": 11897 }, { "epoch": 3.729780564263323, "grad_norm": 12.336709022521973, "learning_rate": 1.109891608568416e-06, "loss": 1.9188, "step": 11898 }, { "epoch": 3.7300940438871475, "grad_norm": 9.395471572875977, "learning_rate": 1.107527341987622e-06, "loss": 1.9267, "step": 11899 }, { "epoch": 3.730407523510972, "grad_norm": 6.240561008453369, "learning_rate": 1.1051648795384167e-06, "loss": 2.1032, "step": 11900 }, { "epoch": 3.7307210031347964, "grad_norm": 11.791426658630371, "learning_rate": 1.1028042242816811e-06, "loss": 2.1902, "step": 11901 }, { "epoch": 3.731034482758621, "grad_norm": 7.924000263214111, "learning_rate": 1.1004453792759547e-06, "loss": 2.2723, "step": 11902 }, { "epoch": 3.7313479623824453, "grad_norm": 9.424219131469727, "learning_rate": 1.098088347577432e-06, "loss": 2.3908, "step": 11903 }, { "epoch": 3.7316614420062697, "grad_norm": 7.5416460037231445, "learning_rate": 1.0957331322399575e-06, "loss": 2.0521, "step": 11904 }, { "epoch": 3.731974921630094, "grad_norm": 7.891406059265137, "learning_rate": 1.0933797363150225e-06, "loss": 2.0275, "step": 11905 }, { "epoch": 3.7322884012539186, "grad_norm": 9.418791770935059, "learning_rate": 1.0910281628517601e-06, "loss": 1.7307, "step": 11906 }, { "epoch": 3.732601880877743, "grad_norm": 9.191510200500488, "learning_rate": 1.0886784148969457e-06, "loss": 2.177, "step": 11907 }, { "epoch": 3.7329153605015675, "grad_norm": 12.091560363769531, "learning_rate": 1.0863304954949856e-06, "loss": 1.9554, "step": 11908 }, { "epoch": 3.733228840125392, "grad_norm": 8.623270034790039, "learning_rate": 1.0839844076879186e-06, "loss": 2.0674, "step": 11909 }, { "epoch": 3.7335423197492164, "grad_norm": 8.856292724609375, "learning_rate": 1.0816401545154097e-06, "loss": 2.1713, "step": 11910 }, { "epoch": 3.733855799373041, "grad_norm": 47.55047607421875, "learning_rate": 1.0792977390147474e-06, "loss": 2.2183, "step": 11911 }, { "epoch": 3.7341692789968652, "grad_norm": 9.250981330871582, "learning_rate": 1.0769571642208404e-06, "loss": 2.1518, "step": 11912 }, { "epoch": 3.7344827586206897, "grad_norm": 9.835676193237305, "learning_rate": 1.0746184331662106e-06, "loss": 2.3378, "step": 11913 }, { "epoch": 3.734796238244514, "grad_norm": 18.650583267211914, "learning_rate": 1.072281548880992e-06, "loss": 2.0404, "step": 11914 }, { "epoch": 3.7351097178683386, "grad_norm": 6.585302829742432, "learning_rate": 1.0699465143929257e-06, "loss": 2.1768, "step": 11915 }, { "epoch": 3.735423197492163, "grad_norm": 8.610756874084473, "learning_rate": 1.0676133327273563e-06, "loss": 2.0128, "step": 11916 }, { "epoch": 3.7357366771159874, "grad_norm": 100.36659240722656, "learning_rate": 1.0652820069072264e-06, "loss": 2.4216, "step": 11917 }, { "epoch": 3.736050156739812, "grad_norm": 5.343726634979248, "learning_rate": 1.062952539953078e-06, "loss": 2.4781, "step": 11918 }, { "epoch": 3.7363636363636363, "grad_norm": 11.59766674041748, "learning_rate": 1.0606249348830402e-06, "loss": 2.1185, "step": 11919 }, { "epoch": 3.7366771159874608, "grad_norm": 6.221990585327148, "learning_rate": 1.0582991947128324e-06, "loss": 2.7201, "step": 11920 }, { "epoch": 3.736990595611285, "grad_norm": 11.677096366882324, "learning_rate": 1.055975322455757e-06, "loss": 2.5184, "step": 11921 }, { "epoch": 3.7373040752351097, "grad_norm": 6.609655380249023, "learning_rate": 1.0536533211226954e-06, "loss": 2.2162, "step": 11922 }, { "epoch": 3.737617554858934, "grad_norm": 14.637845993041992, "learning_rate": 1.0513331937221075e-06, "loss": 2.0968, "step": 11923 }, { "epoch": 3.7379310344827585, "grad_norm": 35.554935455322266, "learning_rate": 1.049014943260023e-06, "loss": 2.1474, "step": 11924 }, { "epoch": 3.738244514106583, "grad_norm": 10.789631843566895, "learning_rate": 1.0466985727400406e-06, "loss": 2.1436, "step": 11925 }, { "epoch": 3.7385579937304074, "grad_norm": 6.4490251541137695, "learning_rate": 1.0443840851633227e-06, "loss": 2.0549, "step": 11926 }, { "epoch": 3.738871473354232, "grad_norm": 14.543988227844238, "learning_rate": 1.042071483528593e-06, "loss": 2.0452, "step": 11927 }, { "epoch": 3.7391849529780563, "grad_norm": 9.555737495422363, "learning_rate": 1.0397607708321302e-06, "loss": 1.9675, "step": 11928 }, { "epoch": 3.7394984326018808, "grad_norm": 6.71116828918457, "learning_rate": 1.0374519500677688e-06, "loss": 2.1818, "step": 11929 }, { "epoch": 3.739811912225705, "grad_norm": 7.908589839935303, "learning_rate": 1.0351450242268885e-06, "loss": 2.1953, "step": 11930 }, { "epoch": 3.7401253918495296, "grad_norm": 12.734683990478516, "learning_rate": 1.0328399962984155e-06, "loss": 2.3702, "step": 11931 }, { "epoch": 3.740438871473354, "grad_norm": 5.850962162017822, "learning_rate": 1.0305368692688175e-06, "loss": 2.0694, "step": 11932 }, { "epoch": 3.7407523510971785, "grad_norm": 8.24632453918457, "learning_rate": 1.0282356461220978e-06, "loss": 2.2298, "step": 11933 }, { "epoch": 3.741065830721003, "grad_norm": 15.722797393798828, "learning_rate": 1.0259363298397942e-06, "loss": 1.7837, "step": 11934 }, { "epoch": 3.7413793103448274, "grad_norm": 7.637423992156982, "learning_rate": 1.0236389234009728e-06, "loss": 2.1387, "step": 11935 }, { "epoch": 3.741692789968652, "grad_norm": 16.73200035095215, "learning_rate": 1.0213434297822275e-06, "loss": 2.0429, "step": 11936 }, { "epoch": 3.7420062695924763, "grad_norm": 11.35887622833252, "learning_rate": 1.0190498519576715e-06, "loss": 2.0452, "step": 11937 }, { "epoch": 3.7423197492163007, "grad_norm": 9.096086502075195, "learning_rate": 1.0167581928989373e-06, "loss": 2.2276, "step": 11938 }, { "epoch": 3.742633228840125, "grad_norm": 8.275463104248047, "learning_rate": 1.01446845557517e-06, "loss": 2.3436, "step": 11939 }, { "epoch": 3.7429467084639496, "grad_norm": 7.6157379150390625, "learning_rate": 1.0121806429530263e-06, "loss": 2.4603, "step": 11940 }, { "epoch": 3.743260188087774, "grad_norm": 8.301525115966797, "learning_rate": 1.009894757996668e-06, "loss": 1.9047, "step": 11941 }, { "epoch": 3.7435736677115985, "grad_norm": 8.64967155456543, "learning_rate": 1.0076108036677603e-06, "loss": 2.2024, "step": 11942 }, { "epoch": 3.743887147335423, "grad_norm": 57.8975944519043, "learning_rate": 1.0053287829254664e-06, "loss": 2.0049, "step": 11943 }, { "epoch": 3.7442006269592474, "grad_norm": 13.911728858947754, "learning_rate": 1.0030486987264436e-06, "loss": 2.2962, "step": 11944 }, { "epoch": 3.7445141065830723, "grad_norm": 8.647785186767578, "learning_rate": 1.0007705540248434e-06, "loss": 2.6317, "step": 11945 }, { "epoch": 3.7448275862068967, "grad_norm": 5.570077419281006, "learning_rate": 9.984943517723003e-07, "loss": 1.9771, "step": 11946 }, { "epoch": 3.745141065830721, "grad_norm": 7.486064910888672, "learning_rate": 9.962200949179346e-07, "loss": 1.9251, "step": 11947 }, { "epoch": 3.7454545454545456, "grad_norm": 10.714069366455078, "learning_rate": 9.939477864083447e-07, "loss": 2.2553, "step": 11948 }, { "epoch": 3.74576802507837, "grad_norm": 31.733442306518555, "learning_rate": 9.916774291876052e-07, "loss": 1.9828, "step": 11949 }, { "epoch": 3.7460815047021945, "grad_norm": 6.2514729499816895, "learning_rate": 9.89409026197264e-07, "loss": 2.1186, "step": 11950 }, { "epoch": 3.746394984326019, "grad_norm": 8.483033180236816, "learning_rate": 9.871425803763345e-07, "loss": 2.0998, "step": 11951 }, { "epoch": 3.7467084639498434, "grad_norm": 46.883724212646484, "learning_rate": 9.848780946612962e-07, "loss": 2.2536, "step": 11952 }, { "epoch": 3.747021943573668, "grad_norm": 8.046271324157715, "learning_rate": 9.826155719860884e-07, "loss": 2.1581, "step": 11953 }, { "epoch": 3.7473354231974922, "grad_norm": 6.872846603393555, "learning_rate": 9.803550152821064e-07, "loss": 2.0501, "step": 11954 }, { "epoch": 3.7476489028213167, "grad_norm": 6.081361770629883, "learning_rate": 9.780964274781984e-07, "loss": 2.1348, "step": 11955 }, { "epoch": 3.747962382445141, "grad_norm": 4.825142860412598, "learning_rate": 9.758398115006637e-07, "loss": 2.0885, "step": 11956 }, { "epoch": 3.7482758620689656, "grad_norm": 8.373165130615234, "learning_rate": 9.735851702732449e-07, "loss": 1.8751, "step": 11957 }, { "epoch": 3.74858934169279, "grad_norm": 127.41792297363281, "learning_rate": 9.713325067171255e-07, "loss": 2.2399, "step": 11958 }, { "epoch": 3.7489028213166145, "grad_norm": 8.200927734375, "learning_rate": 9.69081823750929e-07, "loss": 2.6334, "step": 11959 }, { "epoch": 3.749216300940439, "grad_norm": 10.676941871643066, "learning_rate": 9.66833124290709e-07, "loss": 1.9404, "step": 11960 }, { "epoch": 3.7495297805642633, "grad_norm": 11.820837020874023, "learning_rate": 9.645864112499543e-07, "loss": 2.5413, "step": 11961 }, { "epoch": 3.749843260188088, "grad_norm": 8.776460647583008, "learning_rate": 9.623416875395763e-07, "loss": 2.1115, "step": 11962 }, { "epoch": 3.750156739811912, "grad_norm": 9.670048713684082, "learning_rate": 9.600989560679098e-07, "loss": 2.2322, "step": 11963 }, { "epoch": 3.7504702194357367, "grad_norm": 7.556646347045898, "learning_rate": 9.578582197407088e-07, "loss": 1.9967, "step": 11964 }, { "epoch": 3.750783699059561, "grad_norm": 5.065824031829834, "learning_rate": 9.55619481461141e-07, "loss": 2.065, "step": 11965 }, { "epoch": 3.7510971786833855, "grad_norm": 10.450499534606934, "learning_rate": 9.533827441297882e-07, "loss": 1.8422, "step": 11966 }, { "epoch": 3.75141065830721, "grad_norm": 6.769338607788086, "learning_rate": 9.511480106446375e-07, "loss": 1.9484, "step": 11967 }, { "epoch": 3.7517241379310344, "grad_norm": 8.089903831481934, "learning_rate": 9.489152839010799e-07, "loss": 1.9368, "step": 11968 }, { "epoch": 3.752037617554859, "grad_norm": 6.693366050720215, "learning_rate": 9.46684566791907e-07, "loss": 2.1209, "step": 11969 }, { "epoch": 3.7523510971786833, "grad_norm": 6.809491157531738, "learning_rate": 9.44455862207306e-07, "loss": 2.2697, "step": 11970 }, { "epoch": 3.7523510971786833, "eval_loss": 2.2757768630981445, "eval_runtime": 21.0741, "eval_samples_per_second": 127.502, "eval_steps_per_second": 7.972, "step": 11970 }, { "epoch": 3.7526645768025078, "grad_norm": 9.806812286376953, "learning_rate": 9.422291730348565e-07, "loss": 1.9761, "step": 11971 }, { "epoch": 3.752978056426332, "grad_norm": 6.890644550323486, "learning_rate": 9.400045021595288e-07, "loss": 1.8985, "step": 11972 }, { "epoch": 3.7532915360501566, "grad_norm": 9.695711135864258, "learning_rate": 9.377818524636762e-07, "loss": 2.0586, "step": 11973 }, { "epoch": 3.753605015673981, "grad_norm": 16.372920989990234, "learning_rate": 9.355612268270339e-07, "loss": 2.1239, "step": 11974 }, { "epoch": 3.7539184952978055, "grad_norm": 6.470925807952881, "learning_rate": 9.333426281267144e-07, "loss": 1.9586, "step": 11975 }, { "epoch": 3.75423197492163, "grad_norm": 14.075987815856934, "learning_rate": 9.311260592372045e-07, "loss": 2.1622, "step": 11976 }, { "epoch": 3.7545454545454544, "grad_norm": 7.386596202850342, "learning_rate": 9.28911523030361e-07, "loss": 2.5574, "step": 11977 }, { "epoch": 3.754858934169279, "grad_norm": 4.890467166900635, "learning_rate": 9.266990223754069e-07, "loss": 1.9994, "step": 11978 }, { "epoch": 3.7551724137931033, "grad_norm": 14.262545585632324, "learning_rate": 9.244885601389278e-07, "loss": 2.6237, "step": 11979 }, { "epoch": 3.7554858934169277, "grad_norm": 8.304709434509277, "learning_rate": 9.222801391848688e-07, "loss": 2.1043, "step": 11980 }, { "epoch": 3.7557993730407526, "grad_norm": 8.47544002532959, "learning_rate": 9.200737623745296e-07, "loss": 2.2656, "step": 11981 }, { "epoch": 3.756112852664577, "grad_norm": 35.56737518310547, "learning_rate": 9.178694325665611e-07, "loss": 2.3625, "step": 11982 }, { "epoch": 3.7564263322884015, "grad_norm": 15.459229469299316, "learning_rate": 9.156671526169641e-07, "loss": 2.6239, "step": 11983 }, { "epoch": 3.756739811912226, "grad_norm": 23.40398597717285, "learning_rate": 9.134669253790814e-07, "loss": 2.1565, "step": 11984 }, { "epoch": 3.7570532915360504, "grad_norm": 7.809870719909668, "learning_rate": 9.112687537035971e-07, "loss": 2.0064, "step": 11985 }, { "epoch": 3.757366771159875, "grad_norm": 42.5857048034668, "learning_rate": 9.090726404385319e-07, "loss": 2.3624, "step": 11986 }, { "epoch": 3.7576802507836993, "grad_norm": 4.706591606140137, "learning_rate": 9.068785884292383e-07, "loss": 1.9954, "step": 11987 }, { "epoch": 3.7579937304075237, "grad_norm": 10.213624000549316, "learning_rate": 9.046866005184016e-07, "loss": 2.1867, "step": 11988 }, { "epoch": 3.758307210031348, "grad_norm": 7.902956008911133, "learning_rate": 9.024966795460297e-07, "loss": 2.2495, "step": 11989 }, { "epoch": 3.7586206896551726, "grad_norm": 7.9587202072143555, "learning_rate": 9.003088283494532e-07, "loss": 1.7333, "step": 11990 }, { "epoch": 3.758934169278997, "grad_norm": 7.767221450805664, "learning_rate": 8.981230497633214e-07, "loss": 2.0968, "step": 11991 }, { "epoch": 3.7592476489028215, "grad_norm": 13.436649322509766, "learning_rate": 8.959393466195973e-07, "loss": 2.0835, "step": 11992 }, { "epoch": 3.759561128526646, "grad_norm": 11.943163871765137, "learning_rate": 8.93757721747557e-07, "loss": 1.9976, "step": 11993 }, { "epoch": 3.7598746081504704, "grad_norm": 5.1110429763793945, "learning_rate": 8.915781779737825e-07, "loss": 2.2461, "step": 11994 }, { "epoch": 3.760188087774295, "grad_norm": 5.27134370803833, "learning_rate": 8.894007181221595e-07, "loss": 2.0397, "step": 11995 }, { "epoch": 3.7605015673981192, "grad_norm": 8.853761672973633, "learning_rate": 8.872253450138732e-07, "loss": 2.0236, "step": 11996 }, { "epoch": 3.7608150470219437, "grad_norm": 15.243996620178223, "learning_rate": 8.850520614674063e-07, "loss": 2.3271, "step": 11997 }, { "epoch": 3.761128526645768, "grad_norm": 13.289353370666504, "learning_rate": 8.828808702985325e-07, "loss": 2.0538, "step": 11998 }, { "epoch": 3.7614420062695926, "grad_norm": 27.484540939331055, "learning_rate": 8.807117743203178e-07, "loss": 2.4888, "step": 11999 }, { "epoch": 3.761755485893417, "grad_norm": 5.090723037719727, "learning_rate": 8.785447763431101e-07, "loss": 2.1684, "step": 12000 }, { "epoch": 3.7620689655172415, "grad_norm": 48.98764419555664, "learning_rate": 8.763798791745413e-07, "loss": 2.6244, "step": 12001 }, { "epoch": 3.762382445141066, "grad_norm": 6.001077651977539, "learning_rate": 8.742170856195195e-07, "loss": 1.9215, "step": 12002 }, { "epoch": 3.7626959247648903, "grad_norm": 5.38381290435791, "learning_rate": 8.720563984802283e-07, "loss": 1.8865, "step": 12003 }, { "epoch": 3.763009404388715, "grad_norm": 11.211763381958008, "learning_rate": 8.69897820556124e-07, "loss": 2.3791, "step": 12004 }, { "epoch": 3.7633228840125392, "grad_norm": 21.13630485534668, "learning_rate": 8.67741354643927e-07, "loss": 2.1127, "step": 12005 }, { "epoch": 3.7636363636363637, "grad_norm": 16.204261779785156, "learning_rate": 8.655870035376232e-07, "loss": 2.0926, "step": 12006 }, { "epoch": 3.763949843260188, "grad_norm": 5.810258388519287, "learning_rate": 8.634347700284576e-07, "loss": 2.3769, "step": 12007 }, { "epoch": 3.7642633228840126, "grad_norm": 68.94987487792969, "learning_rate": 8.612846569049324e-07, "loss": 2.0451, "step": 12008 }, { "epoch": 3.764576802507837, "grad_norm": 73.12967681884766, "learning_rate": 8.591366669528009e-07, "loss": 2.159, "step": 12009 }, { "epoch": 3.7648902821316614, "grad_norm": 9.658553123474121, "learning_rate": 8.569908029550686e-07, "loss": 2.3364, "step": 12010 }, { "epoch": 3.765203761755486, "grad_norm": 12.12109088897705, "learning_rate": 8.54847067691984e-07, "loss": 2.1135, "step": 12011 }, { "epoch": 3.7655172413793103, "grad_norm": 6.217269420623779, "learning_rate": 8.527054639410379e-07, "loss": 1.9475, "step": 12012 }, { "epoch": 3.7658307210031348, "grad_norm": 6.688963890075684, "learning_rate": 8.505659944769601e-07, "loss": 2.0078, "step": 12013 }, { "epoch": 3.766144200626959, "grad_norm": 20.765239715576172, "learning_rate": 8.48428662071715e-07, "loss": 2.0206, "step": 12014 }, { "epoch": 3.7664576802507836, "grad_norm": 6.464234352111816, "learning_rate": 8.462934694944977e-07, "loss": 1.966, "step": 12015 }, { "epoch": 3.766771159874608, "grad_norm": 8.193171501159668, "learning_rate": 8.441604195117315e-07, "loss": 2.1923, "step": 12016 }, { "epoch": 3.7670846394984325, "grad_norm": 6.022193431854248, "learning_rate": 8.42029514887063e-07, "loss": 2.0647, "step": 12017 }, { "epoch": 3.767398119122257, "grad_norm": 7.550471782684326, "learning_rate": 8.399007583813598e-07, "loss": 2.2453, "step": 12018 }, { "epoch": 3.7677115987460814, "grad_norm": 7.338099479675293, "learning_rate": 8.377741527527053e-07, "loss": 1.945, "step": 12019 }, { "epoch": 3.768025078369906, "grad_norm": 7.573056221008301, "learning_rate": 8.356497007563988e-07, "loss": 2.134, "step": 12020 }, { "epoch": 3.7683385579937303, "grad_norm": 8.249102592468262, "learning_rate": 8.33527405144946e-07, "loss": 2.1384, "step": 12021 }, { "epoch": 3.7686520376175547, "grad_norm": 5.72969913482666, "learning_rate": 8.31407268668061e-07, "loss": 2.1345, "step": 12022 }, { "epoch": 3.768965517241379, "grad_norm": 6.584782600402832, "learning_rate": 8.292892940726591e-07, "loss": 2.0999, "step": 12023 }, { "epoch": 3.7692789968652036, "grad_norm": 68.34925842285156, "learning_rate": 8.271734841028553e-07, "loss": 1.9973, "step": 12024 }, { "epoch": 3.769592476489028, "grad_norm": 6.560789585113525, "learning_rate": 8.250598414999589e-07, "loss": 1.9014, "step": 12025 }, { "epoch": 3.7699059561128525, "grad_norm": 15.62242317199707, "learning_rate": 8.22948369002474e-07, "loss": 2.2811, "step": 12026 }, { "epoch": 3.770219435736677, "grad_norm": 8.71703052520752, "learning_rate": 8.208390693460899e-07, "loss": 2.0746, "step": 12027 }, { "epoch": 3.7705329153605014, "grad_norm": 25.248292922973633, "learning_rate": 8.187319452636821e-07, "loss": 2.3598, "step": 12028 }, { "epoch": 3.770846394984326, "grad_norm": 50.23675537109375, "learning_rate": 8.166269994853074e-07, "loss": 2.147, "step": 12029 }, { "epoch": 3.7711598746081503, "grad_norm": 5.818582534790039, "learning_rate": 8.14524234738199e-07, "loss": 2.4084, "step": 12030 }, { "epoch": 3.7714733542319747, "grad_norm": 7.863297462463379, "learning_rate": 8.12423653746767e-07, "loss": 2.0181, "step": 12031 }, { "epoch": 3.771786833855799, "grad_norm": 5.4827117919921875, "learning_rate": 8.103252592325897e-07, "loss": 1.8537, "step": 12032 }, { "epoch": 3.7721003134796236, "grad_norm": 9.130858421325684, "learning_rate": 8.082290539144136e-07, "loss": 2.0663, "step": 12033 }, { "epoch": 3.772413793103448, "grad_norm": 8.150651931762695, "learning_rate": 8.061350405081484e-07, "loss": 2.1355, "step": 12034 }, { "epoch": 3.7727272727272725, "grad_norm": 5.222731590270996, "learning_rate": 8.040432217268629e-07, "loss": 2.0666, "step": 12035 }, { "epoch": 3.773040752351097, "grad_norm": 10.511653900146484, "learning_rate": 8.01953600280786e-07, "loss": 2.1759, "step": 12036 }, { "epoch": 3.7733542319749214, "grad_norm": 10.827237129211426, "learning_rate": 7.998661788772957e-07, "loss": 2.2746, "step": 12037 }, { "epoch": 3.773667711598746, "grad_norm": 29.851741790771484, "learning_rate": 7.977809602209216e-07, "loss": 2.2482, "step": 12038 }, { "epoch": 3.7739811912225707, "grad_norm": 5.676889896392822, "learning_rate": 7.956979470133389e-07, "loss": 2.3076, "step": 12039 }, { "epoch": 3.774294670846395, "grad_norm": 8.307246208190918, "learning_rate": 7.936171419533653e-07, "loss": 2.1216, "step": 12040 }, { "epoch": 3.7746081504702196, "grad_norm": 23.052202224731445, "learning_rate": 7.915385477369567e-07, "loss": 2.6813, "step": 12041 }, { "epoch": 3.774921630094044, "grad_norm": 6.265366077423096, "learning_rate": 7.894621670572069e-07, "loss": 2.077, "step": 12042 }, { "epoch": 3.7752351097178685, "grad_norm": 5.935598373413086, "learning_rate": 7.873880026043401e-07, "loss": 2.0226, "step": 12043 }, { "epoch": 3.775548589341693, "grad_norm": 8.344632148742676, "learning_rate": 7.853160570657089e-07, "loss": 2.3396, "step": 12044 }, { "epoch": 3.7758620689655173, "grad_norm": 8.808268547058105, "learning_rate": 7.832463331257917e-07, "loss": 2.391, "step": 12045 }, { "epoch": 3.776175548589342, "grad_norm": 7.171977996826172, "learning_rate": 7.811788334661871e-07, "loss": 1.9444, "step": 12046 }, { "epoch": 3.7764890282131662, "grad_norm": 7.894704341888428, "learning_rate": 7.791135607656147e-07, "loss": 1.9262, "step": 12047 }, { "epoch": 3.7768025078369907, "grad_norm": 6.142268657684326, "learning_rate": 7.770505176999066e-07, "loss": 2.0588, "step": 12048 }, { "epoch": 3.777115987460815, "grad_norm": 5.009681224822998, "learning_rate": 7.749897069420062e-07, "loss": 2.115, "step": 12049 }, { "epoch": 3.7774294670846396, "grad_norm": 8.85492992401123, "learning_rate": 7.729311311619653e-07, "loss": 2.1494, "step": 12050 }, { "epoch": 3.777742946708464, "grad_norm": 11.017219543457031, "learning_rate": 7.708747930269397e-07, "loss": 2.4652, "step": 12051 }, { "epoch": 3.7780564263322884, "grad_norm": 13.954697608947754, "learning_rate": 7.688206952011862e-07, "loss": 2.1812, "step": 12052 }, { "epoch": 3.778369905956113, "grad_norm": 8.303092002868652, "learning_rate": 7.667688403460585e-07, "loss": 1.968, "step": 12053 }, { "epoch": 3.7786833855799373, "grad_norm": 12.601238250732422, "learning_rate": 7.647192311200052e-07, "loss": 2.1347, "step": 12054 }, { "epoch": 3.7789968652037618, "grad_norm": 8.12341022491455, "learning_rate": 7.62671870178564e-07, "loss": 1.7707, "step": 12055 }, { "epoch": 3.779310344827586, "grad_norm": 5.623566150665283, "learning_rate": 7.606267601743614e-07, "loss": 2.0391, "step": 12056 }, { "epoch": 3.7796238244514107, "grad_norm": 119.61457061767578, "learning_rate": 7.585839037571049e-07, "loss": 2.6625, "step": 12057 }, { "epoch": 3.779937304075235, "grad_norm": 14.340103149414062, "learning_rate": 7.565433035735859e-07, "loss": 1.8781, "step": 12058 }, { "epoch": 3.7802507836990595, "grad_norm": 8.142243385314941, "learning_rate": 7.545049622676695e-07, "loss": 2.1181, "step": 12059 }, { "epoch": 3.780564263322884, "grad_norm": 8.767485618591309, "learning_rate": 7.524688824802953e-07, "loss": 2.3199, "step": 12060 }, { "epoch": 3.7808777429467084, "grad_norm": 6.969453811645508, "learning_rate": 7.504350668494725e-07, "loss": 2.1447, "step": 12061 }, { "epoch": 3.781191222570533, "grad_norm": 6.663456916809082, "learning_rate": 7.484035180102764e-07, "loss": 2.0641, "step": 12062 }, { "epoch": 3.7815047021943573, "grad_norm": 5.366084098815918, "learning_rate": 7.463742385948469e-07, "loss": 1.7662, "step": 12063 }, { "epoch": 3.7818181818181817, "grad_norm": 8.051939010620117, "learning_rate": 7.443472312323824e-07, "loss": 2.1459, "step": 12064 }, { "epoch": 3.782131661442006, "grad_norm": 64.05875396728516, "learning_rate": 7.423224985491373e-07, "loss": 2.1569, "step": 12065 }, { "epoch": 3.7824451410658306, "grad_norm": 7.194041728973389, "learning_rate": 7.403000431684193e-07, "loss": 2.0264, "step": 12066 }, { "epoch": 3.782758620689655, "grad_norm": 7.515738010406494, "learning_rate": 7.382798677105857e-07, "loss": 1.9122, "step": 12067 }, { "epoch": 3.7830721003134795, "grad_norm": 370.6310119628906, "learning_rate": 7.362619747930386e-07, "loss": 2.2196, "step": 12068 }, { "epoch": 3.783385579937304, "grad_norm": 33.17133331298828, "learning_rate": 7.342463670302258e-07, "loss": 2.0863, "step": 12069 }, { "epoch": 3.7836990595611284, "grad_norm": 7.970648765563965, "learning_rate": 7.322330470336314e-07, "loss": 2.0541, "step": 12070 }, { "epoch": 3.784012539184953, "grad_norm": 5.17788553237915, "learning_rate": 7.302220174117763e-07, "loss": 2.0797, "step": 12071 }, { "epoch": 3.7843260188087773, "grad_norm": 9.0812406539917, "learning_rate": 7.282132807702144e-07, "loss": 2.2659, "step": 12072 }, { "epoch": 3.7846394984326017, "grad_norm": 10.055680274963379, "learning_rate": 7.262068397115274e-07, "loss": 1.9984, "step": 12073 }, { "epoch": 3.7849529780564266, "grad_norm": 8.626397132873535, "learning_rate": 7.242026968353253e-07, "loss": 1.9952, "step": 12074 }, { "epoch": 3.785266457680251, "grad_norm": 4.632978916168213, "learning_rate": 7.222008547382381e-07, "loss": 2.0913, "step": 12075 }, { "epoch": 3.7855799373040755, "grad_norm": 21.0430850982666, "learning_rate": 7.202013160139159e-07, "loss": 1.9015, "step": 12076 }, { "epoch": 3.7858934169279, "grad_norm": 7.245023727416992, "learning_rate": 7.182040832530244e-07, "loss": 2.0541, "step": 12077 }, { "epoch": 3.7862068965517244, "grad_norm": 17.91461944580078, "learning_rate": 7.162091590432407e-07, "loss": 2.1662, "step": 12078 }, { "epoch": 3.786520376175549, "grad_norm": 45.133514404296875, "learning_rate": 7.142165459692518e-07, "loss": 1.9461, "step": 12079 }, { "epoch": 3.7868338557993733, "grad_norm": 27.943946838378906, "learning_rate": 7.122262466127513e-07, "loss": 2.0678, "step": 12080 }, { "epoch": 3.7871473354231977, "grad_norm": 7.9868550300598145, "learning_rate": 7.102382635524335e-07, "loss": 2.1433, "step": 12081 }, { "epoch": 3.787460815047022, "grad_norm": 57.153072357177734, "learning_rate": 7.082525993639916e-07, "loss": 2.6811, "step": 12082 }, { "epoch": 3.7877742946708466, "grad_norm": 7.297682285308838, "learning_rate": 7.062692566201151e-07, "loss": 2.2624, "step": 12083 }, { "epoch": 3.788087774294671, "grad_norm": 33.20769500732422, "learning_rate": 7.04288237890485e-07, "loss": 2.0655, "step": 12084 }, { "epoch": 3.7884012539184955, "grad_norm": 9.395768165588379, "learning_rate": 7.02309545741773e-07, "loss": 2.1363, "step": 12085 }, { "epoch": 3.78871473354232, "grad_norm": 6.241046905517578, "learning_rate": 7.003331827376345e-07, "loss": 1.8178, "step": 12086 }, { "epoch": 3.7890282131661444, "grad_norm": 6.647274971008301, "learning_rate": 6.983591514387081e-07, "loss": 1.8615, "step": 12087 }, { "epoch": 3.789341692789969, "grad_norm": 6.465008735656738, "learning_rate": 6.963874544026109e-07, "loss": 2.0674, "step": 12088 }, { "epoch": 3.7896551724137932, "grad_norm": 6.660305023193359, "learning_rate": 6.944180941839359e-07, "loss": 2.2031, "step": 12089 }, { "epoch": 3.7899686520376177, "grad_norm": 9.300021171569824, "learning_rate": 6.924510733342485e-07, "loss": 2.2168, "step": 12090 }, { "epoch": 3.790282131661442, "grad_norm": 7.482845306396484, "learning_rate": 6.904863944020834e-07, "loss": 2.4919, "step": 12091 }, { "epoch": 3.7905956112852666, "grad_norm": 7.324012279510498, "learning_rate": 6.885240599329402e-07, "loss": 2.3051, "step": 12092 }, { "epoch": 3.790909090909091, "grad_norm": 8.410722732543945, "learning_rate": 6.865640724692815e-07, "loss": 1.8931, "step": 12093 }, { "epoch": 3.7912225705329154, "grad_norm": 8.654428482055664, "learning_rate": 6.846064345505296e-07, "loss": 2.1884, "step": 12094 }, { "epoch": 3.79153605015674, "grad_norm": 6.320605754852295, "learning_rate": 6.826511487130608e-07, "loss": 2.22, "step": 12095 }, { "epoch": 3.7918495297805643, "grad_norm": 11.265193939208984, "learning_rate": 6.806982174902065e-07, "loss": 2.3107, "step": 12096 }, { "epoch": 3.7921630094043888, "grad_norm": 9.12255573272705, "learning_rate": 6.787476434122462e-07, "loss": 2.0684, "step": 12097 }, { "epoch": 3.792476489028213, "grad_norm": 41.28211212158203, "learning_rate": 6.767994290064048e-07, "loss": 2.0291, "step": 12098 }, { "epoch": 3.7927899686520377, "grad_norm": 7.255666255950928, "learning_rate": 6.748535767968503e-07, "loss": 1.8814, "step": 12099 }, { "epoch": 3.793103448275862, "grad_norm": 6.9552001953125, "learning_rate": 6.729100893046897e-07, "loss": 2.1422, "step": 12100 }, { "epoch": 3.7934169278996865, "grad_norm": 7.678218364715576, "learning_rate": 6.709689690479687e-07, "loss": 1.8189, "step": 12101 }, { "epoch": 3.793730407523511, "grad_norm": 6.097766876220703, "learning_rate": 6.690302185416627e-07, "loss": 1.9816, "step": 12102 }, { "epoch": 3.7940438871473354, "grad_norm": 7.718010902404785, "learning_rate": 6.67093840297679e-07, "loss": 2.0883, "step": 12103 }, { "epoch": 3.79435736677116, "grad_norm": 6.740142345428467, "learning_rate": 6.651598368248494e-07, "loss": 1.923, "step": 12104 }, { "epoch": 3.7946708463949843, "grad_norm": 5.810882568359375, "learning_rate": 6.632282106289307e-07, "loss": 2.1465, "step": 12105 }, { "epoch": 3.7949843260188088, "grad_norm": 5.611645698547363, "learning_rate": 6.612989642125977e-07, "loss": 1.9056, "step": 12106 }, { "epoch": 3.795297805642633, "grad_norm": 5.709554195404053, "learning_rate": 6.593721000754449e-07, "loss": 1.8997, "step": 12107 }, { "epoch": 3.7956112852664576, "grad_norm": 104.28226470947266, "learning_rate": 6.574476207139771e-07, "loss": 2.3435, "step": 12108 }, { "epoch": 3.795924764890282, "grad_norm": 8.518017768859863, "learning_rate": 6.555255286216111e-07, "loss": 2.1666, "step": 12109 }, { "epoch": 3.7962382445141065, "grad_norm": 12.420493125915527, "learning_rate": 6.536058262886699e-07, "loss": 2.3659, "step": 12110 }, { "epoch": 3.796551724137931, "grad_norm": 5.67540168762207, "learning_rate": 6.516885162023801e-07, "loss": 1.9012, "step": 12111 }, { "epoch": 3.7968652037617554, "grad_norm": 9.20692253112793, "learning_rate": 6.497736008468703e-07, "loss": 1.9501, "step": 12112 }, { "epoch": 3.79717868338558, "grad_norm": 22.431560516357422, "learning_rate": 6.478610827031648e-07, "loss": 1.9566, "step": 12113 }, { "epoch": 3.7974921630094043, "grad_norm": 35.908721923828125, "learning_rate": 6.459509642491826e-07, "loss": 2.3181, "step": 12114 }, { "epoch": 3.7978056426332287, "grad_norm": 6.3816423416137695, "learning_rate": 6.440432479597333e-07, "loss": 2.1148, "step": 12115 }, { "epoch": 3.798119122257053, "grad_norm": 271.3537292480469, "learning_rate": 6.421379363065142e-07, "loss": 2.597, "step": 12116 }, { "epoch": 3.7984326018808776, "grad_norm": 11.644683837890625, "learning_rate": 6.402350317581083e-07, "loss": 2.0451, "step": 12117 }, { "epoch": 3.798746081504702, "grad_norm": 6.281188488006592, "learning_rate": 6.383345367799784e-07, "loss": 2.1151, "step": 12118 }, { "epoch": 3.7990595611285265, "grad_norm": 5.492709159851074, "learning_rate": 6.364364538344656e-07, "loss": 2.049, "step": 12119 }, { "epoch": 3.799373040752351, "grad_norm": 13.76723861694336, "learning_rate": 6.345407853807864e-07, "loss": 2.2663, "step": 12120 }, { "epoch": 3.7996865203761754, "grad_norm": 5.519260883331299, "learning_rate": 6.326475338750288e-07, "loss": 2.1115, "step": 12121 }, { "epoch": 3.8, "grad_norm": 6.912147045135498, "learning_rate": 6.307567017701482e-07, "loss": 1.952, "step": 12122 }, { "epoch": 3.8003134796238243, "grad_norm": 7.388167858123779, "learning_rate": 6.288682915159685e-07, "loss": 1.9121, "step": 12123 }, { "epoch": 3.8006269592476487, "grad_norm": 8.161437034606934, "learning_rate": 6.269823055591726e-07, "loss": 2.2269, "step": 12124 }, { "epoch": 3.800940438871473, "grad_norm": 9.292981147766113, "learning_rate": 6.250987463433034e-07, "loss": 2.0294, "step": 12125 }, { "epoch": 3.8012539184952976, "grad_norm": 6.878049850463867, "learning_rate": 6.2321761630876e-07, "loss": 2.0568, "step": 12126 }, { "epoch": 3.801567398119122, "grad_norm": 83.84556579589844, "learning_rate": 6.213389178927937e-07, "loss": 2.03, "step": 12127 }, { "epoch": 3.8018808777429465, "grad_norm": 6.2569169998168945, "learning_rate": 6.194626535295059e-07, "loss": 2.0463, "step": 12128 }, { "epoch": 3.802194357366771, "grad_norm": 6.045008182525635, "learning_rate": 6.175888256498436e-07, "loss": 2.18, "step": 12129 }, { "epoch": 3.8025078369905954, "grad_norm": 7.570947170257568, "learning_rate": 6.157174366815979e-07, "loss": 1.6026, "step": 12130 }, { "epoch": 3.80282131661442, "grad_norm": 167.22085571289062, "learning_rate": 6.138484890493989e-07, "loss": 2.5227, "step": 12131 }, { "epoch": 3.8031347962382442, "grad_norm": 4.100212097167969, "learning_rate": 6.119819851747141e-07, "loss": 2.0345, "step": 12132 }, { "epoch": 3.803448275862069, "grad_norm": 6.6342034339904785, "learning_rate": 6.101179274758462e-07, "loss": 2.2195, "step": 12133 }, { "epoch": 3.8037617554858936, "grad_norm": 150.36041259765625, "learning_rate": 6.082563183679263e-07, "loss": 2.2652, "step": 12134 }, { "epoch": 3.804075235109718, "grad_norm": 10.839032173156738, "learning_rate": 6.06397160262914e-07, "loss": 2.0323, "step": 12135 }, { "epoch": 3.8043887147335425, "grad_norm": 24.812395095825195, "learning_rate": 6.045404555695935e-07, "loss": 2.8368, "step": 12136 }, { "epoch": 3.804702194357367, "grad_norm": 120.35324096679688, "learning_rate": 6.026862066935704e-07, "loss": 2.3219, "step": 12137 }, { "epoch": 3.8050156739811913, "grad_norm": 6.914970397949219, "learning_rate": 6.008344160372667e-07, "loss": 2.0033, "step": 12138 }, { "epoch": 3.805329153605016, "grad_norm": 4.8515424728393555, "learning_rate": 5.989850859999227e-07, "loss": 1.9218, "step": 12139 }, { "epoch": 3.80564263322884, "grad_norm": 10.337703704833984, "learning_rate": 5.971382189775885e-07, "loss": 2.4001, "step": 12140 }, { "epoch": 3.8059561128526647, "grad_norm": 8.330846786499023, "learning_rate": 5.952938173631229e-07, "loss": 2.1492, "step": 12141 }, { "epoch": 3.806269592476489, "grad_norm": 9.755550384521484, "learning_rate": 5.934518835461908e-07, "loss": 1.9879, "step": 12142 }, { "epoch": 3.8065830721003135, "grad_norm": 14.728392601013184, "learning_rate": 5.916124199132597e-07, "loss": 2.6272, "step": 12143 }, { "epoch": 3.806896551724138, "grad_norm": 4.949398040771484, "learning_rate": 5.897754288475979e-07, "loss": 1.9406, "step": 12144 }, { "epoch": 3.8072100313479624, "grad_norm": 5.850383281707764, "learning_rate": 5.879409127292685e-07, "loss": 2.1539, "step": 12145 }, { "epoch": 3.807523510971787, "grad_norm": 7.717843532562256, "learning_rate": 5.861088739351287e-07, "loss": 2.0904, "step": 12146 }, { "epoch": 3.8078369905956113, "grad_norm": 6.100648880004883, "learning_rate": 5.842793148388254e-07, "loss": 2.1849, "step": 12147 }, { "epoch": 3.8081504702194358, "grad_norm": 7.629891395568848, "learning_rate": 5.824522378107936e-07, "loss": 2.2774, "step": 12148 }, { "epoch": 3.80846394984326, "grad_norm": 4.877978801727295, "learning_rate": 5.806276452182511e-07, "loss": 2.3464, "step": 12149 }, { "epoch": 3.8087774294670846, "grad_norm": 5.2534050941467285, "learning_rate": 5.788055394251993e-07, "loss": 2.0514, "step": 12150 }, { "epoch": 3.809090909090909, "grad_norm": 10.989152908325195, "learning_rate": 5.769859227924154e-07, "loss": 2.0245, "step": 12151 }, { "epoch": 3.8094043887147335, "grad_norm": 9.333695411682129, "learning_rate": 5.751687976774523e-07, "loss": 2.1814, "step": 12152 }, { "epoch": 3.809717868338558, "grad_norm": 7.784320831298828, "learning_rate": 5.733541664346343e-07, "loss": 2.0346, "step": 12153 }, { "epoch": 3.8100313479623824, "grad_norm": 5.631944179534912, "learning_rate": 5.715420314150547e-07, "loss": 2.1059, "step": 12154 }, { "epoch": 3.810344827586207, "grad_norm": 8.780485153198242, "learning_rate": 5.697323949665742e-07, "loss": 2.0499, "step": 12155 }, { "epoch": 3.8106583072100313, "grad_norm": 6.912583827972412, "learning_rate": 5.679252594338144e-07, "loss": 2.4824, "step": 12156 }, { "epoch": 3.8109717868338557, "grad_norm": 6.673126697540283, "learning_rate": 5.661206271581568e-07, "loss": 2.0574, "step": 12157 }, { "epoch": 3.81128526645768, "grad_norm": 9.058634757995605, "learning_rate": 5.643185004777405e-07, "loss": 2.1747, "step": 12158 }, { "epoch": 3.8115987460815046, "grad_norm": 4.241602420806885, "learning_rate": 5.625188817274569e-07, "loss": 2.3922, "step": 12159 }, { "epoch": 3.811912225705329, "grad_norm": 6.569570541381836, "learning_rate": 5.607217732389503e-07, "loss": 2.0899, "step": 12160 }, { "epoch": 3.8122257053291535, "grad_norm": 6.766139984130859, "learning_rate": 5.589271773406104e-07, "loss": 2.0548, "step": 12161 }, { "epoch": 3.812539184952978, "grad_norm": 5.071262359619141, "learning_rate": 5.571350963575728e-07, "loss": 2.1408, "step": 12162 }, { "epoch": 3.8128526645768024, "grad_norm": 10.048852920532227, "learning_rate": 5.553455326117139e-07, "loss": 2.6881, "step": 12163 }, { "epoch": 3.813166144200627, "grad_norm": 8.327102661132812, "learning_rate": 5.535584884216491e-07, "loss": 2.2188, "step": 12164 }, { "epoch": 3.8134796238244513, "grad_norm": 5.348828315734863, "learning_rate": 5.517739661027297e-07, "loss": 2.1172, "step": 12165 }, { "epoch": 3.8137931034482757, "grad_norm": 44.20182800292969, "learning_rate": 5.499919679670385e-07, "loss": 2.1866, "step": 12166 }, { "epoch": 3.8141065830721, "grad_norm": 7.822056293487549, "learning_rate": 5.482124963233892e-07, "loss": 2.3387, "step": 12167 }, { "epoch": 3.814420062695925, "grad_norm": 5.019737243652344, "learning_rate": 5.464355534773217e-07, "loss": 2.1205, "step": 12168 }, { "epoch": 3.8147335423197495, "grad_norm": 143.85977172851562, "learning_rate": 5.446611417310985e-07, "loss": 1.9546, "step": 12169 }, { "epoch": 3.815047021943574, "grad_norm": 12.709099769592285, "learning_rate": 5.428892633837035e-07, "loss": 2.0685, "step": 12170 }, { "epoch": 3.8153605015673984, "grad_norm": 12.67153263092041, "learning_rate": 5.411199207308396e-07, "loss": 1.9066, "step": 12171 }, { "epoch": 3.815673981191223, "grad_norm": 5.170335292816162, "learning_rate": 5.393531160649221e-07, "loss": 2.1003, "step": 12172 }, { "epoch": 3.8159874608150472, "grad_norm": 15.243967056274414, "learning_rate": 5.375888516750791e-07, "loss": 2.5282, "step": 12173 }, { "epoch": 3.8163009404388717, "grad_norm": 8.597249984741211, "learning_rate": 5.358271298471474e-07, "loss": 1.8955, "step": 12174 }, { "epoch": 3.816614420062696, "grad_norm": 4.055408000946045, "learning_rate": 5.340679528636694e-07, "loss": 2.0695, "step": 12175 }, { "epoch": 3.8169278996865206, "grad_norm": 7.943803787231445, "learning_rate": 5.323113230038899e-07, "loss": 2.117, "step": 12176 }, { "epoch": 3.817241379310345, "grad_norm": 83.83426666259766, "learning_rate": 5.305572425437552e-07, "loss": 1.9867, "step": 12177 }, { "epoch": 3.8175548589341695, "grad_norm": 10.777070045471191, "learning_rate": 5.288057137559066e-07, "loss": 2.0276, "step": 12178 }, { "epoch": 3.817868338557994, "grad_norm": 12.033699035644531, "learning_rate": 5.270567389096807e-07, "loss": 2.015, "step": 12179 }, { "epoch": 3.8181818181818183, "grad_norm": 5.621195316314697, "learning_rate": 5.25310320271104e-07, "loss": 1.9297, "step": 12180 }, { "epoch": 3.818495297805643, "grad_norm": 31.232702255249023, "learning_rate": 5.235664601028911e-07, "loss": 1.9814, "step": 12181 }, { "epoch": 3.8188087774294672, "grad_norm": 7.120140552520752, "learning_rate": 5.218251606644442e-07, "loss": 1.8806, "step": 12182 }, { "epoch": 3.8191222570532917, "grad_norm": 4.324540615081787, "learning_rate": 5.200864242118447e-07, "loss": 2.1923, "step": 12183 }, { "epoch": 3.819435736677116, "grad_norm": 5.9742560386657715, "learning_rate": 5.183502529978548e-07, "loss": 1.7404, "step": 12184 }, { "epoch": 3.8197492163009406, "grad_norm": 7.4850754737854, "learning_rate": 5.166166492719124e-07, "loss": 2.0601, "step": 12185 }, { "epoch": 3.820062695924765, "grad_norm": 12.733499526977539, "learning_rate": 5.14885615280129e-07, "loss": 2.1779, "step": 12186 }, { "epoch": 3.8203761755485894, "grad_norm": 6.8676910400390625, "learning_rate": 5.131571532652885e-07, "loss": 2.0129, "step": 12187 }, { "epoch": 3.820689655172414, "grad_norm": 9.193533897399902, "learning_rate": 5.114312654668397e-07, "loss": 2.2942, "step": 12188 }, { "epoch": 3.8210031347962383, "grad_norm": 150.84127807617188, "learning_rate": 5.097079541208975e-07, "loss": 2.3044, "step": 12189 }, { "epoch": 3.8213166144200628, "grad_norm": 6.830353736877441, "learning_rate": 5.079872214602388e-07, "loss": 1.8443, "step": 12190 }, { "epoch": 3.821630094043887, "grad_norm": 9.869745254516602, "learning_rate": 5.062690697142991e-07, "loss": 2.0262, "step": 12191 }, { "epoch": 3.8219435736677116, "grad_norm": 10.405044555664062, "learning_rate": 5.045535011091693e-07, "loss": 2.1343, "step": 12192 }, { "epoch": 3.822257053291536, "grad_norm": 11.281662940979004, "learning_rate": 5.02840517867596e-07, "loss": 2.7919, "step": 12193 }, { "epoch": 3.8225705329153605, "grad_norm": 6.283397674560547, "learning_rate": 5.011301222089735e-07, "loss": 2.1367, "step": 12194 }, { "epoch": 3.822884012539185, "grad_norm": 10.86446475982666, "learning_rate": 4.994223163493442e-07, "loss": 2.0689, "step": 12195 }, { "epoch": 3.8231974921630094, "grad_norm": 5.753490447998047, "learning_rate": 4.977171025013961e-07, "loss": 2.0576, "step": 12196 }, { "epoch": 3.823510971786834, "grad_norm": 9.706903457641602, "learning_rate": 4.960144828744567e-07, "loss": 2.0664, "step": 12197 }, { "epoch": 3.8238244514106583, "grad_norm": 8.76662826538086, "learning_rate": 4.943144596744956e-07, "loss": 1.9281, "step": 12198 }, { "epoch": 3.8241379310344827, "grad_norm": 6.069772243499756, "learning_rate": 4.926170351041159e-07, "loss": 2.0933, "step": 12199 }, { "epoch": 3.824451410658307, "grad_norm": 8.968098640441895, "learning_rate": 4.909222113625545e-07, "loss": 2.2385, "step": 12200 }, { "epoch": 3.8247648902821316, "grad_norm": 10.437958717346191, "learning_rate": 4.892299906456785e-07, "loss": 2.1399, "step": 12201 }, { "epoch": 3.825078369905956, "grad_norm": 16.756820678710938, "learning_rate": 4.875403751459825e-07, "loss": 2.1679, "step": 12202 }, { "epoch": 3.8253918495297805, "grad_norm": 55.73908233642578, "learning_rate": 4.858533670525861e-07, "loss": 2.0499, "step": 12203 }, { "epoch": 3.825705329153605, "grad_norm": 12.208808898925781, "learning_rate": 4.841689685512299e-07, "loss": 2.4112, "step": 12204 }, { "epoch": 3.8260188087774294, "grad_norm": 13.98959732055664, "learning_rate": 4.82487181824274e-07, "loss": 2.0804, "step": 12205 }, { "epoch": 3.826332288401254, "grad_norm": 7.377553939819336, "learning_rate": 4.808080090506939e-07, "loss": 2.0063, "step": 12206 }, { "epoch": 3.8266457680250783, "grad_norm": 8.18221664428711, "learning_rate": 4.7913145240608e-07, "loss": 2.1644, "step": 12207 }, { "epoch": 3.8269592476489027, "grad_norm": 18.327136993408203, "learning_rate": 4.774575140626317e-07, "loss": 2.3712, "step": 12208 }, { "epoch": 3.827272727272727, "grad_norm": 63.460018157958984, "learning_rate": 4.757861961891561e-07, "loss": 2.2337, "step": 12209 }, { "epoch": 3.8275862068965516, "grad_norm": 7.675913333892822, "learning_rate": 4.741175009510657e-07, "loss": 2.0108, "step": 12210 }, { "epoch": 3.827899686520376, "grad_norm": 96.57157897949219, "learning_rate": 4.7245143051037475e-07, "loss": 2.5631, "step": 12211 }, { "epoch": 3.8282131661442005, "grad_norm": 10.706033706665039, "learning_rate": 4.7078798702569676e-07, "loss": 2.062, "step": 12212 }, { "epoch": 3.828526645768025, "grad_norm": 5.791657447814941, "learning_rate": 4.691271726522409e-07, "loss": 2.0994, "step": 12213 }, { "epoch": 3.8288401253918494, "grad_norm": 5.241148948669434, "learning_rate": 4.6746898954181217e-07, "loss": 2.0418, "step": 12214 }, { "epoch": 3.829153605015674, "grad_norm": 9.076333999633789, "learning_rate": 4.6581343984280466e-07, "loss": 2.2421, "step": 12215 }, { "epoch": 3.8294670846394983, "grad_norm": 5.588638782501221, "learning_rate": 4.6416052570020047e-07, "loss": 2.1749, "step": 12216 }, { "epoch": 3.8297805642633227, "grad_norm": 6.609468936920166, "learning_rate": 4.625102492555675e-07, "loss": 2.2134, "step": 12217 }, { "epoch": 3.830094043887147, "grad_norm": 186.16290283203125, "learning_rate": 4.608626126470561e-07, "loss": 2.3601, "step": 12218 }, { "epoch": 3.8304075235109716, "grad_norm": 5.383998870849609, "learning_rate": 4.59217618009396e-07, "loss": 2.0254, "step": 12219 }, { "epoch": 3.830721003134796, "grad_norm": 4.680808067321777, "learning_rate": 4.5757526747389506e-07, "loss": 2.0178, "step": 12220 }, { "epoch": 3.8310344827586205, "grad_norm": 7.396971702575684, "learning_rate": 4.5593556316843424e-07, "loss": 2.1933, "step": 12221 }, { "epoch": 3.831347962382445, "grad_norm": 11.149065971374512, "learning_rate": 4.5429850721746604e-07, "loss": 2.24, "step": 12222 }, { "epoch": 3.8316614420062693, "grad_norm": 14.328588485717773, "learning_rate": 4.526641017420119e-07, "loss": 1.9881, "step": 12223 }, { "epoch": 3.831974921630094, "grad_norm": 5.239782810211182, "learning_rate": 4.510323488596588e-07, "loss": 2.281, "step": 12224 }, { "epoch": 3.8322884012539182, "grad_norm": 13.881631851196289, "learning_rate": 4.494032506845583e-07, "loss": 2.0289, "step": 12225 }, { "epoch": 3.8326018808777427, "grad_norm": 6.3986968994140625, "learning_rate": 4.4777680932742124e-07, "loss": 2.3035, "step": 12226 }, { "epoch": 3.8329153605015676, "grad_norm": 7.276906967163086, "learning_rate": 4.461530268955161e-07, "loss": 2.1387, "step": 12227 }, { "epoch": 3.833228840125392, "grad_norm": 6.035831451416016, "learning_rate": 4.4453190549266697e-07, "loss": 2.0906, "step": 12228 }, { "epoch": 3.8335423197492164, "grad_norm": 15.118789672851562, "learning_rate": 4.4291344721924976e-07, "loss": 1.8346, "step": 12229 }, { "epoch": 3.833855799373041, "grad_norm": 11.743492126464844, "learning_rate": 4.4129765417219e-07, "loss": 2.0781, "step": 12230 }, { "epoch": 3.8341692789968653, "grad_norm": 6.170013904571533, "learning_rate": 4.396845284449608e-07, "loss": 1.8613, "step": 12231 }, { "epoch": 3.8344827586206898, "grad_norm": 25.096040725708008, "learning_rate": 4.380740721275786e-07, "loss": 2.1051, "step": 12232 }, { "epoch": 3.834796238244514, "grad_norm": 22.2808837890625, "learning_rate": 4.364662873066017e-07, "loss": 2.2782, "step": 12233 }, { "epoch": 3.8351097178683387, "grad_norm": 6.0276570320129395, "learning_rate": 4.348611760651264e-07, "loss": 2.1802, "step": 12234 }, { "epoch": 3.835423197492163, "grad_norm": 10.121356010437012, "learning_rate": 4.3325874048278537e-07, "loss": 2.347, "step": 12235 }, { "epoch": 3.8357366771159875, "grad_norm": 5.481466770172119, "learning_rate": 4.3165898263574593e-07, "loss": 1.9667, "step": 12236 }, { "epoch": 3.836050156739812, "grad_norm": 58.64488983154297, "learning_rate": 4.3006190459670416e-07, "loss": 2.0413, "step": 12237 }, { "epoch": 3.8363636363636364, "grad_norm": 19.534761428833008, "learning_rate": 4.2846750843488523e-07, "loss": 2.1436, "step": 12238 }, { "epoch": 3.836677115987461, "grad_norm": 6.319474220275879, "learning_rate": 4.268757962160386e-07, "loss": 2.3043, "step": 12239 }, { "epoch": 3.8369905956112853, "grad_norm": 29.628721237182617, "learning_rate": 4.252867700024374e-07, "loss": 1.7483, "step": 12240 }, { "epoch": 3.8373040752351097, "grad_norm": 6.422030925750732, "learning_rate": 4.2370043185287397e-07, "loss": 1.9746, "step": 12241 }, { "epoch": 3.837617554858934, "grad_norm": 9.158935546875, "learning_rate": 4.2211678382265836e-07, "loss": 2.1182, "step": 12242 }, { "epoch": 3.8379310344827586, "grad_norm": 7.166996002197266, "learning_rate": 4.205358279636143e-07, "loss": 2.2988, "step": 12243 }, { "epoch": 3.838244514106583, "grad_norm": 6.601081371307373, "learning_rate": 4.189575663240794e-07, "loss": 2.0759, "step": 12244 }, { "epoch": 3.8385579937304075, "grad_norm": 6.280664443969727, "learning_rate": 4.173820009488988e-07, "loss": 2.4094, "step": 12245 }, { "epoch": 3.838871473354232, "grad_norm": 5.6231303215026855, "learning_rate": 4.1580913387942454e-07, "loss": 2.4467, "step": 12246 }, { "epoch": 3.8391849529780564, "grad_norm": 6.663880348205566, "learning_rate": 4.1423896715351323e-07, "loss": 1.7844, "step": 12247 }, { "epoch": 3.839498432601881, "grad_norm": 8.801311492919922, "learning_rate": 4.1267150280552256e-07, "loss": 2.1803, "step": 12248 }, { "epoch": 3.8398119122257053, "grad_norm": 14.236126899719238, "learning_rate": 4.111067428663082e-07, "loss": 2.209, "step": 12249 }, { "epoch": 3.8401253918495297, "grad_norm": 16.15800666809082, "learning_rate": 4.095446893632235e-07, "loss": 2.0062, "step": 12250 }, { "epoch": 3.840438871473354, "grad_norm": 4.983829498291016, "learning_rate": 4.0798534432011317e-07, "loss": 2.3113, "step": 12251 }, { "epoch": 3.8407523510971786, "grad_norm": 29.493921279907227, "learning_rate": 4.0642870975731527e-07, "loss": 2.0138, "step": 12252 }, { "epoch": 3.841065830721003, "grad_norm": 12.004162788391113, "learning_rate": 4.0487478769165396e-07, "loss": 2.1165, "step": 12253 }, { "epoch": 3.8413793103448275, "grad_norm": 6.703609943389893, "learning_rate": 4.033235801364402e-07, "loss": 2.1207, "step": 12254 }, { "epoch": 3.841692789968652, "grad_norm": 9.163110733032227, "learning_rate": 4.017750891014674e-07, "loss": 2.4992, "step": 12255 }, { "epoch": 3.8420062695924764, "grad_norm": 16.743024826049805, "learning_rate": 4.002293165930088e-07, "loss": 1.8748, "step": 12256 }, { "epoch": 3.842319749216301, "grad_norm": 4.951179504394531, "learning_rate": 3.9868626461381707e-07, "loss": 1.904, "step": 12257 }, { "epoch": 3.8426332288401253, "grad_norm": 8.664706230163574, "learning_rate": 3.9714593516311907e-07, "loss": 1.8318, "step": 12258 }, { "epoch": 3.8429467084639497, "grad_norm": 143.24050903320312, "learning_rate": 3.95608330236614e-07, "loss": 1.9982, "step": 12259 }, { "epoch": 3.843260188087774, "grad_norm": 18.860097885131836, "learning_rate": 3.9407345182647133e-07, "loss": 2.0, "step": 12260 }, { "epoch": 3.8435736677115986, "grad_norm": 12.96799373626709, "learning_rate": 3.925413019213281e-07, "loss": 1.9672, "step": 12261 }, { "epoch": 3.8438871473354235, "grad_norm": 6.26040506362915, "learning_rate": 3.9101188250628557e-07, "loss": 2.2213, "step": 12262 }, { "epoch": 3.844200626959248, "grad_norm": 6.775371551513672, "learning_rate": 3.894851955629092e-07, "loss": 2.0754, "step": 12263 }, { "epoch": 3.8445141065830724, "grad_norm": 5.830297470092773, "learning_rate": 3.879612430692223e-07, "loss": 1.8968, "step": 12264 }, { "epoch": 3.844827586206897, "grad_norm": 4.760717868804932, "learning_rate": 3.8644002699970566e-07, "loss": 1.8171, "step": 12265 }, { "epoch": 3.8451410658307212, "grad_norm": 8.86116886138916, "learning_rate": 3.84921549325295e-07, "loss": 2.015, "step": 12266 }, { "epoch": 3.8454545454545457, "grad_norm": 12.608963966369629, "learning_rate": 3.834058120133774e-07, "loss": 2.0447, "step": 12267 }, { "epoch": 3.84576802507837, "grad_norm": 43.80204772949219, "learning_rate": 3.818928170277911e-07, "loss": 2.076, "step": 12268 }, { "epoch": 3.8460815047021946, "grad_norm": 7.239863395690918, "learning_rate": 3.8038256632881963e-07, "loss": 2.2313, "step": 12269 }, { "epoch": 3.846394984326019, "grad_norm": 5.992273807525635, "learning_rate": 3.7887506187319207e-07, "loss": 2.0627, "step": 12270 }, { "epoch": 3.8467084639498434, "grad_norm": 5.619299411773682, "learning_rate": 3.773703056140779e-07, "loss": 2.0158, "step": 12271 }, { "epoch": 3.847021943573668, "grad_norm": 7.53141975402832, "learning_rate": 3.7586829950108787e-07, "loss": 1.9125, "step": 12272 }, { "epoch": 3.8473354231974923, "grad_norm": 8.5288667678833, "learning_rate": 3.743690454802673e-07, "loss": 1.7702, "step": 12273 }, { "epoch": 3.8476489028213168, "grad_norm": 4.713392734527588, "learning_rate": 3.7287254549409846e-07, "loss": 2.3138, "step": 12274 }, { "epoch": 3.847962382445141, "grad_norm": 9.104580879211426, "learning_rate": 3.713788014814934e-07, "loss": 2.0353, "step": 12275 }, { "epoch": 3.8482758620689657, "grad_norm": 7.1483049392700195, "learning_rate": 3.698878153777943e-07, "loss": 2.4541, "step": 12276 }, { "epoch": 3.84858934169279, "grad_norm": 5.717609882354736, "learning_rate": 3.683995891147696e-07, "loss": 1.9972, "step": 12277 }, { "epoch": 3.8489028213166145, "grad_norm": 38.984100341796875, "learning_rate": 3.669141246206126e-07, "loss": 2.103, "step": 12278 }, { "epoch": 3.849216300940439, "grad_norm": 4.129430294036865, "learning_rate": 3.654314238199377e-07, "loss": 1.9847, "step": 12279 }, { "epoch": 3.8495297805642634, "grad_norm": 45.496402740478516, "learning_rate": 3.639514886337786e-07, "loss": 2.0442, "step": 12280 }, { "epoch": 3.849843260188088, "grad_norm": 6.8521857261657715, "learning_rate": 3.62474320979587e-07, "loss": 2.4204, "step": 12281 }, { "epoch": 3.8501567398119123, "grad_norm": 94.84723663330078, "learning_rate": 3.609999227712277e-07, "loss": 2.6019, "step": 12282 }, { "epoch": 3.8504702194357368, "grad_norm": 68.36756896972656, "learning_rate": 3.595282959189775e-07, "loss": 2.3098, "step": 12283 }, { "epoch": 3.850783699059561, "grad_norm": 6.347437381744385, "learning_rate": 3.5805944232952277e-07, "loss": 2.1004, "step": 12284 }, { "epoch": 3.8510971786833856, "grad_norm": 9.063011169433594, "learning_rate": 3.5659336390595654e-07, "loss": 2.0481, "step": 12285 }, { "epoch": 3.85141065830721, "grad_norm": 8.921833992004395, "learning_rate": 3.551300625477763e-07, "loss": 1.9394, "step": 12286 }, { "epoch": 3.8517241379310345, "grad_norm": 8.498764991760254, "learning_rate": 3.536695401508816e-07, "loss": 2.3041, "step": 12287 }, { "epoch": 3.852037617554859, "grad_norm": 8.321864128112793, "learning_rate": 3.5221179860757156e-07, "loss": 2.0203, "step": 12288 }, { "epoch": 3.8523510971786834, "grad_norm": 8.940202713012695, "learning_rate": 3.507568398065414e-07, "loss": 2.1215, "step": 12289 }, { "epoch": 3.852664576802508, "grad_norm": 16.679109573364258, "learning_rate": 3.49304665632883e-07, "loss": 2.1716, "step": 12290 }, { "epoch": 3.8529780564263323, "grad_norm": 6.664984226226807, "learning_rate": 3.478552779680786e-07, "loss": 1.8615, "step": 12291 }, { "epoch": 3.8532915360501567, "grad_norm": 5.228209495544434, "learning_rate": 3.4640867869000036e-07, "loss": 1.8771, "step": 12292 }, { "epoch": 3.853605015673981, "grad_norm": 9.210502624511719, "learning_rate": 3.4496486967290856e-07, "loss": 2.062, "step": 12293 }, { "epoch": 3.8539184952978056, "grad_norm": 11.038147926330566, "learning_rate": 3.435238527874468e-07, "loss": 2.7926, "step": 12294 }, { "epoch": 3.85423197492163, "grad_norm": 17.68816566467285, "learning_rate": 3.4208562990064334e-07, "loss": 2.6712, "step": 12295 }, { "epoch": 3.8545454545454545, "grad_norm": 5.217115879058838, "learning_rate": 3.4065020287590456e-07, "loss": 1.8053, "step": 12296 }, { "epoch": 3.854858934169279, "grad_norm": 9.164361953735352, "learning_rate": 3.392175735730155e-07, "loss": 2.0793, "step": 12297 }, { "epoch": 3.8551724137931034, "grad_norm": 12.501090049743652, "learning_rate": 3.3778774384813555e-07, "loss": 2.5027, "step": 12298 }, { "epoch": 3.855485893416928, "grad_norm": 8.238424301147461, "learning_rate": 3.3636071555379737e-07, "loss": 1.9717, "step": 12299 }, { "epoch": 3.8557993730407523, "grad_norm": 5.817928314208984, "learning_rate": 3.3493649053890325e-07, "loss": 2.3103, "step": 12300 }, { "epoch": 3.8561128526645767, "grad_norm": 5.130991458892822, "learning_rate": 3.335150706487256e-07, "loss": 1.911, "step": 12301 }, { "epoch": 3.856426332288401, "grad_norm": 8.670504570007324, "learning_rate": 3.320964577249003e-07, "loss": 2.2068, "step": 12302 }, { "epoch": 3.8567398119122256, "grad_norm": 9.235809326171875, "learning_rate": 3.306806536054269e-07, "loss": 2.2303, "step": 12303 }, { "epoch": 3.85705329153605, "grad_norm": 47.71424102783203, "learning_rate": 3.292676601246661e-07, "loss": 1.9612, "step": 12304 }, { "epoch": 3.8573667711598745, "grad_norm": 5.3874945640563965, "learning_rate": 3.278574791133363e-07, "loss": 2.1636, "step": 12305 }, { "epoch": 3.857680250783699, "grad_norm": 9.006978988647461, "learning_rate": 3.2645011239851365e-07, "loss": 2.3945, "step": 12306 }, { "epoch": 3.8579937304075234, "grad_norm": 8.97449016571045, "learning_rate": 3.2504556180362664e-07, "loss": 2.0181, "step": 12307 }, { "epoch": 3.858307210031348, "grad_norm": 164.50733947753906, "learning_rate": 3.2364382914845504e-07, "loss": 2.2677, "step": 12308 }, { "epoch": 3.8586206896551722, "grad_norm": 7.754162788391113, "learning_rate": 3.2224491624912843e-07, "loss": 2.3184, "step": 12309 }, { "epoch": 3.8589341692789967, "grad_norm": 10.392531394958496, "learning_rate": 3.208488249181216e-07, "loss": 2.1114, "step": 12310 }, { "epoch": 3.859247648902821, "grad_norm": 6.9014434814453125, "learning_rate": 3.19455556964256e-07, "loss": 2.0248, "step": 12311 }, { "epoch": 3.8595611285266456, "grad_norm": 8.91592025756836, "learning_rate": 3.18065114192693e-07, "loss": 2.1883, "step": 12312 }, { "epoch": 3.85987460815047, "grad_norm": 66.173828125, "learning_rate": 3.1667749840493424e-07, "loss": 2.6351, "step": 12313 }, { "epoch": 3.8601880877742945, "grad_norm": 6.796383380889893, "learning_rate": 3.152927113988186e-07, "loss": 1.9691, "step": 12314 }, { "epoch": 3.860501567398119, "grad_norm": 6.531740665435791, "learning_rate": 3.139107549685194e-07, "loss": 2.0644, "step": 12315 }, { "epoch": 3.8608150470219433, "grad_norm": 126.80338287353516, "learning_rate": 3.125316309045434e-07, "loss": 2.3282, "step": 12316 }, { "epoch": 3.8611285266457678, "grad_norm": 7.126059055328369, "learning_rate": 3.111553409937279e-07, "loss": 1.5979, "step": 12317 }, { "epoch": 3.861442006269592, "grad_norm": 7.986970901489258, "learning_rate": 3.0978188701923705e-07, "loss": 2.0915, "step": 12318 }, { "epoch": 3.8617554858934167, "grad_norm": 38.70000076293945, "learning_rate": 3.084112707605613e-07, "loss": 2.4987, "step": 12319 }, { "epoch": 3.862068965517241, "grad_norm": 7.533967018127441, "learning_rate": 3.0704349399351437e-07, "loss": 2.909, "step": 12320 }, { "epoch": 3.862382445141066, "grad_norm": 10.108235359191895, "learning_rate": 3.0567855849023136e-07, "loss": 2.7036, "step": 12321 }, { "epoch": 3.8626959247648904, "grad_norm": 5.891214847564697, "learning_rate": 3.0431646601916556e-07, "loss": 2.1868, "step": 12322 }, { "epoch": 3.863009404388715, "grad_norm": 6.993232250213623, "learning_rate": 3.0295721834508686e-07, "loss": 1.8781, "step": 12323 }, { "epoch": 3.8633228840125393, "grad_norm": 28.805322647094727, "learning_rate": 3.016008172290799e-07, "loss": 2.3702, "step": 12324 }, { "epoch": 3.8636363636363638, "grad_norm": 9.001139640808105, "learning_rate": 3.0024726442854063e-07, "loss": 2.2023, "step": 12325 }, { "epoch": 3.863949843260188, "grad_norm": 8.320252418518066, "learning_rate": 2.988965616971748e-07, "loss": 1.6873, "step": 12326 }, { "epoch": 3.8642633228840126, "grad_norm": 6.237418174743652, "learning_rate": 2.975487107849953e-07, "loss": 2.1662, "step": 12327 }, { "epoch": 3.864576802507837, "grad_norm": 142.51400756835938, "learning_rate": 2.962037134383211e-07, "loss": 1.983, "step": 12328 }, { "epoch": 3.8648902821316615, "grad_norm": 8.941393852233887, "learning_rate": 2.9486157139977294e-07, "loss": 1.9199, "step": 12329 }, { "epoch": 3.865203761755486, "grad_norm": 13.267260551452637, "learning_rate": 2.9352228640827235e-07, "loss": 2.0376, "step": 12330 }, { "epoch": 3.8655172413793104, "grad_norm": 37.44633483886719, "learning_rate": 2.921858601990396e-07, "loss": 2.1152, "step": 12331 }, { "epoch": 3.865830721003135, "grad_norm": 7.167781829833984, "learning_rate": 2.9085229450358967e-07, "loss": 2.0385, "step": 12332 }, { "epoch": 3.8661442006269593, "grad_norm": 7.920291900634766, "learning_rate": 2.89521591049734e-07, "loss": 2.1272, "step": 12333 }, { "epoch": 3.8664576802507837, "grad_norm": 10.627370834350586, "learning_rate": 2.881937515615732e-07, "loss": 2.1563, "step": 12334 }, { "epoch": 3.866771159874608, "grad_norm": 8.874905586242676, "learning_rate": 2.868687777594983e-07, "loss": 2.1319, "step": 12335 }, { "epoch": 3.8670846394984326, "grad_norm": 13.720527648925781, "learning_rate": 2.855466713601868e-07, "loss": 2.139, "step": 12336 }, { "epoch": 3.867398119122257, "grad_norm": 7.94887638092041, "learning_rate": 2.842274340766016e-07, "loss": 2.0677, "step": 12337 }, { "epoch": 3.8677115987460815, "grad_norm": 21.36457061767578, "learning_rate": 2.829110676179889e-07, "loss": 2.4199, "step": 12338 }, { "epoch": 3.868025078369906, "grad_norm": 7.6656036376953125, "learning_rate": 2.815975736898743e-07, "loss": 2.0483, "step": 12339 }, { "epoch": 3.8683385579937304, "grad_norm": 11.423320770263672, "learning_rate": 2.80286953994062e-07, "loss": 2.1164, "step": 12340 }, { "epoch": 3.868652037617555, "grad_norm": 15.867389678955078, "learning_rate": 2.789792102286323e-07, "loss": 2.3271, "step": 12341 }, { "epoch": 3.8689655172413793, "grad_norm": 5.178081512451172, "learning_rate": 2.7767434408793955e-07, "loss": 2.1405, "step": 12342 }, { "epoch": 3.8692789968652037, "grad_norm": 10.410581588745117, "learning_rate": 2.763723572626087e-07, "loss": 2.0951, "step": 12343 }, { "epoch": 3.869592476489028, "grad_norm": 4.649520397186279, "learning_rate": 2.750732514395363e-07, "loss": 1.894, "step": 12344 }, { "epoch": 3.8699059561128526, "grad_norm": 7.076162815093994, "learning_rate": 2.7377702830188426e-07, "loss": 1.8803, "step": 12345 }, { "epoch": 3.870219435736677, "grad_norm": 4.700688362121582, "learning_rate": 2.7248368952908055e-07, "loss": 2.1876, "step": 12346 }, { "epoch": 3.8705329153605015, "grad_norm": 6.272420883178711, "learning_rate": 2.7119323679681535e-07, "loss": 2.1234, "step": 12347 }, { "epoch": 3.870846394984326, "grad_norm": 10.212081909179688, "learning_rate": 2.6990567177703984e-07, "loss": 1.9393, "step": 12348 }, { "epoch": 3.8711598746081504, "grad_norm": 5.398184299468994, "learning_rate": 2.686209961379646e-07, "loss": 2.1812, "step": 12349 }, { "epoch": 3.871473354231975, "grad_norm": 5.995745658874512, "learning_rate": 2.6733921154405594e-07, "loss": 2.0955, "step": 12350 }, { "epoch": 3.8717868338557992, "grad_norm": 8.727567672729492, "learning_rate": 2.660603196560341e-07, "loss": 2.1357, "step": 12351 }, { "epoch": 3.8721003134796237, "grad_norm": 8.392951965332031, "learning_rate": 2.647843221308721e-07, "loss": 1.9176, "step": 12352 }, { "epoch": 3.872413793103448, "grad_norm": 6.0438337326049805, "learning_rate": 2.635112206217927e-07, "loss": 2.1213, "step": 12353 }, { "epoch": 3.8727272727272726, "grad_norm": 6.206927299499512, "learning_rate": 2.6224101677826544e-07, "loss": 2.1427, "step": 12354 }, { "epoch": 3.873040752351097, "grad_norm": 236.33522033691406, "learning_rate": 2.609737122460082e-07, "loss": 2.4702, "step": 12355 }, { "epoch": 3.873354231974922, "grad_norm": 10.614058494567871, "learning_rate": 2.5970930866697997e-07, "loss": 2.7589, "step": 12356 }, { "epoch": 3.8736677115987463, "grad_norm": 10.405721664428711, "learning_rate": 2.584478076793823e-07, "loss": 2.4142, "step": 12357 }, { "epoch": 3.873981191222571, "grad_norm": 7.396894931793213, "learning_rate": 2.571892109176552e-07, "loss": 2.3265, "step": 12358 }, { "epoch": 3.8742946708463952, "grad_norm": 12.907299995422363, "learning_rate": 2.5593352001247685e-07, "loss": 2.0371, "step": 12359 }, { "epoch": 3.8746081504702197, "grad_norm": 30.53795623779297, "learning_rate": 2.5468073659076e-07, "loss": 2.119, "step": 12360 }, { "epoch": 3.874921630094044, "grad_norm": 9.367634773254395, "learning_rate": 2.5343086227565037e-07, "loss": 1.9648, "step": 12361 }, { "epoch": 3.8752351097178686, "grad_norm": 10.085928916931152, "learning_rate": 2.5218389868652436e-07, "loss": 2.003, "step": 12362 }, { "epoch": 3.875548589341693, "grad_norm": 10.780524253845215, "learning_rate": 2.509398474389882e-07, "loss": 2.3568, "step": 12363 }, { "epoch": 3.8758620689655174, "grad_norm": 11.271638870239258, "learning_rate": 2.496987101448728e-07, "loss": 2.3218, "step": 12364 }, { "epoch": 3.876175548589342, "grad_norm": 390.8479919433594, "learning_rate": 2.48460488412236e-07, "loss": 2.8784, "step": 12365 }, { "epoch": 3.8764890282131663, "grad_norm": 9.336274147033691, "learning_rate": 2.4722518384535653e-07, "loss": 2.1342, "step": 12366 }, { "epoch": 3.8768025078369908, "grad_norm": 6.797178745269775, "learning_rate": 2.4599279804473407e-07, "loss": 2.2765, "step": 12367 }, { "epoch": 3.877115987460815, "grad_norm": 122.47384643554688, "learning_rate": 2.44763332607087e-07, "loss": 2.5503, "step": 12368 }, { "epoch": 3.8774294670846396, "grad_norm": 21.827131271362305, "learning_rate": 2.43536789125349e-07, "loss": 2.0516, "step": 12369 }, { "epoch": 3.877742946708464, "grad_norm": 6.386552333831787, "learning_rate": 2.423131691886682e-07, "loss": 2.3239, "step": 12370 }, { "epoch": 3.8780564263322885, "grad_norm": 87.83345794677734, "learning_rate": 2.410924743824067e-07, "loss": 2.2856, "step": 12371 }, { "epoch": 3.878369905956113, "grad_norm": 9.303428649902344, "learning_rate": 2.398747062881343e-07, "loss": 2.4317, "step": 12372 }, { "epoch": 3.8786833855799374, "grad_norm": 8.866641998291016, "learning_rate": 2.386598664836298e-07, "loss": 2.0685, "step": 12373 }, { "epoch": 3.878996865203762, "grad_norm": 7.986985683441162, "learning_rate": 2.3744795654287834e-07, "loss": 1.8044, "step": 12374 }, { "epoch": 3.8793103448275863, "grad_norm": 7.299438953399658, "learning_rate": 2.3623897803606828e-07, "loss": 2.3872, "step": 12375 }, { "epoch": 3.8796238244514107, "grad_norm": 10.429910659790039, "learning_rate": 2.3503293252959136e-07, "loss": 2.4688, "step": 12376 }, { "epoch": 3.879937304075235, "grad_norm": 6.523449420928955, "learning_rate": 2.338298215860374e-07, "loss": 2.0812, "step": 12377 }, { "epoch": 3.8802507836990596, "grad_norm": 6.559314250946045, "learning_rate": 2.326296467641953e-07, "loss": 2.1181, "step": 12378 }, { "epoch": 3.880564263322884, "grad_norm": 6.979979991912842, "learning_rate": 2.3143240961904927e-07, "loss": 1.8837, "step": 12379 }, { "epoch": 3.8808777429467085, "grad_norm": 5.891752243041992, "learning_rate": 2.302381117017774e-07, "loss": 1.9968, "step": 12380 }, { "epoch": 3.881191222570533, "grad_norm": 122.4478530883789, "learning_rate": 2.2904675455975077e-07, "loss": 2.4714, "step": 12381 }, { "epoch": 3.8815047021943574, "grad_norm": 10.108304023742676, "learning_rate": 2.278583397365286e-07, "loss": 2.2427, "step": 12382 }, { "epoch": 3.881818181818182, "grad_norm": 5.558844089508057, "learning_rate": 2.2667286877185862e-07, "loss": 1.9614, "step": 12383 }, { "epoch": 3.8821316614420063, "grad_norm": 6.55927848815918, "learning_rate": 2.2549034320167501e-07, "loss": 2.0117, "step": 12384 }, { "epoch": 3.8824451410658307, "grad_norm": 10.165188789367676, "learning_rate": 2.243107645580947e-07, "loss": 2.2467, "step": 12385 }, { "epoch": 3.882758620689655, "grad_norm": 5.552518844604492, "learning_rate": 2.2313413436941712e-07, "loss": 2.1315, "step": 12386 }, { "epoch": 3.8830721003134796, "grad_norm": 6.142151355743408, "learning_rate": 2.2196045416012225e-07, "loss": 2.0026, "step": 12387 }, { "epoch": 3.883385579937304, "grad_norm": 8.623331069946289, "learning_rate": 2.2078972545086647e-07, "loss": 2.0891, "step": 12388 }, { "epoch": 3.8836990595611285, "grad_norm": 63.07131576538086, "learning_rate": 2.196219497584834e-07, "loss": 1.9167, "step": 12389 }, { "epoch": 3.884012539184953, "grad_norm": 40.17546844482422, "learning_rate": 2.1845712859597995e-07, "loss": 2.4147, "step": 12390 }, { "epoch": 3.8843260188087774, "grad_norm": 7.367454528808594, "learning_rate": 2.172952634725345e-07, "loss": 2.0986, "step": 12391 }, { "epoch": 3.884639498432602, "grad_norm": 153.94093322753906, "learning_rate": 2.1613635589349756e-07, "loss": 2.5964, "step": 12392 }, { "epoch": 3.8849529780564263, "grad_norm": 6.758372783660889, "learning_rate": 2.149804073603859e-07, "loss": 2.0604, "step": 12393 }, { "epoch": 3.8852664576802507, "grad_norm": 7.149299144744873, "learning_rate": 2.138274193708828e-07, "loss": 2.1315, "step": 12394 }, { "epoch": 3.885579937304075, "grad_norm": 6.6694536209106445, "learning_rate": 2.1267739341883586e-07, "loss": 1.973, "step": 12395 }, { "epoch": 3.8858934169278996, "grad_norm": 7.293829917907715, "learning_rate": 2.115303309942554e-07, "loss": 2.2414, "step": 12396 }, { "epoch": 3.886206896551724, "grad_norm": 7.804798126220703, "learning_rate": 2.103862335833115e-07, "loss": 1.9086, "step": 12397 }, { "epoch": 3.8865203761755485, "grad_norm": 7.323897361755371, "learning_rate": 2.0924510266833286e-07, "loss": 1.9436, "step": 12398 }, { "epoch": 3.886833855799373, "grad_norm": 5.731020450592041, "learning_rate": 2.0810693972780516e-07, "loss": 1.8287, "step": 12399 }, { "epoch": 3.8871473354231973, "grad_norm": 39.48750305175781, "learning_rate": 2.0697174623636795e-07, "loss": 2.0029, "step": 12400 }, { "epoch": 3.887460815047022, "grad_norm": 7.690300464630127, "learning_rate": 2.0583952366481365e-07, "loss": 1.864, "step": 12401 }, { "epoch": 3.8877742946708462, "grad_norm": 5.035684585571289, "learning_rate": 2.047102734800857e-07, "loss": 2.2418, "step": 12402 }, { "epoch": 3.8880877742946707, "grad_norm": 8.786688804626465, "learning_rate": 2.035839971452766e-07, "loss": 1.9506, "step": 12403 }, { "epoch": 3.888401253918495, "grad_norm": 7.7487030029296875, "learning_rate": 2.0246069611962544e-07, "loss": 2.1603, "step": 12404 }, { "epoch": 3.8887147335423196, "grad_norm": 8.400206565856934, "learning_rate": 2.0134037185851662e-07, "loss": 2.1259, "step": 12405 }, { "epoch": 3.889028213166144, "grad_norm": 7.953679084777832, "learning_rate": 2.002230258134777e-07, "loss": 2.0573, "step": 12406 }, { "epoch": 3.8893416927899684, "grad_norm": 6.720125198364258, "learning_rate": 1.9910865943217672e-07, "loss": 2.1677, "step": 12407 }, { "epoch": 3.889655172413793, "grad_norm": 9.31767749786377, "learning_rate": 1.9799727415842323e-07, "loss": 2.3579, "step": 12408 }, { "epoch": 3.8899686520376173, "grad_norm": 9.771011352539062, "learning_rate": 1.968888714321626e-07, "loss": 2.389, "step": 12409 }, { "epoch": 3.8902821316614418, "grad_norm": 5.8086676597595215, "learning_rate": 1.957834526894764e-07, "loss": 2.169, "step": 12410 }, { "epoch": 3.890595611285266, "grad_norm": 7.373335838317871, "learning_rate": 1.9468101936258006e-07, "loss": 1.9799, "step": 12411 }, { "epoch": 3.8909090909090907, "grad_norm": 6.591977596282959, "learning_rate": 1.9358157287982099e-07, "loss": 2.3062, "step": 12412 }, { "epoch": 3.891222570532915, "grad_norm": 8.72375202178955, "learning_rate": 1.9248511466567643e-07, "loss": 2.0428, "step": 12413 }, { "epoch": 3.8915360501567395, "grad_norm": 125.18231964111328, "learning_rate": 1.9139164614075306e-07, "loss": 2.1035, "step": 12414 }, { "epoch": 3.8918495297805644, "grad_norm": 5.802567958831787, "learning_rate": 1.9030116872178317e-07, "loss": 2.0278, "step": 12415 }, { "epoch": 3.892163009404389, "grad_norm": 6.546736717224121, "learning_rate": 1.8921368382162352e-07, "loss": 1.9259, "step": 12416 }, { "epoch": 3.8924764890282133, "grad_norm": 6.053040504455566, "learning_rate": 1.88129192849254e-07, "loss": 2.1339, "step": 12417 }, { "epoch": 3.8927899686520377, "grad_norm": 17.16111946105957, "learning_rate": 1.8704769720977484e-07, "loss": 2.2959, "step": 12418 }, { "epoch": 3.893103448275862, "grad_norm": 9.416359901428223, "learning_rate": 1.859691983044068e-07, "loss": 2.1525, "step": 12419 }, { "epoch": 3.8934169278996866, "grad_norm": 32.17069625854492, "learning_rate": 1.8489369753048686e-07, "loss": 2.2477, "step": 12420 }, { "epoch": 3.893730407523511, "grad_norm": 6.723576068878174, "learning_rate": 1.838211962814679e-07, "loss": 2.0873, "step": 12421 }, { "epoch": 3.8940438871473355, "grad_norm": 7.425007343292236, "learning_rate": 1.827516959469164e-07, "loss": 1.8971, "step": 12422 }, { "epoch": 3.89435736677116, "grad_norm": 8.076188087463379, "learning_rate": 1.816851979125106e-07, "loss": 2.45, "step": 12423 }, { "epoch": 3.8946708463949844, "grad_norm": 8.671245574951172, "learning_rate": 1.8062170356003854e-07, "loss": 2.3328, "step": 12424 }, { "epoch": 3.894984326018809, "grad_norm": 15.397347450256348, "learning_rate": 1.795612142673986e-07, "loss": 2.111, "step": 12425 }, { "epoch": 3.8952978056426333, "grad_norm": 5.021137237548828, "learning_rate": 1.78503731408593e-07, "loss": 2.2996, "step": 12426 }, { "epoch": 3.8956112852664577, "grad_norm": 5.816932201385498, "learning_rate": 1.7744925635373056e-07, "loss": 1.8974, "step": 12427 }, { "epoch": 3.895924764890282, "grad_norm": 6.388751029968262, "learning_rate": 1.7639779046902228e-07, "loss": 1.9187, "step": 12428 }, { "epoch": 3.8962382445141066, "grad_norm": 10.3386812210083, "learning_rate": 1.753493351167801e-07, "loss": 2.241, "step": 12429 }, { "epoch": 3.896551724137931, "grad_norm": 4.957164287567139, "learning_rate": 1.743038916554171e-07, "loss": 2.1414, "step": 12430 }, { "epoch": 3.8968652037617555, "grad_norm": 8.43482494354248, "learning_rate": 1.732614614394418e-07, "loss": 2.2054, "step": 12431 }, { "epoch": 3.89717868338558, "grad_norm": 6.10299015045166, "learning_rate": 1.7222204581946038e-07, "loss": 1.9955, "step": 12432 }, { "epoch": 3.8974921630094044, "grad_norm": 8.415992736816406, "learning_rate": 1.711856461421721e-07, "loss": 2.4639, "step": 12433 }, { "epoch": 3.897805642633229, "grad_norm": 17.298070907592773, "learning_rate": 1.7015226375036942e-07, "loss": 2.0931, "step": 12434 }, { "epoch": 3.8981191222570533, "grad_norm": 7.084414005279541, "learning_rate": 1.6912189998293526e-07, "loss": 2.0908, "step": 12435 }, { "epoch": 3.8984326018808777, "grad_norm": 8.1871919631958, "learning_rate": 1.680945561748412e-07, "loss": 2.0789, "step": 12436 }, { "epoch": 3.898746081504702, "grad_norm": 6.51884651184082, "learning_rate": 1.670702336571467e-07, "loss": 2.1735, "step": 12437 }, { "epoch": 3.8990595611285266, "grad_norm": 7.3128581047058105, "learning_rate": 1.6604893375699594e-07, "loss": 2.0369, "step": 12438 }, { "epoch": 3.899373040752351, "grad_norm": 10.92259407043457, "learning_rate": 1.6503065779761796e-07, "loss": 2.4639, "step": 12439 }, { "epoch": 3.8996865203761755, "grad_norm": 63.86602783203125, "learning_rate": 1.6401540709832242e-07, "loss": 2.0434, "step": 12440 }, { "epoch": 3.9, "grad_norm": 6.510636806488037, "learning_rate": 1.6300318297450102e-07, "loss": 2.0737, "step": 12441 }, { "epoch": 3.9003134796238244, "grad_norm": 11.581326484680176, "learning_rate": 1.6199398673762328e-07, "loss": 2.2675, "step": 12442 }, { "epoch": 3.900626959247649, "grad_norm": 5.930288314819336, "learning_rate": 1.6098781969523524e-07, "loss": 1.9786, "step": 12443 }, { "epoch": 3.9009404388714732, "grad_norm": 10.57120418548584, "learning_rate": 1.5998468315095938e-07, "loss": 2.0055, "step": 12444 }, { "epoch": 3.9012539184952977, "grad_norm": 8.299304962158203, "learning_rate": 1.5898457840448994e-07, "loss": 2.0413, "step": 12445 }, { "epoch": 3.901567398119122, "grad_norm": 6.308582305908203, "learning_rate": 1.5798750675159569e-07, "loss": 2.1148, "step": 12446 }, { "epoch": 3.9018808777429466, "grad_norm": 7.528541564941406, "learning_rate": 1.5699346948411354e-07, "loss": 1.9397, "step": 12447 }, { "epoch": 3.902194357366771, "grad_norm": 5.5431294441223145, "learning_rate": 1.5600246788994938e-07, "loss": 1.9948, "step": 12448 }, { "epoch": 3.9025078369905954, "grad_norm": 6.08336877822876, "learning_rate": 1.5501450325307644e-07, "loss": 2.2364, "step": 12449 }, { "epoch": 3.9028213166144203, "grad_norm": 88.39453125, "learning_rate": 1.5402957685353298e-07, "loss": 2.6213, "step": 12450 }, { "epoch": 3.9031347962382448, "grad_norm": 5.89178991317749, "learning_rate": 1.530476899674202e-07, "loss": 2.0016, "step": 12451 }, { "epoch": 3.903448275862069, "grad_norm": 124.23216247558594, "learning_rate": 1.5206884386690296e-07, "loss": 2.4088, "step": 12452 }, { "epoch": 3.9037617554858937, "grad_norm": 7.824604511260986, "learning_rate": 1.5109303982020402e-07, "loss": 2.289, "step": 12453 }, { "epoch": 3.904075235109718, "grad_norm": 31.46741485595703, "learning_rate": 1.5012027909160675e-07, "loss": 2.2894, "step": 12454 }, { "epoch": 3.9043887147335425, "grad_norm": 8.259088516235352, "learning_rate": 1.4915056294145024e-07, "loss": 1.9871, "step": 12455 }, { "epoch": 3.904702194357367, "grad_norm": 66.3331069946289, "learning_rate": 1.4818389262612948e-07, "loss": 2.3322, "step": 12456 }, { "epoch": 3.9050156739811914, "grad_norm": 5.395639419555664, "learning_rate": 1.4722026939809314e-07, "loss": 1.9555, "step": 12457 }, { "epoch": 3.905329153605016, "grad_norm": 7.655972003936768, "learning_rate": 1.4625969450584198e-07, "loss": 2.5931, "step": 12458 }, { "epoch": 3.9056426332288403, "grad_norm": 5.5763444900512695, "learning_rate": 1.4530216919392742e-07, "loss": 2.2863, "step": 12459 }, { "epoch": 3.9059561128526648, "grad_norm": 5.934469223022461, "learning_rate": 1.44347694702949e-07, "loss": 2.0314, "step": 12460 }, { "epoch": 3.906269592476489, "grad_norm": 11.137063026428223, "learning_rate": 1.4339627226955394e-07, "loss": 2.0541, "step": 12461 }, { "epoch": 3.9065830721003136, "grad_norm": 8.915059089660645, "learning_rate": 1.4244790312643587e-07, "loss": 2.1827, "step": 12462 }, { "epoch": 3.906896551724138, "grad_norm": 5.620582580566406, "learning_rate": 1.4150258850233163e-07, "loss": 2.5181, "step": 12463 }, { "epoch": 3.9072100313479625, "grad_norm": 6.776147365570068, "learning_rate": 1.4056032962202038e-07, "loss": 2.0145, "step": 12464 }, { "epoch": 3.907523510971787, "grad_norm": 5.9330735206604, "learning_rate": 1.3962112770632308e-07, "loss": 2.6215, "step": 12465 }, { "epoch": 3.9078369905956114, "grad_norm": 8.184157371520996, "learning_rate": 1.3868498397209884e-07, "loss": 1.9485, "step": 12466 }, { "epoch": 3.908150470219436, "grad_norm": 6.879455089569092, "learning_rate": 1.3775189963224493e-07, "loss": 1.826, "step": 12467 }, { "epoch": 3.9084639498432603, "grad_norm": 9.796491622924805, "learning_rate": 1.3682187589569573e-07, "loss": 1.9218, "step": 12468 }, { "epoch": 3.9087774294670847, "grad_norm": 10.502157211303711, "learning_rate": 1.35894913967419e-07, "loss": 2.2236, "step": 12469 }, { "epoch": 3.909090909090909, "grad_norm": 10.189908027648926, "learning_rate": 1.349710150484157e-07, "loss": 2.0398, "step": 12470 }, { "epoch": 3.9094043887147336, "grad_norm": 7.004083633422852, "learning_rate": 1.3405018033571887e-07, "loss": 2.1295, "step": 12471 }, { "epoch": 3.909717868338558, "grad_norm": 13.977380752563477, "learning_rate": 1.3313241102239056e-07, "loss": 2.2937, "step": 12472 }, { "epoch": 3.9100313479623825, "grad_norm": 14.094507217407227, "learning_rate": 1.3221770829752233e-07, "loss": 2.387, "step": 12473 }, { "epoch": 3.910344827586207, "grad_norm": 10.971450805664062, "learning_rate": 1.3130607334623125e-07, "loss": 1.9878, "step": 12474 }, { "epoch": 3.9106583072100314, "grad_norm": 9.939473152160645, "learning_rate": 1.3039750734966077e-07, "loss": 2.3272, "step": 12475 }, { "epoch": 3.910971786833856, "grad_norm": 9.1459379196167, "learning_rate": 1.2949201148497792e-07, "loss": 2.1362, "step": 12476 }, { "epoch": 3.9112852664576803, "grad_norm": 8.442465782165527, "learning_rate": 1.285895869253709e-07, "loss": 2.2215, "step": 12477 }, { "epoch": 3.9115987460815047, "grad_norm": 5.189630508422852, "learning_rate": 1.2769023484005049e-07, "loss": 2.1395, "step": 12478 }, { "epoch": 3.911912225705329, "grad_norm": 6.689826488494873, "learning_rate": 1.2679395639424525e-07, "loss": 2.1124, "step": 12479 }, { "epoch": 3.9122257053291536, "grad_norm": 6.800660133361816, "learning_rate": 1.2590075274920206e-07, "loss": 2.0225, "step": 12480 }, { "epoch": 3.912539184952978, "grad_norm": 5.358881950378418, "learning_rate": 1.2501062506218354e-07, "loss": 1.9474, "step": 12481 }, { "epoch": 3.9128526645768025, "grad_norm": 7.679074764251709, "learning_rate": 1.2412357448646716e-07, "loss": 2.0559, "step": 12482 }, { "epoch": 3.913166144200627, "grad_norm": 6.113692283630371, "learning_rate": 1.2323960217134362e-07, "loss": 2.2461, "step": 12483 }, { "epoch": 3.9134796238244514, "grad_norm": 8.154008865356445, "learning_rate": 1.223587092621162e-07, "loss": 2.0321, "step": 12484 }, { "epoch": 3.913793103448276, "grad_norm": 3.832369327545166, "learning_rate": 1.214808969000969e-07, "loss": 2.1519, "step": 12485 }, { "epoch": 3.9141065830721002, "grad_norm": 28.701013565063477, "learning_rate": 1.2060616622260712e-07, "loss": 2.3403, "step": 12486 }, { "epoch": 3.9144200626959247, "grad_norm": 41.665374755859375, "learning_rate": 1.1973451836297612e-07, "loss": 2.1839, "step": 12487 }, { "epoch": 3.914733542319749, "grad_norm": 7.900162220001221, "learning_rate": 1.1886595445053745e-07, "loss": 1.8629, "step": 12488 }, { "epoch": 3.9150470219435736, "grad_norm": 10.815903663635254, "learning_rate": 1.1800047561063094e-07, "loss": 2.164, "step": 12489 }, { "epoch": 3.915360501567398, "grad_norm": 6.596642017364502, "learning_rate": 1.1713808296459794e-07, "loss": 2.2244, "step": 12490 }, { "epoch": 3.9156739811912225, "grad_norm": 5.763444423675537, "learning_rate": 1.162787776297819e-07, "loss": 2.1566, "step": 12491 }, { "epoch": 3.915987460815047, "grad_norm": 6.763387203216553, "learning_rate": 1.1542256071952529e-07, "loss": 2.214, "step": 12492 }, { "epoch": 3.9163009404388713, "grad_norm": 5.740962505340576, "learning_rate": 1.1456943334317044e-07, "loss": 2.1649, "step": 12493 }, { "epoch": 3.916614420062696, "grad_norm": 7.866600036621094, "learning_rate": 1.1371939660605568e-07, "loss": 2.011, "step": 12494 }, { "epoch": 3.91692789968652, "grad_norm": 6.798871040344238, "learning_rate": 1.1287245160951615e-07, "loss": 1.9507, "step": 12495 }, { "epoch": 3.9172413793103447, "grad_norm": 6.266920566558838, "learning_rate": 1.120285994508799e-07, "loss": 2.642, "step": 12496 }, { "epoch": 3.917554858934169, "grad_norm": 99.69943237304688, "learning_rate": 1.1118784122346877e-07, "loss": 2.2228, "step": 12497 }, { "epoch": 3.9178683385579935, "grad_norm": 8.083369255065918, "learning_rate": 1.1035017801659554e-07, "loss": 2.3472, "step": 12498 }, { "epoch": 3.918181818181818, "grad_norm": 5.158436298370361, "learning_rate": 1.095156109155629e-07, "loss": 1.8112, "step": 12499 }, { "epoch": 3.9184952978056424, "grad_norm": 6.657656192779541, "learning_rate": 1.0868414100166314e-07, "loss": 2.0985, "step": 12500 }, { "epoch": 3.918808777429467, "grad_norm": 74.34187316894531, "learning_rate": 1.0785576935217451e-07, "loss": 2.4759, "step": 12501 }, { "epoch": 3.9191222570532913, "grad_norm": 5.3137526512146, "learning_rate": 1.0703049704036128e-07, "loss": 1.9528, "step": 12502 }, { "epoch": 3.9194357366771158, "grad_norm": 53.29677963256836, "learning_rate": 1.0620832513547258e-07, "loss": 2.1393, "step": 12503 }, { "epoch": 3.91974921630094, "grad_norm": 48.30497360229492, "learning_rate": 1.053892547027402e-07, "loss": 2.5779, "step": 12504 }, { "epoch": 3.9200626959247646, "grad_norm": 6.6365227699279785, "learning_rate": 1.0457328680337803e-07, "loss": 1.9805, "step": 12505 }, { "epoch": 3.920376175548589, "grad_norm": 36.281803131103516, "learning_rate": 1.0376042249457929e-07, "loss": 2.053, "step": 12506 }, { "epoch": 3.9206896551724135, "grad_norm": 11.131946563720703, "learning_rate": 1.0295066282951738e-07, "loss": 1.8748, "step": 12507 }, { "epoch": 3.9210031347962384, "grad_norm": 6.464557647705078, "learning_rate": 1.0214400885734194e-07, "loss": 2.5065, "step": 12508 }, { "epoch": 3.921316614420063, "grad_norm": 8.033820152282715, "learning_rate": 1.0134046162317973e-07, "loss": 2.1713, "step": 12509 }, { "epoch": 3.9216300940438873, "grad_norm": 6.372591495513916, "learning_rate": 1.0054002216813158e-07, "loss": 2.5857, "step": 12510 }, { "epoch": 3.9219435736677117, "grad_norm": 10.972809791564941, "learning_rate": 9.974269152927235e-08, "loss": 2.2203, "step": 12511 }, { "epoch": 3.922257053291536, "grad_norm": 34.45088577270508, "learning_rate": 9.894847073964875e-08, "loss": 2.6501, "step": 12512 }, { "epoch": 3.9225705329153606, "grad_norm": 7.934930801391602, "learning_rate": 9.815736082827848e-08, "loss": 1.9839, "step": 12513 }, { "epoch": 3.922884012539185, "grad_norm": 14.807062149047852, "learning_rate": 9.73693628201483e-08, "loss": 2.3996, "step": 12514 }, { "epoch": 3.9231974921630095, "grad_norm": 6.125711441040039, "learning_rate": 9.658447773621349e-08, "loss": 2.6702, "step": 12515 }, { "epoch": 3.923510971786834, "grad_norm": 17.209362030029297, "learning_rate": 9.580270659339614e-08, "loss": 1.9645, "step": 12516 }, { "epoch": 3.9238244514106584, "grad_norm": 10.432965278625488, "learning_rate": 9.502405040458324e-08, "loss": 1.8708, "step": 12517 }, { "epoch": 3.924137931034483, "grad_norm": 7.404596328735352, "learning_rate": 9.424851017862696e-08, "loss": 2.0967, "step": 12518 }, { "epoch": 3.9244514106583073, "grad_norm": 6.521049976348877, "learning_rate": 9.347608692034105e-08, "loss": 2.1951, "step": 12519 }, { "epoch": 3.9247648902821317, "grad_norm": 5.4139204025268555, "learning_rate": 9.270678163050218e-08, "loss": 2.2889, "step": 12520 }, { "epoch": 3.925078369905956, "grad_norm": 110.67821502685547, "learning_rate": 9.194059530584553e-08, "loss": 2.4405, "step": 12521 }, { "epoch": 3.9253918495297806, "grad_norm": 11.404617309570312, "learning_rate": 9.117752893906762e-08, "loss": 2.0541, "step": 12522 }, { "epoch": 3.925705329153605, "grad_norm": 10.486613273620605, "learning_rate": 9.041758351882063e-08, "loss": 2.0703, "step": 12523 }, { "epoch": 3.9260188087774295, "grad_norm": 6.460341930389404, "learning_rate": 8.96607600297142e-08, "loss": 2.066, "step": 12524 }, { "epoch": 3.926332288401254, "grad_norm": 5.010915756225586, "learning_rate": 8.890705945231226e-08, "loss": 2.1086, "step": 12525 }, { "epoch": 3.9266457680250784, "grad_norm": 34.00679016113281, "learning_rate": 8.815648276313343e-08, "loss": 2.6886, "step": 12526 }, { "epoch": 3.926959247648903, "grad_norm": 7.078770637512207, "learning_rate": 8.74090309346487e-08, "loss": 2.0969, "step": 12527 }, { "epoch": 3.9272727272727272, "grad_norm": 85.4335708618164, "learning_rate": 8.666470493528007e-08, "loss": 1.696, "step": 12528 }, { "epoch": 3.9275862068965517, "grad_norm": 11.946975708007812, "learning_rate": 8.59235057294e-08, "loss": 2.1694, "step": 12529 }, { "epoch": 3.927899686520376, "grad_norm": 10.548701286315918, "learning_rate": 8.518543427732951e-08, "loss": 2.1434, "step": 12530 }, { "epoch": 3.9282131661442006, "grad_norm": 12.50023078918457, "learning_rate": 8.445049153533696e-08, "loss": 2.6508, "step": 12531 }, { "epoch": 3.928526645768025, "grad_norm": 29.473112106323242, "learning_rate": 8.371867845563819e-08, "loss": 1.8934, "step": 12532 }, { "epoch": 3.9288401253918495, "grad_norm": 7.067403316497803, "learning_rate": 8.298999598639279e-08, "loss": 2.0923, "step": 12533 }, { "epoch": 3.929153605015674, "grad_norm": 7.969937324523926, "learning_rate": 8.2264445071705e-08, "loss": 2.1042, "step": 12534 }, { "epoch": 3.9294670846394983, "grad_norm": 6.620874404907227, "learning_rate": 8.154202665162147e-08, "loss": 2.4798, "step": 12535 }, { "epoch": 3.929780564263323, "grad_norm": 7.94537878036499, "learning_rate": 8.082274166213016e-08, "loss": 1.9122, "step": 12536 }, { "epoch": 3.9300940438871472, "grad_norm": 6.158256530761719, "learning_rate": 8.010659103515949e-08, "loss": 1.9125, "step": 12537 }, { "epoch": 3.9304075235109717, "grad_norm": 8.336881637573242, "learning_rate": 7.939357569857697e-08, "loss": 2.1798, "step": 12538 }, { "epoch": 3.930721003134796, "grad_norm": 6.519012928009033, "learning_rate": 7.868369657618808e-08, "loss": 2.068, "step": 12539 }, { "epoch": 3.9310344827586206, "grad_norm": 16.63939666748047, "learning_rate": 7.79769545877343e-08, "loss": 2.2961, "step": 12540 }, { "epoch": 3.931347962382445, "grad_norm": 5.6653361320495605, "learning_rate": 7.727335064889264e-08, "loss": 2.2135, "step": 12541 }, { "epoch": 3.9316614420062694, "grad_norm": 30.36586570739746, "learning_rate": 7.657288567127496e-08, "loss": 2.5274, "step": 12542 }, { "epoch": 3.931974921630094, "grad_norm": 6.376472473144531, "learning_rate": 7.58755605624259e-08, "loss": 2.102, "step": 12543 }, { "epoch": 3.9322884012539188, "grad_norm": 20.562931060791016, "learning_rate": 7.518137622582189e-08, "loss": 2.0686, "step": 12544 }, { "epoch": 3.932601880877743, "grad_norm": 11.859230995178223, "learning_rate": 7.449033356086987e-08, "loss": 1.9931, "step": 12545 }, { "epoch": 3.9329153605015676, "grad_norm": 12.697888374328613, "learning_rate": 7.380243346290672e-08, "loss": 2.1877, "step": 12546 }, { "epoch": 3.933228840125392, "grad_norm": 5.3852643966674805, "learning_rate": 7.311767682319754e-08, "loss": 1.9379, "step": 12547 }, { "epoch": 3.9335423197492165, "grad_norm": 6.507474899291992, "learning_rate": 7.243606452893487e-08, "loss": 2.1136, "step": 12548 }, { "epoch": 3.933855799373041, "grad_norm": 6.952459812164307, "learning_rate": 7.175759746323646e-08, "loss": 2.1854, "step": 12549 }, { "epoch": 3.9341692789968654, "grad_norm": 6.807275772094727, "learning_rate": 7.108227650514637e-08, "loss": 1.8278, "step": 12550 }, { "epoch": 3.93448275862069, "grad_norm": 7.330295085906982, "learning_rate": 7.041010252963138e-08, "loss": 1.9954, "step": 12551 }, { "epoch": 3.9347962382445143, "grad_norm": 6.461508274078369, "learning_rate": 6.974107640758176e-08, "loss": 2.1176, "step": 12552 }, { "epoch": 3.9351097178683387, "grad_norm": 35.467281341552734, "learning_rate": 6.907519900580862e-08, "loss": 2.4737, "step": 12553 }, { "epoch": 3.935423197492163, "grad_norm": 40.19991683959961, "learning_rate": 6.841247118704375e-08, "loss": 2.1748, "step": 12554 }, { "epoch": 3.9357366771159876, "grad_norm": 10.807709693908691, "learning_rate": 6.775289380993838e-08, "loss": 2.0338, "step": 12555 }, { "epoch": 3.936050156739812, "grad_norm": 4.8482866287231445, "learning_rate": 6.70964677290617e-08, "loss": 2.1727, "step": 12556 }, { "epoch": 3.9363636363636365, "grad_norm": 7.922297954559326, "learning_rate": 6.644319379490005e-08, "loss": 2.111, "step": 12557 }, { "epoch": 3.936677115987461, "grad_norm": 7.391587257385254, "learning_rate": 6.579307285385555e-08, "loss": 2.2493, "step": 12558 }, { "epoch": 3.9369905956112854, "grad_norm": 10.746057510375977, "learning_rate": 6.514610574824609e-08, "loss": 1.9547, "step": 12559 }, { "epoch": 3.93730407523511, "grad_norm": 4.3314971923828125, "learning_rate": 6.450229331630253e-08, "loss": 2.0319, "step": 12560 }, { "epoch": 3.9376175548589343, "grad_norm": 7.793868064880371, "learning_rate": 6.386163639216819e-08, "loss": 1.9902, "step": 12561 }, { "epoch": 3.9379310344827587, "grad_norm": 7.0031418800354, "learning_rate": 6.322413580589881e-08, "loss": 2.0347, "step": 12562 }, { "epoch": 3.938244514106583, "grad_norm": 7.459324359893799, "learning_rate": 6.258979238346036e-08, "loss": 2.0721, "step": 12563 }, { "epoch": 3.9385579937304076, "grad_norm": 8.908483505249023, "learning_rate": 6.195860694672734e-08, "loss": 2.3647, "step": 12564 }, { "epoch": 3.938871473354232, "grad_norm": 9.04961109161377, "learning_rate": 6.133058031348477e-08, "loss": 3.1299, "step": 12565 }, { "epoch": 3.9391849529780565, "grad_norm": 7.218745231628418, "learning_rate": 6.070571329742342e-08, "loss": 2.0785, "step": 12566 }, { "epoch": 3.939498432601881, "grad_norm": 6.0504302978515625, "learning_rate": 6.008400670814041e-08, "loss": 2.1491, "step": 12567 }, { "epoch": 3.9398119122257054, "grad_norm": 8.154651641845703, "learning_rate": 5.946546135113862e-08, "loss": 2.0469, "step": 12568 }, { "epoch": 3.94012539184953, "grad_norm": 13.227487564086914, "learning_rate": 5.885007802782505e-08, "loss": 2.1935, "step": 12569 }, { "epoch": 3.9404388714733543, "grad_norm": 9.29893970489502, "learning_rate": 5.82378575355097e-08, "loss": 2.1953, "step": 12570 }, { "epoch": 3.9407523510971787, "grad_norm": 5.561811923980713, "learning_rate": 5.762880066740473e-08, "loss": 2.1106, "step": 12571 }, { "epoch": 3.941065830721003, "grad_norm": 8.0753812789917, "learning_rate": 5.702290821262391e-08, "loss": 2.2574, "step": 12572 }, { "epoch": 3.9413793103448276, "grad_norm": 29.043251037597656, "learning_rate": 5.6420180956180127e-08, "loss": 2.3853, "step": 12573 }, { "epoch": 3.941692789968652, "grad_norm": 5.503210067749023, "learning_rate": 5.5820619678985947e-08, "loss": 2.1904, "step": 12574 }, { "epoch": 3.9420062695924765, "grad_norm": 173.61607360839844, "learning_rate": 5.5224225157851916e-08, "loss": 2.4962, "step": 12575 }, { "epoch": 3.942319749216301, "grad_norm": 4.66765832901001, "learning_rate": 5.463099816548578e-08, "loss": 2.0081, "step": 12576 }, { "epoch": 3.9426332288401253, "grad_norm": 14.371882438659668, "learning_rate": 5.4040939470491324e-08, "loss": 2.311, "step": 12577 }, { "epoch": 3.94294670846395, "grad_norm": 6.219605445861816, "learning_rate": 5.345404983736702e-08, "loss": 2.0067, "step": 12578 }, { "epoch": 3.9432601880877742, "grad_norm": 5.749545097351074, "learning_rate": 5.2870330026505715e-08, "loss": 2.0331, "step": 12579 }, { "epoch": 3.9435736677115987, "grad_norm": 8.666352272033691, "learning_rate": 5.2289780794192726e-08, "loss": 1.8925, "step": 12580 }, { "epoch": 3.943887147335423, "grad_norm": 5.19416618347168, "learning_rate": 5.171240289260665e-08, "loss": 1.8636, "step": 12581 }, { "epoch": 3.9442006269592476, "grad_norm": 6.474620819091797, "learning_rate": 5.1138197069816596e-08, "loss": 2.2659, "step": 12582 }, { "epoch": 3.944514106583072, "grad_norm": 35.07327651977539, "learning_rate": 5.056716406978135e-08, "loss": 2.2113, "step": 12583 }, { "epoch": 3.9448275862068964, "grad_norm": 104.28427124023438, "learning_rate": 4.999930463234964e-08, "loss": 1.7822, "step": 12584 }, { "epoch": 3.945141065830721, "grad_norm": 22.719585418701172, "learning_rate": 4.943461949325767e-08, "loss": 1.8319, "step": 12585 }, { "epoch": 3.9454545454545453, "grad_norm": 9.660236358642578, "learning_rate": 4.8873109384129927e-08, "loss": 2.7498, "step": 12586 }, { "epoch": 3.9457680250783698, "grad_norm": 8.51151180267334, "learning_rate": 4.8314775032476114e-08, "loss": 2.437, "step": 12587 }, { "epoch": 3.946081504702194, "grad_norm": 120.54972076416016, "learning_rate": 4.775961716169203e-08, "loss": 2.3623, "step": 12588 }, { "epoch": 3.9463949843260187, "grad_norm": 18.042964935302734, "learning_rate": 4.720763649105814e-08, "loss": 2.1853, "step": 12589 }, { "epoch": 3.946708463949843, "grad_norm": 8.91815185546875, "learning_rate": 4.665883373573765e-08, "loss": 2.0414, "step": 12590 }, { "epoch": 3.9470219435736675, "grad_norm": 8.26434326171875, "learning_rate": 4.6113209606776784e-08, "loss": 2.1252, "step": 12591 }, { "epoch": 3.947335423197492, "grad_norm": 7.893552303314209, "learning_rate": 4.557076481110367e-08, "loss": 2.3171, "step": 12592 }, { "epoch": 3.9476489028213164, "grad_norm": 7.191619873046875, "learning_rate": 4.50315000515264e-08, "loss": 2.3134, "step": 12593 }, { "epoch": 3.947962382445141, "grad_norm": 10.770123481750488, "learning_rate": 4.449541602673413e-08, "loss": 2.3218, "step": 12594 }, { "epoch": 3.9482758620689653, "grad_norm": 12.885184288024902, "learning_rate": 4.396251343129376e-08, "loss": 2.528, "step": 12595 }, { "epoch": 3.9485893416927897, "grad_norm": 38.590267181396484, "learning_rate": 4.3432792955650784e-08, "loss": 2.2173, "step": 12596 }, { "epoch": 3.948902821316614, "grad_norm": 6.402647495269775, "learning_rate": 4.2906255286128126e-08, "loss": 2.1671, "step": 12597 }, { "epoch": 3.9492163009404386, "grad_norm": 7.3495635986328125, "learning_rate": 4.238290110492427e-08, "loss": 2.0152, "step": 12598 }, { "epoch": 3.949529780564263, "grad_norm": 5.208764553070068, "learning_rate": 4.186273109011374e-08, "loss": 2.0295, "step": 12599 }, { "epoch": 3.9498432601880875, "grad_norm": 5.931898593902588, "learning_rate": 4.134574591564494e-08, "loss": 2.0494, "step": 12600 }, { "epoch": 3.950156739811912, "grad_norm": 6.5731987953186035, "learning_rate": 4.083194625134013e-08, "loss": 1.9437, "step": 12601 }, { "epoch": 3.950470219435737, "grad_norm": 7.614995002746582, "learning_rate": 4.032133276289485e-08, "loss": 2.1063, "step": 12602 }, { "epoch": 3.9507836990595613, "grad_norm": 16.59844970703125, "learning_rate": 3.9813906111876024e-08, "loss": 2.0151, "step": 12603 }, { "epoch": 3.9510971786833857, "grad_norm": 12.317241668701172, "learning_rate": 3.930966695572136e-08, "loss": 2.1867, "step": 12604 }, { "epoch": 3.95141065830721, "grad_norm": 268.85595703125, "learning_rate": 3.88086159477391e-08, "loss": 2.4092, "step": 12605 }, { "epoch": 3.9517241379310346, "grad_norm": 4.645162582397461, "learning_rate": 3.831075373710691e-08, "loss": 2.1355, "step": 12606 }, { "epoch": 3.952037617554859, "grad_norm": 48.40354537963867, "learning_rate": 3.781608096887046e-08, "loss": 1.6416, "step": 12607 }, { "epoch": 3.9523510971786835, "grad_norm": 122.00611877441406, "learning_rate": 3.732459828394402e-08, "loss": 2.1404, "step": 12608 }, { "epoch": 3.952664576802508, "grad_norm": 19.70774269104004, "learning_rate": 3.683630631910767e-08, "loss": 2.0815, "step": 12609 }, { "epoch": 3.9529780564263324, "grad_norm": 25.505407333374023, "learning_rate": 3.635120570700784e-08, "loss": 2.032, "step": 12610 }, { "epoch": 3.953291536050157, "grad_norm": 5.594041347503662, "learning_rate": 3.5869297076156495e-08, "loss": 2.0391, "step": 12611 }, { "epoch": 3.9536050156739813, "grad_norm": 17.85814666748047, "learning_rate": 3.539058105092946e-08, "loss": 1.7775, "step": 12612 }, { "epoch": 3.9539184952978057, "grad_norm": 5.719181060791016, "learning_rate": 3.491505825156671e-08, "loss": 2.0172, "step": 12613 }, { "epoch": 3.95423197492163, "grad_norm": 5.699315547943115, "learning_rate": 3.444272929417014e-08, "loss": 2.3082, "step": 12614 }, { "epoch": 3.9545454545454546, "grad_norm": 33.915096282958984, "learning_rate": 3.397359479070467e-08, "loss": 2.1078, "step": 12615 }, { "epoch": 3.954858934169279, "grad_norm": 6.705883026123047, "learning_rate": 3.3507655348995194e-08, "loss": 1.8221, "step": 12616 }, { "epoch": 3.9551724137931035, "grad_norm": 7.71638298034668, "learning_rate": 3.304491157272771e-08, "loss": 1.8982, "step": 12617 }, { "epoch": 3.955485893416928, "grad_norm": 6.711490154266357, "learning_rate": 3.2585364061447635e-08, "loss": 2.0547, "step": 12618 }, { "epoch": 3.9557993730407524, "grad_norm": 23.083044052124023, "learning_rate": 3.21290134105598e-08, "loss": 2.4661, "step": 12619 }, { "epoch": 3.956112852664577, "grad_norm": 18.3245906829834, "learning_rate": 3.167586021132596e-08, "loss": 2.5481, "step": 12620 }, { "epoch": 3.9564263322884012, "grad_norm": 8.039846420288086, "learning_rate": 3.122590505086592e-08, "loss": 2.1501, "step": 12621 }, { "epoch": 3.9567398119122257, "grad_norm": 8.0797119140625, "learning_rate": 3.077914851215585e-08, "loss": 2.0454, "step": 12622 }, { "epoch": 3.95705329153605, "grad_norm": 9.826496124267578, "learning_rate": 3.033559117402746e-08, "loss": 1.9374, "step": 12623 }, { "epoch": 3.9573667711598746, "grad_norm": 4.460193634033203, "learning_rate": 2.98952336111677e-08, "loss": 2.0166, "step": 12624 }, { "epoch": 3.957680250783699, "grad_norm": 7.7192254066467285, "learning_rate": 2.9458076394117686e-08, "loss": 2.2909, "step": 12625 }, { "epoch": 3.9579937304075234, "grad_norm": 5.295419692993164, "learning_rate": 2.902412008927269e-08, "loss": 1.9681, "step": 12626 }, { "epoch": 3.958307210031348, "grad_norm": 8.617871284484863, "learning_rate": 2.8593365258879623e-08, "loss": 2.1526, "step": 12627 }, { "epoch": 3.9586206896551723, "grad_norm": 4.752935409545898, "learning_rate": 2.8165812461038166e-08, "loss": 2.4753, "step": 12628 }, { "epoch": 3.9589341692789968, "grad_norm": 6.697999000549316, "learning_rate": 2.774146224969909e-08, "loss": 2.4493, "step": 12629 }, { "epoch": 3.959247648902821, "grad_norm": 7.052475452423096, "learning_rate": 2.7320315174664547e-08, "loss": 2.4046, "step": 12630 }, { "epoch": 3.9595611285266457, "grad_norm": 6.271543025970459, "learning_rate": 2.690237178158528e-08, "loss": 2.0068, "step": 12631 }, { "epoch": 3.95987460815047, "grad_norm": 6.17680549621582, "learning_rate": 2.6487632611962578e-08, "loss": 2.1685, "step": 12632 }, { "epoch": 3.9601880877742945, "grad_norm": 50.27005386352539, "learning_rate": 2.607609820314494e-08, "loss": 2.1755, "step": 12633 }, { "epoch": 3.960501567398119, "grad_norm": 6.475771427154541, "learning_rate": 2.566776908832974e-08, "loss": 2.187, "step": 12634 }, { "epoch": 3.9608150470219434, "grad_norm": 11.918554306030273, "learning_rate": 2.526264579656129e-08, "loss": 2.0867, "step": 12635 }, { "epoch": 3.961128526645768, "grad_norm": 5.862170219421387, "learning_rate": 2.4860728852729998e-08, "loss": 2.1552, "step": 12636 }, { "epoch": 3.9614420062695928, "grad_norm": 10.571253776550293, "learning_rate": 2.4462018777572093e-08, "loss": 2.0517, "step": 12637 }, { "epoch": 3.961755485893417, "grad_norm": 6.6249871253967285, "learning_rate": 2.40665160876688e-08, "loss": 1.9504, "step": 12638 }, { "epoch": 3.9620689655172416, "grad_norm": 62.728206634521484, "learning_rate": 2.3674221295445775e-08, "loss": 2.1693, "step": 12639 }, { "epoch": 3.962382445141066, "grad_norm": 9.081912994384766, "learning_rate": 2.3285134909173113e-08, "loss": 2.3261, "step": 12640 }, { "epoch": 3.9626959247648905, "grad_norm": 9.547565460205078, "learning_rate": 2.289925743296312e-08, "loss": 1.916, "step": 12641 }, { "epoch": 3.963009404388715, "grad_norm": 11.15211009979248, "learning_rate": 2.2516589366770868e-08, "loss": 2.3421, "step": 12642 }, { "epoch": 3.9633228840125394, "grad_norm": 8.41959285736084, "learning_rate": 2.2137131206393104e-08, "loss": 1.9474, "step": 12643 }, { "epoch": 3.963636363636364, "grad_norm": 6.995291709899902, "learning_rate": 2.1760883443468217e-08, "loss": 2.232, "step": 12644 }, { "epoch": 3.9639498432601883, "grad_norm": 3.9948678016662598, "learning_rate": 2.1387846565474047e-08, "loss": 2.1876, "step": 12645 }, { "epoch": 3.9642633228840127, "grad_norm": 27.997098922729492, "learning_rate": 2.1018021055729256e-08, "loss": 1.8557, "step": 12646 }, { "epoch": 3.964576802507837, "grad_norm": 6.628724098205566, "learning_rate": 2.0651407393391944e-08, "loss": 1.9463, "step": 12647 }, { "epoch": 3.9648902821316616, "grad_norm": 18.096660614013672, "learning_rate": 2.028800605345771e-08, "loss": 2.1047, "step": 12648 }, { "epoch": 3.965203761755486, "grad_norm": 9.180160522460938, "learning_rate": 1.992781750676076e-08, "loss": 1.9812, "step": 12649 }, { "epoch": 3.9655172413793105, "grad_norm": 5.9001264572143555, "learning_rate": 1.957084221997335e-08, "loss": 2.1124, "step": 12650 }, { "epoch": 3.965830721003135, "grad_norm": 5.242674827575684, "learning_rate": 1.921708065560357e-08, "loss": 2.0813, "step": 12651 }, { "epoch": 3.9661442006269594, "grad_norm": 7.923477649688721, "learning_rate": 1.886653327199617e-08, "loss": 2.0065, "step": 12652 }, { "epoch": 3.966457680250784, "grad_norm": 6.5006303787231445, "learning_rate": 1.8519200523331448e-08, "loss": 1.8826, "step": 12653 }, { "epoch": 3.9667711598746083, "grad_norm": 9.397626876831055, "learning_rate": 1.8175082859624438e-08, "loss": 1.9252, "step": 12654 }, { "epoch": 3.9670846394984327, "grad_norm": 7.8473405838012695, "learning_rate": 1.783418072672516e-08, "loss": 2.2635, "step": 12655 }, { "epoch": 3.967398119122257, "grad_norm": 7.3857421875, "learning_rate": 1.7496494566317247e-08, "loss": 2.0029, "step": 12656 }, { "epoch": 3.9677115987460816, "grad_norm": 43.034996032714844, "learning_rate": 1.716202481591739e-08, "loss": 2.252, "step": 12657 }, { "epoch": 3.968025078369906, "grad_norm": 6.622166156768799, "learning_rate": 1.6830771908875054e-08, "loss": 2.0894, "step": 12658 }, { "epoch": 3.9683385579937305, "grad_norm": 6.895655632019043, "learning_rate": 1.650273627437221e-08, "loss": 1.9553, "step": 12659 }, { "epoch": 3.968652037617555, "grad_norm": 137.3245849609375, "learning_rate": 1.6177918337422216e-08, "loss": 2.3331, "step": 12660 }, { "epoch": 3.9689655172413794, "grad_norm": 6.229492664337158, "learning_rate": 1.5856318518868986e-08, "loss": 2.0735, "step": 12661 }, { "epoch": 3.969278996865204, "grad_norm": 8.209566116333008, "learning_rate": 1.55379372353881e-08, "loss": 1.8717, "step": 12662 }, { "epoch": 3.9695924764890282, "grad_norm": 11.834969520568848, "learning_rate": 1.522277489948376e-08, "loss": 2.031, "step": 12663 }, { "epoch": 3.9699059561128527, "grad_norm": 99.6766128540039, "learning_rate": 1.4910831919490997e-08, "loss": 2.1822, "step": 12664 }, { "epoch": 3.970219435736677, "grad_norm": 5.851038932800293, "learning_rate": 1.4602108699572348e-08, "loss": 1.9393, "step": 12665 }, { "epoch": 3.9705329153605016, "grad_norm": 6.284003257751465, "learning_rate": 1.4296605639719518e-08, "loss": 2.1592, "step": 12666 }, { "epoch": 3.970846394984326, "grad_norm": 9.139131546020508, "learning_rate": 1.3994323135751997e-08, "loss": 2.0762, "step": 12667 }, { "epoch": 3.9711598746081505, "grad_norm": 4.112520694732666, "learning_rate": 1.3695261579316776e-08, "loss": 2.0227, "step": 12668 }, { "epoch": 3.971473354231975, "grad_norm": 10.510337829589844, "learning_rate": 1.3399421357886966e-08, "loss": 2.165, "step": 12669 }, { "epoch": 3.9717868338557993, "grad_norm": 9.027532577514648, "learning_rate": 1.3106802854762901e-08, "loss": 2.4172, "step": 12670 }, { "epoch": 3.972100313479624, "grad_norm": 9.17184829711914, "learning_rate": 1.281740644907048e-08, "loss": 1.9049, "step": 12671 }, { "epoch": 3.972413793103448, "grad_norm": 161.7455596923828, "learning_rate": 1.2531232515760328e-08, "loss": 2.4562, "step": 12672 }, { "epoch": 3.9727272727272727, "grad_norm": 5.930420398712158, "learning_rate": 1.2248281425608632e-08, "loss": 2.029, "step": 12673 }, { "epoch": 3.973040752351097, "grad_norm": 7.755592346191406, "learning_rate": 1.1968553545216033e-08, "loss": 1.8434, "step": 12674 }, { "epoch": 3.9733542319749215, "grad_norm": 7.692620754241943, "learning_rate": 1.169204923700651e-08, "loss": 2.5208, "step": 12675 }, { "epoch": 3.973667711598746, "grad_norm": 8.110469818115234, "learning_rate": 1.1418768859227935e-08, "loss": 1.7317, "step": 12676 }, { "epoch": 3.9739811912225704, "grad_norm": 6.956070423126221, "learning_rate": 1.1148712765950975e-08, "loss": 2.265, "step": 12677 }, { "epoch": 3.974294670846395, "grad_norm": 7.516609191894531, "learning_rate": 1.0881881307069075e-08, "loss": 1.9707, "step": 12678 }, { "epoch": 3.9746081504702193, "grad_norm": 10.601871490478516, "learning_rate": 1.0618274828296804e-08, "loss": 2.1305, "step": 12679 }, { "epoch": 3.9749216300940438, "grad_norm": 9.962512016296387, "learning_rate": 1.0357893671171793e-08, "loss": 2.0782, "step": 12680 }, { "epoch": 3.975235109717868, "grad_norm": 6.738135814666748, "learning_rate": 1.0100738173051961e-08, "loss": 2.1863, "step": 12681 }, { "epoch": 3.9755485893416926, "grad_norm": 4.814486503601074, "learning_rate": 9.846808667115515e-09, "loss": 2.0777, "step": 12682 }, { "epoch": 3.975862068965517, "grad_norm": 14.629364967346191, "learning_rate": 9.59610548236234e-09, "loss": 2.1195, "step": 12683 }, { "epoch": 3.9761755485893415, "grad_norm": 98.10295104980469, "learning_rate": 9.348628943610938e-09, "loss": 1.9973, "step": 12684 }, { "epoch": 3.976489028213166, "grad_norm": 17.108505249023438, "learning_rate": 9.104379371500104e-09, "loss": 2.2844, "step": 12685 }, { "epoch": 3.9768025078369904, "grad_norm": 21.097562789916992, "learning_rate": 8.863357082486701e-09, "loss": 2.9584, "step": 12686 }, { "epoch": 3.977115987460815, "grad_norm": 22.398021697998047, "learning_rate": 8.625562388847319e-09, "loss": 2.1317, "step": 12687 }, { "epoch": 3.9774294670846393, "grad_norm": 10.313966751098633, "learning_rate": 8.390995598676067e-09, "loss": 2.2577, "step": 12688 }, { "epoch": 3.9777429467084637, "grad_norm": 17.007793426513672, "learning_rate": 8.159657015885391e-09, "loss": 2.1477, "step": 12689 }, { "epoch": 3.978056426332288, "grad_norm": 108.22034454345703, "learning_rate": 7.931546940204426e-09, "loss": 2.0868, "step": 12690 }, { "epoch": 3.9783699059561126, "grad_norm": 4.729015350341797, "learning_rate": 7.70666566718009e-09, "loss": 1.9611, "step": 12691 }, { "epoch": 3.978683385579937, "grad_norm": 44.864906311035156, "learning_rate": 7.485013488175708e-09, "loss": 2.0948, "step": 12692 }, { "epoch": 3.9789968652037615, "grad_norm": 6.346789836883545, "learning_rate": 7.266590690370446e-09, "loss": 2.0512, "step": 12693 }, { "epoch": 3.979310344827586, "grad_norm": 11.22291088104248, "learning_rate": 7.051397556760153e-09, "loss": 2.0132, "step": 12694 }, { "epoch": 3.9796238244514104, "grad_norm": 27.740875244140625, "learning_rate": 6.839434366155973e-09, "loss": 2.329, "step": 12695 }, { "epoch": 3.9799373040752353, "grad_norm": 11.092616081237793, "learning_rate": 6.63070139318378e-09, "loss": 2.4239, "step": 12696 }, { "epoch": 3.9802507836990597, "grad_norm": 7.52955436706543, "learning_rate": 6.425198908284747e-09, "loss": 2.0991, "step": 12697 }, { "epoch": 3.980564263322884, "grad_norm": 7.220992565155029, "learning_rate": 6.222927177714222e-09, "loss": 1.9008, "step": 12698 }, { "epoch": 3.9808777429467086, "grad_norm": 5.689886569976807, "learning_rate": 6.023886463542294e-09, "loss": 2.3961, "step": 12699 }, { "epoch": 3.981191222570533, "grad_norm": 21.512739181518555, "learning_rate": 5.828077023651846e-09, "loss": 2.4111, "step": 12700 }, { "epoch": 3.9815047021943575, "grad_norm": 6.81117582321167, "learning_rate": 5.6354991117404945e-09, "loss": 2.4326, "step": 12701 }, { "epoch": 3.981818181818182, "grad_norm": 39.43579864501953, "learning_rate": 5.4461529773175425e-09, "loss": 2.1686, "step": 12702 }, { "epoch": 3.9821316614420064, "grad_norm": 8.766960144042969, "learning_rate": 5.2600388657061965e-09, "loss": 1.932, "step": 12703 }, { "epoch": 3.982445141065831, "grad_norm": 276.6195068359375, "learning_rate": 5.077157018041623e-09, "loss": 2.2467, "step": 12704 }, { "epoch": 3.9827586206896552, "grad_norm": 8.210851669311523, "learning_rate": 4.897507671271507e-09, "loss": 1.9024, "step": 12705 }, { "epoch": 3.9830721003134797, "grad_norm": 11.781941413879395, "learning_rate": 4.721091058154936e-09, "loss": 2.1629, "step": 12706 }, { "epoch": 3.983385579937304, "grad_norm": 5.0369977951049805, "learning_rate": 4.547907407263241e-09, "loss": 1.8876, "step": 12707 }, { "epoch": 3.9836990595611286, "grad_norm": 12.46202564239502, "learning_rate": 4.3779569429783234e-09, "loss": 2.5164, "step": 12708 }, { "epoch": 3.984012539184953, "grad_norm": 7.152958869934082, "learning_rate": 4.2112398854934896e-09, "loss": 2.6635, "step": 12709 }, { "epoch": 3.9843260188087775, "grad_norm": 59.67278289794922, "learning_rate": 4.047756450812623e-09, "loss": 2.4921, "step": 12710 }, { "epoch": 3.984639498432602, "grad_norm": 9.979945182800293, "learning_rate": 3.8875068507499e-09, "loss": 1.8948, "step": 12711 }, { "epoch": 3.9849529780564263, "grad_norm": 6.173013687133789, "learning_rate": 3.730491292930072e-09, "loss": 2.0382, "step": 12712 }, { "epoch": 3.985266457680251, "grad_norm": 18.885005950927734, "learning_rate": 3.576709980787352e-09, "loss": 2.1928, "step": 12713 }, { "epoch": 3.9855799373040752, "grad_norm": 9.766221046447754, "learning_rate": 3.4261631135654174e-09, "loss": 2.2393, "step": 12714 }, { "epoch": 3.9858934169278997, "grad_norm": 143.46243286132812, "learning_rate": 3.278850886317686e-09, "loss": 2.8206, "step": 12715 }, { "epoch": 3.986206896551724, "grad_norm": 7.646114349365234, "learning_rate": 3.134773489906762e-09, "loss": 2.157, "step": 12716 }, { "epoch": 3.9865203761755486, "grad_norm": 19.199792861938477, "learning_rate": 2.993931111003601e-09, "loss": 1.9789, "step": 12717 }, { "epoch": 3.986833855799373, "grad_norm": 6.494765281677246, "learning_rate": 2.8563239320880676e-09, "loss": 2.4239, "step": 12718 }, { "epoch": 3.9871473354231974, "grad_norm": 8.833643913269043, "learning_rate": 2.72195213144838e-09, "loss": 2.4146, "step": 12719 }, { "epoch": 3.987460815047022, "grad_norm": 8.307612419128418, "learning_rate": 2.590815883181108e-09, "loss": 2.4042, "step": 12720 }, { "epoch": 3.9877742946708463, "grad_norm": 5.999251842498779, "learning_rate": 2.462915357190343e-09, "loss": 2.1909, "step": 12721 }, { "epoch": 3.9880877742946708, "grad_norm": 11.219511985778809, "learning_rate": 2.3382507191879735e-09, "loss": 1.9983, "step": 12722 }, { "epoch": 3.988401253918495, "grad_norm": 14.71632194519043, "learning_rate": 2.2168221306934077e-09, "loss": 2.1648, "step": 12723 }, { "epoch": 3.9887147335423196, "grad_norm": 8.8840913772583, "learning_rate": 2.0986297490338536e-09, "loss": 2.1218, "step": 12724 }, { "epoch": 3.989028213166144, "grad_norm": 7.461636543273926, "learning_rate": 1.9836737273423724e-09, "loss": 2.4244, "step": 12725 }, { "epoch": 3.9893416927899685, "grad_norm": 6.580917835235596, "learning_rate": 1.871954214560101e-09, "loss": 2.0006, "step": 12726 }, { "epoch": 3.989655172413793, "grad_norm": 5.275368690490723, "learning_rate": 1.7634713554343098e-09, "loss": 1.8579, "step": 12727 }, { "epoch": 3.9899686520376174, "grad_norm": 5.8258891105651855, "learning_rate": 1.6582252905186779e-09, "loss": 2.1853, "step": 12728 }, { "epoch": 3.990282131661442, "grad_norm": 6.9276556968688965, "learning_rate": 1.5562161561735734e-09, "loss": 2.2193, "step": 12729 }, { "epoch": 3.9905956112852663, "grad_norm": 5.113367557525635, "learning_rate": 1.4574440845649406e-09, "loss": 2.0194, "step": 12730 }, { "epoch": 3.990909090909091, "grad_norm": 13.347027778625488, "learning_rate": 1.361909203665135e-09, "loss": 2.1514, "step": 12731 }, { "epoch": 3.9912225705329156, "grad_norm": 5.281754493713379, "learning_rate": 1.269611637252366e-09, "loss": 1.988, "step": 12732 }, { "epoch": 3.99153605015674, "grad_norm": 7.865966320037842, "learning_rate": 1.1805515049098659e-09, "loss": 2.1653, "step": 12733 }, { "epoch": 3.9918495297805645, "grad_norm": 10.334778785705566, "learning_rate": 1.0947289220269996e-09, "loss": 2.0653, "step": 12734 }, { "epoch": 3.992163009404389, "grad_norm": 10.057568550109863, "learning_rate": 1.0121439997984316e-09, "loss": 1.8026, "step": 12735 }, { "epoch": 3.9924764890282134, "grad_norm": 4.623770236968994, "learning_rate": 9.32796845223294e-10, "loss": 2.0219, "step": 12736 }, { "epoch": 3.992789968652038, "grad_norm": 24.63326072692871, "learning_rate": 8.566875611068503e-10, "loss": 2.1233, "step": 12737 }, { "epoch": 3.9931034482758623, "grad_norm": 6.014158248901367, "learning_rate": 7.838162460582776e-10, "loss": 2.0038, "step": 12738 }, { "epoch": 3.9934169278996867, "grad_norm": 5.994048595428467, "learning_rate": 7.14182994492052e-10, "loss": 2.1412, "step": 12739 }, { "epoch": 3.993730407523511, "grad_norm": 100.19117736816406, "learning_rate": 6.477878966271167e-10, "loss": 3.8244, "step": 12740 }, { "epoch": 3.9940438871473356, "grad_norm": 14.947168350219727, "learning_rate": 5.8463103848716e-10, "loss": 2.2512, "step": 12741 }, { "epoch": 3.99435736677116, "grad_norm": 331.1975402832031, "learning_rate": 5.247125019003374e-10, "loss": 2.1486, "step": 12742 }, { "epoch": 3.9946708463949845, "grad_norm": 6.320652961730957, "learning_rate": 4.680323644989937e-10, "loss": 2.0461, "step": 12743 }, { "epoch": 3.994984326018809, "grad_norm": 8.907705307006836, "learning_rate": 4.1459069971938604e-10, "loss": 2.3835, "step": 12744 }, { "epoch": 3.9952978056426334, "grad_norm": 7.159172058105469, "learning_rate": 3.643875768025162e-10, "loss": 2.0062, "step": 12745 }, { "epoch": 3.995611285266458, "grad_norm": 15.475163459777832, "learning_rate": 3.1742306079302065e-10, "loss": 2.0283, "step": 12746 }, { "epoch": 3.9959247648902823, "grad_norm": 5.509530067443848, "learning_rate": 2.7369721253972523e-10, "loss": 2.292, "step": 12747 }, { "epoch": 3.9962382445141067, "grad_norm": 9.327977180480957, "learning_rate": 2.3321008869481296e-10, "loss": 2.2608, "step": 12748 }, { "epoch": 3.996551724137931, "grad_norm": 6.04849100112915, "learning_rate": 1.959617417154891e-10, "loss": 2.0273, "step": 12749 }, { "epoch": 3.9968652037617556, "grad_norm": 35.294132232666016, "learning_rate": 1.6195221986120556e-10, "loss": 2.2501, "step": 12750 }, { "epoch": 3.99717868338558, "grad_norm": 6.0097336769104, "learning_rate": 1.311815671958816e-10, "loss": 2.4435, "step": 12751 }, { "epoch": 3.9974921630094045, "grad_norm": 7.058928489685059, "learning_rate": 1.0364982358707087e-10, "loss": 2.0963, "step": 12752 }, { "epoch": 3.997805642633229, "grad_norm": 8.432992935180664, "learning_rate": 7.935702470568408e-11, "loss": 2.3081, "step": 12753 }, { "epoch": 3.9981191222570533, "grad_norm": 15.352566719055176, "learning_rate": 5.830320202654394e-11, "loss": 2.1357, "step": 12754 }, { "epoch": 3.998432601880878, "grad_norm": 21.441253662109375, "learning_rate": 4.048838282699752e-11, "loss": 2.5543, "step": 12755 }, { "epoch": 3.9987460815047022, "grad_norm": 9.423775672912598, "learning_rate": 2.591259018913661e-11, "loss": 2.3324, "step": 12756 }, { "epoch": 3.9990595611285267, "grad_norm": 9.215910911560059, "learning_rate": 1.4575842997577305e-11, "loss": 2.2925, "step": 12757 }, { "epoch": 3.999373040752351, "grad_norm": 6.387824058532715, "learning_rate": 6.4781559405702186e-12, "loss": 2.0044, "step": 12758 }, { "epoch": 3.9996865203761756, "grad_norm": 8.17207145690918, "learning_rate": 1.6195395097229339e-12, "loss": 2.058, "step": 12759 }, { "epoch": 4.0, "grad_norm": 6.437479496002197, "learning_rate": 0.0, "loss": 2.1064, "step": 12760 } ], "logging_steps": 1, "max_steps": 12760, "num_input_tokens_seen": 0, "num_train_epochs": 4, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 6.87786776519639e+16, "train_batch_size": 16, "trial_name": null, "trial_params": null }