{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.0, "eval_steps": 798, "global_step": 6380, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00031347962382445143, "grad_norm": 4.473039150238037, "learning_rate": 5e-10, "loss": 8.3121, "step": 1 }, { "epoch": 0.00031347962382445143, "eval_loss": 28.89470672607422, "eval_runtime": 20.3862, "eval_samples_per_second": 131.805, "eval_steps_per_second": 8.241, "step": 1 }, { "epoch": 0.0006269592476489029, "grad_norm": 3.657515287399292, "learning_rate": 1e-09, "loss": 7.0439, "step": 2 }, { "epoch": 0.0009404388714733542, "grad_norm": 2.806065559387207, "learning_rate": 1.5e-09, "loss": 6.6257, "step": 3 }, { "epoch": 0.0012539184952978057, "grad_norm": 2.810598611831665, "learning_rate": 2e-09, "loss": 6.5132, "step": 4 }, { "epoch": 0.001567398119122257, "grad_norm": 4.099343299865723, "learning_rate": 2.5e-09, "loss": 8.1055, "step": 5 }, { "epoch": 0.0018808777429467085, "grad_norm": 3.389047384262085, "learning_rate": 3e-09, "loss": 8.7388, "step": 6 }, { "epoch": 0.00219435736677116, "grad_norm": 3.4108972549438477, "learning_rate": 3.5000000000000003e-09, "loss": 7.8426, "step": 7 }, { "epoch": 0.0025078369905956114, "grad_norm": 3.3495447635650635, "learning_rate": 4e-09, "loss": 8.2811, "step": 8 }, { "epoch": 0.0028213166144200625, "grad_norm": 3.485499143600464, "learning_rate": 4.500000000000001e-09, "loss": 7.3836, "step": 9 }, { "epoch": 0.003134796238244514, "grad_norm": 4.053262233734131, "learning_rate": 5e-09, "loss": 9.1597, "step": 10 }, { "epoch": 0.0034482758620689655, "grad_norm": 4.419150352478027, "learning_rate": 5.5000000000000004e-09, "loss": 12.4238, "step": 11 }, { "epoch": 0.003761755485893417, "grad_norm": 3.277268409729004, "learning_rate": 6e-09, "loss": 8.4173, "step": 12 }, { "epoch": 0.0040752351097178684, "grad_norm": 3.8386521339416504, "learning_rate": 6.5e-09, "loss": 8.2201, "step": 13 }, { "epoch": 0.00438871473354232, "grad_norm": 4.788142681121826, "learning_rate": 7.000000000000001e-09, "loss": 11.7976, "step": 14 }, { "epoch": 0.004702194357366771, "grad_norm": 3.0357067584991455, "learning_rate": 7.500000000000001e-09, "loss": 8.5286, "step": 15 }, { "epoch": 0.005015673981191223, "grad_norm": 3.4486613273620605, "learning_rate": 8e-09, "loss": 7.839, "step": 16 }, { "epoch": 0.005329153605015674, "grad_norm": 4.829311370849609, "learning_rate": 8.5e-09, "loss": 9.6776, "step": 17 }, { "epoch": 0.005642633228840125, "grad_norm": 4.219234466552734, "learning_rate": 9.000000000000001e-09, "loss": 7.8809, "step": 18 }, { "epoch": 0.0059561128526645765, "grad_norm": 3.634462833404541, "learning_rate": 9.5e-09, "loss": 9.879, "step": 19 }, { "epoch": 0.006269592476489028, "grad_norm": 3.924381732940674, "learning_rate": 1e-08, "loss": 9.0212, "step": 20 }, { "epoch": 0.0065830721003134795, "grad_norm": 3.562445640563965, "learning_rate": 1.05e-08, "loss": 8.593, "step": 21 }, { "epoch": 0.006896551724137931, "grad_norm": 2.989351749420166, "learning_rate": 1.1000000000000001e-08, "loss": 6.3834, "step": 22 }, { "epoch": 0.007210031347962382, "grad_norm": 3.755312919616699, "learning_rate": 1.15e-08, "loss": 9.2207, "step": 23 }, { "epoch": 0.007523510971786834, "grad_norm": 3.771742582321167, "learning_rate": 1.2e-08, "loss": 9.2504, "step": 24 }, { "epoch": 0.007836990595611285, "grad_norm": 2.664695978164673, "learning_rate": 1.2500000000000001e-08, "loss": 7.0286, "step": 25 }, { "epoch": 0.008150470219435737, "grad_norm": 2.916489601135254, "learning_rate": 1.3e-08, "loss": 7.5708, "step": 26 }, { "epoch": 0.008463949843260187, "grad_norm": 3.4173834323883057, "learning_rate": 1.3500000000000002e-08, "loss": 7.0245, "step": 27 }, { "epoch": 0.00877742946708464, "grad_norm": 4.134421348571777, "learning_rate": 1.4000000000000001e-08, "loss": 8.218, "step": 28 }, { "epoch": 0.00909090909090909, "grad_norm": 2.478123188018799, "learning_rate": 1.45e-08, "loss": 6.6974, "step": 29 }, { "epoch": 0.009404388714733543, "grad_norm": 4.880210876464844, "learning_rate": 1.5000000000000002e-08, "loss": 13.0694, "step": 30 }, { "epoch": 0.009717868338557993, "grad_norm": 3.532501220703125, "learning_rate": 1.55e-08, "loss": 6.8898, "step": 31 }, { "epoch": 0.010031347962382446, "grad_norm": 3.8485097885131836, "learning_rate": 1.6e-08, "loss": 10.3871, "step": 32 }, { "epoch": 0.010344827586206896, "grad_norm": 4.742996692657471, "learning_rate": 1.6500000000000002e-08, "loss": 10.987, "step": 33 }, { "epoch": 0.010658307210031349, "grad_norm": 3.595039129257202, "learning_rate": 1.7e-08, "loss": 9.6221, "step": 34 }, { "epoch": 0.0109717868338558, "grad_norm": 3.5569448471069336, "learning_rate": 1.75e-08, "loss": 6.5841, "step": 35 }, { "epoch": 0.01128526645768025, "grad_norm": 4.763426303863525, "learning_rate": 1.8000000000000002e-08, "loss": 7.6671, "step": 36 }, { "epoch": 0.011598746081504702, "grad_norm": 3.6167099475860596, "learning_rate": 1.8500000000000004e-08, "loss": 10.4234, "step": 37 }, { "epoch": 0.011912225705329153, "grad_norm": 4.075372219085693, "learning_rate": 1.9e-08, "loss": 11.068, "step": 38 }, { "epoch": 0.012225705329153605, "grad_norm": 6.240333080291748, "learning_rate": 1.95e-08, "loss": 17.4708, "step": 39 }, { "epoch": 0.012539184952978056, "grad_norm": 3.7709567546844482, "learning_rate": 2e-08, "loss": 8.2594, "step": 40 }, { "epoch": 0.012852664576802508, "grad_norm": 2.885645866394043, "learning_rate": 2.0500000000000005e-08, "loss": 6.7495, "step": 41 }, { "epoch": 0.013166144200626959, "grad_norm": 3.3022165298461914, "learning_rate": 2.1e-08, "loss": 7.0996, "step": 42 }, { "epoch": 0.013479623824451411, "grad_norm": 3.8264548778533936, "learning_rate": 2.15e-08, "loss": 8.166, "step": 43 }, { "epoch": 0.013793103448275862, "grad_norm": 4.580362796783447, "learning_rate": 2.2000000000000002e-08, "loss": 14.0935, "step": 44 }, { "epoch": 0.014106583072100314, "grad_norm": 4.131213665008545, "learning_rate": 2.25e-08, "loss": 12.5754, "step": 45 }, { "epoch": 0.014420062695924765, "grad_norm": 3.5261685848236084, "learning_rate": 2.3e-08, "loss": 7.9894, "step": 46 }, { "epoch": 0.014733542319749215, "grad_norm": 3.0584185123443604, "learning_rate": 2.3500000000000002e-08, "loss": 6.4833, "step": 47 }, { "epoch": 0.015047021943573668, "grad_norm": 3.39809513092041, "learning_rate": 2.4e-08, "loss": 8.1705, "step": 48 }, { "epoch": 0.015360501567398118, "grad_norm": 4.006636619567871, "learning_rate": 2.45e-08, "loss": 8.5717, "step": 49 }, { "epoch": 0.01567398119122257, "grad_norm": 5.995611190795898, "learning_rate": 2.5000000000000002e-08, "loss": 8.4349, "step": 50 }, { "epoch": 0.01598746081504702, "grad_norm": 3.6572229862213135, "learning_rate": 2.5500000000000003e-08, "loss": 7.9315, "step": 51 }, { "epoch": 0.016300940438871474, "grad_norm": 3.6359832286834717, "learning_rate": 2.6e-08, "loss": 8.9688, "step": 52 }, { "epoch": 0.016614420062695926, "grad_norm": 2.8677127361297607, "learning_rate": 2.6500000000000002e-08, "loss": 6.3906, "step": 53 }, { "epoch": 0.016927899686520375, "grad_norm": 2.7862868309020996, "learning_rate": 2.7000000000000004e-08, "loss": 7.7946, "step": 54 }, { "epoch": 0.017241379310344827, "grad_norm": 3.8618948459625244, "learning_rate": 2.75e-08, "loss": 7.8249, "step": 55 }, { "epoch": 0.01755485893416928, "grad_norm": 3.185887336730957, "learning_rate": 2.8000000000000003e-08, "loss": 7.2935, "step": 56 }, { "epoch": 0.017868338557993732, "grad_norm": 2.7501015663146973, "learning_rate": 2.8500000000000004e-08, "loss": 6.9817, "step": 57 }, { "epoch": 0.01818181818181818, "grad_norm": 3.1869773864746094, "learning_rate": 2.9e-08, "loss": 7.383, "step": 58 }, { "epoch": 0.018495297805642633, "grad_norm": 5.212038993835449, "learning_rate": 2.9500000000000003e-08, "loss": 14.1096, "step": 59 }, { "epoch": 0.018808777429467086, "grad_norm": 3.6909945011138916, "learning_rate": 3.0000000000000004e-08, "loss": 9.0418, "step": 60 }, { "epoch": 0.019122257053291535, "grad_norm": 4.0268120765686035, "learning_rate": 3.05e-08, "loss": 11.0808, "step": 61 }, { "epoch": 0.019435736677115987, "grad_norm": 4.155067443847656, "learning_rate": 3.1e-08, "loss": 10.4174, "step": 62 }, { "epoch": 0.01974921630094044, "grad_norm": 4.077535629272461, "learning_rate": 3.1500000000000004e-08, "loss": 8.0036, "step": 63 }, { "epoch": 0.02006269592476489, "grad_norm": 3.3879809379577637, "learning_rate": 3.2e-08, "loss": 7.2178, "step": 64 }, { "epoch": 0.02037617554858934, "grad_norm": 4.345617771148682, "learning_rate": 3.25e-08, "loss": 11.3773, "step": 65 }, { "epoch": 0.020689655172413793, "grad_norm": 3.596099853515625, "learning_rate": 3.3000000000000004e-08, "loss": 7.7855, "step": 66 }, { "epoch": 0.021003134796238245, "grad_norm": 2.798985004425049, "learning_rate": 3.35e-08, "loss": 6.8671, "step": 67 }, { "epoch": 0.021316614420062698, "grad_norm": 3.921792507171631, "learning_rate": 3.4e-08, "loss": 11.5327, "step": 68 }, { "epoch": 0.021630094043887146, "grad_norm": 3.9658515453338623, "learning_rate": 3.4500000000000005e-08, "loss": 7.4312, "step": 69 }, { "epoch": 0.0219435736677116, "grad_norm": 3.8455910682678223, "learning_rate": 3.5e-08, "loss": 11.7665, "step": 70 }, { "epoch": 0.02225705329153605, "grad_norm": 3.4687187671661377, "learning_rate": 3.550000000000001e-08, "loss": 9.1295, "step": 71 }, { "epoch": 0.0225705329153605, "grad_norm": 3.2690117359161377, "learning_rate": 3.6000000000000005e-08, "loss": 8.421, "step": 72 }, { "epoch": 0.022884012539184952, "grad_norm": 4.495233535766602, "learning_rate": 3.65e-08, "loss": 8.9433, "step": 73 }, { "epoch": 0.023197492163009405, "grad_norm": 3.4736380577087402, "learning_rate": 3.700000000000001e-08, "loss": 7.5205, "step": 74 }, { "epoch": 0.023510971786833857, "grad_norm": 3.021930456161499, "learning_rate": 3.7500000000000005e-08, "loss": 6.0582, "step": 75 }, { "epoch": 0.023824451410658306, "grad_norm": 3.8124265670776367, "learning_rate": 3.8e-08, "loss": 9.6513, "step": 76 }, { "epoch": 0.02413793103448276, "grad_norm": 2.6047370433807373, "learning_rate": 3.850000000000001e-08, "loss": 6.0856, "step": 77 }, { "epoch": 0.02445141065830721, "grad_norm": 4.412445545196533, "learning_rate": 3.9e-08, "loss": 9.3832, "step": 78 }, { "epoch": 0.024764890282131663, "grad_norm": 4.0916547775268555, "learning_rate": 3.950000000000001e-08, "loss": 10.2509, "step": 79 }, { "epoch": 0.025078369905956112, "grad_norm": 3.16266131401062, "learning_rate": 4e-08, "loss": 8.3115, "step": 80 }, { "epoch": 0.025391849529780564, "grad_norm": 3.2230985164642334, "learning_rate": 4.05e-08, "loss": 8.4255, "step": 81 }, { "epoch": 0.025705329153605017, "grad_norm": 4.53424072265625, "learning_rate": 4.100000000000001e-08, "loss": 9.7329, "step": 82 }, { "epoch": 0.026018808777429465, "grad_norm": 2.889455795288086, "learning_rate": 4.15e-08, "loss": 7.6362, "step": 83 }, { "epoch": 0.026332288401253918, "grad_norm": 4.101494789123535, "learning_rate": 4.2e-08, "loss": 9.3998, "step": 84 }, { "epoch": 0.02664576802507837, "grad_norm": 3.166928291320801, "learning_rate": 4.2500000000000003e-08, "loss": 8.9123, "step": 85 }, { "epoch": 0.026959247648902823, "grad_norm": 4.550684452056885, "learning_rate": 4.3e-08, "loss": 10.9083, "step": 86 }, { "epoch": 0.02727272727272727, "grad_norm": 3.8120524883270264, "learning_rate": 4.35e-08, "loss": 11.0573, "step": 87 }, { "epoch": 0.027586206896551724, "grad_norm": 3.2697463035583496, "learning_rate": 4.4000000000000004e-08, "loss": 6.9374, "step": 88 }, { "epoch": 0.027899686520376176, "grad_norm": 4.755700588226318, "learning_rate": 4.45e-08, "loss": 9.1937, "step": 89 }, { "epoch": 0.02821316614420063, "grad_norm": 2.45084547996521, "learning_rate": 4.5e-08, "loss": 8.1863, "step": 90 }, { "epoch": 0.028526645768025077, "grad_norm": 4.488142490386963, "learning_rate": 4.5500000000000004e-08, "loss": 10.0932, "step": 91 }, { "epoch": 0.02884012539184953, "grad_norm": 2.933330774307251, "learning_rate": 4.6e-08, "loss": 6.199, "step": 92 }, { "epoch": 0.029153605015673982, "grad_norm": 3.5675182342529297, "learning_rate": 4.65e-08, "loss": 11.0667, "step": 93 }, { "epoch": 0.02946708463949843, "grad_norm": 3.300973653793335, "learning_rate": 4.7000000000000004e-08, "loss": 8.7929, "step": 94 }, { "epoch": 0.029780564263322883, "grad_norm": 4.447729587554932, "learning_rate": 4.75e-08, "loss": 13.4648, "step": 95 }, { "epoch": 0.030094043887147336, "grad_norm": 3.820695638656616, "learning_rate": 4.8e-08, "loss": 8.4362, "step": 96 }, { "epoch": 0.030407523510971788, "grad_norm": 3.2529191970825195, "learning_rate": 4.8500000000000004e-08, "loss": 7.8372, "step": 97 }, { "epoch": 0.030721003134796237, "grad_norm": 5.665775775909424, "learning_rate": 4.9e-08, "loss": 13.5665, "step": 98 }, { "epoch": 0.03103448275862069, "grad_norm": 3.9593966007232666, "learning_rate": 4.9500000000000006e-08, "loss": 8.0411, "step": 99 }, { "epoch": 0.03134796238244514, "grad_norm": 3.611478567123413, "learning_rate": 5.0000000000000004e-08, "loss": 7.8446, "step": 100 }, { "epoch": 0.031661442006269594, "grad_norm": 3.8838236331939697, "learning_rate": 5.05e-08, "loss": 8.1889, "step": 101 }, { "epoch": 0.03197492163009404, "grad_norm": 3.930360794067383, "learning_rate": 5.100000000000001e-08, "loss": 11.6122, "step": 102 }, { "epoch": 0.0322884012539185, "grad_norm": 3.1364479064941406, "learning_rate": 5.1500000000000005e-08, "loss": 7.8931, "step": 103 }, { "epoch": 0.03260188087774295, "grad_norm": 4.427796840667725, "learning_rate": 5.2e-08, "loss": 10.2633, "step": 104 }, { "epoch": 0.032915360501567396, "grad_norm": 4.472553730010986, "learning_rate": 5.250000000000001e-08, "loss": 9.7383, "step": 105 }, { "epoch": 0.03322884012539185, "grad_norm": 3.7383227348327637, "learning_rate": 5.3000000000000005e-08, "loss": 10.8243, "step": 106 }, { "epoch": 0.0335423197492163, "grad_norm": 4.171073913574219, "learning_rate": 5.35e-08, "loss": 9.0764, "step": 107 }, { "epoch": 0.03385579937304075, "grad_norm": 3.8834450244903564, "learning_rate": 5.400000000000001e-08, "loss": 11.7289, "step": 108 }, { "epoch": 0.034169278996865206, "grad_norm": 5.5527753829956055, "learning_rate": 5.4500000000000005e-08, "loss": 13.2347, "step": 109 }, { "epoch": 0.034482758620689655, "grad_norm": 3.0170059204101562, "learning_rate": 5.5e-08, "loss": 8.0964, "step": 110 }, { "epoch": 0.034796238244514104, "grad_norm": 4.954318046569824, "learning_rate": 5.550000000000001e-08, "loss": 9.1267, "step": 111 }, { "epoch": 0.03510971786833856, "grad_norm": 5.236219882965088, "learning_rate": 5.6000000000000005e-08, "loss": 13.9724, "step": 112 }, { "epoch": 0.03542319749216301, "grad_norm": 3.203052520751953, "learning_rate": 5.65e-08, "loss": 7.3345, "step": 113 }, { "epoch": 0.035736677115987464, "grad_norm": 3.730330228805542, "learning_rate": 5.700000000000001e-08, "loss": 11.3994, "step": 114 }, { "epoch": 0.03605015673981191, "grad_norm": 3.654137372970581, "learning_rate": 5.7500000000000005e-08, "loss": 7.6924, "step": 115 }, { "epoch": 0.03636363636363636, "grad_norm": 3.058237314224243, "learning_rate": 5.8e-08, "loss": 6.7775, "step": 116 }, { "epoch": 0.03667711598746082, "grad_norm": 3.9933547973632812, "learning_rate": 5.850000000000001e-08, "loss": 9.6667, "step": 117 }, { "epoch": 0.03699059561128527, "grad_norm": 4.147707462310791, "learning_rate": 5.9000000000000006e-08, "loss": 7.6471, "step": 118 }, { "epoch": 0.037304075235109715, "grad_norm": 4.296032428741455, "learning_rate": 5.950000000000001e-08, "loss": 13.4849, "step": 119 }, { "epoch": 0.03761755485893417, "grad_norm": 3.6536829471588135, "learning_rate": 6.000000000000001e-08, "loss": 10.2448, "step": 120 }, { "epoch": 0.03793103448275862, "grad_norm": 3.77724552154541, "learning_rate": 6.05e-08, "loss": 8.568, "step": 121 }, { "epoch": 0.03824451410658307, "grad_norm": 3.5914738178253174, "learning_rate": 6.1e-08, "loss": 7.9479, "step": 122 }, { "epoch": 0.038557993730407525, "grad_norm": 8.78665828704834, "learning_rate": 6.15e-08, "loss": 7.3475, "step": 123 }, { "epoch": 0.038871473354231974, "grad_norm": 3.715463399887085, "learning_rate": 6.2e-08, "loss": 8.8613, "step": 124 }, { "epoch": 0.03918495297805643, "grad_norm": 3.4635043144226074, "learning_rate": 6.250000000000001e-08, "loss": 7.8405, "step": 125 }, { "epoch": 0.03949843260188088, "grad_norm": 3.9361000061035156, "learning_rate": 6.300000000000001e-08, "loss": 7.8298, "step": 126 }, { "epoch": 0.03981191222570533, "grad_norm": 3.872082233428955, "learning_rate": 6.35e-08, "loss": 11.5712, "step": 127 }, { "epoch": 0.04012539184952978, "grad_norm": 3.6292521953582764, "learning_rate": 6.4e-08, "loss": 8.9746, "step": 128 }, { "epoch": 0.04043887147335423, "grad_norm": 3.303443193435669, "learning_rate": 6.45e-08, "loss": 7.3576, "step": 129 }, { "epoch": 0.04075235109717868, "grad_norm": 2.838657855987549, "learning_rate": 6.5e-08, "loss": 8.1253, "step": 130 }, { "epoch": 0.04106583072100314, "grad_norm": 3.875380039215088, "learning_rate": 6.550000000000001e-08, "loss": 9.9537, "step": 131 }, { "epoch": 0.041379310344827586, "grad_norm": 3.1287903785705566, "learning_rate": 6.600000000000001e-08, "loss": 8.871, "step": 132 }, { "epoch": 0.041692789968652035, "grad_norm": 3.414733648300171, "learning_rate": 6.65e-08, "loss": 6.58, "step": 133 }, { "epoch": 0.04200626959247649, "grad_norm": 3.1217563152313232, "learning_rate": 6.7e-08, "loss": 7.591, "step": 134 }, { "epoch": 0.04231974921630094, "grad_norm": 4.4469733238220215, "learning_rate": 6.75e-08, "loss": 9.0155, "step": 135 }, { "epoch": 0.042633228840125395, "grad_norm": 3.728848695755005, "learning_rate": 6.8e-08, "loss": 8.98, "step": 136 }, { "epoch": 0.042946708463949844, "grad_norm": 4.085875988006592, "learning_rate": 6.850000000000001e-08, "loss": 9.7549, "step": 137 }, { "epoch": 0.04326018808777429, "grad_norm": 2.9348063468933105, "learning_rate": 6.900000000000001e-08, "loss": 7.0279, "step": 138 }, { "epoch": 0.04357366771159875, "grad_norm": 3.2348380088806152, "learning_rate": 6.950000000000001e-08, "loss": 8.1917, "step": 139 }, { "epoch": 0.0438871473354232, "grad_norm": 3.731046438217163, "learning_rate": 7e-08, "loss": 8.1675, "step": 140 }, { "epoch": 0.044200626959247646, "grad_norm": 2.7796294689178467, "learning_rate": 7.05e-08, "loss": 7.1764, "step": 141 }, { "epoch": 0.0445141065830721, "grad_norm": 3.467015027999878, "learning_rate": 7.100000000000001e-08, "loss": 7.7669, "step": 142 }, { "epoch": 0.04482758620689655, "grad_norm": 4.046057224273682, "learning_rate": 7.150000000000001e-08, "loss": 11.0822, "step": 143 }, { "epoch": 0.045141065830721, "grad_norm": 4.335888385772705, "learning_rate": 7.200000000000001e-08, "loss": 9.0507, "step": 144 }, { "epoch": 0.045454545454545456, "grad_norm": 3.208477258682251, "learning_rate": 7.250000000000001e-08, "loss": 8.5856, "step": 145 }, { "epoch": 0.045768025078369905, "grad_norm": 4.974661350250244, "learning_rate": 7.3e-08, "loss": 13.4032, "step": 146 }, { "epoch": 0.04608150470219436, "grad_norm": 5.451123237609863, "learning_rate": 7.35e-08, "loss": 11.6684, "step": 147 }, { "epoch": 0.04639498432601881, "grad_norm": 5.630722999572754, "learning_rate": 7.400000000000001e-08, "loss": 14.3619, "step": 148 }, { "epoch": 0.04670846394984326, "grad_norm": 3.298954725265503, "learning_rate": 7.450000000000001e-08, "loss": 7.9049, "step": 149 }, { "epoch": 0.047021943573667714, "grad_norm": 6.209390640258789, "learning_rate": 7.500000000000001e-08, "loss": 12.7281, "step": 150 }, { "epoch": 0.04733542319749216, "grad_norm": 4.096133232116699, "learning_rate": 7.550000000000001e-08, "loss": 10.6689, "step": 151 }, { "epoch": 0.04764890282131661, "grad_norm": 3.8314168453216553, "learning_rate": 7.6e-08, "loss": 9.3038, "step": 152 }, { "epoch": 0.04796238244514107, "grad_norm": 4.284617900848389, "learning_rate": 7.65e-08, "loss": 11.3222, "step": 153 }, { "epoch": 0.04827586206896552, "grad_norm": 3.8246536254882812, "learning_rate": 7.700000000000001e-08, "loss": 9.8774, "step": 154 }, { "epoch": 0.048589341692789965, "grad_norm": 3.011291265487671, "learning_rate": 7.750000000000001e-08, "loss": 6.99, "step": 155 }, { "epoch": 0.04890282131661442, "grad_norm": 3.5155868530273438, "learning_rate": 7.8e-08, "loss": 10.0299, "step": 156 }, { "epoch": 0.04921630094043887, "grad_norm": 3.2512242794036865, "learning_rate": 7.85e-08, "loss": 7.9809, "step": 157 }, { "epoch": 0.049529780564263326, "grad_norm": 3.8495590686798096, "learning_rate": 7.900000000000002e-08, "loss": 8.4488, "step": 158 }, { "epoch": 0.049843260188087775, "grad_norm": 4.043474197387695, "learning_rate": 7.950000000000002e-08, "loss": 7.3039, "step": 159 }, { "epoch": 0.050156739811912224, "grad_norm": 3.638864755630493, "learning_rate": 8e-08, "loss": 12.0005, "step": 160 }, { "epoch": 0.05047021943573668, "grad_norm": 3.9693477153778076, "learning_rate": 8.05e-08, "loss": 5.8005, "step": 161 }, { "epoch": 0.05078369905956113, "grad_norm": 4.173606872558594, "learning_rate": 8.1e-08, "loss": 14.1363, "step": 162 }, { "epoch": 0.05109717868338558, "grad_norm": 3.1908795833587646, "learning_rate": 8.15e-08, "loss": 8.1196, "step": 163 }, { "epoch": 0.05141065830721003, "grad_norm": 2.548922061920166, "learning_rate": 8.200000000000002e-08, "loss": 6.7172, "step": 164 }, { "epoch": 0.05172413793103448, "grad_norm": 3.2565834522247314, "learning_rate": 8.25e-08, "loss": 7.6421, "step": 165 }, { "epoch": 0.05203761755485893, "grad_norm": 4.065040111541748, "learning_rate": 8.3e-08, "loss": 9.3676, "step": 166 }, { "epoch": 0.05235109717868339, "grad_norm": 3.38653826713562, "learning_rate": 8.35e-08, "loss": 8.983, "step": 167 }, { "epoch": 0.052664576802507836, "grad_norm": 3.8572115898132324, "learning_rate": 8.4e-08, "loss": 8.1122, "step": 168 }, { "epoch": 0.05297805642633229, "grad_norm": 3.2368323802948, "learning_rate": 8.45e-08, "loss": 8.4592, "step": 169 }, { "epoch": 0.05329153605015674, "grad_norm": 3.422175884246826, "learning_rate": 8.500000000000001e-08, "loss": 9.0227, "step": 170 }, { "epoch": 0.05360501567398119, "grad_norm": 4.226418972015381, "learning_rate": 8.55e-08, "loss": 10.3479, "step": 171 }, { "epoch": 0.053918495297805645, "grad_norm": 4.698319435119629, "learning_rate": 8.6e-08, "loss": 9.7526, "step": 172 }, { "epoch": 0.054231974921630094, "grad_norm": 3.448711395263672, "learning_rate": 8.65e-08, "loss": 9.1493, "step": 173 }, { "epoch": 0.05454545454545454, "grad_norm": 3.3972206115722656, "learning_rate": 8.7e-08, "loss": 6.7097, "step": 174 }, { "epoch": 0.054858934169279, "grad_norm": 3.6650757789611816, "learning_rate": 8.750000000000001e-08, "loss": 10.019, "step": 175 }, { "epoch": 0.05517241379310345, "grad_norm": 3.743401288986206, "learning_rate": 8.800000000000001e-08, "loss": 9.1552, "step": 176 }, { "epoch": 0.055485893416927896, "grad_norm": 3.6061248779296875, "learning_rate": 8.85e-08, "loss": 8.0651, "step": 177 }, { "epoch": 0.05579937304075235, "grad_norm": 3.2070910930633545, "learning_rate": 8.9e-08, "loss": 8.7052, "step": 178 }, { "epoch": 0.0561128526645768, "grad_norm": 4.579248428344727, "learning_rate": 8.95e-08, "loss": 8.3114, "step": 179 }, { "epoch": 0.05642633228840126, "grad_norm": 4.0575690269470215, "learning_rate": 9e-08, "loss": 8.8177, "step": 180 }, { "epoch": 0.056739811912225706, "grad_norm": 3.667318820953369, "learning_rate": 9.050000000000001e-08, "loss": 9.8964, "step": 181 }, { "epoch": 0.057053291536050155, "grad_norm": 4.42782735824585, "learning_rate": 9.100000000000001e-08, "loss": 10.0656, "step": 182 }, { "epoch": 0.05736677115987461, "grad_norm": 3.8123972415924072, "learning_rate": 9.15e-08, "loss": 10.2998, "step": 183 }, { "epoch": 0.05768025078369906, "grad_norm": 4.3586745262146, "learning_rate": 9.2e-08, "loss": 8.8101, "step": 184 }, { "epoch": 0.05799373040752351, "grad_norm": 3.785413980484009, "learning_rate": 9.25e-08, "loss": 9.7424, "step": 185 }, { "epoch": 0.058307210031347964, "grad_norm": 3.5556790828704834, "learning_rate": 9.3e-08, "loss": 10.6945, "step": 186 }, { "epoch": 0.05862068965517241, "grad_norm": 4.004053592681885, "learning_rate": 9.350000000000001e-08, "loss": 9.8506, "step": 187 }, { "epoch": 0.05893416927899686, "grad_norm": 5.22344970703125, "learning_rate": 9.400000000000001e-08, "loss": 12.796, "step": 188 }, { "epoch": 0.05924764890282132, "grad_norm": 3.7727713584899902, "learning_rate": 9.45e-08, "loss": 7.7093, "step": 189 }, { "epoch": 0.05956112852664577, "grad_norm": 3.474813461303711, "learning_rate": 9.5e-08, "loss": 8.4177, "step": 190 }, { "epoch": 0.05987460815047022, "grad_norm": 3.4020941257476807, "learning_rate": 9.55e-08, "loss": 10.3753, "step": 191 }, { "epoch": 0.06018808777429467, "grad_norm": 4.846778392791748, "learning_rate": 9.6e-08, "loss": 9.5903, "step": 192 }, { "epoch": 0.06050156739811912, "grad_norm": 4.567539691925049, "learning_rate": 9.650000000000001e-08, "loss": 9.7159, "step": 193 }, { "epoch": 0.060815047021943576, "grad_norm": 3.0759778022766113, "learning_rate": 9.700000000000001e-08, "loss": 7.0748, "step": 194 }, { "epoch": 0.061128526645768025, "grad_norm": 3.4420666694641113, "learning_rate": 9.75e-08, "loss": 7.8556, "step": 195 }, { "epoch": 0.061442006269592474, "grad_norm": 3.7455813884735107, "learning_rate": 9.8e-08, "loss": 7.4662, "step": 196 }, { "epoch": 0.06175548589341693, "grad_norm": 3.7451171875, "learning_rate": 9.85e-08, "loss": 8.2137, "step": 197 }, { "epoch": 0.06206896551724138, "grad_norm": 3.513235330581665, "learning_rate": 9.900000000000001e-08, "loss": 8.2557, "step": 198 }, { "epoch": 0.06238244514106583, "grad_norm": 4.297177314758301, "learning_rate": 9.950000000000001e-08, "loss": 10.0288, "step": 199 }, { "epoch": 0.06269592476489028, "grad_norm": 4.298759460449219, "learning_rate": 1.0000000000000001e-07, "loss": 10.5449, "step": 200 }, { "epoch": 0.06300940438871473, "grad_norm": 3.4075138568878174, "learning_rate": 1.005e-07, "loss": 7.7592, "step": 201 }, { "epoch": 0.06332288401253919, "grad_norm": 3.19218373298645, "learning_rate": 1.01e-07, "loss": 7.7692, "step": 202 }, { "epoch": 0.06363636363636363, "grad_norm": 4.200982093811035, "learning_rate": 1.015e-07, "loss": 8.7216, "step": 203 }, { "epoch": 0.06394984326018809, "grad_norm": 3.4024147987365723, "learning_rate": 1.0200000000000001e-07, "loss": 7.9235, "step": 204 }, { "epoch": 0.06426332288401254, "grad_norm": 3.751147747039795, "learning_rate": 1.0250000000000001e-07, "loss": 10.3086, "step": 205 }, { "epoch": 0.064576802507837, "grad_norm": 3.3769607543945312, "learning_rate": 1.0300000000000001e-07, "loss": 7.629, "step": 206 }, { "epoch": 0.06489028213166144, "grad_norm": 3.3052303791046143, "learning_rate": 1.0350000000000001e-07, "loss": 9.5893, "step": 207 }, { "epoch": 0.0652037617554859, "grad_norm": 4.127976894378662, "learning_rate": 1.04e-07, "loss": 12.7904, "step": 208 }, { "epoch": 0.06551724137931035, "grad_norm": 3.8568077087402344, "learning_rate": 1.045e-07, "loss": 9.6994, "step": 209 }, { "epoch": 0.06583072100313479, "grad_norm": 3.236760377883911, "learning_rate": 1.0500000000000001e-07, "loss": 8.0482, "step": 210 }, { "epoch": 0.06614420062695925, "grad_norm": 4.277583122253418, "learning_rate": 1.0550000000000001e-07, "loss": 8.1819, "step": 211 }, { "epoch": 0.0664576802507837, "grad_norm": 3.3786604404449463, "learning_rate": 1.0600000000000001e-07, "loss": 7.0545, "step": 212 }, { "epoch": 0.06677115987460815, "grad_norm": 4.007237911224365, "learning_rate": 1.0650000000000001e-07, "loss": 9.9794, "step": 213 }, { "epoch": 0.0670846394984326, "grad_norm": 4.616070747375488, "learning_rate": 1.07e-07, "loss": 11.1093, "step": 214 }, { "epoch": 0.06739811912225706, "grad_norm": 3.0655412673950195, "learning_rate": 1.075e-07, "loss": 9.5085, "step": 215 }, { "epoch": 0.0677115987460815, "grad_norm": 3.05678391456604, "learning_rate": 1.0800000000000001e-07, "loss": 6.3335, "step": 216 }, { "epoch": 0.06802507836990596, "grad_norm": 2.941502094268799, "learning_rate": 1.0850000000000001e-07, "loss": 7.2431, "step": 217 }, { "epoch": 0.06833855799373041, "grad_norm": 3.810181140899658, "learning_rate": 1.0900000000000001e-07, "loss": 9.8534, "step": 218 }, { "epoch": 0.06865203761755485, "grad_norm": 3.007319450378418, "learning_rate": 1.0950000000000001e-07, "loss": 9.2206, "step": 219 }, { "epoch": 0.06896551724137931, "grad_norm": 3.6502251625061035, "learning_rate": 1.1e-07, "loss": 7.9423, "step": 220 }, { "epoch": 0.06927899686520377, "grad_norm": 3.4484660625457764, "learning_rate": 1.1050000000000002e-07, "loss": 8.6399, "step": 221 }, { "epoch": 0.06959247648902821, "grad_norm": 3.3999218940734863, "learning_rate": 1.1100000000000001e-07, "loss": 8.1017, "step": 222 }, { "epoch": 0.06990595611285266, "grad_norm": 3.0518319606781006, "learning_rate": 1.1150000000000001e-07, "loss": 7.0595, "step": 223 }, { "epoch": 0.07021943573667712, "grad_norm": 2.940213918685913, "learning_rate": 1.1200000000000001e-07, "loss": 7.2927, "step": 224 }, { "epoch": 0.07053291536050156, "grad_norm": 4.038904190063477, "learning_rate": 1.1250000000000001e-07, "loss": 9.1408, "step": 225 }, { "epoch": 0.07084639498432602, "grad_norm": 2.5958571434020996, "learning_rate": 1.13e-07, "loss": 6.325, "step": 226 }, { "epoch": 0.07115987460815047, "grad_norm": 3.0129003524780273, "learning_rate": 1.1350000000000002e-07, "loss": 6.963, "step": 227 }, { "epoch": 0.07147335423197493, "grad_norm": 2.9724435806274414, "learning_rate": 1.1400000000000001e-07, "loss": 8.6995, "step": 228 }, { "epoch": 0.07178683385579937, "grad_norm": 3.7590036392211914, "learning_rate": 1.1450000000000001e-07, "loss": 7.9243, "step": 229 }, { "epoch": 0.07210031347962383, "grad_norm": 3.5400218963623047, "learning_rate": 1.1500000000000001e-07, "loss": 6.981, "step": 230 }, { "epoch": 0.07241379310344828, "grad_norm": 2.9733407497406006, "learning_rate": 1.1550000000000001e-07, "loss": 7.1957, "step": 231 }, { "epoch": 0.07272727272727272, "grad_norm": 3.4962241649627686, "learning_rate": 1.16e-07, "loss": 10.1103, "step": 232 }, { "epoch": 0.07304075235109718, "grad_norm": 3.234700918197632, "learning_rate": 1.1650000000000002e-07, "loss": 7.5018, "step": 233 }, { "epoch": 0.07335423197492164, "grad_norm": 3.2388176918029785, "learning_rate": 1.1700000000000002e-07, "loss": 9.0859, "step": 234 }, { "epoch": 0.07366771159874608, "grad_norm": 3.4157941341400146, "learning_rate": 1.1750000000000001e-07, "loss": 7.8399, "step": 235 }, { "epoch": 0.07398119122257053, "grad_norm": 3.5833077430725098, "learning_rate": 1.1800000000000001e-07, "loss": 7.5087, "step": 236 }, { "epoch": 0.07429467084639499, "grad_norm": 4.769215106964111, "learning_rate": 1.1850000000000001e-07, "loss": 8.7621, "step": 237 }, { "epoch": 0.07460815047021943, "grad_norm": 3.941755771636963, "learning_rate": 1.1900000000000002e-07, "loss": 13.078, "step": 238 }, { "epoch": 0.07492163009404389, "grad_norm": 3.798787832260132, "learning_rate": 1.195e-07, "loss": 9.7572, "step": 239 }, { "epoch": 0.07523510971786834, "grad_norm": 4.713004112243652, "learning_rate": 1.2000000000000002e-07, "loss": 13.3091, "step": 240 }, { "epoch": 0.07554858934169278, "grad_norm": 3.408093214035034, "learning_rate": 1.205e-07, "loss": 7.7499, "step": 241 }, { "epoch": 0.07586206896551724, "grad_norm": 4.062272071838379, "learning_rate": 1.21e-07, "loss": 9.7988, "step": 242 }, { "epoch": 0.0761755485893417, "grad_norm": 4.664130210876465, "learning_rate": 1.215e-07, "loss": 9.4779, "step": 243 }, { "epoch": 0.07648902821316614, "grad_norm": 3.5159950256347656, "learning_rate": 1.22e-07, "loss": 8.8292, "step": 244 }, { "epoch": 0.0768025078369906, "grad_norm": 3.664834499359131, "learning_rate": 1.2250000000000002e-07, "loss": 8.0994, "step": 245 }, { "epoch": 0.07711598746081505, "grad_norm": 4.765824794769287, "learning_rate": 1.23e-07, "loss": 8.5952, "step": 246 }, { "epoch": 0.07742946708463949, "grad_norm": 3.9765498638153076, "learning_rate": 1.2350000000000001e-07, "loss": 9.6514, "step": 247 }, { "epoch": 0.07774294670846395, "grad_norm": 3.6185436248779297, "learning_rate": 1.24e-07, "loss": 8.74, "step": 248 }, { "epoch": 0.0780564263322884, "grad_norm": 3.6925275325775146, "learning_rate": 1.245e-07, "loss": 7.0158, "step": 249 }, { "epoch": 0.07836990595611286, "grad_norm": 3.878559112548828, "learning_rate": 1.2500000000000002e-07, "loss": 8.7004, "step": 250 }, { "epoch": 0.0786833855799373, "grad_norm": 4.704588413238525, "learning_rate": 1.255e-07, "loss": 9.9536, "step": 251 }, { "epoch": 0.07899686520376176, "grad_norm": 4.152799129486084, "learning_rate": 1.2600000000000002e-07, "loss": 9.6934, "step": 252 }, { "epoch": 0.07931034482758621, "grad_norm": 3.1007423400878906, "learning_rate": 1.265e-07, "loss": 7.1812, "step": 253 }, { "epoch": 0.07962382445141065, "grad_norm": 3.872610092163086, "learning_rate": 1.27e-07, "loss": 8.3354, "step": 254 }, { "epoch": 0.07993730407523511, "grad_norm": 3.334554433822632, "learning_rate": 1.275e-07, "loss": 9.6324, "step": 255 }, { "epoch": 0.08025078369905957, "grad_norm": 3.413045883178711, "learning_rate": 1.28e-07, "loss": 8.7391, "step": 256 }, { "epoch": 0.08056426332288401, "grad_norm": 3.656658887863159, "learning_rate": 1.2850000000000002e-07, "loss": 10.2712, "step": 257 }, { "epoch": 0.08087774294670846, "grad_norm": 3.1734461784362793, "learning_rate": 1.29e-07, "loss": 7.0908, "step": 258 }, { "epoch": 0.08119122257053292, "grad_norm": 5.781704425811768, "learning_rate": 1.2950000000000001e-07, "loss": 8.8973, "step": 259 }, { "epoch": 0.08150470219435736, "grad_norm": 3.3216211795806885, "learning_rate": 1.3e-07, "loss": 7.9809, "step": 260 }, { "epoch": 0.08181818181818182, "grad_norm": 3.4130654335021973, "learning_rate": 1.305e-07, "loss": 10.6107, "step": 261 }, { "epoch": 0.08213166144200627, "grad_norm": 3.198608160018921, "learning_rate": 1.3100000000000002e-07, "loss": 7.9311, "step": 262 }, { "epoch": 0.08244514106583072, "grad_norm": 3.2649126052856445, "learning_rate": 1.315e-07, "loss": 6.9358, "step": 263 }, { "epoch": 0.08275862068965517, "grad_norm": 3.805680513381958, "learning_rate": 1.3200000000000002e-07, "loss": 8.7701, "step": 264 }, { "epoch": 0.08307210031347963, "grad_norm": 3.8234152793884277, "learning_rate": 1.325e-07, "loss": 7.1507, "step": 265 }, { "epoch": 0.08338557993730407, "grad_norm": 4.14029598236084, "learning_rate": 1.33e-07, "loss": 10.002, "step": 266 }, { "epoch": 0.08369905956112852, "grad_norm": 3.70548677444458, "learning_rate": 1.3350000000000002e-07, "loss": 7.0396, "step": 267 }, { "epoch": 0.08401253918495298, "grad_norm": 4.227056503295898, "learning_rate": 1.34e-07, "loss": 10.8817, "step": 268 }, { "epoch": 0.08432601880877742, "grad_norm": 3.6237289905548096, "learning_rate": 1.3450000000000002e-07, "loss": 8.0917, "step": 269 }, { "epoch": 0.08463949843260188, "grad_norm": 2.858159303665161, "learning_rate": 1.35e-07, "loss": 6.4492, "step": 270 }, { "epoch": 0.08495297805642633, "grad_norm": 3.7545058727264404, "learning_rate": 1.3550000000000002e-07, "loss": 7.2091, "step": 271 }, { "epoch": 0.08526645768025079, "grad_norm": 3.4545750617980957, "learning_rate": 1.36e-07, "loss": 6.7025, "step": 272 }, { "epoch": 0.08557993730407523, "grad_norm": 2.947838068008423, "learning_rate": 1.365e-07, "loss": 6.9652, "step": 273 }, { "epoch": 0.08589341692789969, "grad_norm": 4.7777605056762695, "learning_rate": 1.3700000000000002e-07, "loss": 10.056, "step": 274 }, { "epoch": 0.08620689655172414, "grad_norm": 2.519544839859009, "learning_rate": 1.375e-07, "loss": 6.0347, "step": 275 }, { "epoch": 0.08652037617554859, "grad_norm": 4.664981842041016, "learning_rate": 1.3800000000000002e-07, "loss": 10.88, "step": 276 }, { "epoch": 0.08683385579937304, "grad_norm": 4.797813892364502, "learning_rate": 1.385e-07, "loss": 10.9775, "step": 277 }, { "epoch": 0.0871473354231975, "grad_norm": 3.1882078647613525, "learning_rate": 1.3900000000000001e-07, "loss": 7.7702, "step": 278 }, { "epoch": 0.08746081504702194, "grad_norm": 4.286925792694092, "learning_rate": 1.3950000000000002e-07, "loss": 10.9283, "step": 279 }, { "epoch": 0.0877742946708464, "grad_norm": 3.3588335514068604, "learning_rate": 1.4e-07, "loss": 7.7342, "step": 280 }, { "epoch": 0.08808777429467085, "grad_norm": 2.775568962097168, "learning_rate": 1.4050000000000002e-07, "loss": 6.9301, "step": 281 }, { "epoch": 0.08840125391849529, "grad_norm": 4.123598575592041, "learning_rate": 1.41e-07, "loss": 10.0182, "step": 282 }, { "epoch": 0.08871473354231975, "grad_norm": 4.854518413543701, "learning_rate": 1.4150000000000002e-07, "loss": 12.1932, "step": 283 }, { "epoch": 0.0890282131661442, "grad_norm": 2.8263025283813477, "learning_rate": 1.4200000000000003e-07, "loss": 7.8395, "step": 284 }, { "epoch": 0.08934169278996865, "grad_norm": 3.2582967281341553, "learning_rate": 1.425e-07, "loss": 7.4587, "step": 285 }, { "epoch": 0.0896551724137931, "grad_norm": 3.1975438594818115, "learning_rate": 1.4300000000000002e-07, "loss": 8.037, "step": 286 }, { "epoch": 0.08996865203761756, "grad_norm": 3.0459468364715576, "learning_rate": 1.435e-07, "loss": 7.1786, "step": 287 }, { "epoch": 0.090282131661442, "grad_norm": 4.034944534301758, "learning_rate": 1.4400000000000002e-07, "loss": 8.2916, "step": 288 }, { "epoch": 0.09059561128526646, "grad_norm": 5.0129499435424805, "learning_rate": 1.445e-07, "loss": 11.0094, "step": 289 }, { "epoch": 0.09090909090909091, "grad_norm": 4.1220903396606445, "learning_rate": 1.4500000000000001e-07, "loss": 8.5799, "step": 290 }, { "epoch": 0.09122257053291537, "grad_norm": 3.6785085201263428, "learning_rate": 1.4550000000000003e-07, "loss": 9.5831, "step": 291 }, { "epoch": 0.09153605015673981, "grad_norm": 3.44356369972229, "learning_rate": 1.46e-07, "loss": 7.9535, "step": 292 }, { "epoch": 0.09184952978056427, "grad_norm": 4.070002555847168, "learning_rate": 1.4650000000000002e-07, "loss": 11.4279, "step": 293 }, { "epoch": 0.09216300940438872, "grad_norm": 3.0748281478881836, "learning_rate": 1.47e-07, "loss": 7.3921, "step": 294 }, { "epoch": 0.09247648902821316, "grad_norm": 2.973388671875, "learning_rate": 1.4750000000000002e-07, "loss": 7.6754, "step": 295 }, { "epoch": 0.09278996865203762, "grad_norm": 3.4642834663391113, "learning_rate": 1.4800000000000003e-07, "loss": 8.1933, "step": 296 }, { "epoch": 0.09310344827586207, "grad_norm": 4.090353012084961, "learning_rate": 1.485e-07, "loss": 7.926, "step": 297 }, { "epoch": 0.09341692789968652, "grad_norm": 3.8677382469177246, "learning_rate": 1.4900000000000002e-07, "loss": 7.5799, "step": 298 }, { "epoch": 0.09373040752351097, "grad_norm": 3.7805559635162354, "learning_rate": 1.495e-07, "loss": 9.6573, "step": 299 }, { "epoch": 0.09404388714733543, "grad_norm": 3.8729629516601562, "learning_rate": 1.5000000000000002e-07, "loss": 10.3939, "step": 300 }, { "epoch": 0.09435736677115987, "grad_norm": 4.191401481628418, "learning_rate": 1.505e-07, "loss": 11.4678, "step": 301 }, { "epoch": 0.09467084639498433, "grad_norm": 3.4422829151153564, "learning_rate": 1.5100000000000002e-07, "loss": 7.9499, "step": 302 }, { "epoch": 0.09498432601880878, "grad_norm": 3.8708252906799316, "learning_rate": 1.5150000000000003e-07, "loss": 9.9911, "step": 303 }, { "epoch": 0.09529780564263322, "grad_norm": 3.080570936203003, "learning_rate": 1.52e-07, "loss": 7.4803, "step": 304 }, { "epoch": 0.09561128526645768, "grad_norm": 3.9624032974243164, "learning_rate": 1.5250000000000002e-07, "loss": 10.5125, "step": 305 }, { "epoch": 0.09592476489028214, "grad_norm": 4.132358551025391, "learning_rate": 1.53e-07, "loss": 10.1554, "step": 306 }, { "epoch": 0.09623824451410658, "grad_norm": 3.146574020385742, "learning_rate": 1.5350000000000002e-07, "loss": 6.7796, "step": 307 }, { "epoch": 0.09655172413793103, "grad_norm": 3.1497974395751953, "learning_rate": 1.5400000000000003e-07, "loss": 8.9689, "step": 308 }, { "epoch": 0.09686520376175549, "grad_norm": 4.384578227996826, "learning_rate": 1.5450000000000001e-07, "loss": 12.2934, "step": 309 }, { "epoch": 0.09717868338557993, "grad_norm": 3.5674219131469727, "learning_rate": 1.5500000000000002e-07, "loss": 10.6616, "step": 310 }, { "epoch": 0.09749216300940439, "grad_norm": 4.463435649871826, "learning_rate": 1.555e-07, "loss": 7.4585, "step": 311 }, { "epoch": 0.09780564263322884, "grad_norm": 4.1657609939575195, "learning_rate": 1.56e-07, "loss": 8.5832, "step": 312 }, { "epoch": 0.0981191222570533, "grad_norm": 5.202960968017578, "learning_rate": 1.5650000000000003e-07, "loss": 12.9104, "step": 313 }, { "epoch": 0.09843260188087774, "grad_norm": 4.485203742980957, "learning_rate": 1.57e-07, "loss": 7.8665, "step": 314 }, { "epoch": 0.0987460815047022, "grad_norm": 3.092864751815796, "learning_rate": 1.575e-07, "loss": 8.4677, "step": 315 }, { "epoch": 0.09905956112852665, "grad_norm": 3.2465710639953613, "learning_rate": 1.5800000000000004e-07, "loss": 7.498, "step": 316 }, { "epoch": 0.0993730407523511, "grad_norm": 5.711244106292725, "learning_rate": 1.585e-07, "loss": 12.3313, "step": 317 }, { "epoch": 0.09968652037617555, "grad_norm": 3.7532050609588623, "learning_rate": 1.5900000000000003e-07, "loss": 10.8737, "step": 318 }, { "epoch": 0.1, "grad_norm": 4.172240257263184, "learning_rate": 1.595e-07, "loss": 8.8672, "step": 319 }, { "epoch": 0.10031347962382445, "grad_norm": 3.577533721923828, "learning_rate": 1.6e-07, "loss": 11.197, "step": 320 }, { "epoch": 0.1006269592476489, "grad_norm": 3.8616108894348145, "learning_rate": 1.605e-07, "loss": 9.6396, "step": 321 }, { "epoch": 0.10094043887147336, "grad_norm": 5.926366329193115, "learning_rate": 1.61e-07, "loss": 12.75, "step": 322 }, { "epoch": 0.1012539184952978, "grad_norm": 3.613895893096924, "learning_rate": 1.6150000000000004e-07, "loss": 7.4106, "step": 323 }, { "epoch": 0.10156739811912226, "grad_norm": 3.902071952819824, "learning_rate": 1.62e-07, "loss": 11.4143, "step": 324 }, { "epoch": 0.10188087774294671, "grad_norm": 2.8985440731048584, "learning_rate": 1.625e-07, "loss": 7.3166, "step": 325 }, { "epoch": 0.10219435736677115, "grad_norm": 3.545724391937256, "learning_rate": 1.63e-07, "loss": 7.8454, "step": 326 }, { "epoch": 0.10250783699059561, "grad_norm": 3.259948253631592, "learning_rate": 1.635e-07, "loss": 8.4287, "step": 327 }, { "epoch": 0.10282131661442007, "grad_norm": 3.207242012023926, "learning_rate": 1.6400000000000004e-07, "loss": 7.234, "step": 328 }, { "epoch": 0.10313479623824451, "grad_norm": 3.2320971488952637, "learning_rate": 1.645e-07, "loss": 7.4103, "step": 329 }, { "epoch": 0.10344827586206896, "grad_norm": 4.266717910766602, "learning_rate": 1.65e-07, "loss": 10.7557, "step": 330 }, { "epoch": 0.10376175548589342, "grad_norm": 4.703216075897217, "learning_rate": 1.655e-07, "loss": 13.6921, "step": 331 }, { "epoch": 0.10407523510971786, "grad_norm": 2.98171329498291, "learning_rate": 1.66e-07, "loss": 6.4217, "step": 332 }, { "epoch": 0.10438871473354232, "grad_norm": 4.266834735870361, "learning_rate": 1.6650000000000004e-07, "loss": 8.7814, "step": 333 }, { "epoch": 0.10470219435736677, "grad_norm": 3.13301420211792, "learning_rate": 1.67e-07, "loss": 7.5465, "step": 334 }, { "epoch": 0.10501567398119123, "grad_norm": 3.93803071975708, "learning_rate": 1.675e-07, "loss": 9.7471, "step": 335 }, { "epoch": 0.10532915360501567, "grad_norm": 3.124204397201538, "learning_rate": 1.68e-07, "loss": 8.2321, "step": 336 }, { "epoch": 0.10564263322884013, "grad_norm": 3.1418778896331787, "learning_rate": 1.685e-07, "loss": 8.8142, "step": 337 }, { "epoch": 0.10595611285266458, "grad_norm": 3.865973472595215, "learning_rate": 1.69e-07, "loss": 10.8355, "step": 338 }, { "epoch": 0.10626959247648902, "grad_norm": 3.5527279376983643, "learning_rate": 1.695e-07, "loss": 7.6143, "step": 339 }, { "epoch": 0.10658307210031348, "grad_norm": 3.353538751602173, "learning_rate": 1.7000000000000001e-07, "loss": 9.4709, "step": 340 }, { "epoch": 0.10689655172413794, "grad_norm": 3.7409756183624268, "learning_rate": 1.705e-07, "loss": 11.7136, "step": 341 }, { "epoch": 0.10721003134796238, "grad_norm": 4.518954277038574, "learning_rate": 1.71e-07, "loss": 8.8255, "step": 342 }, { "epoch": 0.10752351097178683, "grad_norm": 4.619668960571289, "learning_rate": 1.715e-07, "loss": 12.3065, "step": 343 }, { "epoch": 0.10783699059561129, "grad_norm": 3.260843276977539, "learning_rate": 1.72e-07, "loss": 9.3232, "step": 344 }, { "epoch": 0.10815047021943573, "grad_norm": 3.9291486740112305, "learning_rate": 1.7250000000000002e-07, "loss": 9.9018, "step": 345 }, { "epoch": 0.10846394984326019, "grad_norm": 4.316411972045898, "learning_rate": 1.73e-07, "loss": 8.0697, "step": 346 }, { "epoch": 0.10877742946708464, "grad_norm": 3.2413835525512695, "learning_rate": 1.735e-07, "loss": 8.5543, "step": 347 }, { "epoch": 0.10909090909090909, "grad_norm": 2.881052017211914, "learning_rate": 1.74e-07, "loss": 6.481, "step": 348 }, { "epoch": 0.10940438871473354, "grad_norm": 3.0879931449890137, "learning_rate": 1.745e-07, "loss": 9.835, "step": 349 }, { "epoch": 0.109717868338558, "grad_norm": 3.5517656803131104, "learning_rate": 1.7500000000000002e-07, "loss": 8.4705, "step": 350 }, { "epoch": 0.11003134796238244, "grad_norm": 3.764478921890259, "learning_rate": 1.755e-07, "loss": 9.458, "step": 351 }, { "epoch": 0.1103448275862069, "grad_norm": 5.203456878662109, "learning_rate": 1.7600000000000001e-07, "loss": 11.9955, "step": 352 }, { "epoch": 0.11065830721003135, "grad_norm": 4.887335777282715, "learning_rate": 1.765e-07, "loss": 11.2742, "step": 353 }, { "epoch": 0.11097178683385579, "grad_norm": 3.6964869499206543, "learning_rate": 1.77e-07, "loss": 8.1842, "step": 354 }, { "epoch": 0.11128526645768025, "grad_norm": 3.0974836349487305, "learning_rate": 1.775e-07, "loss": 7.7249, "step": 355 }, { "epoch": 0.1115987460815047, "grad_norm": 3.3595480918884277, "learning_rate": 1.78e-07, "loss": 7.27, "step": 356 }, { "epoch": 0.11191222570532916, "grad_norm": 3.602126359939575, "learning_rate": 1.7850000000000002e-07, "loss": 10.0521, "step": 357 }, { "epoch": 0.1122257053291536, "grad_norm": 3.0182366371154785, "learning_rate": 1.79e-07, "loss": 8.1576, "step": 358 }, { "epoch": 0.11253918495297806, "grad_norm": 7.666667461395264, "learning_rate": 1.795e-07, "loss": 17.8031, "step": 359 }, { "epoch": 0.11285266457680251, "grad_norm": 3.1000211238861084, "learning_rate": 1.8e-07, "loss": 8.0788, "step": 360 }, { "epoch": 0.11316614420062696, "grad_norm": 3.830988883972168, "learning_rate": 1.805e-07, "loss": 10.3451, "step": 361 }, { "epoch": 0.11347962382445141, "grad_norm": 3.502091407775879, "learning_rate": 1.8100000000000002e-07, "loss": 7.3287, "step": 362 }, { "epoch": 0.11379310344827587, "grad_norm": 3.8298943042755127, "learning_rate": 1.815e-07, "loss": 7.3115, "step": 363 }, { "epoch": 0.11410658307210031, "grad_norm": 3.5254364013671875, "learning_rate": 1.8200000000000002e-07, "loss": 10.1726, "step": 364 }, { "epoch": 0.11442006269592477, "grad_norm": 2.8689956665039062, "learning_rate": 1.825e-07, "loss": 5.9041, "step": 365 }, { "epoch": 0.11473354231974922, "grad_norm": 3.9380292892456055, "learning_rate": 1.83e-07, "loss": 8.7651, "step": 366 }, { "epoch": 0.11504702194357366, "grad_norm": 4.940433502197266, "learning_rate": 1.8350000000000002e-07, "loss": 13.3309, "step": 367 }, { "epoch": 0.11536050156739812, "grad_norm": 3.1874752044677734, "learning_rate": 1.84e-07, "loss": 7.9231, "step": 368 }, { "epoch": 0.11567398119122257, "grad_norm": 3.361844778060913, "learning_rate": 1.8450000000000002e-07, "loss": 8.2631, "step": 369 }, { "epoch": 0.11598746081504702, "grad_norm": 4.501951217651367, "learning_rate": 1.85e-07, "loss": 11.7963, "step": 370 }, { "epoch": 0.11630094043887147, "grad_norm": 3.1025617122650146, "learning_rate": 1.8550000000000001e-07, "loss": 7.3833, "step": 371 }, { "epoch": 0.11661442006269593, "grad_norm": 3.30342960357666, "learning_rate": 1.86e-07, "loss": 8.7881, "step": 372 }, { "epoch": 0.11692789968652037, "grad_norm": 3.470710515975952, "learning_rate": 1.865e-07, "loss": 7.7015, "step": 373 }, { "epoch": 0.11724137931034483, "grad_norm": 3.2457809448242188, "learning_rate": 1.8700000000000002e-07, "loss": 8.8045, "step": 374 }, { "epoch": 0.11755485893416928, "grad_norm": 3.0047662258148193, "learning_rate": 1.875e-07, "loss": 8.3349, "step": 375 }, { "epoch": 0.11786833855799372, "grad_norm": 3.3185179233551025, "learning_rate": 1.8800000000000002e-07, "loss": 6.7955, "step": 376 }, { "epoch": 0.11818181818181818, "grad_norm": 3.7260663509368896, "learning_rate": 1.885e-07, "loss": 8.8171, "step": 377 }, { "epoch": 0.11849529780564264, "grad_norm": 3.9109082221984863, "learning_rate": 1.89e-07, "loss": 9.8559, "step": 378 }, { "epoch": 0.11880877742946709, "grad_norm": 3.2869834899902344, "learning_rate": 1.8950000000000002e-07, "loss": 11.0736, "step": 379 }, { "epoch": 0.11912225705329153, "grad_norm": 4.131401538848877, "learning_rate": 1.9e-07, "loss": 11.9681, "step": 380 }, { "epoch": 0.11943573667711599, "grad_norm": 3.241637945175171, "learning_rate": 1.9050000000000002e-07, "loss": 9.0305, "step": 381 }, { "epoch": 0.11974921630094044, "grad_norm": 3.635667324066162, "learning_rate": 1.91e-07, "loss": 9.6321, "step": 382 }, { "epoch": 0.12006269592476489, "grad_norm": 2.901669979095459, "learning_rate": 1.9150000000000001e-07, "loss": 7.8995, "step": 383 }, { "epoch": 0.12037617554858934, "grad_norm": 3.7463796138763428, "learning_rate": 1.92e-07, "loss": 7.9248, "step": 384 }, { "epoch": 0.1206896551724138, "grad_norm": 3.811286211013794, "learning_rate": 1.925e-07, "loss": 7.6093, "step": 385 }, { "epoch": 0.12100313479623824, "grad_norm": 3.3971803188323975, "learning_rate": 1.9300000000000002e-07, "loss": 7.9361, "step": 386 }, { "epoch": 0.1213166144200627, "grad_norm": 3.552212953567505, "learning_rate": 1.935e-07, "loss": 9.3616, "step": 387 }, { "epoch": 0.12163009404388715, "grad_norm": 3.006312608718872, "learning_rate": 1.9400000000000002e-07, "loss": 8.1449, "step": 388 }, { "epoch": 0.1219435736677116, "grad_norm": 3.3143630027770996, "learning_rate": 1.945e-07, "loss": 8.2126, "step": 389 }, { "epoch": 0.12225705329153605, "grad_norm": 3.8813610076904297, "learning_rate": 1.95e-07, "loss": 10.1397, "step": 390 }, { "epoch": 0.1225705329153605, "grad_norm": 2.9321842193603516, "learning_rate": 1.9550000000000002e-07, "loss": 6.3588, "step": 391 }, { "epoch": 0.12288401253918495, "grad_norm": 3.752800941467285, "learning_rate": 1.96e-07, "loss": 8.3612, "step": 392 }, { "epoch": 0.1231974921630094, "grad_norm": 3.759476661682129, "learning_rate": 1.9650000000000002e-07, "loss": 8.3461, "step": 393 }, { "epoch": 0.12351097178683386, "grad_norm": 2.937373161315918, "learning_rate": 1.97e-07, "loss": 6.9189, "step": 394 }, { "epoch": 0.1238244514106583, "grad_norm": 3.327111005783081, "learning_rate": 1.9750000000000001e-07, "loss": 9.0745, "step": 395 }, { "epoch": 0.12413793103448276, "grad_norm": 2.807039976119995, "learning_rate": 1.9800000000000003e-07, "loss": 6.5768, "step": 396 }, { "epoch": 0.12445141065830721, "grad_norm": 4.174190044403076, "learning_rate": 1.985e-07, "loss": 14.798, "step": 397 }, { "epoch": 0.12476489028213165, "grad_norm": 3.8052163124084473, "learning_rate": 1.9900000000000002e-07, "loss": 7.4751, "step": 398 }, { "epoch": 0.1250783699059561, "grad_norm": 3.2749545574188232, "learning_rate": 1.995e-07, "loss": 6.8837, "step": 399 }, { "epoch": 0.12539184952978055, "grad_norm": 5.2682695388793945, "learning_rate": 2.0000000000000002e-07, "loss": 17.3949, "step": 400 }, { "epoch": 0.12570532915360502, "grad_norm": 3.6819639205932617, "learning_rate": 2.005e-07, "loss": 8.8539, "step": 401 }, { "epoch": 0.12601880877742946, "grad_norm": 3.480807304382324, "learning_rate": 2.01e-07, "loss": 7.9894, "step": 402 }, { "epoch": 0.1263322884012539, "grad_norm": 4.27021598815918, "learning_rate": 2.0150000000000002e-07, "loss": 10.9579, "step": 403 }, { "epoch": 0.12664576802507838, "grad_norm": 3.2545857429504395, "learning_rate": 2.02e-07, "loss": 7.1311, "step": 404 }, { "epoch": 0.12695924764890282, "grad_norm": 3.5351126194000244, "learning_rate": 2.0250000000000002e-07, "loss": 6.8761, "step": 405 }, { "epoch": 0.12727272727272726, "grad_norm": 3.6219284534454346, "learning_rate": 2.03e-07, "loss": 8.3192, "step": 406 }, { "epoch": 0.12758620689655173, "grad_norm": 3.1589441299438477, "learning_rate": 2.0350000000000002e-07, "loss": 7.9498, "step": 407 }, { "epoch": 0.12789968652037617, "grad_norm": 3.862013101577759, "learning_rate": 2.0400000000000003e-07, "loss": 12.2729, "step": 408 }, { "epoch": 0.1282131661442006, "grad_norm": 3.1884796619415283, "learning_rate": 2.045e-07, "loss": 6.5746, "step": 409 }, { "epoch": 0.12852664576802508, "grad_norm": 2.733370065689087, "learning_rate": 2.0500000000000002e-07, "loss": 6.8392, "step": 410 }, { "epoch": 0.12884012539184952, "grad_norm": 4.183924198150635, "learning_rate": 2.055e-07, "loss": 9.7881, "step": 411 }, { "epoch": 0.129153605015674, "grad_norm": 4.186039924621582, "learning_rate": 2.0600000000000002e-07, "loss": 7.9172, "step": 412 }, { "epoch": 0.12946708463949844, "grad_norm": 4.129965305328369, "learning_rate": 2.0650000000000003e-07, "loss": 8.3458, "step": 413 }, { "epoch": 0.12978056426332288, "grad_norm": 3.3639910221099854, "learning_rate": 2.0700000000000001e-07, "loss": 7.8052, "step": 414 }, { "epoch": 0.13009404388714735, "grad_norm": 3.8352103233337402, "learning_rate": 2.0750000000000003e-07, "loss": 10.0176, "step": 415 }, { "epoch": 0.1304075235109718, "grad_norm": 3.672574281692505, "learning_rate": 2.08e-07, "loss": 10.4668, "step": 416 }, { "epoch": 0.13072100313479623, "grad_norm": 2.9237465858459473, "learning_rate": 2.0850000000000002e-07, "loss": 5.7591, "step": 417 }, { "epoch": 0.1310344827586207, "grad_norm": 4.392944812774658, "learning_rate": 2.09e-07, "loss": 10.9199, "step": 418 }, { "epoch": 0.13134796238244514, "grad_norm": 4.115564823150635, "learning_rate": 2.0950000000000002e-07, "loss": 7.3922, "step": 419 }, { "epoch": 0.13166144200626959, "grad_norm": 3.255119562149048, "learning_rate": 2.1000000000000003e-07, "loss": 7.2558, "step": 420 }, { "epoch": 0.13197492163009406, "grad_norm": 2.9578206539154053, "learning_rate": 2.105e-07, "loss": 8.0258, "step": 421 }, { "epoch": 0.1322884012539185, "grad_norm": 3.8864405155181885, "learning_rate": 2.1100000000000002e-07, "loss": 9.2595, "step": 422 }, { "epoch": 0.13260188087774294, "grad_norm": 3.2227253913879395, "learning_rate": 2.115e-07, "loss": 7.2088, "step": 423 }, { "epoch": 0.1329153605015674, "grad_norm": 3.5899605751037598, "learning_rate": 2.1200000000000002e-07, "loss": 8.3448, "step": 424 }, { "epoch": 0.13322884012539185, "grad_norm": 3.725222110748291, "learning_rate": 2.1250000000000003e-07, "loss": 8.4943, "step": 425 }, { "epoch": 0.1335423197492163, "grad_norm": 4.701845645904541, "learning_rate": 2.1300000000000001e-07, "loss": 12.9665, "step": 426 }, { "epoch": 0.13385579937304076, "grad_norm": 3.239241123199463, "learning_rate": 2.1350000000000003e-07, "loss": 7.3724, "step": 427 }, { "epoch": 0.1341692789968652, "grad_norm": 2.888547420501709, "learning_rate": 2.14e-07, "loss": 7.2136, "step": 428 }, { "epoch": 0.13448275862068965, "grad_norm": 3.1289238929748535, "learning_rate": 2.1450000000000002e-07, "loss": 8.2064, "step": 429 }, { "epoch": 0.13479623824451412, "grad_norm": 3.992865800857544, "learning_rate": 2.15e-07, "loss": 8.4046, "step": 430 }, { "epoch": 0.13510971786833856, "grad_norm": 3.2021284103393555, "learning_rate": 2.1550000000000002e-07, "loss": 12.1033, "step": 431 }, { "epoch": 0.135423197492163, "grad_norm": 3.114281177520752, "learning_rate": 2.1600000000000003e-07, "loss": 7.0616, "step": 432 }, { "epoch": 0.13573667711598747, "grad_norm": 3.175734043121338, "learning_rate": 2.165e-07, "loss": 7.1838, "step": 433 }, { "epoch": 0.1360501567398119, "grad_norm": 3.4493837356567383, "learning_rate": 2.1700000000000002e-07, "loss": 7.9487, "step": 434 }, { "epoch": 0.13636363636363635, "grad_norm": 3.04085373878479, "learning_rate": 2.175e-07, "loss": 7.6261, "step": 435 }, { "epoch": 0.13667711598746082, "grad_norm": 4.71550989151001, "learning_rate": 2.1800000000000002e-07, "loss": 13.7869, "step": 436 }, { "epoch": 0.13699059561128527, "grad_norm": 3.4197592735290527, "learning_rate": 2.1850000000000003e-07, "loss": 7.7423, "step": 437 }, { "epoch": 0.1373040752351097, "grad_norm": 3.3364362716674805, "learning_rate": 2.1900000000000002e-07, "loss": 8.9017, "step": 438 }, { "epoch": 0.13761755485893418, "grad_norm": 3.6165454387664795, "learning_rate": 2.1950000000000003e-07, "loss": 7.4735, "step": 439 }, { "epoch": 0.13793103448275862, "grad_norm": 5.154653072357178, "learning_rate": 2.2e-07, "loss": 14.2502, "step": 440 }, { "epoch": 0.13824451410658306, "grad_norm": 3.3679981231689453, "learning_rate": 2.2050000000000002e-07, "loss": 7.7449, "step": 441 }, { "epoch": 0.13855799373040753, "grad_norm": 3.918351888656616, "learning_rate": 2.2100000000000003e-07, "loss": 10.1961, "step": 442 }, { "epoch": 0.13887147335423197, "grad_norm": 3.199449300765991, "learning_rate": 2.2150000000000002e-07, "loss": 7.5792, "step": 443 }, { "epoch": 0.13918495297805641, "grad_norm": 3.322082757949829, "learning_rate": 2.2200000000000003e-07, "loss": 7.9956, "step": 444 }, { "epoch": 0.13949843260188088, "grad_norm": 3.1004014015197754, "learning_rate": 2.2250000000000001e-07, "loss": 7.0897, "step": 445 }, { "epoch": 0.13981191222570533, "grad_norm": 3.9262382984161377, "learning_rate": 2.2300000000000002e-07, "loss": 9.4931, "step": 446 }, { "epoch": 0.14012539184952977, "grad_norm": 3.103273868560791, "learning_rate": 2.235e-07, "loss": 6.7228, "step": 447 }, { "epoch": 0.14043887147335424, "grad_norm": 3.1082265377044678, "learning_rate": 2.2400000000000002e-07, "loss": 8.3279, "step": 448 }, { "epoch": 0.14075235109717868, "grad_norm": 3.3036298751831055, "learning_rate": 2.2450000000000003e-07, "loss": 7.875, "step": 449 }, { "epoch": 0.14106583072100312, "grad_norm": 3.8810064792633057, "learning_rate": 2.2500000000000002e-07, "loss": 8.3544, "step": 450 }, { "epoch": 0.1413793103448276, "grad_norm": 4.209482192993164, "learning_rate": 2.2550000000000003e-07, "loss": 9.1295, "step": 451 }, { "epoch": 0.14169278996865203, "grad_norm": 3.419205665588379, "learning_rate": 2.26e-07, "loss": 6.9125, "step": 452 }, { "epoch": 0.1420062695924765, "grad_norm": 3.985367774963379, "learning_rate": 2.2650000000000002e-07, "loss": 11.5769, "step": 453 }, { "epoch": 0.14231974921630094, "grad_norm": 3.542846202850342, "learning_rate": 2.2700000000000003e-07, "loss": 8.6733, "step": 454 }, { "epoch": 0.1426332288401254, "grad_norm": 3.4622037410736084, "learning_rate": 2.2750000000000002e-07, "loss": 9.0828, "step": 455 }, { "epoch": 0.14294670846394986, "grad_norm": 3.1978418827056885, "learning_rate": 2.2800000000000003e-07, "loss": 6.9524, "step": 456 }, { "epoch": 0.1432601880877743, "grad_norm": 3.3328914642333984, "learning_rate": 2.2850000000000001e-07, "loss": 7.4014, "step": 457 }, { "epoch": 0.14357366771159874, "grad_norm": 3.5065202713012695, "learning_rate": 2.2900000000000003e-07, "loss": 9.0877, "step": 458 }, { "epoch": 0.1438871473354232, "grad_norm": 6.033443927764893, "learning_rate": 2.2950000000000004e-07, "loss": 15.9271, "step": 459 }, { "epoch": 0.14420062695924765, "grad_norm": 4.368375301361084, "learning_rate": 2.3000000000000002e-07, "loss": 11.6588, "step": 460 }, { "epoch": 0.1445141065830721, "grad_norm": 3.0983519554138184, "learning_rate": 2.3050000000000003e-07, "loss": 6.6645, "step": 461 }, { "epoch": 0.14482758620689656, "grad_norm": 3.8745956420898438, "learning_rate": 2.3100000000000002e-07, "loss": 10.457, "step": 462 }, { "epoch": 0.145141065830721, "grad_norm": 3.9520504474639893, "learning_rate": 2.3150000000000003e-07, "loss": 10.4561, "step": 463 }, { "epoch": 0.14545454545454545, "grad_norm": 8.04711627960205, "learning_rate": 2.32e-07, "loss": 19.9331, "step": 464 }, { "epoch": 0.14576802507836992, "grad_norm": 4.129817485809326, "learning_rate": 2.3250000000000002e-07, "loss": 8.8334, "step": 465 }, { "epoch": 0.14608150470219436, "grad_norm": 4.049325466156006, "learning_rate": 2.3300000000000003e-07, "loss": 8.772, "step": 466 }, { "epoch": 0.1463949843260188, "grad_norm": 4.55631160736084, "learning_rate": 2.3350000000000002e-07, "loss": 13.48, "step": 467 }, { "epoch": 0.14670846394984327, "grad_norm": 3.488584518432617, "learning_rate": 2.3400000000000003e-07, "loss": 10.4798, "step": 468 }, { "epoch": 0.1470219435736677, "grad_norm": 5.771771430969238, "learning_rate": 2.3450000000000002e-07, "loss": 11.7797, "step": 469 }, { "epoch": 0.14733542319749215, "grad_norm": 3.116539716720581, "learning_rate": 2.3500000000000003e-07, "loss": 6.873, "step": 470 }, { "epoch": 0.14764890282131662, "grad_norm": 2.576658248901367, "learning_rate": 2.3550000000000004e-07, "loss": 7.21, "step": 471 }, { "epoch": 0.14796238244514107, "grad_norm": 3.1541895866394043, "learning_rate": 2.3600000000000002e-07, "loss": 7.4172, "step": 472 }, { "epoch": 0.1482758620689655, "grad_norm": 4.149017810821533, "learning_rate": 2.3650000000000003e-07, "loss": 13.571, "step": 473 }, { "epoch": 0.14858934169278998, "grad_norm": 4.210467338562012, "learning_rate": 2.3700000000000002e-07, "loss": 11.6499, "step": 474 }, { "epoch": 0.14890282131661442, "grad_norm": 5.193135738372803, "learning_rate": 2.3750000000000003e-07, "loss": 17.2001, "step": 475 }, { "epoch": 0.14921630094043886, "grad_norm": 3.6134843826293945, "learning_rate": 2.3800000000000004e-07, "loss": 6.4639, "step": 476 }, { "epoch": 0.14952978056426333, "grad_norm": 5.032776355743408, "learning_rate": 2.385e-07, "loss": 13.8185, "step": 477 }, { "epoch": 0.14984326018808777, "grad_norm": 4.616049766540527, "learning_rate": 2.39e-07, "loss": 10.5861, "step": 478 }, { "epoch": 0.15015673981191222, "grad_norm": 3.1708860397338867, "learning_rate": 2.395e-07, "loss": 7.8613, "step": 479 }, { "epoch": 0.15047021943573669, "grad_norm": 4.432284355163574, "learning_rate": 2.4000000000000003e-07, "loss": 10.7543, "step": 480 }, { "epoch": 0.15078369905956113, "grad_norm": 3.439668655395508, "learning_rate": 2.405e-07, "loss": 7.4266, "step": 481 }, { "epoch": 0.15109717868338557, "grad_norm": 4.03348445892334, "learning_rate": 2.41e-07, "loss": 11.5268, "step": 482 }, { "epoch": 0.15141065830721004, "grad_norm": 3.4068636894226074, "learning_rate": 2.4150000000000004e-07, "loss": 7.1183, "step": 483 }, { "epoch": 0.15172413793103448, "grad_norm": 4.846498966217041, "learning_rate": 2.42e-07, "loss": 11.0191, "step": 484 }, { "epoch": 0.15203761755485892, "grad_norm": 5.300605773925781, "learning_rate": 2.425e-07, "loss": 9.6552, "step": 485 }, { "epoch": 0.1523510971786834, "grad_norm": 4.646756649017334, "learning_rate": 2.43e-07, "loss": 13.7772, "step": 486 }, { "epoch": 0.15266457680250783, "grad_norm": 3.7247793674468994, "learning_rate": 2.4350000000000003e-07, "loss": 8.6148, "step": 487 }, { "epoch": 0.15297805642633228, "grad_norm": 3.396324634552002, "learning_rate": 2.44e-07, "loss": 11.5777, "step": 488 }, { "epoch": 0.15329153605015675, "grad_norm": 3.7958171367645264, "learning_rate": 2.445e-07, "loss": 8.7151, "step": 489 }, { "epoch": 0.1536050156739812, "grad_norm": 3.312392473220825, "learning_rate": 2.4500000000000004e-07, "loss": 8.2649, "step": 490 }, { "epoch": 0.15391849529780563, "grad_norm": 2.4669861793518066, "learning_rate": 2.455e-07, "loss": 6.4529, "step": 491 }, { "epoch": 0.1542319749216301, "grad_norm": 5.108162879943848, "learning_rate": 2.46e-07, "loss": 14.7079, "step": 492 }, { "epoch": 0.15454545454545454, "grad_norm": 3.430584192276001, "learning_rate": 2.465e-07, "loss": 8.4653, "step": 493 }, { "epoch": 0.15485893416927898, "grad_norm": 3.0151519775390625, "learning_rate": 2.4700000000000003e-07, "loss": 6.9225, "step": 494 }, { "epoch": 0.15517241379310345, "grad_norm": 4.027904510498047, "learning_rate": 2.475e-07, "loss": 8.5251, "step": 495 }, { "epoch": 0.1554858934169279, "grad_norm": 3.500556707382202, "learning_rate": 2.48e-07, "loss": 9.9441, "step": 496 }, { "epoch": 0.15579937304075236, "grad_norm": 3.032302141189575, "learning_rate": 2.4850000000000003e-07, "loss": 6.9624, "step": 497 }, { "epoch": 0.1561128526645768, "grad_norm": 4.127997875213623, "learning_rate": 2.49e-07, "loss": 9.5575, "step": 498 }, { "epoch": 0.15642633228840125, "grad_norm": 3.3242340087890625, "learning_rate": 2.495e-07, "loss": 7.8737, "step": 499 }, { "epoch": 0.15673981191222572, "grad_norm": 3.514852285385132, "learning_rate": 2.5000000000000004e-07, "loss": 7.5601, "step": 500 }, { "epoch": 0.15705329153605016, "grad_norm": 3.9590742588043213, "learning_rate": 2.505e-07, "loss": 8.2909, "step": 501 }, { "epoch": 0.1573667711598746, "grad_norm": 3.0089352130889893, "learning_rate": 2.51e-07, "loss": 8.9848, "step": 502 }, { "epoch": 0.15768025078369907, "grad_norm": 4.294312477111816, "learning_rate": 2.515e-07, "loss": 10.3314, "step": 503 }, { "epoch": 0.1579937304075235, "grad_norm": 4.103808879852295, "learning_rate": 2.5200000000000003e-07, "loss": 9.262, "step": 504 }, { "epoch": 0.15830721003134796, "grad_norm": 3.4118666648864746, "learning_rate": 2.525e-07, "loss": 7.6306, "step": 505 }, { "epoch": 0.15862068965517243, "grad_norm": 3.144343137741089, "learning_rate": 2.53e-07, "loss": 9.7107, "step": 506 }, { "epoch": 0.15893416927899687, "grad_norm": 3.416949987411499, "learning_rate": 2.5350000000000004e-07, "loss": 7.1229, "step": 507 }, { "epoch": 0.1592476489028213, "grad_norm": 3.30198335647583, "learning_rate": 2.54e-07, "loss": 8.9435, "step": 508 }, { "epoch": 0.15956112852664578, "grad_norm": 3.7316057682037354, "learning_rate": 2.545e-07, "loss": 8.5075, "step": 509 }, { "epoch": 0.15987460815047022, "grad_norm": 3.468623638153076, "learning_rate": 2.55e-07, "loss": 8.5576, "step": 510 }, { "epoch": 0.16018808777429466, "grad_norm": 3.579921245574951, "learning_rate": 2.5550000000000003e-07, "loss": 8.5891, "step": 511 }, { "epoch": 0.16050156739811913, "grad_norm": 2.719144105911255, "learning_rate": 2.56e-07, "loss": 7.0097, "step": 512 }, { "epoch": 0.16081504702194357, "grad_norm": 3.503176689147949, "learning_rate": 2.565e-07, "loss": 9.2646, "step": 513 }, { "epoch": 0.16112852664576802, "grad_norm": 3.4942426681518555, "learning_rate": 2.5700000000000004e-07, "loss": 8.3605, "step": 514 }, { "epoch": 0.1614420062695925, "grad_norm": 3.6798880100250244, "learning_rate": 2.575e-07, "loss": 7.8945, "step": 515 }, { "epoch": 0.16175548589341693, "grad_norm": 2.953244924545288, "learning_rate": 2.58e-07, "loss": 6.319, "step": 516 }, { "epoch": 0.16206896551724137, "grad_norm": 4.936113357543945, "learning_rate": 2.5850000000000004e-07, "loss": 13.3042, "step": 517 }, { "epoch": 0.16238244514106584, "grad_norm": 4.359250545501709, "learning_rate": 2.5900000000000003e-07, "loss": 8.4494, "step": 518 }, { "epoch": 0.16269592476489028, "grad_norm": 3.1021082401275635, "learning_rate": 2.595e-07, "loss": 8.4413, "step": 519 }, { "epoch": 0.16300940438871472, "grad_norm": 3.71295166015625, "learning_rate": 2.6e-07, "loss": 9.4076, "step": 520 }, { "epoch": 0.1633228840125392, "grad_norm": 3.4278814792633057, "learning_rate": 2.6050000000000004e-07, "loss": 8.9707, "step": 521 }, { "epoch": 0.16363636363636364, "grad_norm": 4.640440464019775, "learning_rate": 2.61e-07, "loss": 12.3595, "step": 522 }, { "epoch": 0.16394984326018808, "grad_norm": 4.346306800842285, "learning_rate": 2.615e-07, "loss": 11.8391, "step": 523 }, { "epoch": 0.16426332288401255, "grad_norm": 3.444598436355591, "learning_rate": 2.6200000000000004e-07, "loss": 8.5561, "step": 524 }, { "epoch": 0.164576802507837, "grad_norm": 3.9495432376861572, "learning_rate": 2.6250000000000003e-07, "loss": 11.1607, "step": 525 }, { "epoch": 0.16489028213166143, "grad_norm": 3.858132839202881, "learning_rate": 2.63e-07, "loss": 8.4288, "step": 526 }, { "epoch": 0.1652037617554859, "grad_norm": 4.880698204040527, "learning_rate": 2.635e-07, "loss": 17.5724, "step": 527 }, { "epoch": 0.16551724137931034, "grad_norm": 5.0397162437438965, "learning_rate": 2.6400000000000003e-07, "loss": 10.3173, "step": 528 }, { "epoch": 0.16583072100313478, "grad_norm": 3.1536805629730225, "learning_rate": 2.645e-07, "loss": 7.8392, "step": 529 }, { "epoch": 0.16614420062695925, "grad_norm": 2.8486595153808594, "learning_rate": 2.65e-07, "loss": 7.3892, "step": 530 }, { "epoch": 0.1664576802507837, "grad_norm": 3.6035428047180176, "learning_rate": 2.6550000000000004e-07, "loss": 8.3855, "step": 531 }, { "epoch": 0.16677115987460814, "grad_norm": 3.0228993892669678, "learning_rate": 2.66e-07, "loss": 7.0978, "step": 532 }, { "epoch": 0.1670846394984326, "grad_norm": 3.3057663440704346, "learning_rate": 2.665e-07, "loss": 8.4415, "step": 533 }, { "epoch": 0.16739811912225705, "grad_norm": 3.630830764770508, "learning_rate": 2.6700000000000005e-07, "loss": 7.5833, "step": 534 }, { "epoch": 0.1677115987460815, "grad_norm": 4.190485954284668, "learning_rate": 2.6750000000000003e-07, "loss": 6.7527, "step": 535 }, { "epoch": 0.16802507836990596, "grad_norm": 4.3880486488342285, "learning_rate": 2.68e-07, "loss": 10.2591, "step": 536 }, { "epoch": 0.1683385579937304, "grad_norm": 3.767117738723755, "learning_rate": 2.685e-07, "loss": 6.5155, "step": 537 }, { "epoch": 0.16865203761755485, "grad_norm": 4.589762210845947, "learning_rate": 2.6900000000000004e-07, "loss": 11.3011, "step": 538 }, { "epoch": 0.16896551724137931, "grad_norm": 2.933931350708008, "learning_rate": 2.695e-07, "loss": 7.7405, "step": 539 }, { "epoch": 0.16927899686520376, "grad_norm": 4.077798843383789, "learning_rate": 2.7e-07, "loss": 7.7766, "step": 540 }, { "epoch": 0.16959247648902823, "grad_norm": 4.123392581939697, "learning_rate": 2.7050000000000005e-07, "loss": 8.437, "step": 541 }, { "epoch": 0.16990595611285267, "grad_norm": 3.4657938480377197, "learning_rate": 2.7100000000000003e-07, "loss": 8.0518, "step": 542 }, { "epoch": 0.1702194357366771, "grad_norm": 3.4541122913360596, "learning_rate": 2.715e-07, "loss": 8.7644, "step": 543 }, { "epoch": 0.17053291536050158, "grad_norm": 4.604305267333984, "learning_rate": 2.72e-07, "loss": 10.5831, "step": 544 }, { "epoch": 0.17084639498432602, "grad_norm": 3.2504711151123047, "learning_rate": 2.7250000000000004e-07, "loss": 8.0843, "step": 545 }, { "epoch": 0.17115987460815046, "grad_norm": 3.421630382537842, "learning_rate": 2.73e-07, "loss": 7.3776, "step": 546 }, { "epoch": 0.17147335423197493, "grad_norm": 3.5182442665100098, "learning_rate": 2.735e-07, "loss": 7.4899, "step": 547 }, { "epoch": 0.17178683385579938, "grad_norm": 3.437476396560669, "learning_rate": 2.7400000000000004e-07, "loss": 8.6969, "step": 548 }, { "epoch": 0.17210031347962382, "grad_norm": 3.5565860271453857, "learning_rate": 2.7450000000000003e-07, "loss": 7.6131, "step": 549 }, { "epoch": 0.1724137931034483, "grad_norm": 4.052206516265869, "learning_rate": 2.75e-07, "loss": 10.5389, "step": 550 }, { "epoch": 0.17272727272727273, "grad_norm": 4.916564464569092, "learning_rate": 2.7550000000000005e-07, "loss": 13.2927, "step": 551 }, { "epoch": 0.17304075235109717, "grad_norm": 2.8071060180664062, "learning_rate": 2.7600000000000004e-07, "loss": 7.181, "step": 552 }, { "epoch": 0.17335423197492164, "grad_norm": 3.762843370437622, "learning_rate": 2.765e-07, "loss": 9.2195, "step": 553 }, { "epoch": 0.17366771159874608, "grad_norm": 4.453097820281982, "learning_rate": 2.77e-07, "loss": 11.1995, "step": 554 }, { "epoch": 0.17398119122257052, "grad_norm": 3.8368940353393555, "learning_rate": 2.7750000000000004e-07, "loss": 10.7022, "step": 555 }, { "epoch": 0.174294670846395, "grad_norm": 3.128065586090088, "learning_rate": 2.7800000000000003e-07, "loss": 7.396, "step": 556 }, { "epoch": 0.17460815047021944, "grad_norm": 3.999136447906494, "learning_rate": 2.785e-07, "loss": 9.3729, "step": 557 }, { "epoch": 0.17492163009404388, "grad_norm": 3.374950408935547, "learning_rate": 2.7900000000000005e-07, "loss": 6.9749, "step": 558 }, { "epoch": 0.17523510971786835, "grad_norm": 3.5031139850616455, "learning_rate": 2.7950000000000003e-07, "loss": 8.5784, "step": 559 }, { "epoch": 0.1755485893416928, "grad_norm": 3.912729263305664, "learning_rate": 2.8e-07, "loss": 8.9176, "step": 560 }, { "epoch": 0.17586206896551723, "grad_norm": 4.987295627593994, "learning_rate": 2.805e-07, "loss": 10.5289, "step": 561 }, { "epoch": 0.1761755485893417, "grad_norm": 5.127312183380127, "learning_rate": 2.8100000000000004e-07, "loss": 12.005, "step": 562 }, { "epoch": 0.17648902821316614, "grad_norm": 2.887765884399414, "learning_rate": 2.815e-07, "loss": 8.8117, "step": 563 }, { "epoch": 0.17680250783699059, "grad_norm": 3.11470890045166, "learning_rate": 2.82e-07, "loss": 9.1879, "step": 564 }, { "epoch": 0.17711598746081506, "grad_norm": 3.491626739501953, "learning_rate": 2.8250000000000005e-07, "loss": 8.4029, "step": 565 }, { "epoch": 0.1774294670846395, "grad_norm": 3.3320202827453613, "learning_rate": 2.8300000000000003e-07, "loss": 7.258, "step": 566 }, { "epoch": 0.17774294670846394, "grad_norm": 3.7208235263824463, "learning_rate": 2.835e-07, "loss": 9.4842, "step": 567 }, { "epoch": 0.1780564263322884, "grad_norm": 3.597604274749756, "learning_rate": 2.8400000000000005e-07, "loss": 8.5819, "step": 568 }, { "epoch": 0.17836990595611285, "grad_norm": 3.1107091903686523, "learning_rate": 2.8450000000000004e-07, "loss": 7.8507, "step": 569 }, { "epoch": 0.1786833855799373, "grad_norm": 2.587881565093994, "learning_rate": 2.85e-07, "loss": 7.2284, "step": 570 }, { "epoch": 0.17899686520376176, "grad_norm": 2.9738922119140625, "learning_rate": 2.855e-07, "loss": 8.5117, "step": 571 }, { "epoch": 0.1793103448275862, "grad_norm": 3.6646111011505127, "learning_rate": 2.8600000000000005e-07, "loss": 8.4424, "step": 572 }, { "epoch": 0.17962382445141065, "grad_norm": 3.749894380569458, "learning_rate": 2.8650000000000003e-07, "loss": 8.9965, "step": 573 }, { "epoch": 0.17993730407523512, "grad_norm": 2.9281930923461914, "learning_rate": 2.87e-07, "loss": 6.6291, "step": 574 }, { "epoch": 0.18025078369905956, "grad_norm": 5.141971111297607, "learning_rate": 2.8750000000000005e-07, "loss": 11.6456, "step": 575 }, { "epoch": 0.180564263322884, "grad_norm": 3.058166980743408, "learning_rate": 2.8800000000000004e-07, "loss": 7.7736, "step": 576 }, { "epoch": 0.18087774294670847, "grad_norm": 3.9422664642333984, "learning_rate": 2.885e-07, "loss": 11.6602, "step": 577 }, { "epoch": 0.1811912225705329, "grad_norm": 4.253895282745361, "learning_rate": 2.89e-07, "loss": 9.2913, "step": 578 }, { "epoch": 0.18150470219435735, "grad_norm": 3.428746461868286, "learning_rate": 2.8950000000000004e-07, "loss": 7.2803, "step": 579 }, { "epoch": 0.18181818181818182, "grad_norm": 2.983686685562134, "learning_rate": 2.9000000000000003e-07, "loss": 7.1757, "step": 580 }, { "epoch": 0.18213166144200627, "grad_norm": 3.508849859237671, "learning_rate": 2.905e-07, "loss": 8.3138, "step": 581 }, { "epoch": 0.18244514106583073, "grad_norm": 3.7815003395080566, "learning_rate": 2.9100000000000005e-07, "loss": 6.8592, "step": 582 }, { "epoch": 0.18275862068965518, "grad_norm": 3.492091417312622, "learning_rate": 2.9150000000000004e-07, "loss": 9.769, "step": 583 }, { "epoch": 0.18307210031347962, "grad_norm": 3.9170730113983154, "learning_rate": 2.92e-07, "loss": 6.7468, "step": 584 }, { "epoch": 0.1833855799373041, "grad_norm": 3.778157949447632, "learning_rate": 2.9250000000000006e-07, "loss": 7.6754, "step": 585 }, { "epoch": 0.18369905956112853, "grad_norm": 3.811721086502075, "learning_rate": 2.9300000000000004e-07, "loss": 9.9387, "step": 586 }, { "epoch": 0.18401253918495297, "grad_norm": 4.163631916046143, "learning_rate": 2.9350000000000003e-07, "loss": 8.2871, "step": 587 }, { "epoch": 0.18432601880877744, "grad_norm": 3.4269614219665527, "learning_rate": 2.94e-07, "loss": 7.114, "step": 588 }, { "epoch": 0.18463949843260188, "grad_norm": 3.990867853164673, "learning_rate": 2.9450000000000005e-07, "loss": 9.8128, "step": 589 }, { "epoch": 0.18495297805642633, "grad_norm": 3.4142584800720215, "learning_rate": 2.9500000000000003e-07, "loss": 7.7801, "step": 590 }, { "epoch": 0.1852664576802508, "grad_norm": 3.636627674102783, "learning_rate": 2.955e-07, "loss": 10.9509, "step": 591 }, { "epoch": 0.18557993730407524, "grad_norm": 4.59920072555542, "learning_rate": 2.9600000000000006e-07, "loss": 10.6065, "step": 592 }, { "epoch": 0.18589341692789968, "grad_norm": 3.2188518047332764, "learning_rate": 2.9650000000000004e-07, "loss": 7.1557, "step": 593 }, { "epoch": 0.18620689655172415, "grad_norm": 3.7860870361328125, "learning_rate": 2.97e-07, "loss": 7.7989, "step": 594 }, { "epoch": 0.1865203761755486, "grad_norm": 4.25967264175415, "learning_rate": 2.975e-07, "loss": 12.088, "step": 595 }, { "epoch": 0.18683385579937303, "grad_norm": 2.7574448585510254, "learning_rate": 2.9800000000000005e-07, "loss": 5.9382, "step": 596 }, { "epoch": 0.1871473354231975, "grad_norm": 3.17261004447937, "learning_rate": 2.9850000000000003e-07, "loss": 7.0805, "step": 597 }, { "epoch": 0.18746081504702194, "grad_norm": 4.3467793464660645, "learning_rate": 2.99e-07, "loss": 11.9778, "step": 598 }, { "epoch": 0.1877742946708464, "grad_norm": 4.424931526184082, "learning_rate": 2.9950000000000005e-07, "loss": 8.564, "step": 599 }, { "epoch": 0.18808777429467086, "grad_norm": 3.811779737472534, "learning_rate": 3.0000000000000004e-07, "loss": 7.9092, "step": 600 }, { "epoch": 0.1884012539184953, "grad_norm": 4.103071689605713, "learning_rate": 3.005e-07, "loss": 9.164, "step": 601 }, { "epoch": 0.18871473354231974, "grad_norm": 3.3931884765625, "learning_rate": 3.01e-07, "loss": 7.6855, "step": 602 }, { "epoch": 0.1890282131661442, "grad_norm": 3.917874574661255, "learning_rate": 3.0150000000000005e-07, "loss": 9.6698, "step": 603 }, { "epoch": 0.18934169278996865, "grad_norm": 4.343447685241699, "learning_rate": 3.0200000000000003e-07, "loss": 8.4255, "step": 604 }, { "epoch": 0.1896551724137931, "grad_norm": 4.0894927978515625, "learning_rate": 3.025e-07, "loss": 11.4714, "step": 605 }, { "epoch": 0.18996865203761756, "grad_norm": 4.6894073486328125, "learning_rate": 3.0300000000000005e-07, "loss": 10.4595, "step": 606 }, { "epoch": 0.190282131661442, "grad_norm": 2.9865052700042725, "learning_rate": 3.035e-07, "loss": 6.7574, "step": 607 }, { "epoch": 0.19059561128526645, "grad_norm": 3.2500112056732178, "learning_rate": 3.04e-07, "loss": 7.7867, "step": 608 }, { "epoch": 0.19090909090909092, "grad_norm": 3.837984323501587, "learning_rate": 3.0450000000000006e-07, "loss": 8.9823, "step": 609 }, { "epoch": 0.19122257053291536, "grad_norm": 5.894450664520264, "learning_rate": 3.0500000000000004e-07, "loss": 15.5915, "step": 610 }, { "epoch": 0.1915360501567398, "grad_norm": 3.60793137550354, "learning_rate": 3.0550000000000003e-07, "loss": 7.3988, "step": 611 }, { "epoch": 0.19184952978056427, "grad_norm": 3.7032573223114014, "learning_rate": 3.06e-07, "loss": 7.1858, "step": 612 }, { "epoch": 0.1921630094043887, "grad_norm": 3.515413522720337, "learning_rate": 3.0650000000000005e-07, "loss": 8.9439, "step": 613 }, { "epoch": 0.19247648902821315, "grad_norm": 4.349193096160889, "learning_rate": 3.0700000000000004e-07, "loss": 10.2601, "step": 614 }, { "epoch": 0.19278996865203762, "grad_norm": 3.940768241882324, "learning_rate": 3.075e-07, "loss": 10.1987, "step": 615 }, { "epoch": 0.19310344827586207, "grad_norm": 2.720353841781616, "learning_rate": 3.0800000000000006e-07, "loss": 6.7377, "step": 616 }, { "epoch": 0.1934169278996865, "grad_norm": 3.8752973079681396, "learning_rate": 3.085e-07, "loss": 8.5383, "step": 617 }, { "epoch": 0.19373040752351098, "grad_norm": 3.49699330329895, "learning_rate": 3.0900000000000003e-07, "loss": 8.1884, "step": 618 }, { "epoch": 0.19404388714733542, "grad_norm": 3.857430934906006, "learning_rate": 3.095e-07, "loss": 8.9309, "step": 619 }, { "epoch": 0.19435736677115986, "grad_norm": 3.196040153503418, "learning_rate": 3.1000000000000005e-07, "loss": 7.5568, "step": 620 }, { "epoch": 0.19467084639498433, "grad_norm": 3.7832367420196533, "learning_rate": 3.1050000000000003e-07, "loss": 9.6145, "step": 621 }, { "epoch": 0.19498432601880877, "grad_norm": 3.3154568672180176, "learning_rate": 3.11e-07, "loss": 7.3691, "step": 622 }, { "epoch": 0.19529780564263322, "grad_norm": 5.462655067443848, "learning_rate": 3.1150000000000006e-07, "loss": 12.7679, "step": 623 }, { "epoch": 0.19561128526645769, "grad_norm": 4.170969009399414, "learning_rate": 3.12e-07, "loss": 11.6538, "step": 624 }, { "epoch": 0.19592476489028213, "grad_norm": 5.751269817352295, "learning_rate": 3.125e-07, "loss": 13.9954, "step": 625 }, { "epoch": 0.1962382445141066, "grad_norm": 4.067015171051025, "learning_rate": 3.1300000000000006e-07, "loss": 6.5656, "step": 626 }, { "epoch": 0.19655172413793104, "grad_norm": 3.084688186645508, "learning_rate": 3.1350000000000005e-07, "loss": 8.2132, "step": 627 }, { "epoch": 0.19686520376175548, "grad_norm": 2.9886057376861572, "learning_rate": 3.14e-07, "loss": 6.7756, "step": 628 }, { "epoch": 0.19717868338557995, "grad_norm": 3.593235969543457, "learning_rate": 3.145e-07, "loss": 8.3709, "step": 629 }, { "epoch": 0.1974921630094044, "grad_norm": 3.363783597946167, "learning_rate": 3.15e-07, "loss": 10.0533, "step": 630 }, { "epoch": 0.19780564263322883, "grad_norm": 4.221756935119629, "learning_rate": 3.1550000000000004e-07, "loss": 11.3121, "step": 631 }, { "epoch": 0.1981191222570533, "grad_norm": 4.736835956573486, "learning_rate": 3.160000000000001e-07, "loss": 12.493, "step": 632 }, { "epoch": 0.19843260188087775, "grad_norm": 3.1203339099884033, "learning_rate": 3.165e-07, "loss": 7.1578, "step": 633 }, { "epoch": 0.1987460815047022, "grad_norm": 3.5580482482910156, "learning_rate": 3.17e-07, "loss": 7.742, "step": 634 }, { "epoch": 0.19905956112852666, "grad_norm": 3.4418790340423584, "learning_rate": 3.1750000000000003e-07, "loss": 9.445, "step": 635 }, { "epoch": 0.1993730407523511, "grad_norm": 3.2233057022094727, "learning_rate": 3.1800000000000007e-07, "loss": 8.102, "step": 636 }, { "epoch": 0.19968652037617554, "grad_norm": 3.335604190826416, "learning_rate": 3.1850000000000005e-07, "loss": 6.6202, "step": 637 }, { "epoch": 0.2, "grad_norm": 2.8534433841705322, "learning_rate": 3.19e-07, "loss": 7.2855, "step": 638 }, { "epoch": 0.20031347962382445, "grad_norm": 3.358614683151245, "learning_rate": 3.195e-07, "loss": 9.9256, "step": 639 }, { "epoch": 0.2006269592476489, "grad_norm": 3.0653884410858154, "learning_rate": 3.2e-07, "loss": 6.773, "step": 640 }, { "epoch": 0.20094043887147336, "grad_norm": 3.5670981407165527, "learning_rate": 3.2050000000000004e-07, "loss": 10.0116, "step": 641 }, { "epoch": 0.2012539184952978, "grad_norm": 3.6908740997314453, "learning_rate": 3.21e-07, "loss": 9.244, "step": 642 }, { "epoch": 0.20156739811912225, "grad_norm": 4.45837926864624, "learning_rate": 3.215e-07, "loss": 7.975, "step": 643 }, { "epoch": 0.20188087774294672, "grad_norm": 3.217262029647827, "learning_rate": 3.22e-07, "loss": 7.2521, "step": 644 }, { "epoch": 0.20219435736677116, "grad_norm": 3.4965760707855225, "learning_rate": 3.2250000000000004e-07, "loss": 7.2706, "step": 645 }, { "epoch": 0.2025078369905956, "grad_norm": 3.6670267581939697, "learning_rate": 3.2300000000000007e-07, "loss": 7.8659, "step": 646 }, { "epoch": 0.20282131661442007, "grad_norm": 3.4865596294403076, "learning_rate": 3.235e-07, "loss": 8.019, "step": 647 }, { "epoch": 0.2031347962382445, "grad_norm": 3.0378615856170654, "learning_rate": 3.24e-07, "loss": 8.2733, "step": 648 }, { "epoch": 0.20344827586206896, "grad_norm": 4.101233005523682, "learning_rate": 3.2450000000000003e-07, "loss": 8.759, "step": 649 }, { "epoch": 0.20376175548589343, "grad_norm": 3.027646064758301, "learning_rate": 3.25e-07, "loss": 7.4447, "step": 650 }, { "epoch": 0.20407523510971787, "grad_norm": 4.034641742706299, "learning_rate": 3.2550000000000005e-07, "loss": 10.331, "step": 651 }, { "epoch": 0.2043887147335423, "grad_norm": 3.4955785274505615, "learning_rate": 3.26e-07, "loss": 8.214, "step": 652 }, { "epoch": 0.20470219435736678, "grad_norm": 3.91501784324646, "learning_rate": 3.265e-07, "loss": 8.5537, "step": 653 }, { "epoch": 0.20501567398119122, "grad_norm": 3.4765992164611816, "learning_rate": 3.27e-07, "loss": 8.4838, "step": 654 }, { "epoch": 0.20532915360501566, "grad_norm": 2.9636855125427246, "learning_rate": 3.2750000000000004e-07, "loss": 7.3184, "step": 655 }, { "epoch": 0.20564263322884013, "grad_norm": 3.7056450843811035, "learning_rate": 3.280000000000001e-07, "loss": 8.578, "step": 656 }, { "epoch": 0.20595611285266457, "grad_norm": 5.734414100646973, "learning_rate": 3.285e-07, "loss": 19.8354, "step": 657 }, { "epoch": 0.20626959247648902, "grad_norm": 4.2713303565979, "learning_rate": 3.29e-07, "loss": 12.1838, "step": 658 }, { "epoch": 0.2065830721003135, "grad_norm": 3.7173306941986084, "learning_rate": 3.2950000000000003e-07, "loss": 8.0242, "step": 659 }, { "epoch": 0.20689655172413793, "grad_norm": 3.0381829738616943, "learning_rate": 3.3e-07, "loss": 7.6178, "step": 660 }, { "epoch": 0.20721003134796237, "grad_norm": 3.788667678833008, "learning_rate": 3.3050000000000005e-07, "loss": 9.3861, "step": 661 }, { "epoch": 0.20752351097178684, "grad_norm": 4.688927173614502, "learning_rate": 3.31e-07, "loss": 9.3514, "step": 662 }, { "epoch": 0.20783699059561128, "grad_norm": 4.714510440826416, "learning_rate": 3.315e-07, "loss": 11.6471, "step": 663 }, { "epoch": 0.20815047021943572, "grad_norm": 3.319627046585083, "learning_rate": 3.32e-07, "loss": 6.9164, "step": 664 }, { "epoch": 0.2084639498432602, "grad_norm": 4.469924449920654, "learning_rate": 3.3250000000000005e-07, "loss": 10.5332, "step": 665 }, { "epoch": 0.20877742946708464, "grad_norm": 3.59578537940979, "learning_rate": 3.330000000000001e-07, "loss": 7.8451, "step": 666 }, { "epoch": 0.20909090909090908, "grad_norm": 3.3554725646972656, "learning_rate": 3.335e-07, "loss": 8.2171, "step": 667 }, { "epoch": 0.20940438871473355, "grad_norm": 3.555884599685669, "learning_rate": 3.34e-07, "loss": 9.0251, "step": 668 }, { "epoch": 0.209717868338558, "grad_norm": 3.4159555435180664, "learning_rate": 3.3450000000000004e-07, "loss": 8.1913, "step": 669 }, { "epoch": 0.21003134796238246, "grad_norm": 3.6287314891815186, "learning_rate": 3.35e-07, "loss": 7.8455, "step": 670 }, { "epoch": 0.2103448275862069, "grad_norm": 5.647161960601807, "learning_rate": 3.3550000000000006e-07, "loss": 14.64, "step": 671 }, { "epoch": 0.21065830721003134, "grad_norm": 3.4719080924987793, "learning_rate": 3.36e-07, "loss": 8.0084, "step": 672 }, { "epoch": 0.2109717868338558, "grad_norm": 3.0406198501586914, "learning_rate": 3.3650000000000003e-07, "loss": 8.0089, "step": 673 }, { "epoch": 0.21128526645768025, "grad_norm": 3.4766266345977783, "learning_rate": 3.37e-07, "loss": 8.2851, "step": 674 }, { "epoch": 0.2115987460815047, "grad_norm": 4.269450664520264, "learning_rate": 3.3750000000000005e-07, "loss": 8.4858, "step": 675 }, { "epoch": 0.21191222570532917, "grad_norm": 4.416221618652344, "learning_rate": 3.38e-07, "loss": 8.1825, "step": 676 }, { "epoch": 0.2122257053291536, "grad_norm": 3.905118465423584, "learning_rate": 3.385e-07, "loss": 8.2382, "step": 677 }, { "epoch": 0.21253918495297805, "grad_norm": 3.801635980606079, "learning_rate": 3.39e-07, "loss": 8.5147, "step": 678 }, { "epoch": 0.21285266457680252, "grad_norm": 3.1189794540405273, "learning_rate": 3.3950000000000004e-07, "loss": 8.2367, "step": 679 }, { "epoch": 0.21316614420062696, "grad_norm": 3.4691970348358154, "learning_rate": 3.4000000000000003e-07, "loss": 8.0238, "step": 680 }, { "epoch": 0.2134796238244514, "grad_norm": 2.9724643230438232, "learning_rate": 3.405e-07, "loss": 7.9417, "step": 681 }, { "epoch": 0.21379310344827587, "grad_norm": 3.1940643787384033, "learning_rate": 3.41e-07, "loss": 8.5166, "step": 682 }, { "epoch": 0.21410658307210031, "grad_norm": 3.5346059799194336, "learning_rate": 3.4150000000000003e-07, "loss": 7.2445, "step": 683 }, { "epoch": 0.21442006269592476, "grad_norm": 3.558433771133423, "learning_rate": 3.42e-07, "loss": 6.4688, "step": 684 }, { "epoch": 0.21473354231974923, "grad_norm": 3.95771861076355, "learning_rate": 3.4250000000000006e-07, "loss": 9.9655, "step": 685 }, { "epoch": 0.21504702194357367, "grad_norm": 3.4017562866210938, "learning_rate": 3.43e-07, "loss": 8.1479, "step": 686 }, { "epoch": 0.2153605015673981, "grad_norm": 3.312255859375, "learning_rate": 3.435e-07, "loss": 7.3946, "step": 687 }, { "epoch": 0.21567398119122258, "grad_norm": 3.0924856662750244, "learning_rate": 3.44e-07, "loss": 7.1558, "step": 688 }, { "epoch": 0.21598746081504702, "grad_norm": 4.0401482582092285, "learning_rate": 3.4450000000000005e-07, "loss": 8.9632, "step": 689 }, { "epoch": 0.21630094043887146, "grad_norm": 3.2518508434295654, "learning_rate": 3.4500000000000003e-07, "loss": 7.0506, "step": 690 }, { "epoch": 0.21661442006269593, "grad_norm": 3.050063133239746, "learning_rate": 3.455e-07, "loss": 7.7591, "step": 691 }, { "epoch": 0.21692789968652038, "grad_norm": 4.7669997215271, "learning_rate": 3.46e-07, "loss": 9.7782, "step": 692 }, { "epoch": 0.21724137931034482, "grad_norm": 3.25658917427063, "learning_rate": 3.4650000000000004e-07, "loss": 7.9233, "step": 693 }, { "epoch": 0.2175548589341693, "grad_norm": 2.796995162963867, "learning_rate": 3.47e-07, "loss": 6.1858, "step": 694 }, { "epoch": 0.21786833855799373, "grad_norm": 3.696262836456299, "learning_rate": 3.4750000000000006e-07, "loss": 8.2826, "step": 695 }, { "epoch": 0.21818181818181817, "grad_norm": 4.308631896972656, "learning_rate": 3.48e-07, "loss": 12.5166, "step": 696 }, { "epoch": 0.21849529780564264, "grad_norm": 2.7071919441223145, "learning_rate": 3.4850000000000003e-07, "loss": 6.3807, "step": 697 }, { "epoch": 0.21880877742946708, "grad_norm": 4.019406795501709, "learning_rate": 3.49e-07, "loss": 11.1741, "step": 698 }, { "epoch": 0.21912225705329152, "grad_norm": 4.2864580154418945, "learning_rate": 3.4950000000000005e-07, "loss": 9.779, "step": 699 }, { "epoch": 0.219435736677116, "grad_norm": 3.9045979976654053, "learning_rate": 3.5000000000000004e-07, "loss": 6.9983, "step": 700 }, { "epoch": 0.21974921630094044, "grad_norm": 3.1046054363250732, "learning_rate": 3.505e-07, "loss": 7.6675, "step": 701 }, { "epoch": 0.22006269592476488, "grad_norm": 3.561466693878174, "learning_rate": 3.51e-07, "loss": 5.8858, "step": 702 }, { "epoch": 0.22037617554858935, "grad_norm": 3.2833735942840576, "learning_rate": 3.5150000000000004e-07, "loss": 6.8125, "step": 703 }, { "epoch": 0.2206896551724138, "grad_norm": 5.263795375823975, "learning_rate": 3.5200000000000003e-07, "loss": 11.6207, "step": 704 }, { "epoch": 0.22100313479623823, "grad_norm": 3.4245877265930176, "learning_rate": 3.525e-07, "loss": 8.0098, "step": 705 }, { "epoch": 0.2213166144200627, "grad_norm": 3.0912983417510986, "learning_rate": 3.53e-07, "loss": 7.7124, "step": 706 }, { "epoch": 0.22163009404388714, "grad_norm": 2.501897096633911, "learning_rate": 3.5350000000000004e-07, "loss": 6.3126, "step": 707 }, { "epoch": 0.22194357366771159, "grad_norm": 2.732363224029541, "learning_rate": 3.54e-07, "loss": 8.4596, "step": 708 }, { "epoch": 0.22225705329153606, "grad_norm": 2.866844892501831, "learning_rate": 3.5450000000000006e-07, "loss": 7.4127, "step": 709 }, { "epoch": 0.2225705329153605, "grad_norm": 3.562100410461426, "learning_rate": 3.55e-07, "loss": 10.6676, "step": 710 }, { "epoch": 0.22288401253918494, "grad_norm": 2.7351932525634766, "learning_rate": 3.5550000000000003e-07, "loss": 6.7216, "step": 711 }, { "epoch": 0.2231974921630094, "grad_norm": 2.929007053375244, "learning_rate": 3.56e-07, "loss": 7.5504, "step": 712 }, { "epoch": 0.22351097178683385, "grad_norm": 2.9847233295440674, "learning_rate": 3.5650000000000005e-07, "loss": 8.5679, "step": 713 }, { "epoch": 0.22382445141065832, "grad_norm": 3.290311336517334, "learning_rate": 3.5700000000000003e-07, "loss": 8.5614, "step": 714 }, { "epoch": 0.22413793103448276, "grad_norm": 3.176608085632324, "learning_rate": 3.575e-07, "loss": 7.1585, "step": 715 }, { "epoch": 0.2244514106583072, "grad_norm": 5.026858806610107, "learning_rate": 3.58e-07, "loss": 15.6788, "step": 716 }, { "epoch": 0.22476489028213167, "grad_norm": 2.588078260421753, "learning_rate": 3.5850000000000004e-07, "loss": 6.1721, "step": 717 }, { "epoch": 0.22507836990595612, "grad_norm": 3.75956654548645, "learning_rate": 3.59e-07, "loss": 9.8582, "step": 718 }, { "epoch": 0.22539184952978056, "grad_norm": 3.7098045349121094, "learning_rate": 3.5950000000000006e-07, "loss": 7.7692, "step": 719 }, { "epoch": 0.22570532915360503, "grad_norm": 3.2847020626068115, "learning_rate": 3.6e-07, "loss": 7.2862, "step": 720 }, { "epoch": 0.22601880877742947, "grad_norm": 4.180566310882568, "learning_rate": 3.6050000000000003e-07, "loss": 8.2584, "step": 721 }, { "epoch": 0.2263322884012539, "grad_norm": 3.886276960372925, "learning_rate": 3.61e-07, "loss": 8.167, "step": 722 }, { "epoch": 0.22664576802507838, "grad_norm": 3.5251870155334473, "learning_rate": 3.6150000000000005e-07, "loss": 8.091, "step": 723 }, { "epoch": 0.22695924764890282, "grad_norm": 3.8147459030151367, "learning_rate": 3.6200000000000004e-07, "loss": 8.1781, "step": 724 }, { "epoch": 0.22727272727272727, "grad_norm": 4.299187660217285, "learning_rate": 3.625e-07, "loss": 10.4068, "step": 725 }, { "epoch": 0.22758620689655173, "grad_norm": 4.2207255363464355, "learning_rate": 3.63e-07, "loss": 12.8631, "step": 726 }, { "epoch": 0.22789968652037618, "grad_norm": 4.004030704498291, "learning_rate": 3.6350000000000005e-07, "loss": 8.9691, "step": 727 }, { "epoch": 0.22821316614420062, "grad_norm": 4.197298526763916, "learning_rate": 3.6400000000000003e-07, "loss": 8.0809, "step": 728 }, { "epoch": 0.2285266457680251, "grad_norm": 4.978548049926758, "learning_rate": 3.6450000000000007e-07, "loss": 11.308, "step": 729 }, { "epoch": 0.22884012539184953, "grad_norm": 3.4799234867095947, "learning_rate": 3.65e-07, "loss": 8.5875, "step": 730 }, { "epoch": 0.22915360501567397, "grad_norm": 3.4249789714813232, "learning_rate": 3.6550000000000004e-07, "loss": 9.2064, "step": 731 }, { "epoch": 0.22946708463949844, "grad_norm": 2.496309995651245, "learning_rate": 3.66e-07, "loss": 6.6208, "step": 732 }, { "epoch": 0.22978056426332288, "grad_norm": 4.72249698638916, "learning_rate": 3.6650000000000006e-07, "loss": 7.5853, "step": 733 }, { "epoch": 0.23009404388714733, "grad_norm": 3.36556339263916, "learning_rate": 3.6700000000000004e-07, "loss": 6.5957, "step": 734 }, { "epoch": 0.2304075235109718, "grad_norm": 5.186474800109863, "learning_rate": 3.6750000000000003e-07, "loss": 19.8754, "step": 735 }, { "epoch": 0.23072100313479624, "grad_norm": 3.4509902000427246, "learning_rate": 3.68e-07, "loss": 8.3037, "step": 736 }, { "epoch": 0.23103448275862068, "grad_norm": 3.9138832092285156, "learning_rate": 3.6850000000000005e-07, "loss": 10.7141, "step": 737 }, { "epoch": 0.23134796238244515, "grad_norm": 4.275444030761719, "learning_rate": 3.6900000000000004e-07, "loss": 12.6663, "step": 738 }, { "epoch": 0.2316614420062696, "grad_norm": 2.733919858932495, "learning_rate": 3.695e-07, "loss": 7.8312, "step": 739 }, { "epoch": 0.23197492163009403, "grad_norm": 3.960728645324707, "learning_rate": 3.7e-07, "loss": 7.1807, "step": 740 }, { "epoch": 0.2322884012539185, "grad_norm": 3.9499926567077637, "learning_rate": 3.7050000000000004e-07, "loss": 8.1952, "step": 741 }, { "epoch": 0.23260188087774294, "grad_norm": 3.199276924133301, "learning_rate": 3.7100000000000003e-07, "loss": 8.8693, "step": 742 }, { "epoch": 0.2329153605015674, "grad_norm": 3.465644121170044, "learning_rate": 3.7150000000000006e-07, "loss": 8.3512, "step": 743 }, { "epoch": 0.23322884012539186, "grad_norm": 3.957305431365967, "learning_rate": 3.72e-07, "loss": 12.9835, "step": 744 }, { "epoch": 0.2335423197492163, "grad_norm": 4.241058349609375, "learning_rate": 3.7250000000000003e-07, "loss": 11.0323, "step": 745 }, { "epoch": 0.23385579937304074, "grad_norm": 3.187009572982788, "learning_rate": 3.73e-07, "loss": 8.9532, "step": 746 }, { "epoch": 0.2341692789968652, "grad_norm": 3.3454670906066895, "learning_rate": 3.7350000000000006e-07, "loss": 8.9809, "step": 747 }, { "epoch": 0.23448275862068965, "grad_norm": 4.55251407623291, "learning_rate": 3.7400000000000004e-07, "loss": 8.4751, "step": 748 }, { "epoch": 0.2347962382445141, "grad_norm": 3.5339887142181396, "learning_rate": 3.745e-07, "loss": 10.128, "step": 749 }, { "epoch": 0.23510971786833856, "grad_norm": 3.775538206100464, "learning_rate": 3.75e-07, "loss": 8.0116, "step": 750 }, { "epoch": 0.235423197492163, "grad_norm": 3.8628170490264893, "learning_rate": 3.7550000000000005e-07, "loss": 8.7356, "step": 751 }, { "epoch": 0.23573667711598745, "grad_norm": 3.2559432983398438, "learning_rate": 3.7600000000000003e-07, "loss": 7.3864, "step": 752 }, { "epoch": 0.23605015673981192, "grad_norm": 3.832037925720215, "learning_rate": 3.7650000000000007e-07, "loss": 10.6071, "step": 753 }, { "epoch": 0.23636363636363636, "grad_norm": 3.030069589614868, "learning_rate": 3.77e-07, "loss": 8.176, "step": 754 }, { "epoch": 0.23667711598746083, "grad_norm": 3.3893072605133057, "learning_rate": 3.7750000000000004e-07, "loss": 7.696, "step": 755 }, { "epoch": 0.23699059561128527, "grad_norm": 3.5850815773010254, "learning_rate": 3.78e-07, "loss": 8.6539, "step": 756 }, { "epoch": 0.2373040752351097, "grad_norm": 3.6070199012756348, "learning_rate": 3.7850000000000006e-07, "loss": 9.2962, "step": 757 }, { "epoch": 0.23761755485893418, "grad_norm": 3.547919273376465, "learning_rate": 3.7900000000000005e-07, "loss": 10.0205, "step": 758 }, { "epoch": 0.23793103448275862, "grad_norm": 3.6676182746887207, "learning_rate": 3.7950000000000003e-07, "loss": 7.0754, "step": 759 }, { "epoch": 0.23824451410658307, "grad_norm": 4.715800762176514, "learning_rate": 3.8e-07, "loss": 13.3314, "step": 760 }, { "epoch": 0.23855799373040754, "grad_norm": 3.9477193355560303, "learning_rate": 3.8050000000000005e-07, "loss": 8.7171, "step": 761 }, { "epoch": 0.23887147335423198, "grad_norm": 4.012876987457275, "learning_rate": 3.8100000000000004e-07, "loss": 8.9844, "step": 762 }, { "epoch": 0.23918495297805642, "grad_norm": 3.585956335067749, "learning_rate": 3.815000000000001e-07, "loss": 8.8991, "step": 763 }, { "epoch": 0.2394984326018809, "grad_norm": 3.68974232673645, "learning_rate": 3.82e-07, "loss": 8.506, "step": 764 }, { "epoch": 0.23981191222570533, "grad_norm": 3.4120147228240967, "learning_rate": 3.8250000000000004e-07, "loss": 7.4377, "step": 765 }, { "epoch": 0.24012539184952977, "grad_norm": 2.9373185634613037, "learning_rate": 3.8300000000000003e-07, "loss": 6.9149, "step": 766 }, { "epoch": 0.24043887147335424, "grad_norm": 6.263192176818848, "learning_rate": 3.8350000000000007e-07, "loss": 15.8049, "step": 767 }, { "epoch": 0.24075235109717869, "grad_norm": 3.48296856880188, "learning_rate": 3.84e-07, "loss": 7.886, "step": 768 }, { "epoch": 0.24106583072100313, "grad_norm": 3.5808095932006836, "learning_rate": 3.8450000000000004e-07, "loss": 8.8138, "step": 769 }, { "epoch": 0.2413793103448276, "grad_norm": 4.42182731628418, "learning_rate": 3.85e-07, "loss": 11.1607, "step": 770 }, { "epoch": 0.24169278996865204, "grad_norm": 3.0944645404815674, "learning_rate": 3.8550000000000006e-07, "loss": 8.216, "step": 771 }, { "epoch": 0.24200626959247648, "grad_norm": 4.562041282653809, "learning_rate": 3.8600000000000004e-07, "loss": 10.2914, "step": 772 }, { "epoch": 0.24231974921630095, "grad_norm": 3.654736042022705, "learning_rate": 3.865e-07, "loss": 7.9714, "step": 773 }, { "epoch": 0.2426332288401254, "grad_norm": 3.496178150177002, "learning_rate": 3.87e-07, "loss": 8.5865, "step": 774 }, { "epoch": 0.24294670846394983, "grad_norm": 3.801335096359253, "learning_rate": 3.8750000000000005e-07, "loss": 8.9361, "step": 775 }, { "epoch": 0.2432601880877743, "grad_norm": 4.17310094833374, "learning_rate": 3.8800000000000003e-07, "loss": 7.906, "step": 776 }, { "epoch": 0.24357366771159875, "grad_norm": 3.4103686809539795, "learning_rate": 3.8850000000000007e-07, "loss": 8.7571, "step": 777 }, { "epoch": 0.2438871473354232, "grad_norm": 3.6416525840759277, "learning_rate": 3.89e-07, "loss": 7.8949, "step": 778 }, { "epoch": 0.24420062695924766, "grad_norm": 4.429583549499512, "learning_rate": 3.8950000000000004e-07, "loss": 7.3726, "step": 779 }, { "epoch": 0.2445141065830721, "grad_norm": 3.5960888862609863, "learning_rate": 3.9e-07, "loss": 8.3664, "step": 780 }, { "epoch": 0.24482758620689654, "grad_norm": 6.068627834320068, "learning_rate": 3.9050000000000006e-07, "loss": 14.7055, "step": 781 }, { "epoch": 0.245141065830721, "grad_norm": 4.344872951507568, "learning_rate": 3.9100000000000005e-07, "loss": 10.0272, "step": 782 }, { "epoch": 0.24545454545454545, "grad_norm": 3.6392626762390137, "learning_rate": 3.915e-07, "loss": 6.46, "step": 783 }, { "epoch": 0.2457680250783699, "grad_norm": 3.415555238723755, "learning_rate": 3.92e-07, "loss": 8.4829, "step": 784 }, { "epoch": 0.24608150470219436, "grad_norm": 4.941524982452393, "learning_rate": 3.9250000000000005e-07, "loss": 11.3501, "step": 785 }, { "epoch": 0.2463949843260188, "grad_norm": 3.191941738128662, "learning_rate": 3.9300000000000004e-07, "loss": 8.494, "step": 786 }, { "epoch": 0.24670846394984325, "grad_norm": 3.4646642208099365, "learning_rate": 3.935000000000001e-07, "loss": 7.9202, "step": 787 }, { "epoch": 0.24702194357366772, "grad_norm": 3.47352933883667, "learning_rate": 3.94e-07, "loss": 10.1235, "step": 788 }, { "epoch": 0.24733542319749216, "grad_norm": 4.042746543884277, "learning_rate": 3.9450000000000005e-07, "loss": 7.8208, "step": 789 }, { "epoch": 0.2476489028213166, "grad_norm": 3.284029006958008, "learning_rate": 3.9500000000000003e-07, "loss": 7.8496, "step": 790 }, { "epoch": 0.24796238244514107, "grad_norm": 4.137977600097656, "learning_rate": 3.9550000000000007e-07, "loss": 11.9702, "step": 791 }, { "epoch": 0.2482758620689655, "grad_norm": 3.6068592071533203, "learning_rate": 3.9600000000000005e-07, "loss": 6.9573, "step": 792 }, { "epoch": 0.24858934169278996, "grad_norm": 3.264119863510132, "learning_rate": 3.965e-07, "loss": 7.5779, "step": 793 }, { "epoch": 0.24890282131661443, "grad_norm": 3.505993366241455, "learning_rate": 3.97e-07, "loss": 7.7447, "step": 794 }, { "epoch": 0.24921630094043887, "grad_norm": 3.9648635387420654, "learning_rate": 3.9750000000000006e-07, "loss": 11.2637, "step": 795 }, { "epoch": 0.2495297805642633, "grad_norm": 3.1940486431121826, "learning_rate": 3.9800000000000004e-07, "loss": 8.9283, "step": 796 }, { "epoch": 0.24984326018808778, "grad_norm": 3.5564422607421875, "learning_rate": 3.985e-07, "loss": 7.2672, "step": 797 }, { "epoch": 0.2501567398119122, "grad_norm": 3.626217842102051, "learning_rate": 3.99e-07, "loss": 8.51, "step": 798 }, { "epoch": 0.2501567398119122, "eval_loss": 28.84930992126465, "eval_runtime": 20.5251, "eval_samples_per_second": 130.913, "eval_steps_per_second": 8.185, "step": 798 }, { "epoch": 0.25047021943573666, "grad_norm": 3.097006320953369, "learning_rate": 3.9950000000000005e-07, "loss": 7.4537, "step": 799 }, { "epoch": 0.2507836990595611, "grad_norm": 3.5833940505981445, "learning_rate": 4.0000000000000003e-07, "loss": 10.0463, "step": 800 }, { "epoch": 0.2510971786833856, "grad_norm": 4.196984767913818, "learning_rate": 4.0050000000000007e-07, "loss": 7.771, "step": 801 }, { "epoch": 0.25141065830721004, "grad_norm": 4.240922927856445, "learning_rate": 4.01e-07, "loss": 10.3631, "step": 802 }, { "epoch": 0.2517241379310345, "grad_norm": 3.425544023513794, "learning_rate": 4.015e-07, "loss": 6.2936, "step": 803 }, { "epoch": 0.25203761755485893, "grad_norm": 3.8338463306427, "learning_rate": 4.02e-07, "loss": 11.4034, "step": 804 }, { "epoch": 0.25235109717868337, "grad_norm": 3.4138238430023193, "learning_rate": 4.0250000000000006e-07, "loss": 6.9629, "step": 805 }, { "epoch": 0.2526645768025078, "grad_norm": 3.1681156158447266, "learning_rate": 4.0300000000000005e-07, "loss": 7.616, "step": 806 }, { "epoch": 0.2529780564263323, "grad_norm": 13.125308990478516, "learning_rate": 4.035e-07, "loss": 7.871, "step": 807 }, { "epoch": 0.25329153605015675, "grad_norm": 5.784317493438721, "learning_rate": 4.04e-07, "loss": 20.8958, "step": 808 }, { "epoch": 0.2536050156739812, "grad_norm": 3.021087646484375, "learning_rate": 4.0450000000000006e-07, "loss": 7.3132, "step": 809 }, { "epoch": 0.25391849529780564, "grad_norm": 2.8129184246063232, "learning_rate": 4.0500000000000004e-07, "loss": 7.4504, "step": 810 }, { "epoch": 0.2542319749216301, "grad_norm": 3.036388874053955, "learning_rate": 4.055000000000001e-07, "loss": 6.5679, "step": 811 }, { "epoch": 0.2545454545454545, "grad_norm": 3.112868547439575, "learning_rate": 4.06e-07, "loss": 8.2039, "step": 812 }, { "epoch": 0.254858934169279, "grad_norm": 3.948190212249756, "learning_rate": 4.065e-07, "loss": 9.2969, "step": 813 }, { "epoch": 0.25517241379310346, "grad_norm": 3.823413133621216, "learning_rate": 4.0700000000000003e-07, "loss": 8.2135, "step": 814 }, { "epoch": 0.2554858934169279, "grad_norm": 3.8010671138763428, "learning_rate": 4.0750000000000007e-07, "loss": 8.7767, "step": 815 }, { "epoch": 0.25579937304075234, "grad_norm": 2.9464848041534424, "learning_rate": 4.0800000000000005e-07, "loss": 6.3003, "step": 816 }, { "epoch": 0.2561128526645768, "grad_norm": 3.2114903926849365, "learning_rate": 4.085e-07, "loss": 8.0724, "step": 817 }, { "epoch": 0.2564263322884012, "grad_norm": 3.254835844039917, "learning_rate": 4.09e-07, "loss": 7.6632, "step": 818 }, { "epoch": 0.2567398119122257, "grad_norm": 4.548694133758545, "learning_rate": 4.0950000000000006e-07, "loss": 12.1637, "step": 819 }, { "epoch": 0.25705329153605017, "grad_norm": 4.109635829925537, "learning_rate": 4.1000000000000004e-07, "loss": 9.4519, "step": 820 }, { "epoch": 0.2573667711598746, "grad_norm": 4.340848922729492, "learning_rate": 4.105000000000001e-07, "loss": 8.5242, "step": 821 }, { "epoch": 0.25768025078369905, "grad_norm": 3.260594367980957, "learning_rate": 4.11e-07, "loss": 7.83, "step": 822 }, { "epoch": 0.2579937304075235, "grad_norm": 5.5108771324157715, "learning_rate": 4.115e-07, "loss": 10.5884, "step": 823 }, { "epoch": 0.258307210031348, "grad_norm": 3.490274429321289, "learning_rate": 4.1200000000000004e-07, "loss": 8.7728, "step": 824 }, { "epoch": 0.25862068965517243, "grad_norm": 6.578031539916992, "learning_rate": 4.125000000000001e-07, "loss": 10.9723, "step": 825 }, { "epoch": 0.2589341692789969, "grad_norm": 2.774808406829834, "learning_rate": 4.1300000000000006e-07, "loss": 6.9211, "step": 826 }, { "epoch": 0.2592476489028213, "grad_norm": 5.915481090545654, "learning_rate": 4.135e-07, "loss": 9.1981, "step": 827 }, { "epoch": 0.25956112852664576, "grad_norm": 2.8038599491119385, "learning_rate": 4.1400000000000003e-07, "loss": 6.1127, "step": 828 }, { "epoch": 0.2598746081504702, "grad_norm": 3.501262903213501, "learning_rate": 4.1450000000000007e-07, "loss": 6.5746, "step": 829 }, { "epoch": 0.2601880877742947, "grad_norm": 3.5788586139678955, "learning_rate": 4.1500000000000005e-07, "loss": 8.6381, "step": 830 }, { "epoch": 0.26050156739811914, "grad_norm": 5.020801544189453, "learning_rate": 4.155e-07, "loss": 16.7279, "step": 831 }, { "epoch": 0.2608150470219436, "grad_norm": 3.9327170848846436, "learning_rate": 4.16e-07, "loss": 12.7287, "step": 832 }, { "epoch": 0.261128526645768, "grad_norm": 3.2840352058410645, "learning_rate": 4.165e-07, "loss": 6.839, "step": 833 }, { "epoch": 0.26144200626959246, "grad_norm": 3.750833511352539, "learning_rate": 4.1700000000000004e-07, "loss": 11.1242, "step": 834 }, { "epoch": 0.2617554858934169, "grad_norm": 3.46420955657959, "learning_rate": 4.175000000000001e-07, "loss": 9.7763, "step": 835 }, { "epoch": 0.2620689655172414, "grad_norm": 3.476485252380371, "learning_rate": 4.18e-07, "loss": 7.1691, "step": 836 }, { "epoch": 0.26238244514106585, "grad_norm": 4.025050163269043, "learning_rate": 4.185e-07, "loss": 9.9421, "step": 837 }, { "epoch": 0.2626959247648903, "grad_norm": 3.0116124153137207, "learning_rate": 4.1900000000000003e-07, "loss": 6.3215, "step": 838 }, { "epoch": 0.26300940438871473, "grad_norm": 4.668792247772217, "learning_rate": 4.1950000000000007e-07, "loss": 11.0969, "step": 839 }, { "epoch": 0.26332288401253917, "grad_norm": 2.963440418243408, "learning_rate": 4.2000000000000006e-07, "loss": 7.2545, "step": 840 }, { "epoch": 0.2636363636363636, "grad_norm": 4.21551513671875, "learning_rate": 4.205e-07, "loss": 8.2629, "step": 841 }, { "epoch": 0.2639498432601881, "grad_norm": 3.958061456680298, "learning_rate": 4.21e-07, "loss": 9.7021, "step": 842 }, { "epoch": 0.26426332288401255, "grad_norm": 2.9558305740356445, "learning_rate": 4.215e-07, "loss": 7.3453, "step": 843 }, { "epoch": 0.264576802507837, "grad_norm": 4.2135233879089355, "learning_rate": 4.2200000000000005e-07, "loss": 8.955, "step": 844 }, { "epoch": 0.26489028213166144, "grad_norm": 3.562573194503784, "learning_rate": 4.225000000000001e-07, "loss": 7.2757, "step": 845 }, { "epoch": 0.2652037617554859, "grad_norm": 3.2057554721832275, "learning_rate": 4.23e-07, "loss": 8.2123, "step": 846 }, { "epoch": 0.2655172413793103, "grad_norm": 3.901542901992798, "learning_rate": 4.235e-07, "loss": 9.1358, "step": 847 }, { "epoch": 0.2658307210031348, "grad_norm": 3.1347365379333496, "learning_rate": 4.2400000000000004e-07, "loss": 6.5317, "step": 848 }, { "epoch": 0.26614420062695926, "grad_norm": 5.042209625244141, "learning_rate": 4.245000000000001e-07, "loss": 14.4307, "step": 849 }, { "epoch": 0.2664576802507837, "grad_norm": 3.5455801486968994, "learning_rate": 4.2500000000000006e-07, "loss": 8.8807, "step": 850 }, { "epoch": 0.26677115987460814, "grad_norm": 3.4751904010772705, "learning_rate": 4.255e-07, "loss": 8.8506, "step": 851 }, { "epoch": 0.2670846394984326, "grad_norm": 3.6751580238342285, "learning_rate": 4.2600000000000003e-07, "loss": 10.4928, "step": 852 }, { "epoch": 0.267398119122257, "grad_norm": 4.02175760269165, "learning_rate": 4.265e-07, "loss": 6.9068, "step": 853 }, { "epoch": 0.2677115987460815, "grad_norm": 7.32761287689209, "learning_rate": 4.2700000000000005e-07, "loss": 17.0699, "step": 854 }, { "epoch": 0.26802507836990597, "grad_norm": 3.607837200164795, "learning_rate": 4.275000000000001e-07, "loss": 8.021, "step": 855 }, { "epoch": 0.2683385579937304, "grad_norm": 2.8955557346343994, "learning_rate": 4.28e-07, "loss": 8.8195, "step": 856 }, { "epoch": 0.26865203761755485, "grad_norm": 3.2017621994018555, "learning_rate": 4.285e-07, "loss": 7.1978, "step": 857 }, { "epoch": 0.2689655172413793, "grad_norm": 3.7214231491088867, "learning_rate": 4.2900000000000004e-07, "loss": 7.2386, "step": 858 }, { "epoch": 0.26927899686520373, "grad_norm": 3.3874855041503906, "learning_rate": 4.295000000000001e-07, "loss": 7.6419, "step": 859 }, { "epoch": 0.26959247648902823, "grad_norm": 4.008450508117676, "learning_rate": 4.3e-07, "loss": 8.0012, "step": 860 }, { "epoch": 0.2699059561128527, "grad_norm": 3.3975911140441895, "learning_rate": 4.305e-07, "loss": 7.3739, "step": 861 }, { "epoch": 0.2702194357366771, "grad_norm": 3.656843423843384, "learning_rate": 4.3100000000000003e-07, "loss": 8.8982, "step": 862 }, { "epoch": 0.27053291536050156, "grad_norm": 3.2853686809539795, "learning_rate": 4.315e-07, "loss": 7.6732, "step": 863 }, { "epoch": 0.270846394984326, "grad_norm": 4.232872009277344, "learning_rate": 4.3200000000000006e-07, "loss": 10.4912, "step": 864 }, { "epoch": 0.2711598746081505, "grad_norm": 3.6398162841796875, "learning_rate": 4.325e-07, "loss": 9.274, "step": 865 }, { "epoch": 0.27147335423197494, "grad_norm": 2.7679269313812256, "learning_rate": 4.33e-07, "loss": 7.0839, "step": 866 }, { "epoch": 0.2717868338557994, "grad_norm": 3.5400402545928955, "learning_rate": 4.335e-07, "loss": 8.8406, "step": 867 }, { "epoch": 0.2721003134796238, "grad_norm": 3.0687062740325928, "learning_rate": 4.3400000000000005e-07, "loss": 7.36, "step": 868 }, { "epoch": 0.27241379310344827, "grad_norm": 3.573641061782837, "learning_rate": 4.345000000000001e-07, "loss": 8.0488, "step": 869 }, { "epoch": 0.2727272727272727, "grad_norm": 4.390239238739014, "learning_rate": 4.35e-07, "loss": 12.9096, "step": 870 }, { "epoch": 0.2730407523510972, "grad_norm": 3.314115524291992, "learning_rate": 4.355e-07, "loss": 7.1066, "step": 871 }, { "epoch": 0.27335423197492165, "grad_norm": 5.46345853805542, "learning_rate": 4.3600000000000004e-07, "loss": 11.487, "step": 872 }, { "epoch": 0.2736677115987461, "grad_norm": 5.023271083831787, "learning_rate": 4.365e-07, "loss": 10.0573, "step": 873 }, { "epoch": 0.27398119122257053, "grad_norm": 2.962111234664917, "learning_rate": 4.3700000000000006e-07, "loss": 8.829, "step": 874 }, { "epoch": 0.274294670846395, "grad_norm": 3.1559054851531982, "learning_rate": 4.375e-07, "loss": 7.1875, "step": 875 }, { "epoch": 0.2746081504702194, "grad_norm": 3.6950178146362305, "learning_rate": 4.3800000000000003e-07, "loss": 8.4211, "step": 876 }, { "epoch": 0.2749216300940439, "grad_norm": 4.661724090576172, "learning_rate": 4.385e-07, "loss": 11.424, "step": 877 }, { "epoch": 0.27523510971786835, "grad_norm": 3.7680742740631104, "learning_rate": 4.3900000000000005e-07, "loss": 9.1259, "step": 878 }, { "epoch": 0.2755485893416928, "grad_norm": 3.409186601638794, "learning_rate": 4.395000000000001e-07, "loss": 6.5314, "step": 879 }, { "epoch": 0.27586206896551724, "grad_norm": 3.2190394401550293, "learning_rate": 4.4e-07, "loss": 6.4504, "step": 880 }, { "epoch": 0.2761755485893417, "grad_norm": 4.012719631195068, "learning_rate": 4.405e-07, "loss": 9.4085, "step": 881 }, { "epoch": 0.2764890282131661, "grad_norm": 4.595925807952881, "learning_rate": 4.4100000000000004e-07, "loss": 13.1328, "step": 882 }, { "epoch": 0.2768025078369906, "grad_norm": 4.065830230712891, "learning_rate": 4.4150000000000003e-07, "loss": 8.3007, "step": 883 }, { "epoch": 0.27711598746081506, "grad_norm": 4.229946613311768, "learning_rate": 4.4200000000000007e-07, "loss": 11.1843, "step": 884 }, { "epoch": 0.2774294670846395, "grad_norm": 3.8973495960235596, "learning_rate": 4.425e-07, "loss": 7.978, "step": 885 }, { "epoch": 0.27774294670846394, "grad_norm": 2.952505111694336, "learning_rate": 4.4300000000000004e-07, "loss": 8.5853, "step": 886 }, { "epoch": 0.2780564263322884, "grad_norm": 2.839816093444824, "learning_rate": 4.435e-07, "loss": 6.6168, "step": 887 }, { "epoch": 0.27836990595611283, "grad_norm": 4.309621334075928, "learning_rate": 4.4400000000000006e-07, "loss": 10.9452, "step": 888 }, { "epoch": 0.2786833855799373, "grad_norm": 5.065432071685791, "learning_rate": 4.445000000000001e-07, "loss": 9.2203, "step": 889 }, { "epoch": 0.27899686520376177, "grad_norm": 4.246471881866455, "learning_rate": 4.4500000000000003e-07, "loss": 7.9399, "step": 890 }, { "epoch": 0.2793103448275862, "grad_norm": 5.130437850952148, "learning_rate": 4.455e-07, "loss": 11.7158, "step": 891 }, { "epoch": 0.27962382445141065, "grad_norm": 3.858509063720703, "learning_rate": 4.4600000000000005e-07, "loss": 9.7248, "step": 892 }, { "epoch": 0.2799373040752351, "grad_norm": 3.6608774662017822, "learning_rate": 4.4650000000000003e-07, "loss": 12.7871, "step": 893 }, { "epoch": 0.28025078369905954, "grad_norm": 4.863644123077393, "learning_rate": 4.47e-07, "loss": 9.9666, "step": 894 }, { "epoch": 0.28056426332288403, "grad_norm": 3.754490613937378, "learning_rate": 4.475e-07, "loss": 7.9063, "step": 895 }, { "epoch": 0.2808777429467085, "grad_norm": 3.4950146675109863, "learning_rate": 4.4800000000000004e-07, "loss": 8.867, "step": 896 }, { "epoch": 0.2811912225705329, "grad_norm": 3.564337968826294, "learning_rate": 4.485e-07, "loss": 7.4725, "step": 897 }, { "epoch": 0.28150470219435736, "grad_norm": 2.9661481380462646, "learning_rate": 4.4900000000000006e-07, "loss": 6.9609, "step": 898 }, { "epoch": 0.2818181818181818, "grad_norm": 3.3152127265930176, "learning_rate": 4.495e-07, "loss": 7.2334, "step": 899 }, { "epoch": 0.28213166144200624, "grad_norm": 3.4572389125823975, "learning_rate": 4.5000000000000003e-07, "loss": 7.8584, "step": 900 }, { "epoch": 0.28244514106583074, "grad_norm": 3.333373546600342, "learning_rate": 4.505e-07, "loss": 6.6077, "step": 901 }, { "epoch": 0.2827586206896552, "grad_norm": 4.405636310577393, "learning_rate": 4.5100000000000005e-07, "loss": 11.1708, "step": 902 }, { "epoch": 0.2830721003134796, "grad_norm": 3.0989999771118164, "learning_rate": 4.5150000000000004e-07, "loss": 7.3427, "step": 903 }, { "epoch": 0.28338557993730407, "grad_norm": 3.1780450344085693, "learning_rate": 4.52e-07, "loss": 7.8806, "step": 904 }, { "epoch": 0.2836990595611285, "grad_norm": 3.923069953918457, "learning_rate": 4.525e-07, "loss": 9.7537, "step": 905 }, { "epoch": 0.284012539184953, "grad_norm": 4.220728874206543, "learning_rate": 4.5300000000000005e-07, "loss": 9.528, "step": 906 }, { "epoch": 0.28432601880877745, "grad_norm": 3.3002758026123047, "learning_rate": 4.5350000000000003e-07, "loss": 6.0225, "step": 907 }, { "epoch": 0.2846394984326019, "grad_norm": 4.327773571014404, "learning_rate": 4.5400000000000007e-07, "loss": 8.0762, "step": 908 }, { "epoch": 0.28495297805642633, "grad_norm": 3.128321409225464, "learning_rate": 4.545e-07, "loss": 8.3285, "step": 909 }, { "epoch": 0.2852664576802508, "grad_norm": 3.2514123916625977, "learning_rate": 4.5500000000000004e-07, "loss": 8.2311, "step": 910 }, { "epoch": 0.2855799373040752, "grad_norm": 5.544764518737793, "learning_rate": 4.555e-07, "loss": 11.3571, "step": 911 }, { "epoch": 0.2858934169278997, "grad_norm": 5.904582977294922, "learning_rate": 4.5600000000000006e-07, "loss": 18.3895, "step": 912 }, { "epoch": 0.28620689655172415, "grad_norm": 3.38208270072937, "learning_rate": 4.5650000000000004e-07, "loss": 9.4236, "step": 913 }, { "epoch": 0.2865203761755486, "grad_norm": 3.433018445968628, "learning_rate": 4.5700000000000003e-07, "loss": 11.2537, "step": 914 }, { "epoch": 0.28683385579937304, "grad_norm": 4.306755542755127, "learning_rate": 4.575e-07, "loss": 11.701, "step": 915 }, { "epoch": 0.2871473354231975, "grad_norm": 3.296084403991699, "learning_rate": 4.5800000000000005e-07, "loss": 6.5184, "step": 916 }, { "epoch": 0.2874608150470219, "grad_norm": 4.437682628631592, "learning_rate": 4.5850000000000004e-07, "loss": 8.7946, "step": 917 }, { "epoch": 0.2877742946708464, "grad_norm": 4.028353214263916, "learning_rate": 4.5900000000000007e-07, "loss": 9.0222, "step": 918 }, { "epoch": 0.28808777429467086, "grad_norm": 2.981740951538086, "learning_rate": 4.595e-07, "loss": 6.3495, "step": 919 }, { "epoch": 0.2884012539184953, "grad_norm": 3.5462889671325684, "learning_rate": 4.6000000000000004e-07, "loss": 9.6387, "step": 920 }, { "epoch": 0.28871473354231975, "grad_norm": 3.0416839122772217, "learning_rate": 4.6050000000000003e-07, "loss": 5.743, "step": 921 }, { "epoch": 0.2890282131661442, "grad_norm": 2.538724660873413, "learning_rate": 4.6100000000000006e-07, "loss": 6.5413, "step": 922 }, { "epoch": 0.28934169278996863, "grad_norm": 3.508538007736206, "learning_rate": 4.615e-07, "loss": 9.6875, "step": 923 }, { "epoch": 0.2896551724137931, "grad_norm": 3.4180221557617188, "learning_rate": 4.6200000000000003e-07, "loss": 7.9855, "step": 924 }, { "epoch": 0.28996865203761757, "grad_norm": 3.5554192066192627, "learning_rate": 4.625e-07, "loss": 7.1513, "step": 925 }, { "epoch": 0.290282131661442, "grad_norm": 3.860269784927368, "learning_rate": 4.6300000000000006e-07, "loss": 9.0043, "step": 926 }, { "epoch": 0.29059561128526645, "grad_norm": 3.593224287033081, "learning_rate": 4.6350000000000004e-07, "loss": 8.9827, "step": 927 }, { "epoch": 0.2909090909090909, "grad_norm": 4.743045330047607, "learning_rate": 4.64e-07, "loss": 12.2761, "step": 928 }, { "epoch": 0.29122257053291534, "grad_norm": 3.506476640701294, "learning_rate": 4.645e-07, "loss": 9.6425, "step": 929 }, { "epoch": 0.29153605015673983, "grad_norm": 3.420703887939453, "learning_rate": 4.6500000000000005e-07, "loss": 7.238, "step": 930 }, { "epoch": 0.2918495297805643, "grad_norm": 8.075421333312988, "learning_rate": 4.6550000000000003e-07, "loss": 19.2958, "step": 931 }, { "epoch": 0.2921630094043887, "grad_norm": 4.11147928237915, "learning_rate": 4.6600000000000007e-07, "loss": 7.9719, "step": 932 }, { "epoch": 0.29247648902821316, "grad_norm": 4.024860382080078, "learning_rate": 4.665e-07, "loss": 8.5926, "step": 933 }, { "epoch": 0.2927899686520376, "grad_norm": 3.3219480514526367, "learning_rate": 4.6700000000000004e-07, "loss": 8.4054, "step": 934 }, { "epoch": 0.29310344827586204, "grad_norm": 3.3816304206848145, "learning_rate": 4.675e-07, "loss": 6.9335, "step": 935 }, { "epoch": 0.29341692789968654, "grad_norm": 3.370476484298706, "learning_rate": 4.6800000000000006e-07, "loss": 8.5852, "step": 936 }, { "epoch": 0.293730407523511, "grad_norm": 3.3092455863952637, "learning_rate": 4.6850000000000005e-07, "loss": 8.164, "step": 937 }, { "epoch": 0.2940438871473354, "grad_norm": 3.3833985328674316, "learning_rate": 4.6900000000000003e-07, "loss": 7.7537, "step": 938 }, { "epoch": 0.29435736677115987, "grad_norm": 4.879644870758057, "learning_rate": 4.695e-07, "loss": 9.2243, "step": 939 }, { "epoch": 0.2946708463949843, "grad_norm": 3.931847333908081, "learning_rate": 4.7000000000000005e-07, "loss": 9.3823, "step": 940 }, { "epoch": 0.29498432601880875, "grad_norm": 3.1454243659973145, "learning_rate": 4.7050000000000004e-07, "loss": 6.3691, "step": 941 }, { "epoch": 0.29529780564263325, "grad_norm": 3.438483953475952, "learning_rate": 4.710000000000001e-07, "loss": 7.0824, "step": 942 }, { "epoch": 0.2956112852664577, "grad_norm": 3.97963809967041, "learning_rate": 4.715e-07, "loss": 9.5606, "step": 943 }, { "epoch": 0.29592476489028213, "grad_norm": 5.063857078552246, "learning_rate": 4.7200000000000004e-07, "loss": 12.8541, "step": 944 }, { "epoch": 0.2962382445141066, "grad_norm": 4.647433280944824, "learning_rate": 4.7250000000000003e-07, "loss": 11.1892, "step": 945 }, { "epoch": 0.296551724137931, "grad_norm": 4.020174026489258, "learning_rate": 4.7300000000000007e-07, "loss": 9.2035, "step": 946 }, { "epoch": 0.29686520376175546, "grad_norm": 3.719456434249878, "learning_rate": 4.7350000000000005e-07, "loss": 7.5178, "step": 947 }, { "epoch": 0.29717868338557996, "grad_norm": 3.1977577209472656, "learning_rate": 4.7400000000000004e-07, "loss": 6.5933, "step": 948 }, { "epoch": 0.2974921630094044, "grad_norm": 2.9651682376861572, "learning_rate": 4.745e-07, "loss": 6.2385, "step": 949 }, { "epoch": 0.29780564263322884, "grad_norm": 4.075620174407959, "learning_rate": 4.7500000000000006e-07, "loss": 13.3818, "step": 950 }, { "epoch": 0.2981191222570533, "grad_norm": 4.676611423492432, "learning_rate": 4.7550000000000004e-07, "loss": 13.5246, "step": 951 }, { "epoch": 0.2984326018808777, "grad_norm": 4.3568291664123535, "learning_rate": 4.760000000000001e-07, "loss": 10.122, "step": 952 }, { "epoch": 0.2987460815047022, "grad_norm": 3.4638936519622803, "learning_rate": 4.765e-07, "loss": 9.9878, "step": 953 }, { "epoch": 0.29905956112852666, "grad_norm": 3.4983904361724854, "learning_rate": 4.77e-07, "loss": 7.968, "step": 954 }, { "epoch": 0.2993730407523511, "grad_norm": 2.9241857528686523, "learning_rate": 4.775000000000001e-07, "loss": 6.1837, "step": 955 }, { "epoch": 0.29968652037617555, "grad_norm": 5.027895927429199, "learning_rate": 4.78e-07, "loss": 10.7868, "step": 956 }, { "epoch": 0.3, "grad_norm": 3.0963079929351807, "learning_rate": 4.785000000000001e-07, "loss": 9.2105, "step": 957 }, { "epoch": 0.30031347962382443, "grad_norm": 3.367241144180298, "learning_rate": 4.79e-07, "loss": 8.6549, "step": 958 }, { "epoch": 0.30062695924764893, "grad_norm": 3.1908528804779053, "learning_rate": 4.795e-07, "loss": 9.3335, "step": 959 }, { "epoch": 0.30094043887147337, "grad_norm": 3.797089099884033, "learning_rate": 4.800000000000001e-07, "loss": 10.5798, "step": 960 }, { "epoch": 0.3012539184952978, "grad_norm": 3.816810131072998, "learning_rate": 4.805000000000001e-07, "loss": 8.424, "step": 961 }, { "epoch": 0.30156739811912225, "grad_norm": 2.98754620552063, "learning_rate": 4.81e-07, "loss": 6.592, "step": 962 }, { "epoch": 0.3018808777429467, "grad_norm": 7.078648090362549, "learning_rate": 4.815000000000001e-07, "loss": 20.1471, "step": 963 }, { "epoch": 0.30219435736677114, "grad_norm": 3.7833642959594727, "learning_rate": 4.82e-07, "loss": 8.8326, "step": 964 }, { "epoch": 0.30250783699059564, "grad_norm": 2.790283203125, "learning_rate": 4.825e-07, "loss": 6.6206, "step": 965 }, { "epoch": 0.3028213166144201, "grad_norm": 4.206464767456055, "learning_rate": 4.830000000000001e-07, "loss": 10.1851, "step": 966 }, { "epoch": 0.3031347962382445, "grad_norm": 4.659297943115234, "learning_rate": 4.835e-07, "loss": 11.0335, "step": 967 }, { "epoch": 0.30344827586206896, "grad_norm": 4.047898769378662, "learning_rate": 4.84e-07, "loss": 10.0668, "step": 968 }, { "epoch": 0.3037617554858934, "grad_norm": 4.826545715332031, "learning_rate": 4.845000000000001e-07, "loss": 9.4059, "step": 969 }, { "epoch": 0.30407523510971785, "grad_norm": 3.3105831146240234, "learning_rate": 4.85e-07, "loss": 9.7796, "step": 970 }, { "epoch": 0.30438871473354234, "grad_norm": 3.3956053256988525, "learning_rate": 4.855e-07, "loss": 8.4835, "step": 971 }, { "epoch": 0.3047021943573668, "grad_norm": 3.9917337894439697, "learning_rate": 4.86e-07, "loss": 12.5467, "step": 972 }, { "epoch": 0.3050156739811912, "grad_norm": 2.869931221008301, "learning_rate": 4.865e-07, "loss": 7.641, "step": 973 }, { "epoch": 0.30532915360501567, "grad_norm": 4.201666355133057, "learning_rate": 4.870000000000001e-07, "loss": 8.2918, "step": 974 }, { "epoch": 0.3056426332288401, "grad_norm": 3.3518385887145996, "learning_rate": 4.875000000000001e-07, "loss": 10.0857, "step": 975 }, { "epoch": 0.30595611285266455, "grad_norm": 3.7927677631378174, "learning_rate": 4.88e-07, "loss": 8.8694, "step": 976 }, { "epoch": 0.30626959247648905, "grad_norm": 2.8095505237579346, "learning_rate": 4.885000000000001e-07, "loss": 6.6404, "step": 977 }, { "epoch": 0.3065830721003135, "grad_norm": 3.649327516555786, "learning_rate": 4.89e-07, "loss": 8.3927, "step": 978 }, { "epoch": 0.30689655172413793, "grad_norm": 3.178901195526123, "learning_rate": 4.895e-07, "loss": 7.3435, "step": 979 }, { "epoch": 0.3072100313479624, "grad_norm": 3.161583185195923, "learning_rate": 4.900000000000001e-07, "loss": 8.2137, "step": 980 }, { "epoch": 0.3075235109717868, "grad_norm": 3.098173141479492, "learning_rate": 4.905000000000001e-07, "loss": 7.279, "step": 981 }, { "epoch": 0.30783699059561126, "grad_norm": 2.761258363723755, "learning_rate": 4.91e-07, "loss": 6.147, "step": 982 }, { "epoch": 0.30815047021943576, "grad_norm": 2.6040921211242676, "learning_rate": 4.915000000000001e-07, "loss": 6.242, "step": 983 }, { "epoch": 0.3084639498432602, "grad_norm": 3.172119617462158, "learning_rate": 4.92e-07, "loss": 6.6438, "step": 984 }, { "epoch": 0.30877742946708464, "grad_norm": 3.180616617202759, "learning_rate": 4.925e-07, "loss": 7.9243, "step": 985 }, { "epoch": 0.3090909090909091, "grad_norm": 5.080252170562744, "learning_rate": 4.93e-07, "loss": 13.2643, "step": 986 }, { "epoch": 0.3094043887147335, "grad_norm": 3.6815905570983887, "learning_rate": 4.935e-07, "loss": 10.5387, "step": 987 }, { "epoch": 0.30971786833855797, "grad_norm": 5.263372898101807, "learning_rate": 4.940000000000001e-07, "loss": 16.638, "step": 988 }, { "epoch": 0.31003134796238246, "grad_norm": 2.8743481636047363, "learning_rate": 4.945000000000001e-07, "loss": 9.2944, "step": 989 }, { "epoch": 0.3103448275862069, "grad_norm": 5.4028096199035645, "learning_rate": 4.95e-07, "loss": 13.0038, "step": 990 }, { "epoch": 0.31065830721003135, "grad_norm": 3.379136323928833, "learning_rate": 4.955e-07, "loss": 7.5712, "step": 991 }, { "epoch": 0.3109717868338558, "grad_norm": 3.4020957946777344, "learning_rate": 4.96e-07, "loss": 8.871, "step": 992 }, { "epoch": 0.31128526645768023, "grad_norm": 4.882453918457031, "learning_rate": 4.965e-07, "loss": 12.0715, "step": 993 }, { "epoch": 0.31159874608150473, "grad_norm": 5.079671859741211, "learning_rate": 4.970000000000001e-07, "loss": 14.1268, "step": 994 }, { "epoch": 0.31191222570532917, "grad_norm": 4.953401565551758, "learning_rate": 4.975000000000001e-07, "loss": 9.5667, "step": 995 }, { "epoch": 0.3122257053291536, "grad_norm": 3.9187986850738525, "learning_rate": 4.98e-07, "loss": 8.5076, "step": 996 }, { "epoch": 0.31253918495297806, "grad_norm": 3.059863805770874, "learning_rate": 4.985000000000001e-07, "loss": 6.8895, "step": 997 }, { "epoch": 0.3128526645768025, "grad_norm": 3.27205228805542, "learning_rate": 4.99e-07, "loss": 7.4944, "step": 998 }, { "epoch": 0.31316614420062694, "grad_norm": 4.175319194793701, "learning_rate": 4.995e-07, "loss": 10.0083, "step": 999 }, { "epoch": 0.31347962382445144, "grad_norm": 4.775247097015381, "learning_rate": 5.000000000000001e-07, "loss": 10.5517, "step": 1000 }, { "epoch": 0.3137931034482759, "grad_norm": 4.07278299331665, "learning_rate": 5.005e-07, "loss": 10.0079, "step": 1001 }, { "epoch": 0.3141065830721003, "grad_norm": 3.5604565143585205, "learning_rate": 5.01e-07, "loss": 9.4462, "step": 1002 }, { "epoch": 0.31442006269592476, "grad_norm": 4.0364990234375, "learning_rate": 5.015000000000001e-07, "loss": 10.3931, "step": 1003 }, { "epoch": 0.3147335423197492, "grad_norm": 4.032431125640869, "learning_rate": 5.02e-07, "loss": 8.9061, "step": 1004 }, { "epoch": 0.31504702194357365, "grad_norm": 3.2184982299804688, "learning_rate": 5.025000000000001e-07, "loss": 7.8915, "step": 1005 }, { "epoch": 0.31536050156739814, "grad_norm": 3.669436454772949, "learning_rate": 5.03e-07, "loss": 8.4085, "step": 1006 }, { "epoch": 0.3156739811912226, "grad_norm": 3.7609059810638428, "learning_rate": 5.035e-07, "loss": 10.8993, "step": 1007 }, { "epoch": 0.315987460815047, "grad_norm": 3.2515149116516113, "learning_rate": 5.040000000000001e-07, "loss": 10.0338, "step": 1008 }, { "epoch": 0.31630094043887147, "grad_norm": 4.648274898529053, "learning_rate": 5.045000000000001e-07, "loss": 11.952, "step": 1009 }, { "epoch": 0.3166144200626959, "grad_norm": 3.8987960815429688, "learning_rate": 5.05e-07, "loss": 8.06, "step": 1010 }, { "epoch": 0.31692789968652035, "grad_norm": 3.405416965484619, "learning_rate": 5.055e-07, "loss": 6.5823, "step": 1011 }, { "epoch": 0.31724137931034485, "grad_norm": 4.502613067626953, "learning_rate": 5.06e-07, "loss": 13.0741, "step": 1012 }, { "epoch": 0.3175548589341693, "grad_norm": 3.1743547916412354, "learning_rate": 5.065e-07, "loss": 6.6114, "step": 1013 }, { "epoch": 0.31786833855799373, "grad_norm": 3.0800437927246094, "learning_rate": 5.070000000000001e-07, "loss": 7.7446, "step": 1014 }, { "epoch": 0.3181818181818182, "grad_norm": 4.586092472076416, "learning_rate": 5.075000000000001e-07, "loss": 12.149, "step": 1015 }, { "epoch": 0.3184952978056426, "grad_norm": 4.496054649353027, "learning_rate": 5.08e-07, "loss": 9.6337, "step": 1016 }, { "epoch": 0.31880877742946706, "grad_norm": 2.7630536556243896, "learning_rate": 5.085000000000001e-07, "loss": 6.5648, "step": 1017 }, { "epoch": 0.31912225705329156, "grad_norm": 4.153172016143799, "learning_rate": 5.09e-07, "loss": 8.3805, "step": 1018 }, { "epoch": 0.319435736677116, "grad_norm": 3.8220574855804443, "learning_rate": 5.095000000000001e-07, "loss": 8.8453, "step": 1019 }, { "epoch": 0.31974921630094044, "grad_norm": 3.7293951511383057, "learning_rate": 5.1e-07, "loss": 9.2166, "step": 1020 }, { "epoch": 0.3200626959247649, "grad_norm": 3.5755627155303955, "learning_rate": 5.105e-07, "loss": 8.9008, "step": 1021 }, { "epoch": 0.3203761755485893, "grad_norm": 3.9862473011016846, "learning_rate": 5.110000000000001e-07, "loss": 11.2926, "step": 1022 }, { "epoch": 0.32068965517241377, "grad_norm": 3.1247386932373047, "learning_rate": 5.115000000000001e-07, "loss": 7.0044, "step": 1023 }, { "epoch": 0.32100313479623827, "grad_norm": 3.4025156497955322, "learning_rate": 5.12e-07, "loss": 8.6599, "step": 1024 }, { "epoch": 0.3213166144200627, "grad_norm": 3.8938844203948975, "learning_rate": 5.125e-07, "loss": 7.4781, "step": 1025 }, { "epoch": 0.32163009404388715, "grad_norm": 3.694965124130249, "learning_rate": 5.13e-07, "loss": 8.2075, "step": 1026 }, { "epoch": 0.3219435736677116, "grad_norm": 3.9770612716674805, "learning_rate": 5.135e-07, "loss": 6.7056, "step": 1027 }, { "epoch": 0.32225705329153603, "grad_norm": 3.3346900939941406, "learning_rate": 5.140000000000001e-07, "loss": 6.8695, "step": 1028 }, { "epoch": 0.3225705329153605, "grad_norm": 4.177748680114746, "learning_rate": 5.145000000000001e-07, "loss": 11.471, "step": 1029 }, { "epoch": 0.322884012539185, "grad_norm": 2.610774517059326, "learning_rate": 5.15e-07, "loss": 7.1106, "step": 1030 }, { "epoch": 0.3231974921630094, "grad_norm": 4.45445442199707, "learning_rate": 5.155e-07, "loss": 9.852, "step": 1031 }, { "epoch": 0.32351097178683386, "grad_norm": 3.7908337116241455, "learning_rate": 5.16e-07, "loss": 7.8356, "step": 1032 }, { "epoch": 0.3238244514106583, "grad_norm": 4.410386562347412, "learning_rate": 5.165e-07, "loss": 9.0971, "step": 1033 }, { "epoch": 0.32413793103448274, "grad_norm": 3.542137384414673, "learning_rate": 5.170000000000001e-07, "loss": 11.263, "step": 1034 }, { "epoch": 0.32445141065830724, "grad_norm": 4.1727681159973145, "learning_rate": 5.175e-07, "loss": 13.1233, "step": 1035 }, { "epoch": 0.3247648902821317, "grad_norm": 3.6058053970336914, "learning_rate": 5.180000000000001e-07, "loss": 9.9525, "step": 1036 }, { "epoch": 0.3250783699059561, "grad_norm": 4.465059757232666, "learning_rate": 5.185000000000001e-07, "loss": 9.9752, "step": 1037 }, { "epoch": 0.32539184952978056, "grad_norm": 4.481200695037842, "learning_rate": 5.19e-07, "loss": 9.631, "step": 1038 }, { "epoch": 0.325705329153605, "grad_norm": 4.167693138122559, "learning_rate": 5.195000000000001e-07, "loss": 10.6723, "step": 1039 }, { "epoch": 0.32601880877742945, "grad_norm": 4.743021011352539, "learning_rate": 5.2e-07, "loss": 10.3528, "step": 1040 }, { "epoch": 0.32633228840125394, "grad_norm": 3.419243812561035, "learning_rate": 5.205e-07, "loss": 6.2905, "step": 1041 }, { "epoch": 0.3266457680250784, "grad_norm": 3.0460777282714844, "learning_rate": 5.210000000000001e-07, "loss": 6.7207, "step": 1042 }, { "epoch": 0.32695924764890283, "grad_norm": 3.9700310230255127, "learning_rate": 5.215000000000001e-07, "loss": 8.2999, "step": 1043 }, { "epoch": 0.32727272727272727, "grad_norm": 3.008124351501465, "learning_rate": 5.22e-07, "loss": 7.1742, "step": 1044 }, { "epoch": 0.3275862068965517, "grad_norm": 3.2137064933776855, "learning_rate": 5.225e-07, "loss": 6.8929, "step": 1045 }, { "epoch": 0.32789968652037615, "grad_norm": 4.896831035614014, "learning_rate": 5.23e-07, "loss": 11.7279, "step": 1046 }, { "epoch": 0.32821316614420065, "grad_norm": 3.5546038150787354, "learning_rate": 5.235e-07, "loss": 7.8732, "step": 1047 }, { "epoch": 0.3285266457680251, "grad_norm": 3.5171022415161133, "learning_rate": 5.240000000000001e-07, "loss": 8.6413, "step": 1048 }, { "epoch": 0.32884012539184954, "grad_norm": 3.2662153244018555, "learning_rate": 5.245e-07, "loss": 8.4025, "step": 1049 }, { "epoch": 0.329153605015674, "grad_norm": 4.3712005615234375, "learning_rate": 5.250000000000001e-07, "loss": 8.6959, "step": 1050 }, { "epoch": 0.3294670846394984, "grad_norm": 3.437635660171509, "learning_rate": 5.255e-07, "loss": 9.4805, "step": 1051 }, { "epoch": 0.32978056426332286, "grad_norm": 4.1790876388549805, "learning_rate": 5.26e-07, "loss": 8.4143, "step": 1052 }, { "epoch": 0.33009404388714736, "grad_norm": 3.861347198486328, "learning_rate": 5.265000000000001e-07, "loss": 7.995, "step": 1053 }, { "epoch": 0.3304075235109718, "grad_norm": 4.775915622711182, "learning_rate": 5.27e-07, "loss": 13.2288, "step": 1054 }, { "epoch": 0.33072100313479624, "grad_norm": 3.08349347114563, "learning_rate": 5.275e-07, "loss": 8.4152, "step": 1055 }, { "epoch": 0.3310344827586207, "grad_norm": 4.555473804473877, "learning_rate": 5.280000000000001e-07, "loss": 9.4847, "step": 1056 }, { "epoch": 0.3313479623824451, "grad_norm": 4.14658260345459, "learning_rate": 5.285000000000001e-07, "loss": 9.6713, "step": 1057 }, { "epoch": 0.33166144200626957, "grad_norm": 3.729527711868286, "learning_rate": 5.29e-07, "loss": 9.5404, "step": 1058 }, { "epoch": 0.33197492163009407, "grad_norm": 4.110138893127441, "learning_rate": 5.295e-07, "loss": 7.4485, "step": 1059 }, { "epoch": 0.3322884012539185, "grad_norm": 3.4189460277557373, "learning_rate": 5.3e-07, "loss": 9.1945, "step": 1060 }, { "epoch": 0.33260188087774295, "grad_norm": 4.435842990875244, "learning_rate": 5.305e-07, "loss": 9.3644, "step": 1061 }, { "epoch": 0.3329153605015674, "grad_norm": 3.913818836212158, "learning_rate": 5.310000000000001e-07, "loss": 10.439, "step": 1062 }, { "epoch": 0.33322884012539183, "grad_norm": 5.079927921295166, "learning_rate": 5.315000000000001e-07, "loss": 11.0371, "step": 1063 }, { "epoch": 0.3335423197492163, "grad_norm": 3.028818368911743, "learning_rate": 5.32e-07, "loss": 6.7478, "step": 1064 }, { "epoch": 0.3338557993730408, "grad_norm": 2.9255435466766357, "learning_rate": 5.325e-07, "loss": 6.9333, "step": 1065 }, { "epoch": 0.3341692789968652, "grad_norm": 3.4844298362731934, "learning_rate": 5.33e-07, "loss": 8.3449, "step": 1066 }, { "epoch": 0.33448275862068966, "grad_norm": 4.424320697784424, "learning_rate": 5.335000000000001e-07, "loss": 11.034, "step": 1067 }, { "epoch": 0.3347962382445141, "grad_norm": 4.413206577301025, "learning_rate": 5.340000000000001e-07, "loss": 9.8029, "step": 1068 }, { "epoch": 0.33510971786833854, "grad_norm": 3.4747884273529053, "learning_rate": 5.345e-07, "loss": 8.9154, "step": 1069 }, { "epoch": 0.335423197492163, "grad_norm": 3.6665916442871094, "learning_rate": 5.350000000000001e-07, "loss": 7.1011, "step": 1070 }, { "epoch": 0.3357366771159875, "grad_norm": 2.8998830318450928, "learning_rate": 5.355e-07, "loss": 7.7268, "step": 1071 }, { "epoch": 0.3360501567398119, "grad_norm": 3.9192614555358887, "learning_rate": 5.36e-07, "loss": 10.9592, "step": 1072 }, { "epoch": 0.33636363636363636, "grad_norm": 2.7648117542266846, "learning_rate": 5.365000000000001e-07, "loss": 7.8875, "step": 1073 }, { "epoch": 0.3366771159874608, "grad_norm": 4.6224751472473145, "learning_rate": 5.37e-07, "loss": 10.7205, "step": 1074 }, { "epoch": 0.33699059561128525, "grad_norm": 3.7434327602386475, "learning_rate": 5.375e-07, "loss": 9.4757, "step": 1075 }, { "epoch": 0.3373040752351097, "grad_norm": 2.9543545246124268, "learning_rate": 5.380000000000001e-07, "loss": 6.5248, "step": 1076 }, { "epoch": 0.3376175548589342, "grad_norm": 4.7394185066223145, "learning_rate": 5.385000000000001e-07, "loss": 16.7804, "step": 1077 }, { "epoch": 0.33793103448275863, "grad_norm": 3.574186325073242, "learning_rate": 5.39e-07, "loss": 7.1065, "step": 1078 }, { "epoch": 0.33824451410658307, "grad_norm": 2.6284937858581543, "learning_rate": 5.395e-07, "loss": 6.8801, "step": 1079 }, { "epoch": 0.3385579937304075, "grad_norm": 3.36057710647583, "learning_rate": 5.4e-07, "loss": 10.2646, "step": 1080 }, { "epoch": 0.33887147335423196, "grad_norm": 4.6945624351501465, "learning_rate": 5.405000000000001e-07, "loss": 12.4643, "step": 1081 }, { "epoch": 0.33918495297805645, "grad_norm": 3.1167168617248535, "learning_rate": 5.410000000000001e-07, "loss": 6.7228, "step": 1082 }, { "epoch": 0.3394984326018809, "grad_norm": 3.513232469558716, "learning_rate": 5.415e-07, "loss": 9.1163, "step": 1083 }, { "epoch": 0.33981191222570534, "grad_norm": 5.232317924499512, "learning_rate": 5.420000000000001e-07, "loss": 11.4712, "step": 1084 }, { "epoch": 0.3401253918495298, "grad_norm": 3.7332653999328613, "learning_rate": 5.425e-07, "loss": 7.6378, "step": 1085 }, { "epoch": 0.3404388714733542, "grad_norm": 4.457612037658691, "learning_rate": 5.43e-07, "loss": 11.9674, "step": 1086 }, { "epoch": 0.34075235109717866, "grad_norm": 3.196615695953369, "learning_rate": 5.435000000000001e-07, "loss": 8.66, "step": 1087 }, { "epoch": 0.34106583072100316, "grad_norm": 4.207455635070801, "learning_rate": 5.44e-07, "loss": 8.5547, "step": 1088 }, { "epoch": 0.3413793103448276, "grad_norm": 3.803297519683838, "learning_rate": 5.445e-07, "loss": 8.1143, "step": 1089 }, { "epoch": 0.34169278996865204, "grad_norm": 3.9146108627319336, "learning_rate": 5.450000000000001e-07, "loss": 8.9048, "step": 1090 }, { "epoch": 0.3420062695924765, "grad_norm": 3.0338478088378906, "learning_rate": 5.455e-07, "loss": 7.1311, "step": 1091 }, { "epoch": 0.34231974921630093, "grad_norm": 3.301464557647705, "learning_rate": 5.46e-07, "loss": 7.2672, "step": 1092 }, { "epoch": 0.34263322884012537, "grad_norm": 3.3662900924682617, "learning_rate": 5.465e-07, "loss": 9.0924, "step": 1093 }, { "epoch": 0.34294670846394987, "grad_norm": 3.946822166442871, "learning_rate": 5.47e-07, "loss": 10.3786, "step": 1094 }, { "epoch": 0.3432601880877743, "grad_norm": 3.5544939041137695, "learning_rate": 5.475e-07, "loss": 7.2466, "step": 1095 }, { "epoch": 0.34357366771159875, "grad_norm": 3.7304489612579346, "learning_rate": 5.480000000000001e-07, "loss": 9.4879, "step": 1096 }, { "epoch": 0.3438871473354232, "grad_norm": 4.137592315673828, "learning_rate": 5.485e-07, "loss": 8.2506, "step": 1097 }, { "epoch": 0.34420062695924764, "grad_norm": 3.8458902835845947, "learning_rate": 5.490000000000001e-07, "loss": 10.3391, "step": 1098 }, { "epoch": 0.3445141065830721, "grad_norm": 3.944636106491089, "learning_rate": 5.495e-07, "loss": 10.2381, "step": 1099 }, { "epoch": 0.3448275862068966, "grad_norm": 2.935556411743164, "learning_rate": 5.5e-07, "loss": 6.7504, "step": 1100 }, { "epoch": 0.345141065830721, "grad_norm": 2.839891195297241, "learning_rate": 5.505000000000001e-07, "loss": 6.2424, "step": 1101 }, { "epoch": 0.34545454545454546, "grad_norm": 3.0521023273468018, "learning_rate": 5.510000000000001e-07, "loss": 8.1019, "step": 1102 }, { "epoch": 0.3457680250783699, "grad_norm": 2.7549774646759033, "learning_rate": 5.515e-07, "loss": 6.9908, "step": 1103 }, { "epoch": 0.34608150470219434, "grad_norm": 2.460770845413208, "learning_rate": 5.520000000000001e-07, "loss": 6.4588, "step": 1104 }, { "epoch": 0.3463949843260188, "grad_norm": 3.0610549449920654, "learning_rate": 5.525e-07, "loss": 7.1453, "step": 1105 }, { "epoch": 0.3467084639498433, "grad_norm": 3.7821202278137207, "learning_rate": 5.53e-07, "loss": 8.4413, "step": 1106 }, { "epoch": 0.3470219435736677, "grad_norm": 3.272524356842041, "learning_rate": 5.535000000000001e-07, "loss": 7.5479, "step": 1107 }, { "epoch": 0.34733542319749217, "grad_norm": 4.5854082107543945, "learning_rate": 5.54e-07, "loss": 12.7431, "step": 1108 }, { "epoch": 0.3476489028213166, "grad_norm": 3.382383346557617, "learning_rate": 5.545e-07, "loss": 8.0938, "step": 1109 }, { "epoch": 0.34796238244514105, "grad_norm": 4.590867042541504, "learning_rate": 5.550000000000001e-07, "loss": 9.5862, "step": 1110 }, { "epoch": 0.3482758620689655, "grad_norm": 3.3720743656158447, "learning_rate": 5.555e-07, "loss": 7.8215, "step": 1111 }, { "epoch": 0.34858934169279, "grad_norm": 3.0238683223724365, "learning_rate": 5.560000000000001e-07, "loss": 8.9804, "step": 1112 }, { "epoch": 0.34890282131661443, "grad_norm": 3.9639229774475098, "learning_rate": 5.565e-07, "loss": 9.2357, "step": 1113 }, { "epoch": 0.3492163009404389, "grad_norm": 2.6964738368988037, "learning_rate": 5.57e-07, "loss": 7.0609, "step": 1114 }, { "epoch": 0.3495297805642633, "grad_norm": 4.804142475128174, "learning_rate": 5.575000000000001e-07, "loss": 13.9845, "step": 1115 }, { "epoch": 0.34984326018808776, "grad_norm": 4.389916896820068, "learning_rate": 5.580000000000001e-07, "loss": 7.7403, "step": 1116 }, { "epoch": 0.3501567398119122, "grad_norm": 4.200212478637695, "learning_rate": 5.585e-07, "loss": 9.193, "step": 1117 }, { "epoch": 0.3504702194357367, "grad_norm": 4.732844829559326, "learning_rate": 5.590000000000001e-07, "loss": 13.0463, "step": 1118 }, { "epoch": 0.35078369905956114, "grad_norm": 2.878286123275757, "learning_rate": 5.595e-07, "loss": 6.3315, "step": 1119 }, { "epoch": 0.3510971786833856, "grad_norm": 4.683932781219482, "learning_rate": 5.6e-07, "loss": 10.1464, "step": 1120 }, { "epoch": 0.35141065830721, "grad_norm": 3.83774995803833, "learning_rate": 5.605000000000001e-07, "loss": 8.5081, "step": 1121 }, { "epoch": 0.35172413793103446, "grad_norm": 3.030034303665161, "learning_rate": 5.61e-07, "loss": 6.6782, "step": 1122 }, { "epoch": 0.35203761755485896, "grad_norm": 3.7640645503997803, "learning_rate": 5.615e-07, "loss": 7.6323, "step": 1123 }, { "epoch": 0.3523510971786834, "grad_norm": 3.636522054672241, "learning_rate": 5.620000000000001e-07, "loss": 8.9132, "step": 1124 }, { "epoch": 0.35266457680250785, "grad_norm": 5.256652355194092, "learning_rate": 5.625e-07, "loss": 13.0476, "step": 1125 }, { "epoch": 0.3529780564263323, "grad_norm": 3.7114644050598145, "learning_rate": 5.63e-07, "loss": 7.7855, "step": 1126 }, { "epoch": 0.35329153605015673, "grad_norm": 4.469037055969238, "learning_rate": 5.635e-07, "loss": 10.8488, "step": 1127 }, { "epoch": 0.35360501567398117, "grad_norm": 4.428511619567871, "learning_rate": 5.64e-07, "loss": 10.8417, "step": 1128 }, { "epoch": 0.35391849529780567, "grad_norm": 3.272402048110962, "learning_rate": 5.645000000000001e-07, "loss": 7.0908, "step": 1129 }, { "epoch": 0.3542319749216301, "grad_norm": 4.514960765838623, "learning_rate": 5.650000000000001e-07, "loss": 11.1969, "step": 1130 }, { "epoch": 0.35454545454545455, "grad_norm": 3.0566494464874268, "learning_rate": 5.655e-07, "loss": 6.9586, "step": 1131 }, { "epoch": 0.354858934169279, "grad_norm": 4.796217441558838, "learning_rate": 5.660000000000001e-07, "loss": 12.4509, "step": 1132 }, { "epoch": 0.35517241379310344, "grad_norm": 3.8200011253356934, "learning_rate": 5.665e-07, "loss": 7.6372, "step": 1133 }, { "epoch": 0.3554858934169279, "grad_norm": 4.99943733215332, "learning_rate": 5.67e-07, "loss": 16.3218, "step": 1134 }, { "epoch": 0.3557993730407524, "grad_norm": 4.074048042297363, "learning_rate": 5.675000000000001e-07, "loss": 11.7632, "step": 1135 }, { "epoch": 0.3561128526645768, "grad_norm": 3.519336700439453, "learning_rate": 5.680000000000001e-07, "loss": 12.5035, "step": 1136 }, { "epoch": 0.35642633228840126, "grad_norm": 3.5221669673919678, "learning_rate": 5.685e-07, "loss": 8.2566, "step": 1137 }, { "epoch": 0.3567398119122257, "grad_norm": 3.0210797786712646, "learning_rate": 5.690000000000001e-07, "loss": 7.6131, "step": 1138 }, { "epoch": 0.35705329153605014, "grad_norm": 3.848649740219116, "learning_rate": 5.695e-07, "loss": 9.2893, "step": 1139 }, { "epoch": 0.3573667711598746, "grad_norm": 3.4947142601013184, "learning_rate": 5.7e-07, "loss": 10.0174, "step": 1140 }, { "epoch": 0.3576802507836991, "grad_norm": 4.418102741241455, "learning_rate": 5.705e-07, "loss": 10.2442, "step": 1141 }, { "epoch": 0.3579937304075235, "grad_norm": 3.0057661533355713, "learning_rate": 5.71e-07, "loss": 6.8693, "step": 1142 }, { "epoch": 0.35830721003134797, "grad_norm": 5.070820331573486, "learning_rate": 5.715000000000001e-07, "loss": 12.646, "step": 1143 }, { "epoch": 0.3586206896551724, "grad_norm": 3.685112714767456, "learning_rate": 5.720000000000001e-07, "loss": 8.9385, "step": 1144 }, { "epoch": 0.35893416927899685, "grad_norm": 3.591017007827759, "learning_rate": 5.725e-07, "loss": 7.2102, "step": 1145 }, { "epoch": 0.3592476489028213, "grad_norm": 3.66312575340271, "learning_rate": 5.730000000000001e-07, "loss": 8.0567, "step": 1146 }, { "epoch": 0.3595611285266458, "grad_norm": 3.2532527446746826, "learning_rate": 5.735e-07, "loss": 7.5575, "step": 1147 }, { "epoch": 0.35987460815047023, "grad_norm": 2.452608108520508, "learning_rate": 5.74e-07, "loss": 6.1812, "step": 1148 }, { "epoch": 0.3601880877742947, "grad_norm": 4.427051067352295, "learning_rate": 5.745000000000001e-07, "loss": 9.7562, "step": 1149 }, { "epoch": 0.3605015673981191, "grad_norm": 4.048511981964111, "learning_rate": 5.750000000000001e-07, "loss": 7.2065, "step": 1150 }, { "epoch": 0.36081504702194356, "grad_norm": 4.313746929168701, "learning_rate": 5.755e-07, "loss": 12.0608, "step": 1151 }, { "epoch": 0.361128526645768, "grad_norm": 3.703238010406494, "learning_rate": 5.760000000000001e-07, "loss": 8.8626, "step": 1152 }, { "epoch": 0.3614420062695925, "grad_norm": 3.321089744567871, "learning_rate": 5.765e-07, "loss": 7.7751, "step": 1153 }, { "epoch": 0.36175548589341694, "grad_norm": 3.596169948577881, "learning_rate": 5.77e-07, "loss": 7.2633, "step": 1154 }, { "epoch": 0.3620689655172414, "grad_norm": 3.9182851314544678, "learning_rate": 5.775000000000001e-07, "loss": 8.9107, "step": 1155 }, { "epoch": 0.3623824451410658, "grad_norm": 2.809251070022583, "learning_rate": 5.78e-07, "loss": 7.5999, "step": 1156 }, { "epoch": 0.36269592476489027, "grad_norm": 3.9834253787994385, "learning_rate": 5.785e-07, "loss": 10.8149, "step": 1157 }, { "epoch": 0.3630094043887147, "grad_norm": 3.9162814617156982, "learning_rate": 5.790000000000001e-07, "loss": 9.0022, "step": 1158 }, { "epoch": 0.3633228840125392, "grad_norm": 3.7181437015533447, "learning_rate": 5.795e-07, "loss": 12.1042, "step": 1159 }, { "epoch": 0.36363636363636365, "grad_norm": 3.6745083332061768, "learning_rate": 5.800000000000001e-07, "loss": 10.9387, "step": 1160 }, { "epoch": 0.3639498432601881, "grad_norm": 3.6251866817474365, "learning_rate": 5.805e-07, "loss": 9.3794, "step": 1161 }, { "epoch": 0.36426332288401253, "grad_norm": 3.9675116539001465, "learning_rate": 5.81e-07, "loss": 7.4843, "step": 1162 }, { "epoch": 0.364576802507837, "grad_norm": 4.0218915939331055, "learning_rate": 5.815000000000001e-07, "loss": 11.3793, "step": 1163 }, { "epoch": 0.36489028213166147, "grad_norm": 4.011270999908447, "learning_rate": 5.820000000000001e-07, "loss": 8.5104, "step": 1164 }, { "epoch": 0.3652037617554859, "grad_norm": 3.468784809112549, "learning_rate": 5.825e-07, "loss": 7.195, "step": 1165 }, { "epoch": 0.36551724137931035, "grad_norm": 5.364189624786377, "learning_rate": 5.830000000000001e-07, "loss": 19.2203, "step": 1166 }, { "epoch": 0.3658307210031348, "grad_norm": 4.526871681213379, "learning_rate": 5.835e-07, "loss": 11.444, "step": 1167 }, { "epoch": 0.36614420062695924, "grad_norm": 3.9016737937927246, "learning_rate": 5.84e-07, "loss": 11.455, "step": 1168 }, { "epoch": 0.3664576802507837, "grad_norm": 3.691763401031494, "learning_rate": 5.845000000000001e-07, "loss": 8.6588, "step": 1169 }, { "epoch": 0.3667711598746082, "grad_norm": 3.842839241027832, "learning_rate": 5.850000000000001e-07, "loss": 8.317, "step": 1170 }, { "epoch": 0.3670846394984326, "grad_norm": 3.8796255588531494, "learning_rate": 5.855e-07, "loss": 9.903, "step": 1171 }, { "epoch": 0.36739811912225706, "grad_norm": 4.093709468841553, "learning_rate": 5.860000000000001e-07, "loss": 9.6562, "step": 1172 }, { "epoch": 0.3677115987460815, "grad_norm": 4.154341220855713, "learning_rate": 5.865e-07, "loss": 9.7709, "step": 1173 }, { "epoch": 0.36802507836990594, "grad_norm": 5.882456302642822, "learning_rate": 5.870000000000001e-07, "loss": 16.238, "step": 1174 }, { "epoch": 0.3683385579937304, "grad_norm": 4.124493598937988, "learning_rate": 5.875e-07, "loss": 9.4565, "step": 1175 }, { "epoch": 0.3686520376175549, "grad_norm": 3.4049675464630127, "learning_rate": 5.88e-07, "loss": 6.7959, "step": 1176 }, { "epoch": 0.3689655172413793, "grad_norm": 3.196589708328247, "learning_rate": 5.885000000000001e-07, "loss": 6.6816, "step": 1177 }, { "epoch": 0.36927899686520377, "grad_norm": 4.004443645477295, "learning_rate": 5.890000000000001e-07, "loss": 9.0686, "step": 1178 }, { "epoch": 0.3695924764890282, "grad_norm": 3.7830114364624023, "learning_rate": 5.895e-07, "loss": 8.4215, "step": 1179 }, { "epoch": 0.36990595611285265, "grad_norm": 16.382652282714844, "learning_rate": 5.900000000000001e-07, "loss": 8.5894, "step": 1180 }, { "epoch": 0.3702194357366771, "grad_norm": 4.150466442108154, "learning_rate": 5.905e-07, "loss": 9.9268, "step": 1181 }, { "epoch": 0.3705329153605016, "grad_norm": 4.039095401763916, "learning_rate": 5.91e-07, "loss": 7.6783, "step": 1182 }, { "epoch": 0.37084639498432603, "grad_norm": 4.113837242126465, "learning_rate": 5.915000000000001e-07, "loss": 10.9473, "step": 1183 }, { "epoch": 0.3711598746081505, "grad_norm": 4.3714447021484375, "learning_rate": 5.920000000000001e-07, "loss": 8.6596, "step": 1184 }, { "epoch": 0.3714733542319749, "grad_norm": 4.696350574493408, "learning_rate": 5.925e-07, "loss": 11.0915, "step": 1185 }, { "epoch": 0.37178683385579936, "grad_norm": 3.4268805980682373, "learning_rate": 5.930000000000001e-07, "loss": 8.7898, "step": 1186 }, { "epoch": 0.3721003134796238, "grad_norm": 3.7892448902130127, "learning_rate": 5.935e-07, "loss": 7.9897, "step": 1187 }, { "epoch": 0.3724137931034483, "grad_norm": 3.5174129009246826, "learning_rate": 5.94e-07, "loss": 8.9607, "step": 1188 }, { "epoch": 0.37272727272727274, "grad_norm": 3.5492613315582275, "learning_rate": 5.945000000000001e-07, "loss": 8.2248, "step": 1189 }, { "epoch": 0.3730407523510972, "grad_norm": 3.0557780265808105, "learning_rate": 5.95e-07, "loss": 7.4525, "step": 1190 }, { "epoch": 0.3733542319749216, "grad_norm": 3.610521078109741, "learning_rate": 5.955000000000001e-07, "loss": 8.6173, "step": 1191 }, { "epoch": 0.37366771159874607, "grad_norm": 5.024316310882568, "learning_rate": 5.960000000000001e-07, "loss": 9.0102, "step": 1192 }, { "epoch": 0.3739811912225705, "grad_norm": 3.3854024410247803, "learning_rate": 5.965e-07, "loss": 9.3159, "step": 1193 }, { "epoch": 0.374294670846395, "grad_norm": 3.692943572998047, "learning_rate": 5.970000000000001e-07, "loss": 6.6683, "step": 1194 }, { "epoch": 0.37460815047021945, "grad_norm": 4.656247138977051, "learning_rate": 5.975e-07, "loss": 9.6465, "step": 1195 }, { "epoch": 0.3749216300940439, "grad_norm": 2.7945048809051514, "learning_rate": 5.98e-07, "loss": 7.6828, "step": 1196 }, { "epoch": 0.37523510971786833, "grad_norm": 3.6382293701171875, "learning_rate": 5.985000000000001e-07, "loss": 8.2837, "step": 1197 }, { "epoch": 0.3755485893416928, "grad_norm": 4.916844844818115, "learning_rate": 5.990000000000001e-07, "loss": 10.6478, "step": 1198 }, { "epoch": 0.3758620689655172, "grad_norm": 2.6404612064361572, "learning_rate": 5.995e-07, "loss": 5.8349, "step": 1199 }, { "epoch": 0.3761755485893417, "grad_norm": 3.963083505630493, "learning_rate": 6.000000000000001e-07, "loss": 7.5912, "step": 1200 }, { "epoch": 0.37648902821316615, "grad_norm": 3.9304614067077637, "learning_rate": 6.005e-07, "loss": 9.4547, "step": 1201 }, { "epoch": 0.3768025078369906, "grad_norm": 3.4377453327178955, "learning_rate": 6.01e-07, "loss": 9.7902, "step": 1202 }, { "epoch": 0.37711598746081504, "grad_norm": 3.773444414138794, "learning_rate": 6.015000000000001e-07, "loss": 9.5372, "step": 1203 }, { "epoch": 0.3774294670846395, "grad_norm": 3.075242280960083, "learning_rate": 6.02e-07, "loss": 8.3831, "step": 1204 }, { "epoch": 0.3777429467084639, "grad_norm": 3.2014951705932617, "learning_rate": 6.025000000000001e-07, "loss": 7.9814, "step": 1205 }, { "epoch": 0.3780564263322884, "grad_norm": 3.752372980117798, "learning_rate": 6.030000000000001e-07, "loss": 8.109, "step": 1206 }, { "epoch": 0.37836990595611286, "grad_norm": 4.027518272399902, "learning_rate": 6.035e-07, "loss": 7.6091, "step": 1207 }, { "epoch": 0.3786833855799373, "grad_norm": 3.438729763031006, "learning_rate": 6.040000000000001e-07, "loss": 6.8037, "step": 1208 }, { "epoch": 0.37899686520376175, "grad_norm": 4.082209587097168, "learning_rate": 6.045e-07, "loss": 9.9197, "step": 1209 }, { "epoch": 0.3793103448275862, "grad_norm": 5.536364555358887, "learning_rate": 6.05e-07, "loss": 13.404, "step": 1210 }, { "epoch": 0.3796238244514107, "grad_norm": 3.8817527294158936, "learning_rate": 6.055000000000001e-07, "loss": 11.2098, "step": 1211 }, { "epoch": 0.3799373040752351, "grad_norm": 3.6680006980895996, "learning_rate": 6.060000000000001e-07, "loss": 8.0359, "step": 1212 }, { "epoch": 0.38025078369905957, "grad_norm": 4.336513519287109, "learning_rate": 6.065e-07, "loss": 10.526, "step": 1213 }, { "epoch": 0.380564263322884, "grad_norm": 3.7376954555511475, "learning_rate": 6.07e-07, "loss": 9.7812, "step": 1214 }, { "epoch": 0.38087774294670845, "grad_norm": 4.394288539886475, "learning_rate": 6.075e-07, "loss": 10.5658, "step": 1215 }, { "epoch": 0.3811912225705329, "grad_norm": 3.9178755283355713, "learning_rate": 6.08e-07, "loss": 9.1735, "step": 1216 }, { "epoch": 0.3815047021943574, "grad_norm": 3.385538339614868, "learning_rate": 6.085000000000001e-07, "loss": 8.2474, "step": 1217 }, { "epoch": 0.38181818181818183, "grad_norm": 3.3162765502929688, "learning_rate": 6.090000000000001e-07, "loss": 7.6934, "step": 1218 }, { "epoch": 0.3821316614420063, "grad_norm": 4.181975364685059, "learning_rate": 6.095e-07, "loss": 9.205, "step": 1219 }, { "epoch": 0.3824451410658307, "grad_norm": 3.707218647003174, "learning_rate": 6.100000000000001e-07, "loss": 7.6344, "step": 1220 }, { "epoch": 0.38275862068965516, "grad_norm": 3.0025179386138916, "learning_rate": 6.105e-07, "loss": 8.5384, "step": 1221 }, { "epoch": 0.3830721003134796, "grad_norm": 3.2136359214782715, "learning_rate": 6.110000000000001e-07, "loss": 7.3391, "step": 1222 }, { "epoch": 0.3833855799373041, "grad_norm": 3.87955904006958, "learning_rate": 6.115000000000001e-07, "loss": 9.051, "step": 1223 }, { "epoch": 0.38369905956112854, "grad_norm": 2.9862728118896484, "learning_rate": 6.12e-07, "loss": 6.7873, "step": 1224 }, { "epoch": 0.384012539184953, "grad_norm": 4.336016654968262, "learning_rate": 6.125000000000001e-07, "loss": 8.1016, "step": 1225 }, { "epoch": 0.3843260188087774, "grad_norm": 3.270927906036377, "learning_rate": 6.130000000000001e-07, "loss": 9.3506, "step": 1226 }, { "epoch": 0.38463949843260187, "grad_norm": 4.06575870513916, "learning_rate": 6.135e-07, "loss": 8.5931, "step": 1227 }, { "epoch": 0.3849529780564263, "grad_norm": 3.6191561222076416, "learning_rate": 6.140000000000001e-07, "loss": 8.4115, "step": 1228 }, { "epoch": 0.3852664576802508, "grad_norm": 2.8012547492980957, "learning_rate": 6.145e-07, "loss": 7.0199, "step": 1229 }, { "epoch": 0.38557993730407525, "grad_norm": 4.054346084594727, "learning_rate": 6.15e-07, "loss": 9.9699, "step": 1230 }, { "epoch": 0.3858934169278997, "grad_norm": 4.556372165679932, "learning_rate": 6.155000000000001e-07, "loss": 14.4437, "step": 1231 }, { "epoch": 0.38620689655172413, "grad_norm": 4.824500560760498, "learning_rate": 6.160000000000001e-07, "loss": 10.5535, "step": 1232 }, { "epoch": 0.3865203761755486, "grad_norm": 4.935982704162598, "learning_rate": 6.165e-07, "loss": 8.3582, "step": 1233 }, { "epoch": 0.386833855799373, "grad_norm": 8.828557014465332, "learning_rate": 6.17e-07, "loss": 19.3321, "step": 1234 }, { "epoch": 0.3871473354231975, "grad_norm": 2.967073678970337, "learning_rate": 6.175e-07, "loss": 6.7905, "step": 1235 }, { "epoch": 0.38746081504702196, "grad_norm": 4.4314799308776855, "learning_rate": 6.180000000000001e-07, "loss": 10.2915, "step": 1236 }, { "epoch": 0.3877742946708464, "grad_norm": 3.622016191482544, "learning_rate": 6.185000000000001e-07, "loss": 9.6284, "step": 1237 }, { "epoch": 0.38808777429467084, "grad_norm": 3.8630599975585938, "learning_rate": 6.19e-07, "loss": 7.4754, "step": 1238 }, { "epoch": 0.3884012539184953, "grad_norm": 3.7421834468841553, "learning_rate": 6.195000000000001e-07, "loss": 8.5903, "step": 1239 }, { "epoch": 0.3887147335423197, "grad_norm": 3.313210964202881, "learning_rate": 6.200000000000001e-07, "loss": 9.3886, "step": 1240 }, { "epoch": 0.3890282131661442, "grad_norm": 3.029578924179077, "learning_rate": 6.205e-07, "loss": 7.3684, "step": 1241 }, { "epoch": 0.38934169278996866, "grad_norm": 4.0817670822143555, "learning_rate": 6.210000000000001e-07, "loss": 8.6238, "step": 1242 }, { "epoch": 0.3896551724137931, "grad_norm": 3.1163382530212402, "learning_rate": 6.215e-07, "loss": 7.8326, "step": 1243 }, { "epoch": 0.38996865203761755, "grad_norm": 3.418872833251953, "learning_rate": 6.22e-07, "loss": 9.6686, "step": 1244 }, { "epoch": 0.390282131661442, "grad_norm": 3.265856981277466, "learning_rate": 6.225000000000001e-07, "loss": 6.8915, "step": 1245 }, { "epoch": 0.39059561128526643, "grad_norm": 2.5802524089813232, "learning_rate": 6.230000000000001e-07, "loss": 5.6296, "step": 1246 }, { "epoch": 0.39090909090909093, "grad_norm": 3.2665584087371826, "learning_rate": 6.235e-07, "loss": 7.1075, "step": 1247 }, { "epoch": 0.39122257053291537, "grad_norm": 4.159616470336914, "learning_rate": 6.24e-07, "loss": 11.453, "step": 1248 }, { "epoch": 0.3915360501567398, "grad_norm": 4.0924482345581055, "learning_rate": 6.245e-07, "loss": 9.2337, "step": 1249 }, { "epoch": 0.39184952978056425, "grad_norm": 3.432142972946167, "learning_rate": 6.25e-07, "loss": 8.3582, "step": 1250 }, { "epoch": 0.3921630094043887, "grad_norm": 3.7760117053985596, "learning_rate": 6.255e-07, "loss": 10.0285, "step": 1251 }, { "epoch": 0.3924764890282132, "grad_norm": 3.153862237930298, "learning_rate": 6.260000000000001e-07, "loss": 7.4688, "step": 1252 }, { "epoch": 0.39278996865203764, "grad_norm": 3.4726884365081787, "learning_rate": 6.265000000000001e-07, "loss": 9.7888, "step": 1253 }, { "epoch": 0.3931034482758621, "grad_norm": 2.944378614425659, "learning_rate": 6.270000000000001e-07, "loss": 7.6955, "step": 1254 }, { "epoch": 0.3934169278996865, "grad_norm": 3.5531530380249023, "learning_rate": 6.275e-07, "loss": 6.9592, "step": 1255 }, { "epoch": 0.39373040752351096, "grad_norm": 4.495529651641846, "learning_rate": 6.28e-07, "loss": 13.8164, "step": 1256 }, { "epoch": 0.3940438871473354, "grad_norm": 3.5309019088745117, "learning_rate": 6.285000000000001e-07, "loss": 8.0689, "step": 1257 }, { "epoch": 0.3943573667711599, "grad_norm": 5.7128682136535645, "learning_rate": 6.29e-07, "loss": 14.1, "step": 1258 }, { "epoch": 0.39467084639498434, "grad_norm": 2.9314987659454346, "learning_rate": 6.295000000000001e-07, "loss": 6.5813, "step": 1259 }, { "epoch": 0.3949843260188088, "grad_norm": 4.196994781494141, "learning_rate": 6.3e-07, "loss": 11.5749, "step": 1260 }, { "epoch": 0.3952978056426332, "grad_norm": 3.2020435333251953, "learning_rate": 6.305e-07, "loss": 7.9098, "step": 1261 }, { "epoch": 0.39561128526645767, "grad_norm": 4.530463218688965, "learning_rate": 6.310000000000001e-07, "loss": 13.4221, "step": 1262 }, { "epoch": 0.3959247648902821, "grad_norm": 3.823436975479126, "learning_rate": 6.315e-07, "loss": 10.8347, "step": 1263 }, { "epoch": 0.3962382445141066, "grad_norm": 4.340320110321045, "learning_rate": 6.320000000000002e-07, "loss": 10.6129, "step": 1264 }, { "epoch": 0.39655172413793105, "grad_norm": 4.320498466491699, "learning_rate": 6.325000000000001e-07, "loss": 9.6761, "step": 1265 }, { "epoch": 0.3968652037617555, "grad_norm": 3.8908495903015137, "learning_rate": 6.33e-07, "loss": 9.3212, "step": 1266 }, { "epoch": 0.39717868338557993, "grad_norm": 4.79443359375, "learning_rate": 6.335000000000001e-07, "loss": 8.4957, "step": 1267 }, { "epoch": 0.3974921630094044, "grad_norm": 3.5879569053649902, "learning_rate": 6.34e-07, "loss": 7.9517, "step": 1268 }, { "epoch": 0.3978056426332288, "grad_norm": 3.9859533309936523, "learning_rate": 6.345000000000001e-07, "loss": 10.1546, "step": 1269 }, { "epoch": 0.3981191222570533, "grad_norm": 4.457554817199707, "learning_rate": 6.350000000000001e-07, "loss": 11.0071, "step": 1270 }, { "epoch": 0.39843260188087776, "grad_norm": 3.4207284450531006, "learning_rate": 6.355e-07, "loss": 7.5192, "step": 1271 }, { "epoch": 0.3987460815047022, "grad_norm": 2.9599740505218506, "learning_rate": 6.360000000000001e-07, "loss": 8.1417, "step": 1272 }, { "epoch": 0.39905956112852664, "grad_norm": 5.587130069732666, "learning_rate": 6.365000000000001e-07, "loss": 13.4308, "step": 1273 }, { "epoch": 0.3993730407523511, "grad_norm": 4.257179260253906, "learning_rate": 6.370000000000001e-07, "loss": 9.8754, "step": 1274 }, { "epoch": 0.3996865203761755, "grad_norm": 3.8205981254577637, "learning_rate": 6.375e-07, "loss": 7.3584, "step": 1275 }, { "epoch": 0.4, "grad_norm": 4.358394145965576, "learning_rate": 6.38e-07, "loss": 8.0118, "step": 1276 }, { "epoch": 0.40031347962382446, "grad_norm": 3.71220326423645, "learning_rate": 6.385000000000001e-07, "loss": 10.123, "step": 1277 }, { "epoch": 0.4006269592476489, "grad_norm": 3.896798610687256, "learning_rate": 6.39e-07, "loss": 7.8256, "step": 1278 }, { "epoch": 0.40094043887147335, "grad_norm": 3.1705808639526367, "learning_rate": 6.395000000000001e-07, "loss": 7.985, "step": 1279 }, { "epoch": 0.4012539184952978, "grad_norm": 5.2914838790893555, "learning_rate": 6.4e-07, "loss": 10.3964, "step": 1280 }, { "epoch": 0.40156739811912223, "grad_norm": 3.0464425086975098, "learning_rate": 6.405e-07, "loss": 6.6271, "step": 1281 }, { "epoch": 0.40188087774294673, "grad_norm": 4.177611827850342, "learning_rate": 6.410000000000001e-07, "loss": 12.3688, "step": 1282 }, { "epoch": 0.40219435736677117, "grad_norm": 3.6782820224761963, "learning_rate": 6.415e-07, "loss": 7.3459, "step": 1283 }, { "epoch": 0.4025078369905956, "grad_norm": 3.16139817237854, "learning_rate": 6.42e-07, "loss": 7.4459, "step": 1284 }, { "epoch": 0.40282131661442006, "grad_norm": 3.282700538635254, "learning_rate": 6.425000000000001e-07, "loss": 6.6329, "step": 1285 }, { "epoch": 0.4031347962382445, "grad_norm": 3.4450936317443848, "learning_rate": 6.43e-07, "loss": 9.4665, "step": 1286 }, { "epoch": 0.40344827586206894, "grad_norm": 2.994030475616455, "learning_rate": 6.435000000000001e-07, "loss": 6.7493, "step": 1287 }, { "epoch": 0.40376175548589344, "grad_norm": 4.3425822257995605, "learning_rate": 6.44e-07, "loss": 9.9379, "step": 1288 }, { "epoch": 0.4040752351097179, "grad_norm": 4.892250061035156, "learning_rate": 6.445e-07, "loss": 11.3928, "step": 1289 }, { "epoch": 0.4043887147335423, "grad_norm": 3.246832847595215, "learning_rate": 6.450000000000001e-07, "loss": 7.554, "step": 1290 }, { "epoch": 0.40470219435736676, "grad_norm": 4.307924270629883, "learning_rate": 6.455e-07, "loss": 11.633, "step": 1291 }, { "epoch": 0.4050156739811912, "grad_norm": 3.2229695320129395, "learning_rate": 6.460000000000001e-07, "loss": 6.5894, "step": 1292 }, { "epoch": 0.40532915360501565, "grad_norm": 4.491090774536133, "learning_rate": 6.465000000000001e-07, "loss": 12.8065, "step": 1293 }, { "epoch": 0.40564263322884014, "grad_norm": 3.604726552963257, "learning_rate": 6.47e-07, "loss": 7.5955, "step": 1294 }, { "epoch": 0.4059561128526646, "grad_norm": 2.8530771732330322, "learning_rate": 6.475e-07, "loss": 6.6698, "step": 1295 }, { "epoch": 0.406269592476489, "grad_norm": 3.601217746734619, "learning_rate": 6.48e-07, "loss": 7.8026, "step": 1296 }, { "epoch": 0.40658307210031347, "grad_norm": 3.475271701812744, "learning_rate": 6.485000000000001e-07, "loss": 7.1792, "step": 1297 }, { "epoch": 0.4068965517241379, "grad_norm": 3.568582057952881, "learning_rate": 6.490000000000001e-07, "loss": 8.6444, "step": 1298 }, { "epoch": 0.4072100313479624, "grad_norm": 3.6762585639953613, "learning_rate": 6.495e-07, "loss": 7.4771, "step": 1299 }, { "epoch": 0.40752351097178685, "grad_norm": 4.190446376800537, "learning_rate": 6.5e-07, "loss": 12.6181, "step": 1300 }, { "epoch": 0.4078369905956113, "grad_norm": 3.1675896644592285, "learning_rate": 6.505000000000001e-07, "loss": 6.2409, "step": 1301 }, { "epoch": 0.40815047021943573, "grad_norm": 4.8222856521606445, "learning_rate": 6.510000000000001e-07, "loss": 11.7033, "step": 1302 }, { "epoch": 0.4084639498432602, "grad_norm": 4.360724925994873, "learning_rate": 6.515e-07, "loss": 13.757, "step": 1303 }, { "epoch": 0.4087774294670846, "grad_norm": 3.622544288635254, "learning_rate": 6.52e-07, "loss": 9.4418, "step": 1304 }, { "epoch": 0.4090909090909091, "grad_norm": 3.5663022994995117, "learning_rate": 6.525000000000001e-07, "loss": 7.6926, "step": 1305 }, { "epoch": 0.40940438871473356, "grad_norm": 4.033201694488525, "learning_rate": 6.53e-07, "loss": 8.5123, "step": 1306 }, { "epoch": 0.409717868338558, "grad_norm": 3.480999708175659, "learning_rate": 6.535000000000001e-07, "loss": 7.2778, "step": 1307 }, { "epoch": 0.41003134796238244, "grad_norm": 3.1945245265960693, "learning_rate": 6.54e-07, "loss": 7.371, "step": 1308 }, { "epoch": 0.4103448275862069, "grad_norm": 3.4284608364105225, "learning_rate": 6.545e-07, "loss": 7.7181, "step": 1309 }, { "epoch": 0.4106583072100313, "grad_norm": 6.255939483642578, "learning_rate": 6.550000000000001e-07, "loss": 13.9958, "step": 1310 }, { "epoch": 0.4109717868338558, "grad_norm": 3.8257150650024414, "learning_rate": 6.555e-07, "loss": 9.3615, "step": 1311 }, { "epoch": 0.41128526645768027, "grad_norm": 5.354909420013428, "learning_rate": 6.560000000000002e-07, "loss": 10.8434, "step": 1312 }, { "epoch": 0.4115987460815047, "grad_norm": 4.636758804321289, "learning_rate": 6.565000000000001e-07, "loss": 10.2746, "step": 1313 }, { "epoch": 0.41191222570532915, "grad_norm": 4.180166721343994, "learning_rate": 6.57e-07, "loss": 10.871, "step": 1314 }, { "epoch": 0.4122257053291536, "grad_norm": 4.635682106018066, "learning_rate": 6.575000000000001e-07, "loss": 9.849, "step": 1315 }, { "epoch": 0.41253918495297803, "grad_norm": 3.4844233989715576, "learning_rate": 6.58e-07, "loss": 10.2331, "step": 1316 }, { "epoch": 0.41285266457680253, "grad_norm": 2.8132760524749756, "learning_rate": 6.585000000000001e-07, "loss": 6.6116, "step": 1317 }, { "epoch": 0.413166144200627, "grad_norm": 3.6411001682281494, "learning_rate": 6.590000000000001e-07, "loss": 9.5894, "step": 1318 }, { "epoch": 0.4134796238244514, "grad_norm": 3.4417355060577393, "learning_rate": 6.595e-07, "loss": 8.5706, "step": 1319 }, { "epoch": 0.41379310344827586, "grad_norm": 4.843432903289795, "learning_rate": 6.6e-07, "loss": 8.9815, "step": 1320 }, { "epoch": 0.4141065830721003, "grad_norm": 4.353475093841553, "learning_rate": 6.605000000000001e-07, "loss": 7.6671, "step": 1321 }, { "epoch": 0.41442006269592474, "grad_norm": 3.1126198768615723, "learning_rate": 6.610000000000001e-07, "loss": 7.3491, "step": 1322 }, { "epoch": 0.41473354231974924, "grad_norm": 2.952798366546631, "learning_rate": 6.615e-07, "loss": 6.6309, "step": 1323 }, { "epoch": 0.4150470219435737, "grad_norm": 3.8614158630371094, "learning_rate": 6.62e-07, "loss": 7.1688, "step": 1324 }, { "epoch": 0.4153605015673981, "grad_norm": 3.264158248901367, "learning_rate": 6.625000000000001e-07, "loss": 7.4095, "step": 1325 }, { "epoch": 0.41567398119122256, "grad_norm": 4.060593605041504, "learning_rate": 6.63e-07, "loss": 11.2259, "step": 1326 }, { "epoch": 0.415987460815047, "grad_norm": 3.954174757003784, "learning_rate": 6.635000000000001e-07, "loss": 9.4012, "step": 1327 }, { "epoch": 0.41630094043887145, "grad_norm": 3.335627555847168, "learning_rate": 6.64e-07, "loss": 7.6803, "step": 1328 }, { "epoch": 0.41661442006269594, "grad_norm": 3.3831787109375, "learning_rate": 6.645000000000001e-07, "loss": 8.848, "step": 1329 }, { "epoch": 0.4169278996865204, "grad_norm": 4.064548492431641, "learning_rate": 6.650000000000001e-07, "loss": 7.7122, "step": 1330 }, { "epoch": 0.41724137931034483, "grad_norm": 3.576258659362793, "learning_rate": 6.655e-07, "loss": 9.0345, "step": 1331 }, { "epoch": 0.41755485893416927, "grad_norm": 4.037924766540527, "learning_rate": 6.660000000000002e-07, "loss": 10.0141, "step": 1332 }, { "epoch": 0.4178683385579937, "grad_norm": 4.519690990447998, "learning_rate": 6.665000000000001e-07, "loss": 7.9251, "step": 1333 }, { "epoch": 0.41818181818181815, "grad_norm": 3.73274827003479, "learning_rate": 6.67e-07, "loss": 9.2519, "step": 1334 }, { "epoch": 0.41849529780564265, "grad_norm": 3.267875909805298, "learning_rate": 6.675000000000001e-07, "loss": 6.2191, "step": 1335 }, { "epoch": 0.4188087774294671, "grad_norm": 3.341024160385132, "learning_rate": 6.68e-07, "loss": 7.4184, "step": 1336 }, { "epoch": 0.41912225705329154, "grad_norm": 2.8991432189941406, "learning_rate": 6.685000000000001e-07, "loss": 7.4803, "step": 1337 }, { "epoch": 0.419435736677116, "grad_norm": 4.8145928382873535, "learning_rate": 6.690000000000001e-07, "loss": 10.2224, "step": 1338 }, { "epoch": 0.4197492163009404, "grad_norm": 3.761445999145508, "learning_rate": 6.695e-07, "loss": 8.4562, "step": 1339 }, { "epoch": 0.4200626959247649, "grad_norm": 3.892420768737793, "learning_rate": 6.7e-07, "loss": 9.1571, "step": 1340 }, { "epoch": 0.42037617554858936, "grad_norm": 5.118911266326904, "learning_rate": 6.705000000000001e-07, "loss": 12.1094, "step": 1341 }, { "epoch": 0.4206896551724138, "grad_norm": 3.2898852825164795, "learning_rate": 6.710000000000001e-07, "loss": 7.4492, "step": 1342 }, { "epoch": 0.42100313479623824, "grad_norm": 2.7074368000030518, "learning_rate": 6.715e-07, "loss": 6.0427, "step": 1343 }, { "epoch": 0.4213166144200627, "grad_norm": 4.675681114196777, "learning_rate": 6.72e-07, "loss": 13.983, "step": 1344 }, { "epoch": 0.4216300940438871, "grad_norm": 3.761418104171753, "learning_rate": 6.725000000000001e-07, "loss": 7.1159, "step": 1345 }, { "epoch": 0.4219435736677116, "grad_norm": 3.7398788928985596, "learning_rate": 6.730000000000001e-07, "loss": 8.9778, "step": 1346 }, { "epoch": 0.42225705329153607, "grad_norm": 4.614981651306152, "learning_rate": 6.735e-07, "loss": 11.8006, "step": 1347 }, { "epoch": 0.4225705329153605, "grad_norm": 4.014955043792725, "learning_rate": 6.74e-07, "loss": 7.9856, "step": 1348 }, { "epoch": 0.42288401253918495, "grad_norm": 4.2952470779418945, "learning_rate": 6.745000000000001e-07, "loss": 8.6141, "step": 1349 }, { "epoch": 0.4231974921630094, "grad_norm": 4.254025459289551, "learning_rate": 6.750000000000001e-07, "loss": 13.2271, "step": 1350 }, { "epoch": 0.42351097178683383, "grad_norm": 3.196575403213501, "learning_rate": 6.755e-07, "loss": 6.6306, "step": 1351 }, { "epoch": 0.42382445141065833, "grad_norm": 4.765541076660156, "learning_rate": 6.76e-07, "loss": 12.8829, "step": 1352 }, { "epoch": 0.4241379310344828, "grad_norm": 4.769258499145508, "learning_rate": 6.765000000000001e-07, "loss": 10.9679, "step": 1353 }, { "epoch": 0.4244514106583072, "grad_norm": 3.533536195755005, "learning_rate": 6.77e-07, "loss": 6.7036, "step": 1354 }, { "epoch": 0.42476489028213166, "grad_norm": 4.482390403747559, "learning_rate": 6.775000000000001e-07, "loss": 9.1064, "step": 1355 }, { "epoch": 0.4250783699059561, "grad_norm": 3.877317428588867, "learning_rate": 6.78e-07, "loss": 9.4161, "step": 1356 }, { "epoch": 0.42539184952978054, "grad_norm": 4.770163059234619, "learning_rate": 6.784999999999999e-07, "loss": 13.1058, "step": 1357 }, { "epoch": 0.42570532915360504, "grad_norm": 3.1010990142822266, "learning_rate": 6.790000000000001e-07, "loss": 7.6001, "step": 1358 }, { "epoch": 0.4260188087774295, "grad_norm": 4.240891456604004, "learning_rate": 6.795e-07, "loss": 10.6859, "step": 1359 }, { "epoch": 0.4263322884012539, "grad_norm": 3.357872486114502, "learning_rate": 6.800000000000001e-07, "loss": 9.1908, "step": 1360 }, { "epoch": 0.42664576802507836, "grad_norm": 4.015371322631836, "learning_rate": 6.805000000000001e-07, "loss": 7.1454, "step": 1361 }, { "epoch": 0.4269592476489028, "grad_norm": 3.248347520828247, "learning_rate": 6.81e-07, "loss": 7.6464, "step": 1362 }, { "epoch": 0.42727272727272725, "grad_norm": 3.975553512573242, "learning_rate": 6.815000000000001e-07, "loss": 9.1993, "step": 1363 }, { "epoch": 0.42758620689655175, "grad_norm": 5.927193641662598, "learning_rate": 6.82e-07, "loss": 9.7591, "step": 1364 }, { "epoch": 0.4278996865203762, "grad_norm": 3.059760808944702, "learning_rate": 6.825000000000001e-07, "loss": 6.5011, "step": 1365 }, { "epoch": 0.42821316614420063, "grad_norm": 3.896256923675537, "learning_rate": 6.830000000000001e-07, "loss": 7.2387, "step": 1366 }, { "epoch": 0.42852664576802507, "grad_norm": 3.3756370544433594, "learning_rate": 6.835e-07, "loss": 7.5067, "step": 1367 }, { "epoch": 0.4288401253918495, "grad_norm": 4.085315704345703, "learning_rate": 6.84e-07, "loss": 7.0529, "step": 1368 }, { "epoch": 0.42915360501567396, "grad_norm": 2.654603958129883, "learning_rate": 6.845000000000001e-07, "loss": 7.7151, "step": 1369 }, { "epoch": 0.42946708463949845, "grad_norm": 4.527529716491699, "learning_rate": 6.850000000000001e-07, "loss": 10.333, "step": 1370 }, { "epoch": 0.4297805642633229, "grad_norm": 3.930635690689087, "learning_rate": 6.855e-07, "loss": 9.8118, "step": 1371 }, { "epoch": 0.43009404388714734, "grad_norm": 4.306950092315674, "learning_rate": 6.86e-07, "loss": 9.8894, "step": 1372 }, { "epoch": 0.4304075235109718, "grad_norm": 3.4206173419952393, "learning_rate": 6.865000000000001e-07, "loss": 8.8873, "step": 1373 }, { "epoch": 0.4307210031347962, "grad_norm": 4.430827617645264, "learning_rate": 6.87e-07, "loss": 9.3459, "step": 1374 }, { "epoch": 0.43103448275862066, "grad_norm": 3.1537222862243652, "learning_rate": 6.875000000000001e-07, "loss": 7.5838, "step": 1375 }, { "epoch": 0.43134796238244516, "grad_norm": 4.20504093170166, "learning_rate": 6.88e-07, "loss": 12.7391, "step": 1376 }, { "epoch": 0.4316614420062696, "grad_norm": 3.653794050216675, "learning_rate": 6.885e-07, "loss": 8.7142, "step": 1377 }, { "epoch": 0.43197492163009404, "grad_norm": 4.225579261779785, "learning_rate": 6.890000000000001e-07, "loss": 10.9947, "step": 1378 }, { "epoch": 0.4322884012539185, "grad_norm": 3.8407626152038574, "learning_rate": 6.895e-07, "loss": 9.6915, "step": 1379 }, { "epoch": 0.43260188087774293, "grad_norm": 3.124218463897705, "learning_rate": 6.900000000000001e-07, "loss": 6.9618, "step": 1380 }, { "epoch": 0.4329153605015674, "grad_norm": 3.4793970584869385, "learning_rate": 6.905000000000001e-07, "loss": 10.846, "step": 1381 }, { "epoch": 0.43322884012539187, "grad_norm": 3.0507242679595947, "learning_rate": 6.91e-07, "loss": 8.2012, "step": 1382 }, { "epoch": 0.4335423197492163, "grad_norm": 3.7710156440734863, "learning_rate": 6.915000000000001e-07, "loss": 9.654, "step": 1383 }, { "epoch": 0.43385579937304075, "grad_norm": 3.3001532554626465, "learning_rate": 6.92e-07, "loss": 6.7221, "step": 1384 }, { "epoch": 0.4341692789968652, "grad_norm": 4.092366695404053, "learning_rate": 6.925000000000001e-07, "loss": 10.6304, "step": 1385 }, { "epoch": 0.43448275862068964, "grad_norm": 3.2531421184539795, "learning_rate": 6.930000000000001e-07, "loss": 7.9767, "step": 1386 }, { "epoch": 0.43479623824451413, "grad_norm": 3.08742094039917, "learning_rate": 6.935e-07, "loss": 7.9412, "step": 1387 }, { "epoch": 0.4351097178683386, "grad_norm": 4.16629695892334, "learning_rate": 6.94e-07, "loss": 7.9362, "step": 1388 }, { "epoch": 0.435423197492163, "grad_norm": 3.6127755641937256, "learning_rate": 6.945000000000001e-07, "loss": 8.0281, "step": 1389 }, { "epoch": 0.43573667711598746, "grad_norm": 3.736285924911499, "learning_rate": 6.950000000000001e-07, "loss": 9.2018, "step": 1390 }, { "epoch": 0.4360501567398119, "grad_norm": 5.327370643615723, "learning_rate": 6.955000000000001e-07, "loss": 10.193, "step": 1391 }, { "epoch": 0.43636363636363634, "grad_norm": 3.426323413848877, "learning_rate": 6.96e-07, "loss": 8.9392, "step": 1392 }, { "epoch": 0.43667711598746084, "grad_norm": 4.5006937980651855, "learning_rate": 6.965000000000001e-07, "loss": 11.3513, "step": 1393 }, { "epoch": 0.4369905956112853, "grad_norm": 3.4478158950805664, "learning_rate": 6.970000000000001e-07, "loss": 8.5876, "step": 1394 }, { "epoch": 0.4373040752351097, "grad_norm": 3.8687758445739746, "learning_rate": 6.975000000000001e-07, "loss": 8.731, "step": 1395 }, { "epoch": 0.43761755485893417, "grad_norm": 3.383852958679199, "learning_rate": 6.98e-07, "loss": 7.7361, "step": 1396 }, { "epoch": 0.4379310344827586, "grad_norm": 3.9315686225891113, "learning_rate": 6.985e-07, "loss": 9.827, "step": 1397 }, { "epoch": 0.43824451410658305, "grad_norm": 3.924696683883667, "learning_rate": 6.990000000000001e-07, "loss": 7.6312, "step": 1398 }, { "epoch": 0.43855799373040755, "grad_norm": 3.3615076541900635, "learning_rate": 6.995e-07, "loss": 7.7659, "step": 1399 }, { "epoch": 0.438871473354232, "grad_norm": 3.5409042835235596, "learning_rate": 7.000000000000001e-07, "loss": 6.6861, "step": 1400 }, { "epoch": 0.43918495297805643, "grad_norm": 3.4962658882141113, "learning_rate": 7.005000000000001e-07, "loss": 8.9106, "step": 1401 }, { "epoch": 0.4394984326018809, "grad_norm": 4.335007667541504, "learning_rate": 7.01e-07, "loss": 7.0246, "step": 1402 }, { "epoch": 0.4398119122257053, "grad_norm": 3.592299222946167, "learning_rate": 7.015000000000001e-07, "loss": 9.6732, "step": 1403 }, { "epoch": 0.44012539184952976, "grad_norm": 5.375201225280762, "learning_rate": 7.02e-07, "loss": 8.865, "step": 1404 }, { "epoch": 0.44043887147335425, "grad_norm": 5.488976955413818, "learning_rate": 7.025000000000002e-07, "loss": 14.8031, "step": 1405 }, { "epoch": 0.4407523510971787, "grad_norm": 3.3288614749908447, "learning_rate": 7.030000000000001e-07, "loss": 8.8429, "step": 1406 }, { "epoch": 0.44106583072100314, "grad_norm": 3.75602388381958, "learning_rate": 7.035e-07, "loss": 7.5947, "step": 1407 }, { "epoch": 0.4413793103448276, "grad_norm": 2.7701900005340576, "learning_rate": 7.040000000000001e-07, "loss": 7.4473, "step": 1408 }, { "epoch": 0.441692789968652, "grad_norm": 4.249903678894043, "learning_rate": 7.045000000000001e-07, "loss": 8.4978, "step": 1409 }, { "epoch": 0.44200626959247646, "grad_norm": 3.3370988368988037, "learning_rate": 7.05e-07, "loss": 6.707, "step": 1410 }, { "epoch": 0.44231974921630096, "grad_norm": 5.536358833312988, "learning_rate": 7.055000000000001e-07, "loss": 11.3998, "step": 1411 }, { "epoch": 0.4426332288401254, "grad_norm": 4.11740779876709, "learning_rate": 7.06e-07, "loss": 8.3269, "step": 1412 }, { "epoch": 0.44294670846394985, "grad_norm": 4.711095809936523, "learning_rate": 7.065000000000001e-07, "loss": 12.2608, "step": 1413 }, { "epoch": 0.4432601880877743, "grad_norm": 3.0188283920288086, "learning_rate": 7.070000000000001e-07, "loss": 7.8679, "step": 1414 }, { "epoch": 0.44357366771159873, "grad_norm": 3.6578001976013184, "learning_rate": 7.075e-07, "loss": 7.9547, "step": 1415 }, { "epoch": 0.44388714733542317, "grad_norm": 3.710144519805908, "learning_rate": 7.08e-07, "loss": 9.1333, "step": 1416 }, { "epoch": 0.44420062695924767, "grad_norm": 3.580639600753784, "learning_rate": 7.085e-07, "loss": 8.6172, "step": 1417 }, { "epoch": 0.4445141065830721, "grad_norm": 3.4201769828796387, "learning_rate": 7.090000000000001e-07, "loss": 7.38, "step": 1418 }, { "epoch": 0.44482758620689655, "grad_norm": 3.8317136764526367, "learning_rate": 7.095e-07, "loss": 7.1807, "step": 1419 }, { "epoch": 0.445141065830721, "grad_norm": 6.48163366317749, "learning_rate": 7.1e-07, "loss": 15.1615, "step": 1420 }, { "epoch": 0.44545454545454544, "grad_norm": 3.3979268074035645, "learning_rate": 7.105000000000001e-07, "loss": 6.6834, "step": 1421 }, { "epoch": 0.4457680250783699, "grad_norm": 3.5724081993103027, "learning_rate": 7.110000000000001e-07, "loss": 8.6715, "step": 1422 }, { "epoch": 0.4460815047021944, "grad_norm": 3.738381862640381, "learning_rate": 7.115000000000001e-07, "loss": 7.0441, "step": 1423 }, { "epoch": 0.4463949843260188, "grad_norm": 4.122878551483154, "learning_rate": 7.12e-07, "loss": 7.2443, "step": 1424 }, { "epoch": 0.44670846394984326, "grad_norm": 3.6147820949554443, "learning_rate": 7.125e-07, "loss": 7.8883, "step": 1425 }, { "epoch": 0.4470219435736677, "grad_norm": 3.86806058883667, "learning_rate": 7.130000000000001e-07, "loss": 8.0316, "step": 1426 }, { "epoch": 0.44733542319749214, "grad_norm": 4.159463882446289, "learning_rate": 7.135e-07, "loss": 7.2277, "step": 1427 }, { "epoch": 0.44764890282131664, "grad_norm": 2.5668580532073975, "learning_rate": 7.140000000000001e-07, "loss": 6.9613, "step": 1428 }, { "epoch": 0.4479623824451411, "grad_norm": 4.1336469650268555, "learning_rate": 7.145000000000001e-07, "loss": 9.4702, "step": 1429 }, { "epoch": 0.4482758620689655, "grad_norm": 6.645508289337158, "learning_rate": 7.15e-07, "loss": 15.8057, "step": 1430 }, { "epoch": 0.44858934169278997, "grad_norm": 4.979160785675049, "learning_rate": 7.155000000000001e-07, "loss": 15.3406, "step": 1431 }, { "epoch": 0.4489028213166144, "grad_norm": 3.6304969787597656, "learning_rate": 7.16e-07, "loss": 8.6424, "step": 1432 }, { "epoch": 0.44921630094043885, "grad_norm": 4.989882469177246, "learning_rate": 7.165000000000001e-07, "loss": 11.6095, "step": 1433 }, { "epoch": 0.44952978056426335, "grad_norm": 3.8582489490509033, "learning_rate": 7.170000000000001e-07, "loss": 8.5874, "step": 1434 }, { "epoch": 0.4498432601880878, "grad_norm": 4.933996200561523, "learning_rate": 7.175e-07, "loss": 11.0086, "step": 1435 }, { "epoch": 0.45015673981191223, "grad_norm": 4.105413913726807, "learning_rate": 7.18e-07, "loss": 10.1414, "step": 1436 }, { "epoch": 0.4504702194357367, "grad_norm": 4.856784820556641, "learning_rate": 7.185e-07, "loss": 10.7219, "step": 1437 }, { "epoch": 0.4507836990595611, "grad_norm": 3.4442543983459473, "learning_rate": 7.190000000000001e-07, "loss": 6.6924, "step": 1438 }, { "epoch": 0.45109717868338556, "grad_norm": 3.018460988998413, "learning_rate": 7.195000000000001e-07, "loss": 8.6876, "step": 1439 }, { "epoch": 0.45141065830721006, "grad_norm": 3.129737377166748, "learning_rate": 7.2e-07, "loss": 7.0111, "step": 1440 }, { "epoch": 0.4517241379310345, "grad_norm": 4.270967483520508, "learning_rate": 7.205000000000001e-07, "loss": 9.3915, "step": 1441 }, { "epoch": 0.45203761755485894, "grad_norm": 3.9212779998779297, "learning_rate": 7.210000000000001e-07, "loss": 10.3279, "step": 1442 }, { "epoch": 0.4523510971786834, "grad_norm": 3.4876608848571777, "learning_rate": 7.215000000000001e-07, "loss": 6.606, "step": 1443 }, { "epoch": 0.4526645768025078, "grad_norm": 3.4680469036102295, "learning_rate": 7.22e-07, "loss": 7.3311, "step": 1444 }, { "epoch": 0.45297805642633227, "grad_norm": 3.6926703453063965, "learning_rate": 7.225e-07, "loss": 7.5257, "step": 1445 }, { "epoch": 0.45329153605015676, "grad_norm": 6.416301727294922, "learning_rate": 7.230000000000001e-07, "loss": 14.6459, "step": 1446 }, { "epoch": 0.4536050156739812, "grad_norm": 4.126363277435303, "learning_rate": 7.235e-07, "loss": 8.369, "step": 1447 }, { "epoch": 0.45391849529780565, "grad_norm": 3.9787609577178955, "learning_rate": 7.240000000000001e-07, "loss": 8.115, "step": 1448 }, { "epoch": 0.4542319749216301, "grad_norm": 3.5446219444274902, "learning_rate": 7.245000000000001e-07, "loss": 7.3203, "step": 1449 }, { "epoch": 0.45454545454545453, "grad_norm": 3.7280256748199463, "learning_rate": 7.25e-07, "loss": 9.8608, "step": 1450 }, { "epoch": 0.454858934169279, "grad_norm": 3.1473164558410645, "learning_rate": 7.255000000000001e-07, "loss": 7.5801, "step": 1451 }, { "epoch": 0.45517241379310347, "grad_norm": 3.243011951446533, "learning_rate": 7.26e-07, "loss": 6.5251, "step": 1452 }, { "epoch": 0.4554858934169279, "grad_norm": 3.912731170654297, "learning_rate": 7.265000000000002e-07, "loss": 9.8108, "step": 1453 }, { "epoch": 0.45579937304075235, "grad_norm": 3.3148555755615234, "learning_rate": 7.270000000000001e-07, "loss": 9.7591, "step": 1454 }, { "epoch": 0.4561128526645768, "grad_norm": 3.5249342918395996, "learning_rate": 7.275e-07, "loss": 7.1523, "step": 1455 }, { "epoch": 0.45642633228840124, "grad_norm": 3.985668420791626, "learning_rate": 7.280000000000001e-07, "loss": 7.9672, "step": 1456 }, { "epoch": 0.4567398119122257, "grad_norm": 3.4252846240997314, "learning_rate": 7.285e-07, "loss": 8.2703, "step": 1457 }, { "epoch": 0.4570532915360502, "grad_norm": 5.43139123916626, "learning_rate": 7.290000000000001e-07, "loss": 9.2351, "step": 1458 }, { "epoch": 0.4573667711598746, "grad_norm": 5.333076000213623, "learning_rate": 7.295000000000001e-07, "loss": 13.8034, "step": 1459 }, { "epoch": 0.45768025078369906, "grad_norm": 3.3597097396850586, "learning_rate": 7.3e-07, "loss": 11.3477, "step": 1460 }, { "epoch": 0.4579937304075235, "grad_norm": 2.962592601776123, "learning_rate": 7.305000000000001e-07, "loss": 8.5858, "step": 1461 }, { "epoch": 0.45830721003134794, "grad_norm": 2.7471611499786377, "learning_rate": 7.310000000000001e-07, "loss": 6.0348, "step": 1462 }, { "epoch": 0.4586206896551724, "grad_norm": 4.322995662689209, "learning_rate": 7.315000000000001e-07, "loss": 9.9356, "step": 1463 }, { "epoch": 0.4589341692789969, "grad_norm": 3.186292886734009, "learning_rate": 7.32e-07, "loss": 8.0499, "step": 1464 }, { "epoch": 0.4592476489028213, "grad_norm": 2.747281789779663, "learning_rate": 7.325e-07, "loss": 6.545, "step": 1465 }, { "epoch": 0.45956112852664577, "grad_norm": 4.15464448928833, "learning_rate": 7.330000000000001e-07, "loss": 8.2943, "step": 1466 }, { "epoch": 0.4598746081504702, "grad_norm": 3.4695398807525635, "learning_rate": 7.335e-07, "loss": 7.7409, "step": 1467 }, { "epoch": 0.46018808777429465, "grad_norm": 3.8150601387023926, "learning_rate": 7.340000000000001e-07, "loss": 9.3436, "step": 1468 }, { "epoch": 0.46050156739811915, "grad_norm": 3.744868278503418, "learning_rate": 7.345000000000001e-07, "loss": 9.7803, "step": 1469 }, { "epoch": 0.4608150470219436, "grad_norm": 3.8790037631988525, "learning_rate": 7.350000000000001e-07, "loss": 8.0119, "step": 1470 }, { "epoch": 0.46112852664576803, "grad_norm": 3.8017265796661377, "learning_rate": 7.355000000000001e-07, "loss": 7.3153, "step": 1471 }, { "epoch": 0.4614420062695925, "grad_norm": 4.845677375793457, "learning_rate": 7.36e-07, "loss": 12.615, "step": 1472 }, { "epoch": 0.4617554858934169, "grad_norm": 4.192208766937256, "learning_rate": 7.365e-07, "loss": 7.3787, "step": 1473 }, { "epoch": 0.46206896551724136, "grad_norm": 4.240219593048096, "learning_rate": 7.370000000000001e-07, "loss": 8.5848, "step": 1474 }, { "epoch": 0.46238244514106586, "grad_norm": 4.029890537261963, "learning_rate": 7.375e-07, "loss": 9.3911, "step": 1475 }, { "epoch": 0.4626959247648903, "grad_norm": 3.331756591796875, "learning_rate": 7.380000000000001e-07, "loss": 6.3178, "step": 1476 }, { "epoch": 0.46300940438871474, "grad_norm": 3.7580671310424805, "learning_rate": 7.385e-07, "loss": 9.2023, "step": 1477 }, { "epoch": 0.4633228840125392, "grad_norm": 4.716964244842529, "learning_rate": 7.39e-07, "loss": 11.6383, "step": 1478 }, { "epoch": 0.4636363636363636, "grad_norm": 3.466538906097412, "learning_rate": 7.395000000000001e-07, "loss": 8.556, "step": 1479 }, { "epoch": 0.46394984326018807, "grad_norm": 3.618819236755371, "learning_rate": 7.4e-07, "loss": 8.6228, "step": 1480 }, { "epoch": 0.46426332288401256, "grad_norm": 3.9370975494384766, "learning_rate": 7.405000000000002e-07, "loss": 8.9446, "step": 1481 }, { "epoch": 0.464576802507837, "grad_norm": 3.3497865200042725, "learning_rate": 7.410000000000001e-07, "loss": 8.4896, "step": 1482 }, { "epoch": 0.46489028213166145, "grad_norm": 3.9014110565185547, "learning_rate": 7.415e-07, "loss": 9.6106, "step": 1483 }, { "epoch": 0.4652037617554859, "grad_norm": 3.2826168537139893, "learning_rate": 7.420000000000001e-07, "loss": 7.3699, "step": 1484 }, { "epoch": 0.46551724137931033, "grad_norm": 4.081614971160889, "learning_rate": 7.425e-07, "loss": 8.7689, "step": 1485 }, { "epoch": 0.4658307210031348, "grad_norm": 3.3145084381103516, "learning_rate": 7.430000000000001e-07, "loss": 7.7831, "step": 1486 }, { "epoch": 0.46614420062695927, "grad_norm": 5.257850170135498, "learning_rate": 7.435000000000001e-07, "loss": 14.4335, "step": 1487 }, { "epoch": 0.4664576802507837, "grad_norm": 3.631011486053467, "learning_rate": 7.44e-07, "loss": 8.0316, "step": 1488 }, { "epoch": 0.46677115987460815, "grad_norm": 3.9170570373535156, "learning_rate": 7.445000000000001e-07, "loss": 7.9331, "step": 1489 }, { "epoch": 0.4670846394984326, "grad_norm": 3.710536003112793, "learning_rate": 7.450000000000001e-07, "loss": 10.0156, "step": 1490 }, { "epoch": 0.46739811912225704, "grad_norm": 3.666468858718872, "learning_rate": 7.455000000000001e-07, "loss": 7.2938, "step": 1491 }, { "epoch": 0.4677115987460815, "grad_norm": 4.270216941833496, "learning_rate": 7.46e-07, "loss": 10.1365, "step": 1492 }, { "epoch": 0.468025078369906, "grad_norm": 4.941109657287598, "learning_rate": 7.465e-07, "loss": 10.8911, "step": 1493 }, { "epoch": 0.4683385579937304, "grad_norm": 3.449632167816162, "learning_rate": 7.470000000000001e-07, "loss": 8.6447, "step": 1494 }, { "epoch": 0.46865203761755486, "grad_norm": 2.9346587657928467, "learning_rate": 7.475e-07, "loss": 6.7992, "step": 1495 }, { "epoch": 0.4689655172413793, "grad_norm": 2.9172539710998535, "learning_rate": 7.480000000000001e-07, "loss": 7.0059, "step": 1496 }, { "epoch": 0.46927899686520375, "grad_norm": 3.648991346359253, "learning_rate": 7.485e-07, "loss": 8.9688, "step": 1497 }, { "epoch": 0.4695924764890282, "grad_norm": 3.0724036693573, "learning_rate": 7.49e-07, "loss": 7.4737, "step": 1498 }, { "epoch": 0.4699059561128527, "grad_norm": 3.7592673301696777, "learning_rate": 7.495000000000001e-07, "loss": 9.0526, "step": 1499 }, { "epoch": 0.4702194357366771, "grad_norm": 3.5673487186431885, "learning_rate": 7.5e-07, "loss": 6.967, "step": 1500 }, { "epoch": 0.47053291536050157, "grad_norm": 4.531203269958496, "learning_rate": 7.505000000000002e-07, "loss": 8.6275, "step": 1501 }, { "epoch": 0.470846394984326, "grad_norm": 4.227149486541748, "learning_rate": 7.510000000000001e-07, "loss": 10.3291, "step": 1502 }, { "epoch": 0.47115987460815045, "grad_norm": 3.532550096511841, "learning_rate": 7.515e-07, "loss": 9.2704, "step": 1503 }, { "epoch": 0.4714733542319749, "grad_norm": 3.5547232627868652, "learning_rate": 7.520000000000001e-07, "loss": 8.9358, "step": 1504 }, { "epoch": 0.4717868338557994, "grad_norm": 2.87225604057312, "learning_rate": 7.525e-07, "loss": 7.2077, "step": 1505 }, { "epoch": 0.47210031347962383, "grad_norm": 3.5084965229034424, "learning_rate": 7.530000000000001e-07, "loss": 7.6485, "step": 1506 }, { "epoch": 0.4724137931034483, "grad_norm": 4.405672550201416, "learning_rate": 7.535000000000001e-07, "loss": 9.7914, "step": 1507 }, { "epoch": 0.4727272727272727, "grad_norm": 3.6262218952178955, "learning_rate": 7.54e-07, "loss": 5.8048, "step": 1508 }, { "epoch": 0.47304075235109716, "grad_norm": 4.285321235656738, "learning_rate": 7.545000000000001e-07, "loss": 8.9892, "step": 1509 }, { "epoch": 0.47335423197492166, "grad_norm": 3.221437692642212, "learning_rate": 7.550000000000001e-07, "loss": 8.5896, "step": 1510 }, { "epoch": 0.4736677115987461, "grad_norm": 3.349771022796631, "learning_rate": 7.555000000000001e-07, "loss": 7.2775, "step": 1511 }, { "epoch": 0.47398119122257054, "grad_norm": 4.374917984008789, "learning_rate": 7.56e-07, "loss": 10.1504, "step": 1512 }, { "epoch": 0.474294670846395, "grad_norm": 4.312491416931152, "learning_rate": 7.565e-07, "loss": 9.6675, "step": 1513 }, { "epoch": 0.4746081504702194, "grad_norm": 4.160287380218506, "learning_rate": 7.570000000000001e-07, "loss": 7.7446, "step": 1514 }, { "epoch": 0.47492163009404387, "grad_norm": 4.5663909912109375, "learning_rate": 7.575000000000001e-07, "loss": 10.4192, "step": 1515 }, { "epoch": 0.47523510971786836, "grad_norm": 3.4202966690063477, "learning_rate": 7.580000000000001e-07, "loss": 7.1886, "step": 1516 }, { "epoch": 0.4755485893416928, "grad_norm": 4.065593719482422, "learning_rate": 7.585e-07, "loss": 8.5182, "step": 1517 }, { "epoch": 0.47586206896551725, "grad_norm": 4.027133464813232, "learning_rate": 7.590000000000001e-07, "loss": 8.4011, "step": 1518 }, { "epoch": 0.4761755485893417, "grad_norm": 3.671494722366333, "learning_rate": 7.595000000000001e-07, "loss": 7.6953, "step": 1519 }, { "epoch": 0.47648902821316613, "grad_norm": 4.113628387451172, "learning_rate": 7.6e-07, "loss": 9.0673, "step": 1520 }, { "epoch": 0.4768025078369906, "grad_norm": 4.1456756591796875, "learning_rate": 7.605000000000002e-07, "loss": 8.4153, "step": 1521 }, { "epoch": 0.47711598746081507, "grad_norm": 4.083249568939209, "learning_rate": 7.610000000000001e-07, "loss": 9.7206, "step": 1522 }, { "epoch": 0.4774294670846395, "grad_norm": 4.154261589050293, "learning_rate": 7.615e-07, "loss": 8.6472, "step": 1523 }, { "epoch": 0.47774294670846396, "grad_norm": 4.028384208679199, "learning_rate": 7.620000000000001e-07, "loss": 7.9002, "step": 1524 }, { "epoch": 0.4780564263322884, "grad_norm": 4.162931442260742, "learning_rate": 7.625e-07, "loss": 9.9488, "step": 1525 }, { "epoch": 0.47836990595611284, "grad_norm": 3.495849132537842, "learning_rate": 7.630000000000001e-07, "loss": 9.5747, "step": 1526 }, { "epoch": 0.4786833855799373, "grad_norm": 3.8134756088256836, "learning_rate": 7.635000000000001e-07, "loss": 8.6538, "step": 1527 }, { "epoch": 0.4789968652037618, "grad_norm": 3.9398064613342285, "learning_rate": 7.64e-07, "loss": 8.3008, "step": 1528 }, { "epoch": 0.4793103448275862, "grad_norm": 3.440537691116333, "learning_rate": 7.645000000000002e-07, "loss": 6.1765, "step": 1529 }, { "epoch": 0.47962382445141066, "grad_norm": 4.416642665863037, "learning_rate": 7.650000000000001e-07, "loss": 9.6123, "step": 1530 }, { "epoch": 0.4799373040752351, "grad_norm": 3.9015026092529297, "learning_rate": 7.655000000000001e-07, "loss": 8.8604, "step": 1531 }, { "epoch": 0.48025078369905955, "grad_norm": 3.634904623031616, "learning_rate": 7.660000000000001e-07, "loss": 8.6111, "step": 1532 }, { "epoch": 0.480564263322884, "grad_norm": 3.1789402961730957, "learning_rate": 7.665e-07, "loss": 8.7108, "step": 1533 }, { "epoch": 0.4808777429467085, "grad_norm": 3.556788206100464, "learning_rate": 7.670000000000001e-07, "loss": 7.9619, "step": 1534 }, { "epoch": 0.48119122257053293, "grad_norm": 3.4481894969940186, "learning_rate": 7.675000000000001e-07, "loss": 8.4755, "step": 1535 }, { "epoch": 0.48150470219435737, "grad_norm": 3.0415053367614746, "learning_rate": 7.68e-07, "loss": 6.9378, "step": 1536 }, { "epoch": 0.4818181818181818, "grad_norm": 3.7473058700561523, "learning_rate": 7.685e-07, "loss": 10.5121, "step": 1537 }, { "epoch": 0.48213166144200625, "grad_norm": 4.390431880950928, "learning_rate": 7.690000000000001e-07, "loss": 8.8402, "step": 1538 }, { "epoch": 0.4824451410658307, "grad_norm": 3.8549649715423584, "learning_rate": 7.695000000000001e-07, "loss": 8.73, "step": 1539 }, { "epoch": 0.4827586206896552, "grad_norm": 5.701721668243408, "learning_rate": 7.7e-07, "loss": 9.7011, "step": 1540 }, { "epoch": 0.48307210031347964, "grad_norm": 3.738424777984619, "learning_rate": 7.705e-07, "loss": 8.1286, "step": 1541 }, { "epoch": 0.4833855799373041, "grad_norm": 3.1514458656311035, "learning_rate": 7.710000000000001e-07, "loss": 7.7808, "step": 1542 }, { "epoch": 0.4836990595611285, "grad_norm": 2.9332237243652344, "learning_rate": 7.715e-07, "loss": 7.3622, "step": 1543 }, { "epoch": 0.48401253918495296, "grad_norm": 3.4852969646453857, "learning_rate": 7.720000000000001e-07, "loss": 7.6669, "step": 1544 }, { "epoch": 0.4843260188087774, "grad_norm": 7.102890968322754, "learning_rate": 7.725e-07, "loss": 7.0981, "step": 1545 }, { "epoch": 0.4846394984326019, "grad_norm": 3.0128979682922363, "learning_rate": 7.73e-07, "loss": 6.5679, "step": 1546 }, { "epoch": 0.48495297805642634, "grad_norm": 3.7915170192718506, "learning_rate": 7.735000000000001e-07, "loss": 7.2064, "step": 1547 }, { "epoch": 0.4852664576802508, "grad_norm": 2.735111713409424, "learning_rate": 7.74e-07, "loss": 6.365, "step": 1548 }, { "epoch": 0.4855799373040752, "grad_norm": 3.264190435409546, "learning_rate": 7.745000000000002e-07, "loss": 8.0042, "step": 1549 }, { "epoch": 0.48589341692789967, "grad_norm": 4.327232837677002, "learning_rate": 7.750000000000001e-07, "loss": 9.7193, "step": 1550 }, { "epoch": 0.4862068965517241, "grad_norm": 3.012519359588623, "learning_rate": 7.755e-07, "loss": 6.3156, "step": 1551 }, { "epoch": 0.4865203761755486, "grad_norm": 4.424871921539307, "learning_rate": 7.760000000000001e-07, "loss": 8.8488, "step": 1552 }, { "epoch": 0.48683385579937305, "grad_norm": 3.7431857585906982, "learning_rate": 7.765e-07, "loss": 8.0868, "step": 1553 }, { "epoch": 0.4871473354231975, "grad_norm": 3.6024186611175537, "learning_rate": 7.770000000000001e-07, "loss": 10.3315, "step": 1554 }, { "epoch": 0.48746081504702193, "grad_norm": 3.3676459789276123, "learning_rate": 7.775000000000001e-07, "loss": 6.5772, "step": 1555 }, { "epoch": 0.4877742946708464, "grad_norm": 4.312803745269775, "learning_rate": 7.78e-07, "loss": 8.3249, "step": 1556 }, { "epoch": 0.4880877742946709, "grad_norm": 4.562367916107178, "learning_rate": 7.785e-07, "loss": 7.975, "step": 1557 }, { "epoch": 0.4884012539184953, "grad_norm": 3.697195053100586, "learning_rate": 7.790000000000001e-07, "loss": 7.3428, "step": 1558 }, { "epoch": 0.48871473354231976, "grad_norm": 4.236298561096191, "learning_rate": 7.795000000000001e-07, "loss": 10.2044, "step": 1559 }, { "epoch": 0.4890282131661442, "grad_norm": 3.0572011470794678, "learning_rate": 7.8e-07, "loss": 6.726, "step": 1560 }, { "epoch": 0.48934169278996864, "grad_norm": 3.513315439224243, "learning_rate": 7.805e-07, "loss": 7.963, "step": 1561 }, { "epoch": 0.4896551724137931, "grad_norm": 4.3329548835754395, "learning_rate": 7.810000000000001e-07, "loss": 7.9515, "step": 1562 }, { "epoch": 0.4899686520376176, "grad_norm": 4.270005226135254, "learning_rate": 7.815000000000001e-07, "loss": 10.5315, "step": 1563 }, { "epoch": 0.490282131661442, "grad_norm": 4.627399921417236, "learning_rate": 7.820000000000001e-07, "loss": 7.8852, "step": 1564 }, { "epoch": 0.49059561128526646, "grad_norm": 3.8905036449432373, "learning_rate": 7.825e-07, "loss": 8.9447, "step": 1565 }, { "epoch": 0.4909090909090909, "grad_norm": 4.1380181312561035, "learning_rate": 7.83e-07, "loss": 9.1815, "step": 1566 }, { "epoch": 0.49122257053291535, "grad_norm": 3.6653265953063965, "learning_rate": 7.835000000000001e-07, "loss": 8.7808, "step": 1567 }, { "epoch": 0.4915360501567398, "grad_norm": 3.8347866535186768, "learning_rate": 7.84e-07, "loss": 8.8488, "step": 1568 }, { "epoch": 0.4918495297805643, "grad_norm": 3.9519078731536865, "learning_rate": 7.845000000000001e-07, "loss": 8.6361, "step": 1569 }, { "epoch": 0.49216300940438873, "grad_norm": 4.5167036056518555, "learning_rate": 7.850000000000001e-07, "loss": 12.1671, "step": 1570 }, { "epoch": 0.49247648902821317, "grad_norm": 4.719188690185547, "learning_rate": 7.855e-07, "loss": 8.6138, "step": 1571 }, { "epoch": 0.4927899686520376, "grad_norm": 4.4409284591674805, "learning_rate": 7.860000000000001e-07, "loss": 9.6423, "step": 1572 }, { "epoch": 0.49310344827586206, "grad_norm": 2.9862163066864014, "learning_rate": 7.865e-07, "loss": 6.1838, "step": 1573 }, { "epoch": 0.4934169278996865, "grad_norm": 4.9992146492004395, "learning_rate": 7.870000000000002e-07, "loss": 11.4322, "step": 1574 }, { "epoch": 0.493730407523511, "grad_norm": 4.074015140533447, "learning_rate": 7.875000000000001e-07, "loss": 9.6542, "step": 1575 }, { "epoch": 0.49404388714733544, "grad_norm": 3.1027674674987793, "learning_rate": 7.88e-07, "loss": 6.8902, "step": 1576 }, { "epoch": 0.4943573667711599, "grad_norm": 4.539718151092529, "learning_rate": 7.885e-07, "loss": 11.3049, "step": 1577 }, { "epoch": 0.4946708463949843, "grad_norm": 4.0925140380859375, "learning_rate": 7.890000000000001e-07, "loss": 8.0069, "step": 1578 }, { "epoch": 0.49498432601880876, "grad_norm": 4.172740936279297, "learning_rate": 7.895000000000001e-07, "loss": 7.8851, "step": 1579 }, { "epoch": 0.4952978056426332, "grad_norm": 3.7345142364501953, "learning_rate": 7.900000000000001e-07, "loss": 9.2792, "step": 1580 }, { "epoch": 0.4956112852664577, "grad_norm": 3.628817081451416, "learning_rate": 7.905e-07, "loss": 6.9987, "step": 1581 }, { "epoch": 0.49592476489028214, "grad_norm": 4.242038249969482, "learning_rate": 7.910000000000001e-07, "loss": 9.6288, "step": 1582 }, { "epoch": 0.4962382445141066, "grad_norm": 5.484152793884277, "learning_rate": 7.915000000000001e-07, "loss": 15.2138, "step": 1583 }, { "epoch": 0.496551724137931, "grad_norm": 4.404213905334473, "learning_rate": 7.920000000000001e-07, "loss": 8.7437, "step": 1584 }, { "epoch": 0.49686520376175547, "grad_norm": 3.855804920196533, "learning_rate": 7.925e-07, "loss": 8.6907, "step": 1585 }, { "epoch": 0.4971786833855799, "grad_norm": 6.219444274902344, "learning_rate": 7.93e-07, "loss": 10.0004, "step": 1586 }, { "epoch": 0.4974921630094044, "grad_norm": 5.004199981689453, "learning_rate": 7.935000000000001e-07, "loss": 12.7213, "step": 1587 }, { "epoch": 0.49780564263322885, "grad_norm": 4.671052932739258, "learning_rate": 7.94e-07, "loss": 11.9351, "step": 1588 }, { "epoch": 0.4981191222570533, "grad_norm": 4.2331438064575195, "learning_rate": 7.945000000000001e-07, "loss": 9.8539, "step": 1589 }, { "epoch": 0.49843260188087773, "grad_norm": 3.8226304054260254, "learning_rate": 7.950000000000001e-07, "loss": 8.1552, "step": 1590 }, { "epoch": 0.4987460815047022, "grad_norm": 4.827603340148926, "learning_rate": 7.955e-07, "loss": 11.9278, "step": 1591 }, { "epoch": 0.4990595611285266, "grad_norm": 4.008607864379883, "learning_rate": 7.960000000000001e-07, "loss": 9.1185, "step": 1592 }, { "epoch": 0.4993730407523511, "grad_norm": 3.332667112350464, "learning_rate": 7.965e-07, "loss": 7.4391, "step": 1593 }, { "epoch": 0.49968652037617556, "grad_norm": 4.7476677894592285, "learning_rate": 7.97e-07, "loss": 7.911, "step": 1594 }, { "epoch": 0.5, "grad_norm": 3.6339681148529053, "learning_rate": 7.975000000000001e-07, "loss": 8.406, "step": 1595 }, { "epoch": 0.5003134796238244, "grad_norm": 3.5920093059539795, "learning_rate": 7.98e-07, "loss": 7.1252, "step": 1596 }, { "epoch": 0.5003134796238244, "eval_loss": 28.69378662109375, "eval_runtime": 20.7573, "eval_samples_per_second": 129.449, "eval_steps_per_second": 8.094, "step": 1596 }, { "epoch": 0.5006269592476489, "grad_norm": 3.7933077812194824, "learning_rate": 7.985000000000001e-07, "loss": 7.7692, "step": 1597 }, { "epoch": 0.5009404388714733, "grad_norm": 3.132864475250244, "learning_rate": 7.990000000000001e-07, "loss": 7.6955, "step": 1598 }, { "epoch": 0.5012539184952978, "grad_norm": 3.2161154747009277, "learning_rate": 7.995e-07, "loss": 7.817, "step": 1599 }, { "epoch": 0.5015673981191222, "grad_norm": 3.270275115966797, "learning_rate": 8.000000000000001e-07, "loss": 8.0105, "step": 1600 }, { "epoch": 0.5018808777429468, "grad_norm": 3.4156415462493896, "learning_rate": 8.005e-07, "loss": 7.2844, "step": 1601 }, { "epoch": 0.5021943573667712, "grad_norm": 3.0608019828796387, "learning_rate": 8.010000000000001e-07, "loss": 7.7339, "step": 1602 }, { "epoch": 0.5025078369905956, "grad_norm": 3.434061050415039, "learning_rate": 8.015000000000001e-07, "loss": 8.1699, "step": 1603 }, { "epoch": 0.5028213166144201, "grad_norm": 3.004152536392212, "learning_rate": 8.02e-07, "loss": 6.8241, "step": 1604 }, { "epoch": 0.5031347962382445, "grad_norm": 3.3219809532165527, "learning_rate": 8.025e-07, "loss": 7.6745, "step": 1605 }, { "epoch": 0.503448275862069, "grad_norm": 4.926275730133057, "learning_rate": 8.03e-07, "loss": 11.5191, "step": 1606 }, { "epoch": 0.5037617554858934, "grad_norm": 3.7799763679504395, "learning_rate": 8.035000000000001e-07, "loss": 8.9354, "step": 1607 }, { "epoch": 0.5040752351097179, "grad_norm": 3.3203251361846924, "learning_rate": 8.04e-07, "loss": 7.2962, "step": 1608 }, { "epoch": 0.5043887147335423, "grad_norm": 4.6373419761657715, "learning_rate": 8.045e-07, "loss": 11.3006, "step": 1609 }, { "epoch": 0.5047021943573667, "grad_norm": 4.2450103759765625, "learning_rate": 8.050000000000001e-07, "loss": 7.3383, "step": 1610 }, { "epoch": 0.5050156739811912, "grad_norm": 3.985623359680176, "learning_rate": 8.055000000000001e-07, "loss": 8.2701, "step": 1611 }, { "epoch": 0.5053291536050156, "grad_norm": 3.601292848587036, "learning_rate": 8.060000000000001e-07, "loss": 8.3857, "step": 1612 }, { "epoch": 0.5056426332288402, "grad_norm": 5.563617706298828, "learning_rate": 8.065e-07, "loss": 14.2597, "step": 1613 }, { "epoch": 0.5059561128526646, "grad_norm": 4.677281856536865, "learning_rate": 8.07e-07, "loss": 9.532, "step": 1614 }, { "epoch": 0.5062695924764891, "grad_norm": 3.749307632446289, "learning_rate": 8.075000000000001e-07, "loss": 6.9486, "step": 1615 }, { "epoch": 0.5065830721003135, "grad_norm": 4.544729232788086, "learning_rate": 8.08e-07, "loss": 8.8941, "step": 1616 }, { "epoch": 0.506896551724138, "grad_norm": 3.591256856918335, "learning_rate": 8.085000000000001e-07, "loss": 8.471, "step": 1617 }, { "epoch": 0.5072100313479624, "grad_norm": 5.5906548500061035, "learning_rate": 8.090000000000001e-07, "loss": 10.5195, "step": 1618 }, { "epoch": 0.5075235109717868, "grad_norm": 4.689754009246826, "learning_rate": 8.095e-07, "loss": 8.9176, "step": 1619 }, { "epoch": 0.5078369905956113, "grad_norm": 4.00483512878418, "learning_rate": 8.100000000000001e-07, "loss": 8.361, "step": 1620 }, { "epoch": 0.5081504702194357, "grad_norm": 3.282796621322632, "learning_rate": 8.105e-07, "loss": 7.4275, "step": 1621 }, { "epoch": 0.5084639498432602, "grad_norm": 4.235138416290283, "learning_rate": 8.110000000000002e-07, "loss": 9.8603, "step": 1622 }, { "epoch": 0.5087774294670846, "grad_norm": 4.118296146392822, "learning_rate": 8.115000000000001e-07, "loss": 8.8349, "step": 1623 }, { "epoch": 0.509090909090909, "grad_norm": 3.592233180999756, "learning_rate": 8.12e-07, "loss": 9.0909, "step": 1624 }, { "epoch": 0.5094043887147336, "grad_norm": 3.8250372409820557, "learning_rate": 8.125000000000001e-07, "loss": 6.3841, "step": 1625 }, { "epoch": 0.509717868338558, "grad_norm": 3.962688446044922, "learning_rate": 8.13e-07, "loss": 7.9938, "step": 1626 }, { "epoch": 0.5100313479623825, "grad_norm": 4.621908664703369, "learning_rate": 8.135000000000001e-07, "loss": 9.8957, "step": 1627 }, { "epoch": 0.5103448275862069, "grad_norm": 3.104255199432373, "learning_rate": 8.140000000000001e-07, "loss": 6.7121, "step": 1628 }, { "epoch": 0.5106583072100314, "grad_norm": 4.97641134262085, "learning_rate": 8.145e-07, "loss": 6.1975, "step": 1629 }, { "epoch": 0.5109717868338558, "grad_norm": 4.67917013168335, "learning_rate": 8.150000000000001e-07, "loss": 13.774, "step": 1630 }, { "epoch": 0.5112852664576802, "grad_norm": 3.4128177165985107, "learning_rate": 8.155000000000001e-07, "loss": 8.6059, "step": 1631 }, { "epoch": 0.5115987460815047, "grad_norm": 4.750978469848633, "learning_rate": 8.160000000000001e-07, "loss": 11.319, "step": 1632 }, { "epoch": 0.5119122257053291, "grad_norm": 3.3610668182373047, "learning_rate": 8.165e-07, "loss": 6.3177, "step": 1633 }, { "epoch": 0.5122257053291536, "grad_norm": 4.7112956047058105, "learning_rate": 8.17e-07, "loss": 10.9001, "step": 1634 }, { "epoch": 0.512539184952978, "grad_norm": 4.46808385848999, "learning_rate": 8.175000000000001e-07, "loss": 12.3748, "step": 1635 }, { "epoch": 0.5128526645768025, "grad_norm": 3.6378445625305176, "learning_rate": 8.18e-07, "loss": 9.1623, "step": 1636 }, { "epoch": 0.513166144200627, "grad_norm": 4.917815685272217, "learning_rate": 8.185000000000001e-07, "loss": 11.5192, "step": 1637 }, { "epoch": 0.5134796238244514, "grad_norm": 3.1884193420410156, "learning_rate": 8.190000000000001e-07, "loss": 6.7961, "step": 1638 }, { "epoch": 0.5137931034482759, "grad_norm": 2.6258697509765625, "learning_rate": 8.195e-07, "loss": 6.3646, "step": 1639 }, { "epoch": 0.5141065830721003, "grad_norm": 4.169460296630859, "learning_rate": 8.200000000000001e-07, "loss": 9.467, "step": 1640 }, { "epoch": 0.5144200626959248, "grad_norm": 5.105269432067871, "learning_rate": 8.205e-07, "loss": 13.4161, "step": 1641 }, { "epoch": 0.5147335423197492, "grad_norm": 3.549649238586426, "learning_rate": 8.210000000000002e-07, "loss": 7.6424, "step": 1642 }, { "epoch": 0.5150470219435737, "grad_norm": 5.451201915740967, "learning_rate": 8.215000000000001e-07, "loss": 10.404, "step": 1643 }, { "epoch": 0.5153605015673981, "grad_norm": 4.793301582336426, "learning_rate": 8.22e-07, "loss": 10.6138, "step": 1644 }, { "epoch": 0.5156739811912225, "grad_norm": 4.120952129364014, "learning_rate": 8.225000000000001e-07, "loss": 9.8365, "step": 1645 }, { "epoch": 0.515987460815047, "grad_norm": 5.643743991851807, "learning_rate": 8.23e-07, "loss": 13.1447, "step": 1646 }, { "epoch": 0.5163009404388714, "grad_norm": 3.094728469848633, "learning_rate": 8.235000000000001e-07, "loss": 6.3482, "step": 1647 }, { "epoch": 0.516614420062696, "grad_norm": 5.247632026672363, "learning_rate": 8.240000000000001e-07, "loss": 12.6728, "step": 1648 }, { "epoch": 0.5169278996865204, "grad_norm": 4.007214069366455, "learning_rate": 8.245e-07, "loss": 9.3752, "step": 1649 }, { "epoch": 0.5172413793103449, "grad_norm": 3.778515577316284, "learning_rate": 8.250000000000001e-07, "loss": 8.7075, "step": 1650 }, { "epoch": 0.5175548589341693, "grad_norm": 2.9957754611968994, "learning_rate": 8.255000000000001e-07, "loss": 6.3369, "step": 1651 }, { "epoch": 0.5178683385579937, "grad_norm": 3.4570438861846924, "learning_rate": 8.260000000000001e-07, "loss": 7.8577, "step": 1652 }, { "epoch": 0.5181818181818182, "grad_norm": 3.390144109725952, "learning_rate": 8.265e-07, "loss": 8.0684, "step": 1653 }, { "epoch": 0.5184952978056426, "grad_norm": 3.808957815170288, "learning_rate": 8.27e-07, "loss": 8.0672, "step": 1654 }, { "epoch": 0.5188087774294671, "grad_norm": 3.636155843734741, "learning_rate": 8.275000000000001e-07, "loss": 9.3789, "step": 1655 }, { "epoch": 0.5191222570532915, "grad_norm": 4.033257007598877, "learning_rate": 8.280000000000001e-07, "loss": 8.0115, "step": 1656 }, { "epoch": 0.519435736677116, "grad_norm": 4.070474624633789, "learning_rate": 8.285e-07, "loss": 8.4374, "step": 1657 }, { "epoch": 0.5197492163009404, "grad_norm": 3.276840925216675, "learning_rate": 8.290000000000001e-07, "loss": 6.5463, "step": 1658 }, { "epoch": 0.5200626959247648, "grad_norm": 3.784796714782715, "learning_rate": 8.295000000000001e-07, "loss": 9.6518, "step": 1659 }, { "epoch": 0.5203761755485894, "grad_norm": 3.181030035018921, "learning_rate": 8.300000000000001e-07, "loss": 7.1291, "step": 1660 }, { "epoch": 0.5206896551724138, "grad_norm": 6.969316005706787, "learning_rate": 8.305e-07, "loss": 13.7243, "step": 1661 }, { "epoch": 0.5210031347962383, "grad_norm": 3.7494585514068604, "learning_rate": 8.31e-07, "loss": 7.9837, "step": 1662 }, { "epoch": 0.5213166144200627, "grad_norm": 4.4347991943359375, "learning_rate": 8.315000000000001e-07, "loss": 9.8557, "step": 1663 }, { "epoch": 0.5216300940438872, "grad_norm": 3.5605807304382324, "learning_rate": 8.32e-07, "loss": 8.4987, "step": 1664 }, { "epoch": 0.5219435736677116, "grad_norm": 3.945876359939575, "learning_rate": 8.325000000000001e-07, "loss": 9.7023, "step": 1665 }, { "epoch": 0.522257053291536, "grad_norm": 3.840888500213623, "learning_rate": 8.33e-07, "loss": 10.5396, "step": 1666 }, { "epoch": 0.5225705329153605, "grad_norm": 5.052422523498535, "learning_rate": 8.335e-07, "loss": 10.9094, "step": 1667 }, { "epoch": 0.5228840125391849, "grad_norm": 3.726701259613037, "learning_rate": 8.340000000000001e-07, "loss": 8.6335, "step": 1668 }, { "epoch": 0.5231974921630094, "grad_norm": 3.632563591003418, "learning_rate": 8.345e-07, "loss": 7.3737, "step": 1669 }, { "epoch": 0.5235109717868338, "grad_norm": 5.3117499351501465, "learning_rate": 8.350000000000002e-07, "loss": 9.7193, "step": 1670 }, { "epoch": 0.5238244514106583, "grad_norm": 4.223608493804932, "learning_rate": 8.355000000000001e-07, "loss": 8.8684, "step": 1671 }, { "epoch": 0.5241379310344828, "grad_norm": 4.2938151359558105, "learning_rate": 8.36e-07, "loss": 7.9856, "step": 1672 }, { "epoch": 0.5244514106583072, "grad_norm": 3.227950096130371, "learning_rate": 8.365000000000001e-07, "loss": 8.0511, "step": 1673 }, { "epoch": 0.5247648902821317, "grad_norm": 3.6287143230438232, "learning_rate": 8.37e-07, "loss": 6.8761, "step": 1674 }, { "epoch": 0.5250783699059561, "grad_norm": 3.5546576976776123, "learning_rate": 8.375000000000001e-07, "loss": 8.3474, "step": 1675 }, { "epoch": 0.5253918495297806, "grad_norm": 4.312641143798828, "learning_rate": 8.380000000000001e-07, "loss": 8.7992, "step": 1676 }, { "epoch": 0.525705329153605, "grad_norm": 4.029053688049316, "learning_rate": 8.385e-07, "loss": 8.269, "step": 1677 }, { "epoch": 0.5260188087774295, "grad_norm": 4.90485143661499, "learning_rate": 8.390000000000001e-07, "loss": 10.3717, "step": 1678 }, { "epoch": 0.5263322884012539, "grad_norm": 3.8778631687164307, "learning_rate": 8.395000000000001e-07, "loss": 11.3539, "step": 1679 }, { "epoch": 0.5266457680250783, "grad_norm": 4.429601669311523, "learning_rate": 8.400000000000001e-07, "loss": 10.0634, "step": 1680 }, { "epoch": 0.5269592476489028, "grad_norm": 5.119757175445557, "learning_rate": 8.405e-07, "loss": 10.9194, "step": 1681 }, { "epoch": 0.5272727272727272, "grad_norm": 8.258536338806152, "learning_rate": 8.41e-07, "loss": 22.845, "step": 1682 }, { "epoch": 0.5275862068965518, "grad_norm": 4.698957920074463, "learning_rate": 8.415000000000001e-07, "loss": 8.8281, "step": 1683 }, { "epoch": 0.5278996865203762, "grad_norm": 3.764852285385132, "learning_rate": 8.42e-07, "loss": 7.234, "step": 1684 }, { "epoch": 0.5282131661442007, "grad_norm": 6.866782188415527, "learning_rate": 8.425000000000001e-07, "loss": 9.7779, "step": 1685 }, { "epoch": 0.5285266457680251, "grad_norm": 4.264431476593018, "learning_rate": 8.43e-07, "loss": 10.7163, "step": 1686 }, { "epoch": 0.5288401253918495, "grad_norm": 5.4701995849609375, "learning_rate": 8.435000000000001e-07, "loss": 11.002, "step": 1687 }, { "epoch": 0.529153605015674, "grad_norm": 4.967133522033691, "learning_rate": 8.440000000000001e-07, "loss": 9.6692, "step": 1688 }, { "epoch": 0.5294670846394984, "grad_norm": 3.6275758743286133, "learning_rate": 8.445e-07, "loss": 6.5804, "step": 1689 }, { "epoch": 0.5297805642633229, "grad_norm": 4.6618170738220215, "learning_rate": 8.450000000000002e-07, "loss": 8.9379, "step": 1690 }, { "epoch": 0.5300940438871473, "grad_norm": 3.9030559062957764, "learning_rate": 8.455000000000001e-07, "loss": 7.8233, "step": 1691 }, { "epoch": 0.5304075235109718, "grad_norm": 3.4137344360351562, "learning_rate": 8.46e-07, "loss": 7.6667, "step": 1692 }, { "epoch": 0.5307210031347962, "grad_norm": 3.2711963653564453, "learning_rate": 8.465000000000001e-07, "loss": 7.6183, "step": 1693 }, { "epoch": 0.5310344827586206, "grad_norm": 5.124555587768555, "learning_rate": 8.47e-07, "loss": 16.8092, "step": 1694 }, { "epoch": 0.5313479623824452, "grad_norm": 3.800910711288452, "learning_rate": 8.475000000000001e-07, "loss": 8.8713, "step": 1695 }, { "epoch": 0.5316614420062696, "grad_norm": 4.457700252532959, "learning_rate": 8.480000000000001e-07, "loss": 8.7624, "step": 1696 }, { "epoch": 0.5319749216300941, "grad_norm": 3.4324142932891846, "learning_rate": 8.485e-07, "loss": 7.641, "step": 1697 }, { "epoch": 0.5322884012539185, "grad_norm": 3.523561477661133, "learning_rate": 8.490000000000002e-07, "loss": 7.0024, "step": 1698 }, { "epoch": 0.532601880877743, "grad_norm": 3.4999146461486816, "learning_rate": 8.495000000000001e-07, "loss": 8.2928, "step": 1699 }, { "epoch": 0.5329153605015674, "grad_norm": 4.3573222160339355, "learning_rate": 8.500000000000001e-07, "loss": 8.0377, "step": 1700 }, { "epoch": 0.5332288401253918, "grad_norm": 3.5811939239501953, "learning_rate": 8.505e-07, "loss": 7.2941, "step": 1701 }, { "epoch": 0.5335423197492163, "grad_norm": 3.876803398132324, "learning_rate": 8.51e-07, "loss": 9.4216, "step": 1702 }, { "epoch": 0.5338557993730407, "grad_norm": 2.872718572616577, "learning_rate": 8.515000000000001e-07, "loss": 6.6356, "step": 1703 }, { "epoch": 0.5341692789968652, "grad_norm": 4.38058614730835, "learning_rate": 8.520000000000001e-07, "loss": 9.6664, "step": 1704 }, { "epoch": 0.5344827586206896, "grad_norm": 4.397758960723877, "learning_rate": 8.525000000000001e-07, "loss": 10.8439, "step": 1705 }, { "epoch": 0.534796238244514, "grad_norm": 3.576199769973755, "learning_rate": 8.53e-07, "loss": 7.8423, "step": 1706 }, { "epoch": 0.5351097178683386, "grad_norm": 3.0761783123016357, "learning_rate": 8.535000000000001e-07, "loss": 7.3114, "step": 1707 }, { "epoch": 0.535423197492163, "grad_norm": 4.053654670715332, "learning_rate": 8.540000000000001e-07, "loss": 8.3691, "step": 1708 }, { "epoch": 0.5357366771159875, "grad_norm": 4.436123847961426, "learning_rate": 8.545e-07, "loss": 12.1063, "step": 1709 }, { "epoch": 0.5360501567398119, "grad_norm": 5.484961986541748, "learning_rate": 8.550000000000002e-07, "loss": 10.3859, "step": 1710 }, { "epoch": 0.5363636363636364, "grad_norm": 4.26839542388916, "learning_rate": 8.555000000000001e-07, "loss": 8.8806, "step": 1711 }, { "epoch": 0.5366771159874608, "grad_norm": 5.425708770751953, "learning_rate": 8.56e-07, "loss": 11.4141, "step": 1712 }, { "epoch": 0.5369905956112853, "grad_norm": 4.29015588760376, "learning_rate": 8.565000000000001e-07, "loss": 9.4032, "step": 1713 }, { "epoch": 0.5373040752351097, "grad_norm": 3.599637508392334, "learning_rate": 8.57e-07, "loss": 8.3692, "step": 1714 }, { "epoch": 0.5376175548589341, "grad_norm": 4.9885358810424805, "learning_rate": 8.575000000000002e-07, "loss": 9.0797, "step": 1715 }, { "epoch": 0.5379310344827586, "grad_norm": 3.783560037612915, "learning_rate": 8.580000000000001e-07, "loss": 7.1796, "step": 1716 }, { "epoch": 0.538244514106583, "grad_norm": 8.088088035583496, "learning_rate": 8.585e-07, "loss": 19.7025, "step": 1717 }, { "epoch": 0.5385579937304075, "grad_norm": 2.911607503890991, "learning_rate": 8.590000000000002e-07, "loss": 6.2524, "step": 1718 }, { "epoch": 0.538871473354232, "grad_norm": 3.7300970554351807, "learning_rate": 8.595000000000001e-07, "loss": 8.4442, "step": 1719 }, { "epoch": 0.5391849529780565, "grad_norm": 3.8973944187164307, "learning_rate": 8.6e-07, "loss": 9.406, "step": 1720 }, { "epoch": 0.5394984326018809, "grad_norm": 4.2937397956848145, "learning_rate": 8.605000000000001e-07, "loss": 6.9903, "step": 1721 }, { "epoch": 0.5398119122257053, "grad_norm": 4.311274528503418, "learning_rate": 8.61e-07, "loss": 7.9766, "step": 1722 }, { "epoch": 0.5401253918495298, "grad_norm": 3.7893688678741455, "learning_rate": 8.615000000000001e-07, "loss": 8.0692, "step": 1723 }, { "epoch": 0.5404388714733542, "grad_norm": 3.4368412494659424, "learning_rate": 8.620000000000001e-07, "loss": 7.6627, "step": 1724 }, { "epoch": 0.5407523510971787, "grad_norm": 4.653461933135986, "learning_rate": 8.625e-07, "loss": 8.2418, "step": 1725 }, { "epoch": 0.5410658307210031, "grad_norm": 3.554748058319092, "learning_rate": 8.63e-07, "loss": 7.7693, "step": 1726 }, { "epoch": 0.5413793103448276, "grad_norm": 3.596076726913452, "learning_rate": 8.635000000000001e-07, "loss": 11.7324, "step": 1727 }, { "epoch": 0.541692789968652, "grad_norm": 5.29031229019165, "learning_rate": 8.640000000000001e-07, "loss": 10.1403, "step": 1728 }, { "epoch": 0.5420062695924764, "grad_norm": 4.6957807540893555, "learning_rate": 8.645e-07, "loss": 10.0583, "step": 1729 }, { "epoch": 0.542319749216301, "grad_norm": 3.0998942852020264, "learning_rate": 8.65e-07, "loss": 7.5078, "step": 1730 }, { "epoch": 0.5426332288401254, "grad_norm": 6.108078479766846, "learning_rate": 8.655000000000001e-07, "loss": 15.5995, "step": 1731 }, { "epoch": 0.5429467084639499, "grad_norm": 4.639706611633301, "learning_rate": 8.66e-07, "loss": 8.0792, "step": 1732 }, { "epoch": 0.5432601880877743, "grad_norm": 3.6627628803253174, "learning_rate": 8.665000000000001e-07, "loss": 8.4311, "step": 1733 }, { "epoch": 0.5435736677115988, "grad_norm": 3.9771621227264404, "learning_rate": 8.67e-07, "loss": 8.8372, "step": 1734 }, { "epoch": 0.5438871473354232, "grad_norm": 3.9450392723083496, "learning_rate": 8.675000000000001e-07, "loss": 10.2992, "step": 1735 }, { "epoch": 0.5442006269592476, "grad_norm": 3.019969940185547, "learning_rate": 8.680000000000001e-07, "loss": 7.7473, "step": 1736 }, { "epoch": 0.5445141065830721, "grad_norm": 5.00993013381958, "learning_rate": 8.685e-07, "loss": 10.7672, "step": 1737 }, { "epoch": 0.5448275862068965, "grad_norm": 4.663172245025635, "learning_rate": 8.690000000000002e-07, "loss": 8.9194, "step": 1738 }, { "epoch": 0.545141065830721, "grad_norm": 5.089305400848389, "learning_rate": 8.695000000000001e-07, "loss": 10.1471, "step": 1739 }, { "epoch": 0.5454545454545454, "grad_norm": 8.746878623962402, "learning_rate": 8.7e-07, "loss": 11.3008, "step": 1740 }, { "epoch": 0.5457680250783699, "grad_norm": 3.6730334758758545, "learning_rate": 8.705000000000001e-07, "loss": 9.8861, "step": 1741 }, { "epoch": 0.5460815047021944, "grad_norm": 3.33270263671875, "learning_rate": 8.71e-07, "loss": 8.7378, "step": 1742 }, { "epoch": 0.5463949843260189, "grad_norm": 3.6083526611328125, "learning_rate": 8.715000000000001e-07, "loss": 7.0979, "step": 1743 }, { "epoch": 0.5467084639498433, "grad_norm": 4.226643085479736, "learning_rate": 8.720000000000001e-07, "loss": 10.3873, "step": 1744 }, { "epoch": 0.5470219435736677, "grad_norm": 5.096066951751709, "learning_rate": 8.725e-07, "loss": 10.6816, "step": 1745 }, { "epoch": 0.5473354231974922, "grad_norm": 5.229349613189697, "learning_rate": 8.73e-07, "loss": 11.0695, "step": 1746 }, { "epoch": 0.5476489028213166, "grad_norm": 3.8301353454589844, "learning_rate": 8.735000000000001e-07, "loss": 8.6425, "step": 1747 }, { "epoch": 0.5479623824451411, "grad_norm": 3.517941951751709, "learning_rate": 8.740000000000001e-07, "loss": 6.2092, "step": 1748 }, { "epoch": 0.5482758620689655, "grad_norm": 5.241222858428955, "learning_rate": 8.745000000000001e-07, "loss": 11.7764, "step": 1749 }, { "epoch": 0.54858934169279, "grad_norm": 5.1557793617248535, "learning_rate": 8.75e-07, "loss": 13.8192, "step": 1750 }, { "epoch": 0.5489028213166144, "grad_norm": 4.522706508636475, "learning_rate": 8.755000000000001e-07, "loss": 9.2798, "step": 1751 }, { "epoch": 0.5492163009404388, "grad_norm": 5.382571697235107, "learning_rate": 8.760000000000001e-07, "loss": 11.1607, "step": 1752 }, { "epoch": 0.5495297805642633, "grad_norm": 3.164684295654297, "learning_rate": 8.765000000000001e-07, "loss": 6.6911, "step": 1753 }, { "epoch": 0.5498432601880878, "grad_norm": 3.7033884525299072, "learning_rate": 8.77e-07, "loss": 7.5505, "step": 1754 }, { "epoch": 0.5501567398119123, "grad_norm": 4.4705424308776855, "learning_rate": 8.775000000000001e-07, "loss": 10.6266, "step": 1755 }, { "epoch": 0.5504702194357367, "grad_norm": 3.576754331588745, "learning_rate": 8.780000000000001e-07, "loss": 7.6148, "step": 1756 }, { "epoch": 0.5507836990595611, "grad_norm": 6.662974834442139, "learning_rate": 8.785e-07, "loss": 15.6583, "step": 1757 }, { "epoch": 0.5510971786833856, "grad_norm": 4.8018364906311035, "learning_rate": 8.790000000000002e-07, "loss": 7.7867, "step": 1758 }, { "epoch": 0.55141065830721, "grad_norm": 3.5706562995910645, "learning_rate": 8.795000000000001e-07, "loss": 7.1267, "step": 1759 }, { "epoch": 0.5517241379310345, "grad_norm": 3.8161635398864746, "learning_rate": 8.8e-07, "loss": 8.0858, "step": 1760 }, { "epoch": 0.5520376175548589, "grad_norm": 4.2101898193359375, "learning_rate": 8.805000000000001e-07, "loss": 8.4719, "step": 1761 }, { "epoch": 0.5523510971786834, "grad_norm": 3.5980849266052246, "learning_rate": 8.81e-07, "loss": 8.5086, "step": 1762 }, { "epoch": 0.5526645768025078, "grad_norm": 4.349874973297119, "learning_rate": 8.815000000000002e-07, "loss": 9.6007, "step": 1763 }, { "epoch": 0.5529780564263322, "grad_norm": 4.534544467926025, "learning_rate": 8.820000000000001e-07, "loss": 9.1542, "step": 1764 }, { "epoch": 0.5532915360501567, "grad_norm": 5.753028392791748, "learning_rate": 8.825e-07, "loss": 13.3121, "step": 1765 }, { "epoch": 0.5536050156739812, "grad_norm": 4.646644592285156, "learning_rate": 8.830000000000001e-07, "loss": 10.1693, "step": 1766 }, { "epoch": 0.5539184952978057, "grad_norm": 4.152806758880615, "learning_rate": 8.835000000000001e-07, "loss": 9.3472, "step": 1767 }, { "epoch": 0.5542319749216301, "grad_norm": 4.156120300292969, "learning_rate": 8.840000000000001e-07, "loss": 8.9546, "step": 1768 }, { "epoch": 0.5545454545454546, "grad_norm": 4.695617198944092, "learning_rate": 8.845000000000001e-07, "loss": 9.1984, "step": 1769 }, { "epoch": 0.554858934169279, "grad_norm": 3.514662742614746, "learning_rate": 8.85e-07, "loss": 6.9892, "step": 1770 }, { "epoch": 0.5551724137931034, "grad_norm": 6.180057048797607, "learning_rate": 8.855000000000001e-07, "loss": 15.8813, "step": 1771 }, { "epoch": 0.5554858934169279, "grad_norm": 3.279160261154175, "learning_rate": 8.860000000000001e-07, "loss": 6.86, "step": 1772 }, { "epoch": 0.5557993730407523, "grad_norm": 4.369133472442627, "learning_rate": 8.865000000000001e-07, "loss": 9.133, "step": 1773 }, { "epoch": 0.5561128526645768, "grad_norm": 3.11153507232666, "learning_rate": 8.87e-07, "loss": 7.517, "step": 1774 }, { "epoch": 0.5564263322884012, "grad_norm": 4.086194038391113, "learning_rate": 8.875000000000001e-07, "loss": 8.7146, "step": 1775 }, { "epoch": 0.5567398119122257, "grad_norm": 4.669196605682373, "learning_rate": 8.880000000000001e-07, "loss": 9.4316, "step": 1776 }, { "epoch": 0.5570532915360502, "grad_norm": 4.278200626373291, "learning_rate": 8.885e-07, "loss": 9.8683, "step": 1777 }, { "epoch": 0.5573667711598747, "grad_norm": 6.216943740844727, "learning_rate": 8.890000000000002e-07, "loss": 12.5523, "step": 1778 }, { "epoch": 0.5576802507836991, "grad_norm": 4.141219139099121, "learning_rate": 8.895000000000001e-07, "loss": 9.5758, "step": 1779 }, { "epoch": 0.5579937304075235, "grad_norm": 5.539915561676025, "learning_rate": 8.900000000000001e-07, "loss": 13.5515, "step": 1780 }, { "epoch": 0.558307210031348, "grad_norm": 4.460906505584717, "learning_rate": 8.905000000000001e-07, "loss": 9.7846, "step": 1781 }, { "epoch": 0.5586206896551724, "grad_norm": 3.381852388381958, "learning_rate": 8.91e-07, "loss": 7.3451, "step": 1782 }, { "epoch": 0.5589341692789969, "grad_norm": 4.935987949371338, "learning_rate": 8.915e-07, "loss": 9.0696, "step": 1783 }, { "epoch": 0.5592476489028213, "grad_norm": 4.209127426147461, "learning_rate": 8.920000000000001e-07, "loss": 8.1537, "step": 1784 }, { "epoch": 0.5595611285266457, "grad_norm": 3.22871994972229, "learning_rate": 8.925e-07, "loss": 7.8978, "step": 1785 }, { "epoch": 0.5598746081504702, "grad_norm": 4.8778510093688965, "learning_rate": 8.930000000000001e-07, "loss": 9.9476, "step": 1786 }, { "epoch": 0.5601880877742946, "grad_norm": 4.3461384773254395, "learning_rate": 8.935000000000001e-07, "loss": 8.4605, "step": 1787 }, { "epoch": 0.5605015673981191, "grad_norm": 4.09140157699585, "learning_rate": 8.94e-07, "loss": 8.6371, "step": 1788 }, { "epoch": 0.5608150470219436, "grad_norm": 3.154021739959717, "learning_rate": 8.945000000000001e-07, "loss": 6.6153, "step": 1789 }, { "epoch": 0.5611285266457681, "grad_norm": 5.04860782623291, "learning_rate": 8.95e-07, "loss": 10.1237, "step": 1790 }, { "epoch": 0.5614420062695925, "grad_norm": 3.98972487449646, "learning_rate": 8.955000000000002e-07, "loss": 7.8599, "step": 1791 }, { "epoch": 0.561755485893417, "grad_norm": 3.527597665786743, "learning_rate": 8.960000000000001e-07, "loss": 8.2753, "step": 1792 }, { "epoch": 0.5620689655172414, "grad_norm": 6.0412139892578125, "learning_rate": 8.965e-07, "loss": 13.1168, "step": 1793 }, { "epoch": 0.5623824451410658, "grad_norm": 4.94082498550415, "learning_rate": 8.97e-07, "loss": 10.4559, "step": 1794 }, { "epoch": 0.5626959247648903, "grad_norm": 4.965950012207031, "learning_rate": 8.975000000000001e-07, "loss": 10.237, "step": 1795 }, { "epoch": 0.5630094043887147, "grad_norm": 4.55895471572876, "learning_rate": 8.980000000000001e-07, "loss": 10.7508, "step": 1796 }, { "epoch": 0.5633228840125392, "grad_norm": 4.183038234710693, "learning_rate": 8.985000000000001e-07, "loss": 7.4745, "step": 1797 }, { "epoch": 0.5636363636363636, "grad_norm": 3.6337876319885254, "learning_rate": 8.99e-07, "loss": 6.8318, "step": 1798 }, { "epoch": 0.563949843260188, "grad_norm": 5.689560890197754, "learning_rate": 8.995000000000001e-07, "loss": 12.2413, "step": 1799 }, { "epoch": 0.5642633228840125, "grad_norm": 3.8097832202911377, "learning_rate": 9.000000000000001e-07, "loss": 7.0029, "step": 1800 }, { "epoch": 0.564576802507837, "grad_norm": 4.039278030395508, "learning_rate": 9.005000000000001e-07, "loss": 9.2906, "step": 1801 }, { "epoch": 0.5648902821316615, "grad_norm": 3.7648391723632812, "learning_rate": 9.01e-07, "loss": 7.9313, "step": 1802 }, { "epoch": 0.5652037617554859, "grad_norm": 4.256327152252197, "learning_rate": 9.015e-07, "loss": 6.9056, "step": 1803 }, { "epoch": 0.5655172413793104, "grad_norm": 4.904327869415283, "learning_rate": 9.020000000000001e-07, "loss": 9.3173, "step": 1804 }, { "epoch": 0.5658307210031348, "grad_norm": 3.761366367340088, "learning_rate": 9.025e-07, "loss": 10.9158, "step": 1805 }, { "epoch": 0.5661442006269592, "grad_norm": 4.512961387634277, "learning_rate": 9.030000000000001e-07, "loss": 9.3172, "step": 1806 }, { "epoch": 0.5664576802507837, "grad_norm": 4.454610824584961, "learning_rate": 9.035000000000001e-07, "loss": 9.806, "step": 1807 }, { "epoch": 0.5667711598746081, "grad_norm": 3.878605842590332, "learning_rate": 9.04e-07, "loss": 7.8319, "step": 1808 }, { "epoch": 0.5670846394984326, "grad_norm": 3.3651275634765625, "learning_rate": 9.045000000000001e-07, "loss": 7.7877, "step": 1809 }, { "epoch": 0.567398119122257, "grad_norm": 4.225704669952393, "learning_rate": 9.05e-07, "loss": 8.8007, "step": 1810 }, { "epoch": 0.5677115987460815, "grad_norm": 4.0525617599487305, "learning_rate": 9.055000000000002e-07, "loss": 8.9639, "step": 1811 }, { "epoch": 0.568025078369906, "grad_norm": 3.8242380619049072, "learning_rate": 9.060000000000001e-07, "loss": 8.4361, "step": 1812 }, { "epoch": 0.5683385579937305, "grad_norm": 6.60628080368042, "learning_rate": 9.065e-07, "loss": 9.6387, "step": 1813 }, { "epoch": 0.5686520376175549, "grad_norm": 4.6100616455078125, "learning_rate": 9.070000000000001e-07, "loss": 8.7753, "step": 1814 }, { "epoch": 0.5689655172413793, "grad_norm": 3.9851086139678955, "learning_rate": 9.075000000000001e-07, "loss": 7.7566, "step": 1815 }, { "epoch": 0.5692789968652038, "grad_norm": 5.08043098449707, "learning_rate": 9.080000000000001e-07, "loss": 12.0445, "step": 1816 }, { "epoch": 0.5695924764890282, "grad_norm": 8.151008605957031, "learning_rate": 9.085000000000001e-07, "loss": 12.3116, "step": 1817 }, { "epoch": 0.5699059561128527, "grad_norm": 5.475290775299072, "learning_rate": 9.09e-07, "loss": 12.5451, "step": 1818 }, { "epoch": 0.5702194357366771, "grad_norm": 3.203784227371216, "learning_rate": 9.095000000000001e-07, "loss": 6.5738, "step": 1819 }, { "epoch": 0.5705329153605015, "grad_norm": 3.322894334793091, "learning_rate": 9.100000000000001e-07, "loss": 7.036, "step": 1820 }, { "epoch": 0.570846394984326, "grad_norm": 5.2033772468566895, "learning_rate": 9.105000000000001e-07, "loss": 11.3802, "step": 1821 }, { "epoch": 0.5711598746081504, "grad_norm": 3.5129075050354004, "learning_rate": 9.11e-07, "loss": 5.7609, "step": 1822 }, { "epoch": 0.5714733542319749, "grad_norm": 3.822200059890747, "learning_rate": 9.115e-07, "loss": 8.4694, "step": 1823 }, { "epoch": 0.5717868338557994, "grad_norm": 5.213927268981934, "learning_rate": 9.120000000000001e-07, "loss": 10.8033, "step": 1824 }, { "epoch": 0.5721003134796239, "grad_norm": 3.970292806625366, "learning_rate": 9.125e-07, "loss": 8.1297, "step": 1825 }, { "epoch": 0.5724137931034483, "grad_norm": 3.255862236022949, "learning_rate": 9.130000000000001e-07, "loss": 6.6128, "step": 1826 }, { "epoch": 0.5727272727272728, "grad_norm": 5.0875139236450195, "learning_rate": 9.135000000000001e-07, "loss": 10.7206, "step": 1827 }, { "epoch": 0.5730407523510972, "grad_norm": 4.447558403015137, "learning_rate": 9.140000000000001e-07, "loss": 8.3421, "step": 1828 }, { "epoch": 0.5733542319749216, "grad_norm": 4.044737815856934, "learning_rate": 9.145000000000001e-07, "loss": 7.2425, "step": 1829 }, { "epoch": 0.5736677115987461, "grad_norm": 2.927945852279663, "learning_rate": 9.15e-07, "loss": 6.6691, "step": 1830 }, { "epoch": 0.5739811912225705, "grad_norm": 3.3857815265655518, "learning_rate": 9.155000000000002e-07, "loss": 6.6235, "step": 1831 }, { "epoch": 0.574294670846395, "grad_norm": 4.590656757354736, "learning_rate": 9.160000000000001e-07, "loss": 11.8911, "step": 1832 }, { "epoch": 0.5746081504702194, "grad_norm": 4.849060535430908, "learning_rate": 9.165e-07, "loss": 9.1651, "step": 1833 }, { "epoch": 0.5749216300940438, "grad_norm": 3.661118984222412, "learning_rate": 9.170000000000001e-07, "loss": 8.3145, "step": 1834 }, { "epoch": 0.5752351097178683, "grad_norm": 4.550508975982666, "learning_rate": 9.175000000000001e-07, "loss": 10.9553, "step": 1835 }, { "epoch": 0.5755485893416928, "grad_norm": 4.856189250946045, "learning_rate": 9.180000000000001e-07, "loss": 10.9995, "step": 1836 }, { "epoch": 0.5758620689655173, "grad_norm": 3.962833881378174, "learning_rate": 9.185000000000001e-07, "loss": 7.6791, "step": 1837 }, { "epoch": 0.5761755485893417, "grad_norm": 3.818290948867798, "learning_rate": 9.19e-07, "loss": 9.4411, "step": 1838 }, { "epoch": 0.5764890282131662, "grad_norm": 4.265783309936523, "learning_rate": 9.195000000000002e-07, "loss": 8.6175, "step": 1839 }, { "epoch": 0.5768025078369906, "grad_norm": 4.399896621704102, "learning_rate": 9.200000000000001e-07, "loss": 8.0593, "step": 1840 }, { "epoch": 0.577115987460815, "grad_norm": 3.487797975540161, "learning_rate": 9.205000000000001e-07, "loss": 7.8994, "step": 1841 }, { "epoch": 0.5774294670846395, "grad_norm": 4.428309440612793, "learning_rate": 9.210000000000001e-07, "loss": 7.2103, "step": 1842 }, { "epoch": 0.5777429467084639, "grad_norm": 4.569345951080322, "learning_rate": 9.215e-07, "loss": 9.6218, "step": 1843 }, { "epoch": 0.5780564263322884, "grad_norm": 3.990476369857788, "learning_rate": 9.220000000000001e-07, "loss": 9.4368, "step": 1844 }, { "epoch": 0.5783699059561128, "grad_norm": 3.2364110946655273, "learning_rate": 9.225000000000001e-07, "loss": 6.377, "step": 1845 }, { "epoch": 0.5786833855799373, "grad_norm": 4.474930763244629, "learning_rate": 9.23e-07, "loss": 8.6538, "step": 1846 }, { "epoch": 0.5789968652037617, "grad_norm": 4.316197872161865, "learning_rate": 9.235000000000001e-07, "loss": 9.8697, "step": 1847 }, { "epoch": 0.5793103448275863, "grad_norm": 3.6721043586730957, "learning_rate": 9.240000000000001e-07, "loss": 7.6369, "step": 1848 }, { "epoch": 0.5796238244514107, "grad_norm": 4.066787242889404, "learning_rate": 9.245000000000001e-07, "loss": 8.1153, "step": 1849 }, { "epoch": 0.5799373040752351, "grad_norm": 4.483874320983887, "learning_rate": 9.25e-07, "loss": 7.9151, "step": 1850 }, { "epoch": 0.5802507836990596, "grad_norm": 6.05754280090332, "learning_rate": 9.255e-07, "loss": 11.5638, "step": 1851 }, { "epoch": 0.580564263322884, "grad_norm": 4.692145824432373, "learning_rate": 9.260000000000001e-07, "loss": 9.0891, "step": 1852 }, { "epoch": 0.5808777429467085, "grad_norm": 4.905162334442139, "learning_rate": 9.265e-07, "loss": 11.009, "step": 1853 }, { "epoch": 0.5811912225705329, "grad_norm": 3.9991772174835205, "learning_rate": 9.270000000000001e-07, "loss": 7.1689, "step": 1854 }, { "epoch": 0.5815047021943573, "grad_norm": 3.545417308807373, "learning_rate": 9.275000000000001e-07, "loss": 8.8178, "step": 1855 }, { "epoch": 0.5818181818181818, "grad_norm": 3.4547829627990723, "learning_rate": 9.28e-07, "loss": 6.4124, "step": 1856 }, { "epoch": 0.5821316614420062, "grad_norm": 5.388890266418457, "learning_rate": 9.285000000000001e-07, "loss": 9.8751, "step": 1857 }, { "epoch": 0.5824451410658307, "grad_norm": 3.773027181625366, "learning_rate": 9.29e-07, "loss": 9.1356, "step": 1858 }, { "epoch": 0.5827586206896552, "grad_norm": 3.524904727935791, "learning_rate": 9.295000000000002e-07, "loss": 7.9023, "step": 1859 }, { "epoch": 0.5830721003134797, "grad_norm": 3.681853771209717, "learning_rate": 9.300000000000001e-07, "loss": 7.6041, "step": 1860 }, { "epoch": 0.5833855799373041, "grad_norm": 4.124235153198242, "learning_rate": 9.305e-07, "loss": 10.3233, "step": 1861 }, { "epoch": 0.5836990595611286, "grad_norm": 4.318082809448242, "learning_rate": 9.310000000000001e-07, "loss": 10.4379, "step": 1862 }, { "epoch": 0.584012539184953, "grad_norm": 3.5927541255950928, "learning_rate": 9.315e-07, "loss": 7.0752, "step": 1863 }, { "epoch": 0.5843260188087774, "grad_norm": 3.7160086631774902, "learning_rate": 9.320000000000001e-07, "loss": 8.0991, "step": 1864 }, { "epoch": 0.5846394984326019, "grad_norm": 4.717742919921875, "learning_rate": 9.325000000000001e-07, "loss": 8.6703, "step": 1865 }, { "epoch": 0.5849529780564263, "grad_norm": 4.089090347290039, "learning_rate": 9.33e-07, "loss": 8.6572, "step": 1866 }, { "epoch": 0.5852664576802508, "grad_norm": 5.522655010223389, "learning_rate": 9.335000000000001e-07, "loss": 9.8966, "step": 1867 }, { "epoch": 0.5855799373040752, "grad_norm": 3.4123260974884033, "learning_rate": 9.340000000000001e-07, "loss": 7.0431, "step": 1868 }, { "epoch": 0.5858934169278996, "grad_norm": 4.160542011260986, "learning_rate": 9.345000000000001e-07, "loss": 8.1701, "step": 1869 }, { "epoch": 0.5862068965517241, "grad_norm": 3.6532208919525146, "learning_rate": 9.35e-07, "loss": 10.5839, "step": 1870 }, { "epoch": 0.5865203761755486, "grad_norm": 4.788308143615723, "learning_rate": 9.355e-07, "loss": 9.6103, "step": 1871 }, { "epoch": 0.5868338557993731, "grad_norm": 3.656114339828491, "learning_rate": 9.360000000000001e-07, "loss": 9.0917, "step": 1872 }, { "epoch": 0.5871473354231975, "grad_norm": 3.7740187644958496, "learning_rate": 9.365000000000001e-07, "loss": 10.3359, "step": 1873 }, { "epoch": 0.587460815047022, "grad_norm": 3.203486680984497, "learning_rate": 9.370000000000001e-07, "loss": 6.0882, "step": 1874 }, { "epoch": 0.5877742946708464, "grad_norm": 4.1461920738220215, "learning_rate": 9.375000000000001e-07, "loss": 10.5749, "step": 1875 }, { "epoch": 0.5880877742946709, "grad_norm": 4.353470802307129, "learning_rate": 9.380000000000001e-07, "loss": 8.5529, "step": 1876 }, { "epoch": 0.5884012539184953, "grad_norm": 4.317278861999512, "learning_rate": 9.385000000000001e-07, "loss": 9.1286, "step": 1877 }, { "epoch": 0.5887147335423197, "grad_norm": 4.176400661468506, "learning_rate": 9.39e-07, "loss": 9.2643, "step": 1878 }, { "epoch": 0.5890282131661442, "grad_norm": 3.544487953186035, "learning_rate": 9.395000000000002e-07, "loss": 7.9035, "step": 1879 }, { "epoch": 0.5893416927899686, "grad_norm": 4.3520731925964355, "learning_rate": 9.400000000000001e-07, "loss": 8.9988, "step": 1880 }, { "epoch": 0.5896551724137931, "grad_norm": 4.328791618347168, "learning_rate": 9.405e-07, "loss": 8.2385, "step": 1881 }, { "epoch": 0.5899686520376175, "grad_norm": 4.532482147216797, "learning_rate": 9.410000000000001e-07, "loss": 10.0327, "step": 1882 }, { "epoch": 0.590282131661442, "grad_norm": 4.309502124786377, "learning_rate": 9.415e-07, "loss": 9.5955, "step": 1883 }, { "epoch": 0.5905956112852665, "grad_norm": 9.216413497924805, "learning_rate": 9.420000000000002e-07, "loss": 17.2207, "step": 1884 }, { "epoch": 0.5909090909090909, "grad_norm": 4.302111625671387, "learning_rate": 9.425000000000001e-07, "loss": 8.134, "step": 1885 }, { "epoch": 0.5912225705329154, "grad_norm": 3.478489637374878, "learning_rate": 9.43e-07, "loss": 7.1565, "step": 1886 }, { "epoch": 0.5915360501567398, "grad_norm": 5.942442893981934, "learning_rate": 9.435000000000002e-07, "loss": 16.1045, "step": 1887 }, { "epoch": 0.5918495297805643, "grad_norm": 4.080247402191162, "learning_rate": 9.440000000000001e-07, "loss": 10.6923, "step": 1888 }, { "epoch": 0.5921630094043887, "grad_norm": 4.32432222366333, "learning_rate": 9.445000000000001e-07, "loss": 8.9229, "step": 1889 }, { "epoch": 0.5924764890282131, "grad_norm": 3.346331834793091, "learning_rate": 9.450000000000001e-07, "loss": 6.9437, "step": 1890 }, { "epoch": 0.5927899686520376, "grad_norm": 8.710538864135742, "learning_rate": 9.455e-07, "loss": 18.3767, "step": 1891 }, { "epoch": 0.593103448275862, "grad_norm": 6.308311462402344, "learning_rate": 9.460000000000001e-07, "loss": 13.8071, "step": 1892 }, { "epoch": 0.5934169278996865, "grad_norm": 4.4475321769714355, "learning_rate": 9.465000000000001e-07, "loss": 8.8487, "step": 1893 }, { "epoch": 0.5937304075235109, "grad_norm": 5.897817611694336, "learning_rate": 9.470000000000001e-07, "loss": 11.9756, "step": 1894 }, { "epoch": 0.5940438871473355, "grad_norm": 5.22391414642334, "learning_rate": 9.475e-07, "loss": 7.3174, "step": 1895 }, { "epoch": 0.5943573667711599, "grad_norm": 4.082650184631348, "learning_rate": 9.480000000000001e-07, "loss": 7.8726, "step": 1896 }, { "epoch": 0.5946708463949844, "grad_norm": 4.346643447875977, "learning_rate": 9.485000000000001e-07, "loss": 8.1732, "step": 1897 }, { "epoch": 0.5949843260188088, "grad_norm": 6.866816997528076, "learning_rate": 9.49e-07, "loss": 12.5577, "step": 1898 }, { "epoch": 0.5952978056426332, "grad_norm": 3.5348422527313232, "learning_rate": 9.495000000000002e-07, "loss": 6.8785, "step": 1899 }, { "epoch": 0.5956112852664577, "grad_norm": 3.0132737159729004, "learning_rate": 9.500000000000001e-07, "loss": 5.57, "step": 1900 }, { "epoch": 0.5959247648902821, "grad_norm": 3.286238193511963, "learning_rate": 9.505e-07, "loss": 8.4462, "step": 1901 }, { "epoch": 0.5962382445141066, "grad_norm": 5.34050989151001, "learning_rate": 9.510000000000001e-07, "loss": 8.19, "step": 1902 }, { "epoch": 0.596551724137931, "grad_norm": 3.889528751373291, "learning_rate": 9.515e-07, "loss": 7.5043, "step": 1903 }, { "epoch": 0.5968652037617554, "grad_norm": 3.9055964946746826, "learning_rate": 9.520000000000002e-07, "loss": 8.6483, "step": 1904 }, { "epoch": 0.5971786833855799, "grad_norm": 3.298340320587158, "learning_rate": 9.525000000000001e-07, "loss": 6.4088, "step": 1905 }, { "epoch": 0.5974921630094044, "grad_norm": 3.5493357181549072, "learning_rate": 9.53e-07, "loss": 7.4659, "step": 1906 }, { "epoch": 0.5978056426332289, "grad_norm": 4.918447494506836, "learning_rate": 9.535000000000002e-07, "loss": 9.6589, "step": 1907 }, { "epoch": 0.5981191222570533, "grad_norm": 3.490692615509033, "learning_rate": 9.54e-07, "loss": 7.0492, "step": 1908 }, { "epoch": 0.5984326018808778, "grad_norm": 4.562289237976074, "learning_rate": 9.545e-07, "loss": 9.4843, "step": 1909 }, { "epoch": 0.5987460815047022, "grad_norm": 5.618401050567627, "learning_rate": 9.550000000000002e-07, "loss": 15.0284, "step": 1910 }, { "epoch": 0.5990595611285267, "grad_norm": 5.000195503234863, "learning_rate": 9.555e-07, "loss": 10.3012, "step": 1911 }, { "epoch": 0.5993730407523511, "grad_norm": 4.634150505065918, "learning_rate": 9.56e-07, "loss": 8.0312, "step": 1912 }, { "epoch": 0.5996865203761755, "grad_norm": 4.144935131072998, "learning_rate": 9.565e-07, "loss": 7.6425, "step": 1913 }, { "epoch": 0.6, "grad_norm": 6.016616344451904, "learning_rate": 9.570000000000001e-07, "loss": 10.0162, "step": 1914 }, { "epoch": 0.6003134796238244, "grad_norm": 4.132763862609863, "learning_rate": 9.575000000000001e-07, "loss": 7.813, "step": 1915 }, { "epoch": 0.6006269592476489, "grad_norm": 5.254676818847656, "learning_rate": 9.58e-07, "loss": 9.9075, "step": 1916 }, { "epoch": 0.6009404388714733, "grad_norm": 3.9892544746398926, "learning_rate": 9.585000000000002e-07, "loss": 8.0493, "step": 1917 }, { "epoch": 0.6012539184952979, "grad_norm": 4.14557409286499, "learning_rate": 9.59e-07, "loss": 8.4252, "step": 1918 }, { "epoch": 0.6015673981191223, "grad_norm": 5.171091079711914, "learning_rate": 9.595e-07, "loss": 9.5018, "step": 1919 }, { "epoch": 0.6018808777429467, "grad_norm": 5.105014801025391, "learning_rate": 9.600000000000001e-07, "loss": 8.7117, "step": 1920 }, { "epoch": 0.6021943573667712, "grad_norm": 4.491677284240723, "learning_rate": 9.605e-07, "loss": 9.022, "step": 1921 }, { "epoch": 0.6025078369905956, "grad_norm": 4.195937156677246, "learning_rate": 9.610000000000002e-07, "loss": 8.2956, "step": 1922 }, { "epoch": 0.6028213166144201, "grad_norm": 2.9639382362365723, "learning_rate": 9.615e-07, "loss": 6.2187, "step": 1923 }, { "epoch": 0.6031347962382445, "grad_norm": 5.277063369750977, "learning_rate": 9.62e-07, "loss": 9.245, "step": 1924 }, { "epoch": 0.603448275862069, "grad_norm": 3.676309585571289, "learning_rate": 9.625e-07, "loss": 7.6557, "step": 1925 }, { "epoch": 0.6037617554858934, "grad_norm": 5.051148891448975, "learning_rate": 9.630000000000001e-07, "loss": 11.8118, "step": 1926 }, { "epoch": 0.6040752351097178, "grad_norm": 4.08391809463501, "learning_rate": 9.635000000000002e-07, "loss": 8.8092, "step": 1927 }, { "epoch": 0.6043887147335423, "grad_norm": 4.592138290405273, "learning_rate": 9.64e-07, "loss": 8.8694, "step": 1928 }, { "epoch": 0.6047021943573667, "grad_norm": 6.512213230133057, "learning_rate": 9.645e-07, "loss": 10.43, "step": 1929 }, { "epoch": 0.6050156739811913, "grad_norm": 5.282738208770752, "learning_rate": 9.65e-07, "loss": 11.6431, "step": 1930 }, { "epoch": 0.6053291536050157, "grad_norm": 3.4257993698120117, "learning_rate": 9.655000000000001e-07, "loss": 7.8864, "step": 1931 }, { "epoch": 0.6056426332288402, "grad_norm": 3.866560459136963, "learning_rate": 9.660000000000002e-07, "loss": 7.6513, "step": 1932 }, { "epoch": 0.6059561128526646, "grad_norm": 4.288784503936768, "learning_rate": 9.665e-07, "loss": 8.4351, "step": 1933 }, { "epoch": 0.606269592476489, "grad_norm": 5.792541980743408, "learning_rate": 9.67e-07, "loss": 9.6173, "step": 1934 }, { "epoch": 0.6065830721003135, "grad_norm": 2.695124626159668, "learning_rate": 9.675e-07, "loss": 6.1397, "step": 1935 }, { "epoch": 0.6068965517241379, "grad_norm": 4.266729354858398, "learning_rate": 9.68e-07, "loss": 7.2699, "step": 1936 }, { "epoch": 0.6072100313479624, "grad_norm": 3.4654881954193115, "learning_rate": 9.685000000000001e-07, "loss": 6.0948, "step": 1937 }, { "epoch": 0.6075235109717868, "grad_norm": 4.1121954917907715, "learning_rate": 9.690000000000002e-07, "loss": 8.0899, "step": 1938 }, { "epoch": 0.6078369905956112, "grad_norm": 4.078955173492432, "learning_rate": 9.695e-07, "loss": 7.1175, "step": 1939 }, { "epoch": 0.6081504702194357, "grad_norm": 7.904345512390137, "learning_rate": 9.7e-07, "loss": 14.9369, "step": 1940 }, { "epoch": 0.6084639498432602, "grad_norm": 3.5593857765197754, "learning_rate": 9.705e-07, "loss": 6.9494, "step": 1941 }, { "epoch": 0.6087774294670847, "grad_norm": 3.992682933807373, "learning_rate": 9.71e-07, "loss": 8.6618, "step": 1942 }, { "epoch": 0.6090909090909091, "grad_norm": 3.1125354766845703, "learning_rate": 9.715000000000001e-07, "loss": 5.9992, "step": 1943 }, { "epoch": 0.6094043887147336, "grad_norm": 3.825096845626831, "learning_rate": 9.72e-07, "loss": 7.6425, "step": 1944 }, { "epoch": 0.609717868338558, "grad_norm": 3.6271681785583496, "learning_rate": 9.725e-07, "loss": 7.7853, "step": 1945 }, { "epoch": 0.6100313479623825, "grad_norm": 6.750822067260742, "learning_rate": 9.73e-07, "loss": 16.4377, "step": 1946 }, { "epoch": 0.6103448275862069, "grad_norm": 3.6327602863311768, "learning_rate": 9.735e-07, "loss": 8.6631, "step": 1947 }, { "epoch": 0.6106583072100313, "grad_norm": 3.878876209259033, "learning_rate": 9.740000000000001e-07, "loss": 6.7651, "step": 1948 }, { "epoch": 0.6109717868338558, "grad_norm": 3.9936046600341797, "learning_rate": 9.745e-07, "loss": 7.4598, "step": 1949 }, { "epoch": 0.6112852664576802, "grad_norm": 3.6358838081359863, "learning_rate": 9.750000000000002e-07, "loss": 6.9454, "step": 1950 }, { "epoch": 0.6115987460815047, "grad_norm": 4.248730182647705, "learning_rate": 9.755e-07, "loss": 8.5024, "step": 1951 }, { "epoch": 0.6119122257053291, "grad_norm": 3.337886095046997, "learning_rate": 9.76e-07, "loss": 6.4199, "step": 1952 }, { "epoch": 0.6122257053291537, "grad_norm": 3.8960623741149902, "learning_rate": 9.765e-07, "loss": 7.2547, "step": 1953 }, { "epoch": 0.6125391849529781, "grad_norm": 3.7697343826293945, "learning_rate": 9.770000000000001e-07, "loss": 8.3044, "step": 1954 }, { "epoch": 0.6128526645768025, "grad_norm": 4.274561882019043, "learning_rate": 9.775000000000002e-07, "loss": 9.8826, "step": 1955 }, { "epoch": 0.613166144200627, "grad_norm": 4.1365461349487305, "learning_rate": 9.78e-07, "loss": 6.5002, "step": 1956 }, { "epoch": 0.6134796238244514, "grad_norm": 4.854091167449951, "learning_rate": 9.785000000000002e-07, "loss": 9.7826, "step": 1957 }, { "epoch": 0.6137931034482759, "grad_norm": 4.134148597717285, "learning_rate": 9.79e-07, "loss": 7.5184, "step": 1958 }, { "epoch": 0.6141065830721003, "grad_norm": 3.301070213317871, "learning_rate": 9.795000000000001e-07, "loss": 7.5854, "step": 1959 }, { "epoch": 0.6144200626959248, "grad_norm": 5.0885820388793945, "learning_rate": 9.800000000000001e-07, "loss": 12.4143, "step": 1960 }, { "epoch": 0.6147335423197492, "grad_norm": 3.778027296066284, "learning_rate": 9.805e-07, "loss": 6.6442, "step": 1961 }, { "epoch": 0.6150470219435736, "grad_norm": 4.023343086242676, "learning_rate": 9.810000000000002e-07, "loss": 9.6106, "step": 1962 }, { "epoch": 0.6153605015673981, "grad_norm": 4.735710144042969, "learning_rate": 9.815e-07, "loss": 8.8034, "step": 1963 }, { "epoch": 0.6156739811912225, "grad_norm": 4.353292942047119, "learning_rate": 9.82e-07, "loss": 9.1107, "step": 1964 }, { "epoch": 0.6159874608150471, "grad_norm": 4.989565849304199, "learning_rate": 9.825000000000001e-07, "loss": 9.7556, "step": 1965 }, { "epoch": 0.6163009404388715, "grad_norm": 3.5673139095306396, "learning_rate": 9.830000000000002e-07, "loss": 6.4914, "step": 1966 }, { "epoch": 0.616614420062696, "grad_norm": 4.158758640289307, "learning_rate": 9.835000000000002e-07, "loss": 7.9705, "step": 1967 }, { "epoch": 0.6169278996865204, "grad_norm": 7.679725170135498, "learning_rate": 9.84e-07, "loss": 9.2285, "step": 1968 }, { "epoch": 0.6172413793103448, "grad_norm": 5.259065628051758, "learning_rate": 9.845e-07, "loss": 12.6733, "step": 1969 }, { "epoch": 0.6175548589341693, "grad_norm": 4.931044578552246, "learning_rate": 9.85e-07, "loss": 9.4881, "step": 1970 }, { "epoch": 0.6178683385579937, "grad_norm": 4.3828125, "learning_rate": 9.855000000000001e-07, "loss": 8.3451, "step": 1971 }, { "epoch": 0.6181818181818182, "grad_norm": 5.446940898895264, "learning_rate": 9.86e-07, "loss": 8.3187, "step": 1972 }, { "epoch": 0.6184952978056426, "grad_norm": 5.320723533630371, "learning_rate": 9.865e-07, "loss": 9.4323, "step": 1973 }, { "epoch": 0.618808777429467, "grad_norm": 4.248854637145996, "learning_rate": 9.87e-07, "loss": 7.6606, "step": 1974 }, { "epoch": 0.6191222570532915, "grad_norm": 4.689657211303711, "learning_rate": 9.875e-07, "loss": 6.7403, "step": 1975 }, { "epoch": 0.6194357366771159, "grad_norm": 4.264988422393799, "learning_rate": 9.880000000000001e-07, "loss": 9.1217, "step": 1976 }, { "epoch": 0.6197492163009405, "grad_norm": 5.480489253997803, "learning_rate": 9.885e-07, "loss": 8.779, "step": 1977 }, { "epoch": 0.6200626959247649, "grad_norm": 4.787638187408447, "learning_rate": 9.890000000000002e-07, "loss": 8.9536, "step": 1978 }, { "epoch": 0.6203761755485894, "grad_norm": 4.072272300720215, "learning_rate": 9.895e-07, "loss": 8.668, "step": 1979 }, { "epoch": 0.6206896551724138, "grad_norm": 3.5888800621032715, "learning_rate": 9.9e-07, "loss": 7.6795, "step": 1980 }, { "epoch": 0.6210031347962383, "grad_norm": 5.570566177368164, "learning_rate": 9.905e-07, "loss": 12.1096, "step": 1981 }, { "epoch": 0.6213166144200627, "grad_norm": 4.423348426818848, "learning_rate": 9.91e-07, "loss": 9.1405, "step": 1982 }, { "epoch": 0.6216300940438871, "grad_norm": 4.6284356117248535, "learning_rate": 9.915000000000002e-07, "loss": 7.2074, "step": 1983 }, { "epoch": 0.6219435736677116, "grad_norm": 4.048377990722656, "learning_rate": 9.92e-07, "loss": 7.6424, "step": 1984 }, { "epoch": 0.622257053291536, "grad_norm": 3.3591959476470947, "learning_rate": 9.925e-07, "loss": 7.673, "step": 1985 }, { "epoch": 0.6225705329153605, "grad_norm": 3.9147262573242188, "learning_rate": 9.93e-07, "loss": 7.3318, "step": 1986 }, { "epoch": 0.6228840125391849, "grad_norm": 3.4501211643218994, "learning_rate": 9.935e-07, "loss": 7.3048, "step": 1987 }, { "epoch": 0.6231974921630095, "grad_norm": 3.9171688556671143, "learning_rate": 9.940000000000001e-07, "loss": 7.2895, "step": 1988 }, { "epoch": 0.6235109717868339, "grad_norm": 4.911311626434326, "learning_rate": 9.945e-07, "loss": 7.0245, "step": 1989 }, { "epoch": 0.6238244514106583, "grad_norm": 3.89208722114563, "learning_rate": 9.950000000000002e-07, "loss": 9.1831, "step": 1990 }, { "epoch": 0.6241379310344828, "grad_norm": 5.180747032165527, "learning_rate": 9.955e-07, "loss": 9.1415, "step": 1991 }, { "epoch": 0.6244514106583072, "grad_norm": 4.259772777557373, "learning_rate": 9.96e-07, "loss": 9.0887, "step": 1992 }, { "epoch": 0.6247648902821317, "grad_norm": 3.5109479427337646, "learning_rate": 9.965000000000001e-07, "loss": 7.4708, "step": 1993 }, { "epoch": 0.6250783699059561, "grad_norm": 3.7264599800109863, "learning_rate": 9.970000000000002e-07, "loss": 8.0124, "step": 1994 }, { "epoch": 0.6253918495297806, "grad_norm": 6.5949249267578125, "learning_rate": 9.975000000000002e-07, "loss": 12.249, "step": 1995 }, { "epoch": 0.625705329153605, "grad_norm": 5.1157450675964355, "learning_rate": 9.98e-07, "loss": 10.3741, "step": 1996 }, { "epoch": 0.6260188087774294, "grad_norm": 4.158992767333984, "learning_rate": 9.985e-07, "loss": 7.6302, "step": 1997 }, { "epoch": 0.6263322884012539, "grad_norm": 5.8107428550720215, "learning_rate": 9.99e-07, "loss": 14.3595, "step": 1998 }, { "epoch": 0.6266457680250783, "grad_norm": 5.154208660125732, "learning_rate": 9.995000000000001e-07, "loss": 8.8518, "step": 1999 }, { "epoch": 0.6269592476489029, "grad_norm": 4.65903902053833, "learning_rate": 1.0000000000000002e-06, "loss": 7.9164, "step": 2000 }, { "epoch": 0.6272727272727273, "grad_norm": 4.164371013641357, "learning_rate": 1.0005e-06, "loss": 8.2315, "step": 2001 }, { "epoch": 0.6275862068965518, "grad_norm": 4.161435604095459, "learning_rate": 1.001e-06, "loss": 9.4845, "step": 2002 }, { "epoch": 0.6278996865203762, "grad_norm": 6.261343002319336, "learning_rate": 1.0015e-06, "loss": 14.0749, "step": 2003 }, { "epoch": 0.6282131661442006, "grad_norm": 6.03980827331543, "learning_rate": 1.002e-06, "loss": 12.252, "step": 2004 }, { "epoch": 0.6285266457680251, "grad_norm": 4.8788862228393555, "learning_rate": 1.0025000000000001e-06, "loss": 8.6274, "step": 2005 }, { "epoch": 0.6288401253918495, "grad_norm": 4.340915679931641, "learning_rate": 1.0030000000000002e-06, "loss": 8.0098, "step": 2006 }, { "epoch": 0.629153605015674, "grad_norm": 4.664811611175537, "learning_rate": 1.0035e-06, "loss": 8.8534, "step": 2007 }, { "epoch": 0.6294670846394984, "grad_norm": 3.558058977127075, "learning_rate": 1.004e-06, "loss": 7.5891, "step": 2008 }, { "epoch": 0.6297805642633229, "grad_norm": 5.759692668914795, "learning_rate": 1.0045e-06, "loss": 10.9539, "step": 2009 }, { "epoch": 0.6300940438871473, "grad_norm": 3.947463035583496, "learning_rate": 1.0050000000000001e-06, "loss": 6.9227, "step": 2010 }, { "epoch": 0.6304075235109717, "grad_norm": 4.328242778778076, "learning_rate": 1.0055000000000002e-06, "loss": 8.2226, "step": 2011 }, { "epoch": 0.6307210031347963, "grad_norm": 5.028584957122803, "learning_rate": 1.006e-06, "loss": 11.2504, "step": 2012 }, { "epoch": 0.6310344827586207, "grad_norm": 5.828540325164795, "learning_rate": 1.0065e-06, "loss": 12.0494, "step": 2013 }, { "epoch": 0.6313479623824452, "grad_norm": 5.526819705963135, "learning_rate": 1.007e-06, "loss": 13.3581, "step": 2014 }, { "epoch": 0.6316614420062696, "grad_norm": 4.654907703399658, "learning_rate": 1.0075e-06, "loss": 9.1679, "step": 2015 }, { "epoch": 0.631974921630094, "grad_norm": 4.357663154602051, "learning_rate": 1.0080000000000001e-06, "loss": 8.1994, "step": 2016 }, { "epoch": 0.6322884012539185, "grad_norm": 5.075762748718262, "learning_rate": 1.0085e-06, "loss": 10.5631, "step": 2017 }, { "epoch": 0.6326018808777429, "grad_norm": 3.278012275695801, "learning_rate": 1.0090000000000002e-06, "loss": 7.5705, "step": 2018 }, { "epoch": 0.6329153605015674, "grad_norm": 3.894144058227539, "learning_rate": 1.0095e-06, "loss": 7.3649, "step": 2019 }, { "epoch": 0.6332288401253918, "grad_norm": 3.737834930419922, "learning_rate": 1.01e-06, "loss": 8.4081, "step": 2020 }, { "epoch": 0.6335423197492163, "grad_norm": 3.789785623550415, "learning_rate": 1.0105000000000001e-06, "loss": 6.3057, "step": 2021 }, { "epoch": 0.6338557993730407, "grad_norm": 3.5118370056152344, "learning_rate": 1.011e-06, "loss": 5.7457, "step": 2022 }, { "epoch": 0.6341692789968651, "grad_norm": 3.056246519088745, "learning_rate": 1.0115000000000002e-06, "loss": 5.266, "step": 2023 }, { "epoch": 0.6344827586206897, "grad_norm": 4.797995567321777, "learning_rate": 1.012e-06, "loss": 7.2787, "step": 2024 }, { "epoch": 0.6347962382445141, "grad_norm": 3.8254222869873047, "learning_rate": 1.0125e-06, "loss": 7.5975, "step": 2025 }, { "epoch": 0.6351097178683386, "grad_norm": 3.7210936546325684, "learning_rate": 1.013e-06, "loss": 7.1766, "step": 2026 }, { "epoch": 0.635423197492163, "grad_norm": 4.653789520263672, "learning_rate": 1.0135000000000001e-06, "loss": 8.7087, "step": 2027 }, { "epoch": 0.6357366771159875, "grad_norm": 5.211480617523193, "learning_rate": 1.0140000000000002e-06, "loss": 11.3491, "step": 2028 }, { "epoch": 0.6360501567398119, "grad_norm": 4.071037292480469, "learning_rate": 1.0145e-06, "loss": 9.2332, "step": 2029 }, { "epoch": 0.6363636363636364, "grad_norm": 3.9774158000946045, "learning_rate": 1.0150000000000002e-06, "loss": 7.939, "step": 2030 }, { "epoch": 0.6366771159874608, "grad_norm": 4.108770370483398, "learning_rate": 1.0155e-06, "loss": 8.5832, "step": 2031 }, { "epoch": 0.6369905956112852, "grad_norm": 3.7771849632263184, "learning_rate": 1.016e-06, "loss": 7.9219, "step": 2032 }, { "epoch": 0.6373040752351097, "grad_norm": 4.883484363555908, "learning_rate": 1.0165000000000001e-06, "loss": 7.4839, "step": 2033 }, { "epoch": 0.6376175548589341, "grad_norm": 5.540393829345703, "learning_rate": 1.0170000000000002e-06, "loss": 11.3517, "step": 2034 }, { "epoch": 0.6379310344827587, "grad_norm": 3.476276397705078, "learning_rate": 1.0175e-06, "loss": 5.4489, "step": 2035 }, { "epoch": 0.6382445141065831, "grad_norm": 4.134171485900879, "learning_rate": 1.018e-06, "loss": 9.0043, "step": 2036 }, { "epoch": 0.6385579937304076, "grad_norm": 3.8782660961151123, "learning_rate": 1.0185e-06, "loss": 7.7151, "step": 2037 }, { "epoch": 0.638871473354232, "grad_norm": 5.3437347412109375, "learning_rate": 1.0190000000000001e-06, "loss": 8.5708, "step": 2038 }, { "epoch": 0.6391849529780564, "grad_norm": 4.901749610900879, "learning_rate": 1.0195000000000001e-06, "loss": 9.5016, "step": 2039 }, { "epoch": 0.6394984326018809, "grad_norm": 9.612685203552246, "learning_rate": 1.02e-06, "loss": 25.8075, "step": 2040 }, { "epoch": 0.6398119122257053, "grad_norm": 4.421562671661377, "learning_rate": 1.0205e-06, "loss": 8.9058, "step": 2041 }, { "epoch": 0.6401253918495298, "grad_norm": 4.578634738922119, "learning_rate": 1.021e-06, "loss": 8.4515, "step": 2042 }, { "epoch": 0.6404388714733542, "grad_norm": 5.676279067993164, "learning_rate": 1.0215e-06, "loss": 12.9783, "step": 2043 }, { "epoch": 0.6407523510971787, "grad_norm": 4.57780647277832, "learning_rate": 1.0220000000000001e-06, "loss": 7.5248, "step": 2044 }, { "epoch": 0.6410658307210031, "grad_norm": 4.466324806213379, "learning_rate": 1.0225e-06, "loss": 9.4472, "step": 2045 }, { "epoch": 0.6413793103448275, "grad_norm": 4.786015510559082, "learning_rate": 1.0230000000000002e-06, "loss": 7.8918, "step": 2046 }, { "epoch": 0.6416927899686521, "grad_norm": 4.978856086730957, "learning_rate": 1.0235e-06, "loss": 8.9204, "step": 2047 }, { "epoch": 0.6420062695924765, "grad_norm": 3.4786014556884766, "learning_rate": 1.024e-06, "loss": 7.3538, "step": 2048 }, { "epoch": 0.642319749216301, "grad_norm": 6.084647178649902, "learning_rate": 1.0245e-06, "loss": 8.8253, "step": 2049 }, { "epoch": 0.6426332288401254, "grad_norm": 5.3138556480407715, "learning_rate": 1.025e-06, "loss": 11.3613, "step": 2050 }, { "epoch": 0.6429467084639499, "grad_norm": 3.573272943496704, "learning_rate": 1.0255000000000002e-06, "loss": 7.1191, "step": 2051 }, { "epoch": 0.6432601880877743, "grad_norm": 4.325623035430908, "learning_rate": 1.026e-06, "loss": 6.4057, "step": 2052 }, { "epoch": 0.6435736677115987, "grad_norm": 6.643167972564697, "learning_rate": 1.0265e-06, "loss": 16.8502, "step": 2053 }, { "epoch": 0.6438871473354232, "grad_norm": 4.033902645111084, "learning_rate": 1.027e-06, "loss": 7.7638, "step": 2054 }, { "epoch": 0.6442006269592476, "grad_norm": 3.4600517749786377, "learning_rate": 1.0275000000000001e-06, "loss": 6.4728, "step": 2055 }, { "epoch": 0.6445141065830721, "grad_norm": 5.4869818687438965, "learning_rate": 1.0280000000000002e-06, "loss": 11.0305, "step": 2056 }, { "epoch": 0.6448275862068965, "grad_norm": 2.9913864135742188, "learning_rate": 1.0285e-06, "loss": 6.2851, "step": 2057 }, { "epoch": 0.645141065830721, "grad_norm": 4.9674296379089355, "learning_rate": 1.0290000000000002e-06, "loss": 10.5091, "step": 2058 }, { "epoch": 0.6454545454545455, "grad_norm": 4.117265224456787, "learning_rate": 1.0295e-06, "loss": 7.9589, "step": 2059 }, { "epoch": 0.64576802507837, "grad_norm": 4.783974647521973, "learning_rate": 1.03e-06, "loss": 10.0756, "step": 2060 }, { "epoch": 0.6460815047021944, "grad_norm": 3.9700472354888916, "learning_rate": 1.0305000000000001e-06, "loss": 8.5293, "step": 2061 }, { "epoch": 0.6463949843260188, "grad_norm": 3.448155641555786, "learning_rate": 1.031e-06, "loss": 6.8082, "step": 2062 }, { "epoch": 0.6467084639498433, "grad_norm": 5.751345157623291, "learning_rate": 1.0315000000000002e-06, "loss": 10.7494, "step": 2063 }, { "epoch": 0.6470219435736677, "grad_norm": 4.36277961730957, "learning_rate": 1.032e-06, "loss": 7.5388, "step": 2064 }, { "epoch": 0.6473354231974922, "grad_norm": 4.9406633377075195, "learning_rate": 1.0325e-06, "loss": 8.7483, "step": 2065 }, { "epoch": 0.6476489028213166, "grad_norm": 3.610996961593628, "learning_rate": 1.033e-06, "loss": 5.993, "step": 2066 }, { "epoch": 0.647962382445141, "grad_norm": 8.09903621673584, "learning_rate": 1.0335000000000001e-06, "loss": 18.3466, "step": 2067 }, { "epoch": 0.6482758620689655, "grad_norm": 4.765588760375977, "learning_rate": 1.0340000000000002e-06, "loss": 9.1752, "step": 2068 }, { "epoch": 0.6485893416927899, "grad_norm": 5.2749199867248535, "learning_rate": 1.0345e-06, "loss": 11.2824, "step": 2069 }, { "epoch": 0.6489028213166145, "grad_norm": 5.744815826416016, "learning_rate": 1.035e-06, "loss": 12.4292, "step": 2070 }, { "epoch": 0.6492163009404389, "grad_norm": 4.47613525390625, "learning_rate": 1.0355e-06, "loss": 7.8614, "step": 2071 }, { "epoch": 0.6495297805642634, "grad_norm": 6.598586082458496, "learning_rate": 1.0360000000000001e-06, "loss": 11.9815, "step": 2072 }, { "epoch": 0.6498432601880878, "grad_norm": 5.2339606285095215, "learning_rate": 1.0365000000000002e-06, "loss": 11.3129, "step": 2073 }, { "epoch": 0.6501567398119122, "grad_norm": 4.304049015045166, "learning_rate": 1.0370000000000002e-06, "loss": 7.8178, "step": 2074 }, { "epoch": 0.6504702194357367, "grad_norm": 5.00603723526001, "learning_rate": 1.0375e-06, "loss": 10.3162, "step": 2075 }, { "epoch": 0.6507836990595611, "grad_norm": 3.639469623565674, "learning_rate": 1.038e-06, "loss": 6.6105, "step": 2076 }, { "epoch": 0.6510971786833856, "grad_norm": 4.230920314788818, "learning_rate": 1.0385e-06, "loss": 7.4984, "step": 2077 }, { "epoch": 0.65141065830721, "grad_norm": 4.557835578918457, "learning_rate": 1.0390000000000001e-06, "loss": 8.9535, "step": 2078 }, { "epoch": 0.6517241379310345, "grad_norm": 4.792398452758789, "learning_rate": 1.0395000000000002e-06, "loss": 8.1143, "step": 2079 }, { "epoch": 0.6520376175548589, "grad_norm": 4.397486209869385, "learning_rate": 1.04e-06, "loss": 9.0494, "step": 2080 }, { "epoch": 0.6523510971786833, "grad_norm": 6.484209060668945, "learning_rate": 1.0405e-06, "loss": 11.6424, "step": 2081 }, { "epoch": 0.6526645768025079, "grad_norm": 5.75565767288208, "learning_rate": 1.041e-06, "loss": 10.6868, "step": 2082 }, { "epoch": 0.6529780564263323, "grad_norm": 4.253330707550049, "learning_rate": 1.0415000000000001e-06, "loss": 8.2279, "step": 2083 }, { "epoch": 0.6532915360501568, "grad_norm": 5.125704288482666, "learning_rate": 1.0420000000000001e-06, "loss": 8.9277, "step": 2084 }, { "epoch": 0.6536050156739812, "grad_norm": 4.079887390136719, "learning_rate": 1.0425e-06, "loss": 7.9315, "step": 2085 }, { "epoch": 0.6539184952978057, "grad_norm": 4.4778361320495605, "learning_rate": 1.0430000000000002e-06, "loss": 8.606, "step": 2086 }, { "epoch": 0.6542319749216301, "grad_norm": 4.556799411773682, "learning_rate": 1.0435e-06, "loss": 6.0047, "step": 2087 }, { "epoch": 0.6545454545454545, "grad_norm": 6.23499870300293, "learning_rate": 1.044e-06, "loss": 11.1014, "step": 2088 }, { "epoch": 0.654858934169279, "grad_norm": 5.879521369934082, "learning_rate": 1.0445000000000001e-06, "loss": 11.7058, "step": 2089 }, { "epoch": 0.6551724137931034, "grad_norm": 6.175170421600342, "learning_rate": 1.045e-06, "loss": 8.7771, "step": 2090 }, { "epoch": 0.6554858934169279, "grad_norm": 4.931567192077637, "learning_rate": 1.0455000000000002e-06, "loss": 10.7524, "step": 2091 }, { "epoch": 0.6557993730407523, "grad_norm": 5.440428733825684, "learning_rate": 1.046e-06, "loss": 7.6057, "step": 2092 }, { "epoch": 0.6561128526645768, "grad_norm": 4.918017387390137, "learning_rate": 1.0465e-06, "loss": 10.069, "step": 2093 }, { "epoch": 0.6564263322884013, "grad_norm": 4.340334892272949, "learning_rate": 1.047e-06, "loss": 8.5315, "step": 2094 }, { "epoch": 0.6567398119122257, "grad_norm": 4.921428680419922, "learning_rate": 1.0475000000000001e-06, "loss": 9.7294, "step": 2095 }, { "epoch": 0.6570532915360502, "grad_norm": 6.337207794189453, "learning_rate": 1.0480000000000002e-06, "loss": 9.9782, "step": 2096 }, { "epoch": 0.6573667711598746, "grad_norm": 4.316909313201904, "learning_rate": 1.0485e-06, "loss": 6.4938, "step": 2097 }, { "epoch": 0.6576802507836991, "grad_norm": 4.156019687652588, "learning_rate": 1.049e-06, "loss": 7.8225, "step": 2098 }, { "epoch": 0.6579937304075235, "grad_norm": 3.3352489471435547, "learning_rate": 1.0495e-06, "loss": 6.4576, "step": 2099 }, { "epoch": 0.658307210031348, "grad_norm": 4.093966007232666, "learning_rate": 1.0500000000000001e-06, "loss": 7.3799, "step": 2100 }, { "epoch": 0.6586206896551724, "grad_norm": 3.7547051906585693, "learning_rate": 1.0505000000000001e-06, "loss": 8.6161, "step": 2101 }, { "epoch": 0.6589341692789968, "grad_norm": 4.8714447021484375, "learning_rate": 1.051e-06, "loss": 8.6568, "step": 2102 }, { "epoch": 0.6592476489028213, "grad_norm": 4.202094078063965, "learning_rate": 1.0515e-06, "loss": 7.711, "step": 2103 }, { "epoch": 0.6595611285266457, "grad_norm": 5.062414169311523, "learning_rate": 1.052e-06, "loss": 8.4271, "step": 2104 }, { "epoch": 0.6598746081504702, "grad_norm": 5.461309909820557, "learning_rate": 1.0525e-06, "loss": 8.2217, "step": 2105 }, { "epoch": 0.6601880877742947, "grad_norm": 4.929152965545654, "learning_rate": 1.0530000000000001e-06, "loss": 7.7316, "step": 2106 }, { "epoch": 0.6605015673981192, "grad_norm": 5.310903072357178, "learning_rate": 1.0535000000000002e-06, "loss": 9.4138, "step": 2107 }, { "epoch": 0.6608150470219436, "grad_norm": 5.781913757324219, "learning_rate": 1.054e-06, "loss": 8.4, "step": 2108 }, { "epoch": 0.661128526645768, "grad_norm": 4.272099494934082, "learning_rate": 1.0545e-06, "loss": 7.6243, "step": 2109 }, { "epoch": 0.6614420062695925, "grad_norm": 3.9875686168670654, "learning_rate": 1.055e-06, "loss": 7.8158, "step": 2110 }, { "epoch": 0.6617554858934169, "grad_norm": 7.379361629486084, "learning_rate": 1.0555e-06, "loss": 13.4704, "step": 2111 }, { "epoch": 0.6620689655172414, "grad_norm": 8.117541313171387, "learning_rate": 1.0560000000000001e-06, "loss": 16.404, "step": 2112 }, { "epoch": 0.6623824451410658, "grad_norm": 3.9060027599334717, "learning_rate": 1.0565e-06, "loss": 5.9239, "step": 2113 }, { "epoch": 0.6626959247648903, "grad_norm": 4.950360298156738, "learning_rate": 1.0570000000000002e-06, "loss": 7.5477, "step": 2114 }, { "epoch": 0.6630094043887147, "grad_norm": 6.759565353393555, "learning_rate": 1.0575e-06, "loss": 13.3141, "step": 2115 }, { "epoch": 0.6633228840125391, "grad_norm": 4.705567836761475, "learning_rate": 1.058e-06, "loss": 10.2262, "step": 2116 }, { "epoch": 0.6636363636363637, "grad_norm": 6.787057876586914, "learning_rate": 1.0585000000000001e-06, "loss": 10.6113, "step": 2117 }, { "epoch": 0.6639498432601881, "grad_norm": 4.983616828918457, "learning_rate": 1.059e-06, "loss": 10.8316, "step": 2118 }, { "epoch": 0.6642633228840126, "grad_norm": 3.8625426292419434, "learning_rate": 1.0595000000000002e-06, "loss": 7.7674, "step": 2119 }, { "epoch": 0.664576802507837, "grad_norm": 4.234352111816406, "learning_rate": 1.06e-06, "loss": 7.263, "step": 2120 }, { "epoch": 0.6648902821316615, "grad_norm": 4.677432537078857, "learning_rate": 1.0605e-06, "loss": 7.1481, "step": 2121 }, { "epoch": 0.6652037617554859, "grad_norm": 4.3010029792785645, "learning_rate": 1.061e-06, "loss": 6.9077, "step": 2122 }, { "epoch": 0.6655172413793103, "grad_norm": 3.675344228744507, "learning_rate": 1.0615000000000001e-06, "loss": 6.9315, "step": 2123 }, { "epoch": 0.6658307210031348, "grad_norm": 5.514710903167725, "learning_rate": 1.0620000000000002e-06, "loss": 11.0089, "step": 2124 }, { "epoch": 0.6661442006269592, "grad_norm": 4.249388694763184, "learning_rate": 1.0625e-06, "loss": 8.537, "step": 2125 }, { "epoch": 0.6664576802507837, "grad_norm": 4.913691997528076, "learning_rate": 1.0630000000000002e-06, "loss": 8.3072, "step": 2126 }, { "epoch": 0.6667711598746081, "grad_norm": 7.471773624420166, "learning_rate": 1.0635e-06, "loss": 9.3104, "step": 2127 }, { "epoch": 0.6670846394984326, "grad_norm": 5.314687728881836, "learning_rate": 1.064e-06, "loss": 9.9356, "step": 2128 }, { "epoch": 0.6673981191222571, "grad_norm": 5.965895175933838, "learning_rate": 1.0645000000000001e-06, "loss": 12.0117, "step": 2129 }, { "epoch": 0.6677115987460815, "grad_norm": 4.372379779815674, "learning_rate": 1.065e-06, "loss": 7.607, "step": 2130 }, { "epoch": 0.668025078369906, "grad_norm": 4.424569606781006, "learning_rate": 1.0655000000000002e-06, "loss": 10.169, "step": 2131 }, { "epoch": 0.6683385579937304, "grad_norm": 3.1224844455718994, "learning_rate": 1.066e-06, "loss": 6.0792, "step": 2132 }, { "epoch": 0.6686520376175549, "grad_norm": 5.132580280303955, "learning_rate": 1.0665e-06, "loss": 9.2276, "step": 2133 }, { "epoch": 0.6689655172413793, "grad_norm": 4.87095832824707, "learning_rate": 1.0670000000000001e-06, "loss": 7.84, "step": 2134 }, { "epoch": 0.6692789968652038, "grad_norm": 5.659262657165527, "learning_rate": 1.0675000000000002e-06, "loss": 8.4573, "step": 2135 }, { "epoch": 0.6695924764890282, "grad_norm": 4.755374908447266, "learning_rate": 1.0680000000000002e-06, "loss": 10.6835, "step": 2136 }, { "epoch": 0.6699059561128526, "grad_norm": 5.089763641357422, "learning_rate": 1.0685e-06, "loss": 7.9605, "step": 2137 }, { "epoch": 0.6702194357366771, "grad_norm": 8.315079689025879, "learning_rate": 1.069e-06, "loss": 11.3606, "step": 2138 }, { "epoch": 0.6705329153605015, "grad_norm": 5.036296367645264, "learning_rate": 1.0695e-06, "loss": 9.4513, "step": 2139 }, { "epoch": 0.670846394984326, "grad_norm": 4.769467353820801, "learning_rate": 1.0700000000000001e-06, "loss": 7.5086, "step": 2140 }, { "epoch": 0.6711598746081505, "grad_norm": 5.184993743896484, "learning_rate": 1.0705000000000002e-06, "loss": 10.644, "step": 2141 }, { "epoch": 0.671473354231975, "grad_norm": 3.7264373302459717, "learning_rate": 1.071e-06, "loss": 6.8944, "step": 2142 }, { "epoch": 0.6717868338557994, "grad_norm": 4.51856803894043, "learning_rate": 1.0715e-06, "loss": 7.1033, "step": 2143 }, { "epoch": 0.6721003134796238, "grad_norm": 4.620069980621338, "learning_rate": 1.072e-06, "loss": 9.0823, "step": 2144 }, { "epoch": 0.6724137931034483, "grad_norm": 4.350764274597168, "learning_rate": 1.0725000000000001e-06, "loss": 7.9142, "step": 2145 }, { "epoch": 0.6727272727272727, "grad_norm": 5.366765022277832, "learning_rate": 1.0730000000000001e-06, "loss": 7.2714, "step": 2146 }, { "epoch": 0.6730407523510972, "grad_norm": 4.46659517288208, "learning_rate": 1.0735000000000002e-06, "loss": 8.6368, "step": 2147 }, { "epoch": 0.6733542319749216, "grad_norm": 4.17587947845459, "learning_rate": 1.074e-06, "loss": 6.8412, "step": 2148 }, { "epoch": 0.673667711598746, "grad_norm": 4.389501571655273, "learning_rate": 1.0745e-06, "loss": 5.929, "step": 2149 }, { "epoch": 0.6739811912225705, "grad_norm": 3.766752004623413, "learning_rate": 1.075e-06, "loss": 7.428, "step": 2150 }, { "epoch": 0.6742946708463949, "grad_norm": 4.460997581481934, "learning_rate": 1.0755000000000001e-06, "loss": 8.5694, "step": 2151 }, { "epoch": 0.6746081504702194, "grad_norm": 3.7466864585876465, "learning_rate": 1.0760000000000002e-06, "loss": 6.5698, "step": 2152 }, { "epoch": 0.6749216300940439, "grad_norm": 5.410111904144287, "learning_rate": 1.0765e-06, "loss": 13.9963, "step": 2153 }, { "epoch": 0.6752351097178684, "grad_norm": 6.007702827453613, "learning_rate": 1.0770000000000002e-06, "loss": 11.1922, "step": 2154 }, { "epoch": 0.6755485893416928, "grad_norm": 3.324204206466675, "learning_rate": 1.0775e-06, "loss": 7.0282, "step": 2155 }, { "epoch": 0.6758620689655173, "grad_norm": 6.3619914054870605, "learning_rate": 1.078e-06, "loss": 8.3813, "step": 2156 }, { "epoch": 0.6761755485893417, "grad_norm": 5.453836441040039, "learning_rate": 1.0785000000000001e-06, "loss": 8.7658, "step": 2157 }, { "epoch": 0.6764890282131661, "grad_norm": 4.859471321105957, "learning_rate": 1.079e-06, "loss": 7.7485, "step": 2158 }, { "epoch": 0.6768025078369906, "grad_norm": 5.14132022857666, "learning_rate": 1.0795000000000002e-06, "loss": 10.0406, "step": 2159 }, { "epoch": 0.677115987460815, "grad_norm": 4.963798522949219, "learning_rate": 1.08e-06, "loss": 9.1328, "step": 2160 }, { "epoch": 0.6774294670846395, "grad_norm": 5.451442718505859, "learning_rate": 1.0805e-06, "loss": 7.1831, "step": 2161 }, { "epoch": 0.6777429467084639, "grad_norm": 4.889834403991699, "learning_rate": 1.0810000000000001e-06, "loss": 8.8367, "step": 2162 }, { "epoch": 0.6780564263322884, "grad_norm": 3.9947681427001953, "learning_rate": 1.0815000000000001e-06, "loss": 6.2821, "step": 2163 }, { "epoch": 0.6783699059561129, "grad_norm": 6.472747325897217, "learning_rate": 1.0820000000000002e-06, "loss": 12.1777, "step": 2164 }, { "epoch": 0.6786833855799373, "grad_norm": 3.9725639820098877, "learning_rate": 1.0825e-06, "loss": 7.2472, "step": 2165 }, { "epoch": 0.6789968652037618, "grad_norm": 4.466004371643066, "learning_rate": 1.083e-06, "loss": 7.1208, "step": 2166 }, { "epoch": 0.6793103448275862, "grad_norm": 8.549521446228027, "learning_rate": 1.0835e-06, "loss": 14.6298, "step": 2167 }, { "epoch": 0.6796238244514107, "grad_norm": 3.372896671295166, "learning_rate": 1.0840000000000001e-06, "loss": 6.0872, "step": 2168 }, { "epoch": 0.6799373040752351, "grad_norm": 4.142289161682129, "learning_rate": 1.0845000000000002e-06, "loss": 6.0638, "step": 2169 }, { "epoch": 0.6802507836990596, "grad_norm": 4.548405170440674, "learning_rate": 1.085e-06, "loss": 9.5781, "step": 2170 }, { "epoch": 0.680564263322884, "grad_norm": 4.637272357940674, "learning_rate": 1.0855e-06, "loss": 6.9768, "step": 2171 }, { "epoch": 0.6808777429467084, "grad_norm": 4.340859889984131, "learning_rate": 1.086e-06, "loss": 6.6182, "step": 2172 }, { "epoch": 0.6811912225705329, "grad_norm": 4.831057548522949, "learning_rate": 1.0865e-06, "loss": 8.176, "step": 2173 }, { "epoch": 0.6815047021943573, "grad_norm": 4.174441337585449, "learning_rate": 1.0870000000000001e-06, "loss": 9.2991, "step": 2174 }, { "epoch": 0.6818181818181818, "grad_norm": 4.800502300262451, "learning_rate": 1.0875000000000002e-06, "loss": 8.0076, "step": 2175 }, { "epoch": 0.6821316614420063, "grad_norm": 5.333560466766357, "learning_rate": 1.088e-06, "loss": 8.8977, "step": 2176 }, { "epoch": 0.6824451410658308, "grad_norm": 4.281090259552002, "learning_rate": 1.0885e-06, "loss": 6.8888, "step": 2177 }, { "epoch": 0.6827586206896552, "grad_norm": 7.420187950134277, "learning_rate": 1.089e-06, "loss": 12.2294, "step": 2178 }, { "epoch": 0.6830721003134796, "grad_norm": 5.143314838409424, "learning_rate": 1.0895000000000001e-06, "loss": 8.4805, "step": 2179 }, { "epoch": 0.6833855799373041, "grad_norm": 5.837392807006836, "learning_rate": 1.0900000000000002e-06, "loss": 9.5465, "step": 2180 }, { "epoch": 0.6836990595611285, "grad_norm": 4.490211009979248, "learning_rate": 1.0905e-06, "loss": 7.1872, "step": 2181 }, { "epoch": 0.684012539184953, "grad_norm": 5.512997150421143, "learning_rate": 1.091e-06, "loss": 10.6141, "step": 2182 }, { "epoch": 0.6843260188087774, "grad_norm": 5.654632568359375, "learning_rate": 1.0915e-06, "loss": 10.529, "step": 2183 }, { "epoch": 0.6846394984326019, "grad_norm": 3.9920573234558105, "learning_rate": 1.092e-06, "loss": 7.4162, "step": 2184 }, { "epoch": 0.6849529780564263, "grad_norm": 9.621570587158203, "learning_rate": 1.0925000000000001e-06, "loss": 17.1348, "step": 2185 }, { "epoch": 0.6852664576802507, "grad_norm": 4.3478007316589355, "learning_rate": 1.093e-06, "loss": 7.5085, "step": 2186 }, { "epoch": 0.6855799373040752, "grad_norm": 4.477773666381836, "learning_rate": 1.0935000000000002e-06, "loss": 8.6712, "step": 2187 }, { "epoch": 0.6858934169278997, "grad_norm": 4.364738941192627, "learning_rate": 1.094e-06, "loss": 8.0755, "step": 2188 }, { "epoch": 0.6862068965517242, "grad_norm": 5.307512283325195, "learning_rate": 1.0945e-06, "loss": 9.8482, "step": 2189 }, { "epoch": 0.6865203761755486, "grad_norm": 5.022215366363525, "learning_rate": 1.095e-06, "loss": 9.1433, "step": 2190 }, { "epoch": 0.6868338557993731, "grad_norm": 5.021544933319092, "learning_rate": 1.0955e-06, "loss": 8.4613, "step": 2191 }, { "epoch": 0.6871473354231975, "grad_norm": 3.7242307662963867, "learning_rate": 1.0960000000000002e-06, "loss": 6.4761, "step": 2192 }, { "epoch": 0.687460815047022, "grad_norm": 4.364880561828613, "learning_rate": 1.0965e-06, "loss": 8.0514, "step": 2193 }, { "epoch": 0.6877742946708464, "grad_norm": 6.270079612731934, "learning_rate": 1.097e-06, "loss": 7.7713, "step": 2194 }, { "epoch": 0.6880877742946708, "grad_norm": 4.493875026702881, "learning_rate": 1.0975e-06, "loss": 7.1808, "step": 2195 }, { "epoch": 0.6884012539184953, "grad_norm": 5.605881214141846, "learning_rate": 1.0980000000000001e-06, "loss": 8.8931, "step": 2196 }, { "epoch": 0.6887147335423197, "grad_norm": 5.781085014343262, "learning_rate": 1.0985000000000002e-06, "loss": 11.2424, "step": 2197 }, { "epoch": 0.6890282131661442, "grad_norm": 6.997844219207764, "learning_rate": 1.099e-06, "loss": 14.7351, "step": 2198 }, { "epoch": 0.6893416927899687, "grad_norm": 5.427860260009766, "learning_rate": 1.0995000000000002e-06, "loss": 8.0223, "step": 2199 }, { "epoch": 0.6896551724137931, "grad_norm": 5.856205940246582, "learning_rate": 1.1e-06, "loss": 10.776, "step": 2200 }, { "epoch": 0.6899686520376176, "grad_norm": 5.698307037353516, "learning_rate": 1.1005e-06, "loss": 10.4188, "step": 2201 }, { "epoch": 0.690282131661442, "grad_norm": 3.5812063217163086, "learning_rate": 1.1010000000000001e-06, "loss": 5.7494, "step": 2202 }, { "epoch": 0.6905956112852665, "grad_norm": 6.817094802856445, "learning_rate": 1.1015000000000002e-06, "loss": 15.1615, "step": 2203 }, { "epoch": 0.6909090909090909, "grad_norm": 5.15540075302124, "learning_rate": 1.1020000000000002e-06, "loss": 7.7424, "step": 2204 }, { "epoch": 0.6912225705329154, "grad_norm": 5.062628746032715, "learning_rate": 1.1025e-06, "loss": 6.6624, "step": 2205 }, { "epoch": 0.6915360501567398, "grad_norm": 4.202182769775391, "learning_rate": 1.103e-06, "loss": 7.4259, "step": 2206 }, { "epoch": 0.6918495297805642, "grad_norm": 4.714681148529053, "learning_rate": 1.1035000000000001e-06, "loss": 9.4869, "step": 2207 }, { "epoch": 0.6921630094043887, "grad_norm": 5.397708415985107, "learning_rate": 1.1040000000000001e-06, "loss": 8.4847, "step": 2208 }, { "epoch": 0.6924764890282131, "grad_norm": 6.186098098754883, "learning_rate": 1.1045000000000002e-06, "loss": 9.6202, "step": 2209 }, { "epoch": 0.6927899686520376, "grad_norm": 5.049215793609619, "learning_rate": 1.105e-06, "loss": 8.4358, "step": 2210 }, { "epoch": 0.6931034482758621, "grad_norm": 4.837051868438721, "learning_rate": 1.1055e-06, "loss": 7.3944, "step": 2211 }, { "epoch": 0.6934169278996866, "grad_norm": 5.781193256378174, "learning_rate": 1.106e-06, "loss": 10.1386, "step": 2212 }, { "epoch": 0.693730407523511, "grad_norm": 4.360887050628662, "learning_rate": 1.1065000000000001e-06, "loss": 7.0723, "step": 2213 }, { "epoch": 0.6940438871473354, "grad_norm": 4.668460845947266, "learning_rate": 1.1070000000000002e-06, "loss": 8.563, "step": 2214 }, { "epoch": 0.6943573667711599, "grad_norm": 5.356121063232422, "learning_rate": 1.1075000000000002e-06, "loss": 9.5143, "step": 2215 }, { "epoch": 0.6946708463949843, "grad_norm": 4.979715824127197, "learning_rate": 1.108e-06, "loss": 8.1418, "step": 2216 }, { "epoch": 0.6949843260188088, "grad_norm": 4.595643520355225, "learning_rate": 1.1085e-06, "loss": 9.48, "step": 2217 }, { "epoch": 0.6952978056426332, "grad_norm": 5.082688331604004, "learning_rate": 1.109e-06, "loss": 7.4409, "step": 2218 }, { "epoch": 0.6956112852664577, "grad_norm": 5.961426258087158, "learning_rate": 1.1095e-06, "loss": 9.6929, "step": 2219 }, { "epoch": 0.6959247648902821, "grad_norm": 4.507264137268066, "learning_rate": 1.1100000000000002e-06, "loss": 9.0472, "step": 2220 }, { "epoch": 0.6962382445141065, "grad_norm": 6.450530052185059, "learning_rate": 1.1105e-06, "loss": 11.311, "step": 2221 }, { "epoch": 0.696551724137931, "grad_norm": 6.884239673614502, "learning_rate": 1.111e-06, "loss": 9.9046, "step": 2222 }, { "epoch": 0.6968652037617555, "grad_norm": 4.918006420135498, "learning_rate": 1.1115e-06, "loss": 7.7746, "step": 2223 }, { "epoch": 0.69717868338558, "grad_norm": 3.9736998081207275, "learning_rate": 1.1120000000000001e-06, "loss": 8.2501, "step": 2224 }, { "epoch": 0.6974921630094044, "grad_norm": 5.021994590759277, "learning_rate": 1.1125000000000001e-06, "loss": 9.8031, "step": 2225 }, { "epoch": 0.6978056426332289, "grad_norm": 4.272434711456299, "learning_rate": 1.113e-06, "loss": 6.6057, "step": 2226 }, { "epoch": 0.6981191222570533, "grad_norm": 4.612613677978516, "learning_rate": 1.1135000000000002e-06, "loss": 8.5593, "step": 2227 }, { "epoch": 0.6984326018808777, "grad_norm": 6.331014633178711, "learning_rate": 1.114e-06, "loss": 8.7987, "step": 2228 }, { "epoch": 0.6987460815047022, "grad_norm": 6.129283905029297, "learning_rate": 1.1145e-06, "loss": 8.4672, "step": 2229 }, { "epoch": 0.6990595611285266, "grad_norm": 4.61047887802124, "learning_rate": 1.1150000000000001e-06, "loss": 9.5344, "step": 2230 }, { "epoch": 0.6993730407523511, "grad_norm": 4.625089168548584, "learning_rate": 1.1155e-06, "loss": 8.4914, "step": 2231 }, { "epoch": 0.6996865203761755, "grad_norm": 7.176419258117676, "learning_rate": 1.1160000000000002e-06, "loss": 10.456, "step": 2232 }, { "epoch": 0.7, "grad_norm": 4.757896900177002, "learning_rate": 1.1165e-06, "loss": 8.0939, "step": 2233 }, { "epoch": 0.7003134796238244, "grad_norm": 4.703888893127441, "learning_rate": 1.117e-06, "loss": 8.9324, "step": 2234 }, { "epoch": 0.700626959247649, "grad_norm": 5.082801342010498, "learning_rate": 1.1175e-06, "loss": 7.9512, "step": 2235 }, { "epoch": 0.7009404388714734, "grad_norm": 7.074932098388672, "learning_rate": 1.1180000000000001e-06, "loss": 10.5051, "step": 2236 }, { "epoch": 0.7012539184952978, "grad_norm": 5.920054912567139, "learning_rate": 1.1185000000000002e-06, "loss": 10.0657, "step": 2237 }, { "epoch": 0.7015673981191223, "grad_norm": 4.017765045166016, "learning_rate": 1.119e-06, "loss": 7.4091, "step": 2238 }, { "epoch": 0.7018808777429467, "grad_norm": 4.405153274536133, "learning_rate": 1.1195e-06, "loss": 8.9839, "step": 2239 }, { "epoch": 0.7021943573667712, "grad_norm": 5.551645755767822, "learning_rate": 1.12e-06, "loss": 8.2674, "step": 2240 }, { "epoch": 0.7025078369905956, "grad_norm": 5.920106887817383, "learning_rate": 1.1205000000000001e-06, "loss": 10.3074, "step": 2241 }, { "epoch": 0.70282131661442, "grad_norm": 6.09196138381958, "learning_rate": 1.1210000000000002e-06, "loss": 9.587, "step": 2242 }, { "epoch": 0.7031347962382445, "grad_norm": 6.698037624359131, "learning_rate": 1.1215000000000002e-06, "loss": 10.1409, "step": 2243 }, { "epoch": 0.7034482758620689, "grad_norm": 6.3669962882995605, "learning_rate": 1.122e-06, "loss": 10.6705, "step": 2244 }, { "epoch": 0.7037617554858934, "grad_norm": 4.820910930633545, "learning_rate": 1.1225e-06, "loss": 10.7634, "step": 2245 }, { "epoch": 0.7040752351097179, "grad_norm": 6.152073860168457, "learning_rate": 1.123e-06, "loss": 9.9762, "step": 2246 }, { "epoch": 0.7043887147335424, "grad_norm": 6.534127235412598, "learning_rate": 1.1235000000000001e-06, "loss": 9.1076, "step": 2247 }, { "epoch": 0.7047021943573668, "grad_norm": 4.584049224853516, "learning_rate": 1.1240000000000002e-06, "loss": 6.3946, "step": 2248 }, { "epoch": 0.7050156739811912, "grad_norm": 6.2237653732299805, "learning_rate": 1.1245e-06, "loss": 10.3125, "step": 2249 }, { "epoch": 0.7053291536050157, "grad_norm": 7.915231227874756, "learning_rate": 1.125e-06, "loss": 11.7329, "step": 2250 }, { "epoch": 0.7056426332288401, "grad_norm": 5.413792610168457, "learning_rate": 1.1255e-06, "loss": 7.3437, "step": 2251 }, { "epoch": 0.7059561128526646, "grad_norm": 6.203031063079834, "learning_rate": 1.126e-06, "loss": 8.8131, "step": 2252 }, { "epoch": 0.706269592476489, "grad_norm": 5.079891681671143, "learning_rate": 1.1265000000000001e-06, "loss": 8.0062, "step": 2253 }, { "epoch": 0.7065830721003135, "grad_norm": 5.417928218841553, "learning_rate": 1.127e-06, "loss": 7.8545, "step": 2254 }, { "epoch": 0.7068965517241379, "grad_norm": 8.853837966918945, "learning_rate": 1.1275000000000002e-06, "loss": 12.7805, "step": 2255 }, { "epoch": 0.7072100313479623, "grad_norm": 5.80386209487915, "learning_rate": 1.128e-06, "loss": 10.7414, "step": 2256 }, { "epoch": 0.7075235109717868, "grad_norm": 5.670801639556885, "learning_rate": 1.1285e-06, "loss": 8.2194, "step": 2257 }, { "epoch": 0.7078369905956113, "grad_norm": 3.81965708732605, "learning_rate": 1.1290000000000001e-06, "loss": 6.5429, "step": 2258 }, { "epoch": 0.7081504702194358, "grad_norm": 5.524906635284424, "learning_rate": 1.1295e-06, "loss": 10.1136, "step": 2259 }, { "epoch": 0.7084639498432602, "grad_norm": 7.865563869476318, "learning_rate": 1.1300000000000002e-06, "loss": 11.7016, "step": 2260 }, { "epoch": 0.7087774294670847, "grad_norm": 4.466341495513916, "learning_rate": 1.1305e-06, "loss": 6.4666, "step": 2261 }, { "epoch": 0.7090909090909091, "grad_norm": 4.703956127166748, "learning_rate": 1.131e-06, "loss": 7.191, "step": 2262 }, { "epoch": 0.7094043887147335, "grad_norm": 4.839312553405762, "learning_rate": 1.1315e-06, "loss": 8.4636, "step": 2263 }, { "epoch": 0.709717868338558, "grad_norm": 6.889242172241211, "learning_rate": 1.1320000000000001e-06, "loss": 9.0166, "step": 2264 }, { "epoch": 0.7100313479623824, "grad_norm": 5.956830978393555, "learning_rate": 1.1325000000000002e-06, "loss": 7.6537, "step": 2265 }, { "epoch": 0.7103448275862069, "grad_norm": 3.948118209838867, "learning_rate": 1.133e-06, "loss": 6.8683, "step": 2266 }, { "epoch": 0.7106583072100313, "grad_norm": 5.981534004211426, "learning_rate": 1.1335000000000002e-06, "loss": 9.1839, "step": 2267 }, { "epoch": 0.7109717868338558, "grad_norm": 6.075943470001221, "learning_rate": 1.134e-06, "loss": 8.9451, "step": 2268 }, { "epoch": 0.7112852664576802, "grad_norm": 7.1982879638671875, "learning_rate": 1.1345000000000001e-06, "loss": 10.2837, "step": 2269 }, { "epoch": 0.7115987460815048, "grad_norm": 4.65557861328125, "learning_rate": 1.1350000000000001e-06, "loss": 7.0759, "step": 2270 }, { "epoch": 0.7119122257053292, "grad_norm": 7.2853007316589355, "learning_rate": 1.1355e-06, "loss": 10.6811, "step": 2271 }, { "epoch": 0.7122257053291536, "grad_norm": 4.625248432159424, "learning_rate": 1.1360000000000002e-06, "loss": 6.6552, "step": 2272 }, { "epoch": 0.7125391849529781, "grad_norm": 5.515805244445801, "learning_rate": 1.1365e-06, "loss": 11.0095, "step": 2273 }, { "epoch": 0.7128526645768025, "grad_norm": 8.214932441711426, "learning_rate": 1.137e-06, "loss": 9.5016, "step": 2274 }, { "epoch": 0.713166144200627, "grad_norm": 4.511641979217529, "learning_rate": 1.1375000000000001e-06, "loss": 9.456, "step": 2275 }, { "epoch": 0.7134796238244514, "grad_norm": 4.926652908325195, "learning_rate": 1.1380000000000002e-06, "loss": 6.4554, "step": 2276 }, { "epoch": 0.7137931034482758, "grad_norm": 5.701582908630371, "learning_rate": 1.1385000000000002e-06, "loss": 7.8086, "step": 2277 }, { "epoch": 0.7141065830721003, "grad_norm": 5.361356735229492, "learning_rate": 1.139e-06, "loss": 5.5772, "step": 2278 }, { "epoch": 0.7144200626959247, "grad_norm": 5.268955707550049, "learning_rate": 1.1395e-06, "loss": 6.9328, "step": 2279 }, { "epoch": 0.7147335423197492, "grad_norm": 6.054418087005615, "learning_rate": 1.14e-06, "loss": 11.8742, "step": 2280 }, { "epoch": 0.7150470219435736, "grad_norm": 6.118603229522705, "learning_rate": 1.1405000000000001e-06, "loss": 8.7145, "step": 2281 }, { "epoch": 0.7153605015673982, "grad_norm": 5.574398517608643, "learning_rate": 1.141e-06, "loss": 8.354, "step": 2282 }, { "epoch": 0.7156739811912226, "grad_norm": 8.102057456970215, "learning_rate": 1.1415000000000002e-06, "loss": 9.6707, "step": 2283 }, { "epoch": 0.715987460815047, "grad_norm": 6.044636249542236, "learning_rate": 1.142e-06, "loss": 8.4545, "step": 2284 }, { "epoch": 0.7163009404388715, "grad_norm": 5.610233783721924, "learning_rate": 1.1425e-06, "loss": 6.494, "step": 2285 }, { "epoch": 0.7166144200626959, "grad_norm": 7.252960681915283, "learning_rate": 1.1430000000000001e-06, "loss": 9.9181, "step": 2286 }, { "epoch": 0.7169278996865204, "grad_norm": 7.041637897491455, "learning_rate": 1.1435e-06, "loss": 9.8605, "step": 2287 }, { "epoch": 0.7172413793103448, "grad_norm": 7.529675006866455, "learning_rate": 1.1440000000000002e-06, "loss": 9.2948, "step": 2288 }, { "epoch": 0.7175548589341693, "grad_norm": 6.867233753204346, "learning_rate": 1.1445e-06, "loss": 10.663, "step": 2289 }, { "epoch": 0.7178683385579937, "grad_norm": 4.269197463989258, "learning_rate": 1.145e-06, "loss": 6.5775, "step": 2290 }, { "epoch": 0.7181818181818181, "grad_norm": 5.0431060791015625, "learning_rate": 1.1455e-06, "loss": 8.0019, "step": 2291 }, { "epoch": 0.7184952978056426, "grad_norm": 5.916482925415039, "learning_rate": 1.1460000000000001e-06, "loss": 9.4506, "step": 2292 }, { "epoch": 0.7188087774294671, "grad_norm": 8.643074035644531, "learning_rate": 1.1465000000000002e-06, "loss": 11.3396, "step": 2293 }, { "epoch": 0.7191222570532916, "grad_norm": 4.569228649139404, "learning_rate": 1.147e-06, "loss": 6.5566, "step": 2294 }, { "epoch": 0.719435736677116, "grad_norm": 5.326075077056885, "learning_rate": 1.1475000000000002e-06, "loss": 8.4556, "step": 2295 }, { "epoch": 0.7197492163009405, "grad_norm": 5.192141532897949, "learning_rate": 1.148e-06, "loss": 8.5234, "step": 2296 }, { "epoch": 0.7200626959247649, "grad_norm": 4.826154708862305, "learning_rate": 1.1485e-06, "loss": 6.925, "step": 2297 }, { "epoch": 0.7203761755485893, "grad_norm": 6.822600841522217, "learning_rate": 1.1490000000000001e-06, "loss": 12.9681, "step": 2298 }, { "epoch": 0.7206896551724138, "grad_norm": 7.6909003257751465, "learning_rate": 1.1495e-06, "loss": 10.0613, "step": 2299 }, { "epoch": 0.7210031347962382, "grad_norm": 6.721874713897705, "learning_rate": 1.1500000000000002e-06, "loss": 11.6405, "step": 2300 }, { "epoch": 0.7213166144200627, "grad_norm": 5.8059163093566895, "learning_rate": 1.1505e-06, "loss": 8.9701, "step": 2301 }, { "epoch": 0.7216300940438871, "grad_norm": 8.074931144714355, "learning_rate": 1.151e-06, "loss": 9.3916, "step": 2302 }, { "epoch": 0.7219435736677116, "grad_norm": 7.6194987297058105, "learning_rate": 1.1515000000000001e-06, "loss": 11.8245, "step": 2303 }, { "epoch": 0.722257053291536, "grad_norm": 4.169922351837158, "learning_rate": 1.1520000000000002e-06, "loss": 6.3245, "step": 2304 }, { "epoch": 0.7225705329153606, "grad_norm": 7.672272682189941, "learning_rate": 1.1525000000000002e-06, "loss": 10.7136, "step": 2305 }, { "epoch": 0.722884012539185, "grad_norm": 6.1154656410217285, "learning_rate": 1.153e-06, "loss": 10.13, "step": 2306 }, { "epoch": 0.7231974921630094, "grad_norm": 5.131744861602783, "learning_rate": 1.1535e-06, "loss": 6.4488, "step": 2307 }, { "epoch": 0.7235109717868339, "grad_norm": 4.942484378814697, "learning_rate": 1.154e-06, "loss": 7.9078, "step": 2308 }, { "epoch": 0.7238244514106583, "grad_norm": 6.5893449783325195, "learning_rate": 1.1545000000000001e-06, "loss": 10.2193, "step": 2309 }, { "epoch": 0.7241379310344828, "grad_norm": 5.4219136238098145, "learning_rate": 1.1550000000000002e-06, "loss": 7.0708, "step": 2310 }, { "epoch": 0.7244514106583072, "grad_norm": 8.25130558013916, "learning_rate": 1.1555e-06, "loss": 10.7496, "step": 2311 }, { "epoch": 0.7247648902821316, "grad_norm": 4.836966514587402, "learning_rate": 1.156e-06, "loss": 6.6638, "step": 2312 }, { "epoch": 0.7250783699059561, "grad_norm": 7.104246139526367, "learning_rate": 1.1565e-06, "loss": 8.5781, "step": 2313 }, { "epoch": 0.7253918495297805, "grad_norm": 5.00009822845459, "learning_rate": 1.157e-06, "loss": 6.4871, "step": 2314 }, { "epoch": 0.725705329153605, "grad_norm": 4.757653713226318, "learning_rate": 1.1575000000000001e-06, "loss": 7.6189, "step": 2315 }, { "epoch": 0.7260188087774294, "grad_norm": 6.125999450683594, "learning_rate": 1.1580000000000002e-06, "loss": 7.1317, "step": 2316 }, { "epoch": 0.726332288401254, "grad_norm": 6.397932052612305, "learning_rate": 1.1585e-06, "loss": 7.738, "step": 2317 }, { "epoch": 0.7266457680250784, "grad_norm": 5.191574573516846, "learning_rate": 1.159e-06, "loss": 7.8031, "step": 2318 }, { "epoch": 0.7269592476489029, "grad_norm": 6.020807266235352, "learning_rate": 1.1595e-06, "loss": 9.9713, "step": 2319 }, { "epoch": 0.7272727272727273, "grad_norm": 4.571470260620117, "learning_rate": 1.1600000000000001e-06, "loss": 5.8561, "step": 2320 }, { "epoch": 0.7275862068965517, "grad_norm": 6.151414394378662, "learning_rate": 1.1605000000000002e-06, "loss": 8.3401, "step": 2321 }, { "epoch": 0.7278996865203762, "grad_norm": 5.342029094696045, "learning_rate": 1.161e-06, "loss": 7.4286, "step": 2322 }, { "epoch": 0.7282131661442006, "grad_norm": 8.054902076721191, "learning_rate": 1.1615000000000002e-06, "loss": 11.2857, "step": 2323 }, { "epoch": 0.7285266457680251, "grad_norm": 8.631732940673828, "learning_rate": 1.162e-06, "loss": 11.7461, "step": 2324 }, { "epoch": 0.7288401253918495, "grad_norm": 5.86944580078125, "learning_rate": 1.1625e-06, "loss": 8.2415, "step": 2325 }, { "epoch": 0.729153605015674, "grad_norm": 6.775248050689697, "learning_rate": 1.1630000000000001e-06, "loss": 9.1465, "step": 2326 }, { "epoch": 0.7294670846394984, "grad_norm": 5.014388561248779, "learning_rate": 1.1635e-06, "loss": 9.4963, "step": 2327 }, { "epoch": 0.7297805642633229, "grad_norm": 6.137758731842041, "learning_rate": 1.1640000000000002e-06, "loss": 7.47, "step": 2328 }, { "epoch": 0.7300940438871474, "grad_norm": 7.857049942016602, "learning_rate": 1.1645e-06, "loss": 12.0102, "step": 2329 }, { "epoch": 0.7304075235109718, "grad_norm": 5.5146002769470215, "learning_rate": 1.165e-06, "loss": 7.7632, "step": 2330 }, { "epoch": 0.7307210031347963, "grad_norm": 5.327179431915283, "learning_rate": 1.1655000000000001e-06, "loss": 9.2846, "step": 2331 }, { "epoch": 0.7310344827586207, "grad_norm": 5.694344997406006, "learning_rate": 1.1660000000000001e-06, "loss": 8.7609, "step": 2332 }, { "epoch": 0.7313479623824451, "grad_norm": 6.549810886383057, "learning_rate": 1.1665000000000002e-06, "loss": 8.8996, "step": 2333 }, { "epoch": 0.7316614420062696, "grad_norm": 4.765985488891602, "learning_rate": 1.167e-06, "loss": 7.4456, "step": 2334 }, { "epoch": 0.731974921630094, "grad_norm": 4.749044418334961, "learning_rate": 1.1675000000000003e-06, "loss": 6.8727, "step": 2335 }, { "epoch": 0.7322884012539185, "grad_norm": 6.0622711181640625, "learning_rate": 1.168e-06, "loss": 8.9842, "step": 2336 }, { "epoch": 0.7326018808777429, "grad_norm": 5.587085723876953, "learning_rate": 1.1685000000000001e-06, "loss": 7.9121, "step": 2337 }, { "epoch": 0.7329153605015674, "grad_norm": 6.6463189125061035, "learning_rate": 1.1690000000000002e-06, "loss": 7.1128, "step": 2338 }, { "epoch": 0.7332288401253918, "grad_norm": 6.599662780761719, "learning_rate": 1.1695e-06, "loss": 8.0378, "step": 2339 }, { "epoch": 0.7335423197492164, "grad_norm": 5.061799049377441, "learning_rate": 1.1700000000000002e-06, "loss": 6.4478, "step": 2340 }, { "epoch": 0.7338557993730408, "grad_norm": 5.785806655883789, "learning_rate": 1.1705e-06, "loss": 10.5814, "step": 2341 }, { "epoch": 0.7341692789968652, "grad_norm": 5.782971382141113, "learning_rate": 1.171e-06, "loss": 6.4085, "step": 2342 }, { "epoch": 0.7344827586206897, "grad_norm": 5.601132392883301, "learning_rate": 1.1715000000000001e-06, "loss": 9.9022, "step": 2343 }, { "epoch": 0.7347962382445141, "grad_norm": 5.016490459442139, "learning_rate": 1.1720000000000002e-06, "loss": 7.146, "step": 2344 }, { "epoch": 0.7351097178683386, "grad_norm": 5.3097310066223145, "learning_rate": 1.1725e-06, "loss": 7.8484, "step": 2345 }, { "epoch": 0.735423197492163, "grad_norm": 5.761673927307129, "learning_rate": 1.173e-06, "loss": 6.9101, "step": 2346 }, { "epoch": 0.7357366771159874, "grad_norm": 6.155106544494629, "learning_rate": 1.1735e-06, "loss": 7.6222, "step": 2347 }, { "epoch": 0.7360501567398119, "grad_norm": 5.324407577514648, "learning_rate": 1.1740000000000001e-06, "loss": 7.18, "step": 2348 }, { "epoch": 0.7363636363636363, "grad_norm": 5.507112503051758, "learning_rate": 1.1745000000000001e-06, "loss": 7.4581, "step": 2349 }, { "epoch": 0.7366771159874608, "grad_norm": 6.135045528411865, "learning_rate": 1.175e-06, "loss": 8.3784, "step": 2350 }, { "epoch": 0.7369905956112852, "grad_norm": 5.607395172119141, "learning_rate": 1.1755e-06, "loss": 8.0636, "step": 2351 }, { "epoch": 0.7373040752351098, "grad_norm": 4.155460834503174, "learning_rate": 1.176e-06, "loss": 7.7036, "step": 2352 }, { "epoch": 0.7376175548589342, "grad_norm": 6.254929542541504, "learning_rate": 1.1765e-06, "loss": 7.4921, "step": 2353 }, { "epoch": 0.7379310344827587, "grad_norm": 7.738031387329102, "learning_rate": 1.1770000000000001e-06, "loss": 8.3682, "step": 2354 }, { "epoch": 0.7382445141065831, "grad_norm": 4.532285213470459, "learning_rate": 1.1775e-06, "loss": 6.7809, "step": 2355 }, { "epoch": 0.7385579937304075, "grad_norm": 5.4931640625, "learning_rate": 1.1780000000000002e-06, "loss": 8.2121, "step": 2356 }, { "epoch": 0.738871473354232, "grad_norm": 5.828548908233643, "learning_rate": 1.1785e-06, "loss": 7.4208, "step": 2357 }, { "epoch": 0.7391849529780564, "grad_norm": 8.793325424194336, "learning_rate": 1.179e-06, "loss": 10.5305, "step": 2358 }, { "epoch": 0.7394984326018809, "grad_norm": 7.540804386138916, "learning_rate": 1.1795e-06, "loss": 10.3455, "step": 2359 }, { "epoch": 0.7398119122257053, "grad_norm": 7.395792007446289, "learning_rate": 1.1800000000000001e-06, "loss": 8.6099, "step": 2360 }, { "epoch": 0.7401253918495297, "grad_norm": 7.034842491149902, "learning_rate": 1.1805000000000002e-06, "loss": 12.9946, "step": 2361 }, { "epoch": 0.7404388714733542, "grad_norm": 6.157832145690918, "learning_rate": 1.181e-06, "loss": 9.0294, "step": 2362 }, { "epoch": 0.7407523510971786, "grad_norm": 5.958568096160889, "learning_rate": 1.1815000000000002e-06, "loss": 9.4815, "step": 2363 }, { "epoch": 0.7410658307210032, "grad_norm": 5.166853427886963, "learning_rate": 1.182e-06, "loss": 7.8122, "step": 2364 }, { "epoch": 0.7413793103448276, "grad_norm": 6.027432441711426, "learning_rate": 1.1825000000000001e-06, "loss": 9.6608, "step": 2365 }, { "epoch": 0.7416927899686521, "grad_norm": 8.379853248596191, "learning_rate": 1.1830000000000002e-06, "loss": 10.2204, "step": 2366 }, { "epoch": 0.7420062695924765, "grad_norm": 6.882002353668213, "learning_rate": 1.1835e-06, "loss": 9.1155, "step": 2367 }, { "epoch": 0.742319749216301, "grad_norm": 5.61342191696167, "learning_rate": 1.1840000000000002e-06, "loss": 9.3164, "step": 2368 }, { "epoch": 0.7426332288401254, "grad_norm": 7.321317195892334, "learning_rate": 1.1845e-06, "loss": 7.8616, "step": 2369 }, { "epoch": 0.7429467084639498, "grad_norm": 5.494234561920166, "learning_rate": 1.185e-06, "loss": 6.7372, "step": 2370 }, { "epoch": 0.7432601880877743, "grad_norm": 6.378052234649658, "learning_rate": 1.1855000000000001e-06, "loss": 8.2405, "step": 2371 }, { "epoch": 0.7435736677115987, "grad_norm": 9.557308197021484, "learning_rate": 1.1860000000000002e-06, "loss": 12.0775, "step": 2372 }, { "epoch": 0.7438871473354232, "grad_norm": 7.571987152099609, "learning_rate": 1.1865000000000002e-06, "loss": 9.8201, "step": 2373 }, { "epoch": 0.7442006269592476, "grad_norm": 5.4964213371276855, "learning_rate": 1.187e-06, "loss": 6.8091, "step": 2374 }, { "epoch": 0.7445141065830722, "grad_norm": 5.78679895401001, "learning_rate": 1.1875e-06, "loss": 7.4889, "step": 2375 }, { "epoch": 0.7448275862068966, "grad_norm": 5.633636474609375, "learning_rate": 1.188e-06, "loss": 6.6415, "step": 2376 }, { "epoch": 0.745141065830721, "grad_norm": 6.5983428955078125, "learning_rate": 1.1885000000000001e-06, "loss": 7.7574, "step": 2377 }, { "epoch": 0.7454545454545455, "grad_norm": 9.485814094543457, "learning_rate": 1.1890000000000002e-06, "loss": 12.1369, "step": 2378 }, { "epoch": 0.7457680250783699, "grad_norm": 6.415633201599121, "learning_rate": 1.1895e-06, "loss": 7.9947, "step": 2379 }, { "epoch": 0.7460815047021944, "grad_norm": 6.807877063751221, "learning_rate": 1.19e-06, "loss": 8.5465, "step": 2380 }, { "epoch": 0.7463949843260188, "grad_norm": 6.27053165435791, "learning_rate": 1.1905e-06, "loss": 7.1279, "step": 2381 }, { "epoch": 0.7467084639498432, "grad_norm": 5.453633785247803, "learning_rate": 1.1910000000000001e-06, "loss": 7.1849, "step": 2382 }, { "epoch": 0.7470219435736677, "grad_norm": 6.075678825378418, "learning_rate": 1.1915000000000002e-06, "loss": 8.3007, "step": 2383 }, { "epoch": 0.7473354231974921, "grad_norm": 10.42825984954834, "learning_rate": 1.1920000000000002e-06, "loss": 10.672, "step": 2384 }, { "epoch": 0.7476489028213166, "grad_norm": 5.732131481170654, "learning_rate": 1.1925e-06, "loss": 6.1683, "step": 2385 }, { "epoch": 0.747962382445141, "grad_norm": 6.508482456207275, "learning_rate": 1.193e-06, "loss": 9.2361, "step": 2386 }, { "epoch": 0.7482758620689656, "grad_norm": 5.837035179138184, "learning_rate": 1.1935e-06, "loss": 7.1001, "step": 2387 }, { "epoch": 0.74858934169279, "grad_norm": 9.754945755004883, "learning_rate": 1.1940000000000001e-06, "loss": 10.2582, "step": 2388 }, { "epoch": 0.7489028213166145, "grad_norm": 5.980183124542236, "learning_rate": 1.1945000000000002e-06, "loss": 8.5382, "step": 2389 }, { "epoch": 0.7492163009404389, "grad_norm": 5.945464134216309, "learning_rate": 1.195e-06, "loss": 9.5149, "step": 2390 }, { "epoch": 0.7495297805642633, "grad_norm": 6.858036041259766, "learning_rate": 1.1955e-06, "loss": 9.1617, "step": 2391 }, { "epoch": 0.7498432601880878, "grad_norm": 7.403122901916504, "learning_rate": 1.196e-06, "loss": 9.4436, "step": 2392 }, { "epoch": 0.7501567398119122, "grad_norm": 5.871950626373291, "learning_rate": 1.1965000000000001e-06, "loss": 7.0343, "step": 2393 }, { "epoch": 0.7504702194357367, "grad_norm": 9.108241081237793, "learning_rate": 1.1970000000000001e-06, "loss": 11.0054, "step": 2394 }, { "epoch": 0.7504702194357367, "eval_loss": 27.862751007080078, "eval_runtime": 20.9206, "eval_samples_per_second": 128.438, "eval_steps_per_second": 8.03, "step": 2394 }, { "epoch": 0.7507836990595611, "grad_norm": 6.919027805328369, "learning_rate": 1.1975e-06, "loss": 8.3136, "step": 2395 }, { "epoch": 0.7510971786833855, "grad_norm": 5.883973598480225, "learning_rate": 1.1980000000000002e-06, "loss": 7.4282, "step": 2396 }, { "epoch": 0.75141065830721, "grad_norm": 6.176295280456543, "learning_rate": 1.1985e-06, "loss": 8.8492, "step": 2397 }, { "epoch": 0.7517241379310344, "grad_norm": 6.810075283050537, "learning_rate": 1.199e-06, "loss": 9.706, "step": 2398 }, { "epoch": 0.752037617554859, "grad_norm": 5.519807815551758, "learning_rate": 1.1995000000000001e-06, "loss": 8.259, "step": 2399 }, { "epoch": 0.7523510971786834, "grad_norm": 4.797094821929932, "learning_rate": 1.2000000000000002e-06, "loss": 7.2727, "step": 2400 }, { "epoch": 0.7526645768025079, "grad_norm": 6.261948108673096, "learning_rate": 1.2005000000000002e-06, "loss": 9.45, "step": 2401 }, { "epoch": 0.7529780564263323, "grad_norm": 5.316954135894775, "learning_rate": 1.201e-06, "loss": 6.7442, "step": 2402 }, { "epoch": 0.7532915360501568, "grad_norm": 6.721187591552734, "learning_rate": 1.2015000000000003e-06, "loss": 7.6002, "step": 2403 }, { "epoch": 0.7536050156739812, "grad_norm": 5.482544898986816, "learning_rate": 1.202e-06, "loss": 6.7844, "step": 2404 }, { "epoch": 0.7539184952978056, "grad_norm": 5.34820556640625, "learning_rate": 1.2025000000000001e-06, "loss": 7.4373, "step": 2405 }, { "epoch": 0.7542319749216301, "grad_norm": 7.171077728271484, "learning_rate": 1.2030000000000002e-06, "loss": 7.6944, "step": 2406 }, { "epoch": 0.7545454545454545, "grad_norm": 5.950673580169678, "learning_rate": 1.2035e-06, "loss": 8.4961, "step": 2407 }, { "epoch": 0.754858934169279, "grad_norm": 7.122339725494385, "learning_rate": 1.204e-06, "loss": 8.6441, "step": 2408 }, { "epoch": 0.7551724137931034, "grad_norm": 8.164888381958008, "learning_rate": 1.2045e-06, "loss": 10.9008, "step": 2409 }, { "epoch": 0.7554858934169278, "grad_norm": 8.712209701538086, "learning_rate": 1.2050000000000001e-06, "loss": 11.4492, "step": 2410 }, { "epoch": 0.7557993730407524, "grad_norm": 5.684695720672607, "learning_rate": 1.2055000000000001e-06, "loss": 7.7799, "step": 2411 }, { "epoch": 0.7561128526645768, "grad_norm": 5.714209079742432, "learning_rate": 1.2060000000000002e-06, "loss": 8.4314, "step": 2412 }, { "epoch": 0.7564263322884013, "grad_norm": 5.888979434967041, "learning_rate": 1.2065e-06, "loss": 6.7525, "step": 2413 }, { "epoch": 0.7567398119122257, "grad_norm": 7.8231425285339355, "learning_rate": 1.207e-06, "loss": 12.5721, "step": 2414 }, { "epoch": 0.7570532915360502, "grad_norm": 6.4952898025512695, "learning_rate": 1.2075e-06, "loss": 10.1876, "step": 2415 }, { "epoch": 0.7573667711598746, "grad_norm": 6.374924182891846, "learning_rate": 1.2080000000000001e-06, "loss": 7.2975, "step": 2416 }, { "epoch": 0.757680250783699, "grad_norm": 5.737400531768799, "learning_rate": 1.2085000000000002e-06, "loss": 8.371, "step": 2417 }, { "epoch": 0.7579937304075235, "grad_norm": 5.425909042358398, "learning_rate": 1.209e-06, "loss": 6.2901, "step": 2418 }, { "epoch": 0.7583072100313479, "grad_norm": 7.00061559677124, "learning_rate": 1.2095e-06, "loss": 6.5716, "step": 2419 }, { "epoch": 0.7586206896551724, "grad_norm": 9.082880020141602, "learning_rate": 1.21e-06, "loss": 9.7631, "step": 2420 }, { "epoch": 0.7589341692789968, "grad_norm": 5.837677478790283, "learning_rate": 1.2105e-06, "loss": 7.0652, "step": 2421 }, { "epoch": 0.7592476489028214, "grad_norm": 10.264467239379883, "learning_rate": 1.2110000000000001e-06, "loss": 13.0527, "step": 2422 }, { "epoch": 0.7595611285266458, "grad_norm": 6.860038757324219, "learning_rate": 1.2115e-06, "loss": 8.581, "step": 2423 }, { "epoch": 0.7598746081504703, "grad_norm": 8.064723014831543, "learning_rate": 1.2120000000000002e-06, "loss": 7.2021, "step": 2424 }, { "epoch": 0.7601880877742947, "grad_norm": 5.945399761199951, "learning_rate": 1.2125e-06, "loss": 8.6018, "step": 2425 }, { "epoch": 0.7605015673981191, "grad_norm": 8.428067207336426, "learning_rate": 1.213e-06, "loss": 9.2994, "step": 2426 }, { "epoch": 0.7608150470219436, "grad_norm": 6.145665168762207, "learning_rate": 1.2135000000000001e-06, "loss": 7.1988, "step": 2427 }, { "epoch": 0.761128526645768, "grad_norm": 7.332749843597412, "learning_rate": 1.214e-06, "loss": 8.2611, "step": 2428 }, { "epoch": 0.7614420062695925, "grad_norm": 13.179669380187988, "learning_rate": 1.2145000000000002e-06, "loss": 15.9689, "step": 2429 }, { "epoch": 0.7617554858934169, "grad_norm": 5.4208197593688965, "learning_rate": 1.215e-06, "loss": 6.9613, "step": 2430 }, { "epoch": 0.7620689655172413, "grad_norm": 5.367392063140869, "learning_rate": 1.2155e-06, "loss": 6.4245, "step": 2431 }, { "epoch": 0.7623824451410658, "grad_norm": 5.860224723815918, "learning_rate": 1.216e-06, "loss": 6.6126, "step": 2432 }, { "epoch": 0.7626959247648902, "grad_norm": 10.526330947875977, "learning_rate": 1.2165000000000001e-06, "loss": 13.7313, "step": 2433 }, { "epoch": 0.7630094043887148, "grad_norm": 5.663205623626709, "learning_rate": 1.2170000000000002e-06, "loss": 8.9684, "step": 2434 }, { "epoch": 0.7633228840125392, "grad_norm": 7.089934825897217, "learning_rate": 1.2175e-06, "loss": 7.9986, "step": 2435 }, { "epoch": 0.7636363636363637, "grad_norm": 5.609120845794678, "learning_rate": 1.2180000000000002e-06, "loss": 6.8416, "step": 2436 }, { "epoch": 0.7639498432601881, "grad_norm": 6.426872730255127, "learning_rate": 1.2185e-06, "loss": 7.9864, "step": 2437 }, { "epoch": 0.7642633228840126, "grad_norm": 8.020556449890137, "learning_rate": 1.219e-06, "loss": 8.7561, "step": 2438 }, { "epoch": 0.764576802507837, "grad_norm": 14.095486640930176, "learning_rate": 1.2195000000000001e-06, "loss": 14.5934, "step": 2439 }, { "epoch": 0.7648902821316614, "grad_norm": 6.615420341491699, "learning_rate": 1.2200000000000002e-06, "loss": 6.5653, "step": 2440 }, { "epoch": 0.7652037617554859, "grad_norm": 6.877127647399902, "learning_rate": 1.2205000000000002e-06, "loss": 9.0697, "step": 2441 }, { "epoch": 0.7655172413793103, "grad_norm": 7.938243389129639, "learning_rate": 1.221e-06, "loss": 7.9308, "step": 2442 }, { "epoch": 0.7658307210031348, "grad_norm": 5.383735656738281, "learning_rate": 1.2215e-06, "loss": 6.4287, "step": 2443 }, { "epoch": 0.7661442006269592, "grad_norm": 12.071537971496582, "learning_rate": 1.2220000000000001e-06, "loss": 9.1741, "step": 2444 }, { "epoch": 0.7664576802507836, "grad_norm": 8.097783088684082, "learning_rate": 1.2225000000000002e-06, "loss": 9.6469, "step": 2445 }, { "epoch": 0.7667711598746082, "grad_norm": 10.031937599182129, "learning_rate": 1.2230000000000002e-06, "loss": 11.3807, "step": 2446 }, { "epoch": 0.7670846394984326, "grad_norm": 6.421173572540283, "learning_rate": 1.2235e-06, "loss": 7.5599, "step": 2447 }, { "epoch": 0.7673981191222571, "grad_norm": 7.706267356872559, "learning_rate": 1.224e-06, "loss": 9.3446, "step": 2448 }, { "epoch": 0.7677115987460815, "grad_norm": 8.22382926940918, "learning_rate": 1.2245e-06, "loss": 10.4658, "step": 2449 }, { "epoch": 0.768025078369906, "grad_norm": 9.478100776672363, "learning_rate": 1.2250000000000001e-06, "loss": 9.6859, "step": 2450 }, { "epoch": 0.7683385579937304, "grad_norm": 10.563023567199707, "learning_rate": 1.2255000000000002e-06, "loss": 9.8905, "step": 2451 }, { "epoch": 0.7686520376175549, "grad_norm": 6.427543640136719, "learning_rate": 1.2260000000000002e-06, "loss": 7.7129, "step": 2452 }, { "epoch": 0.7689655172413793, "grad_norm": 7.049678325653076, "learning_rate": 1.2265e-06, "loss": 8.6615, "step": 2453 }, { "epoch": 0.7692789968652037, "grad_norm": 8.58466911315918, "learning_rate": 1.227e-06, "loss": 8.6398, "step": 2454 }, { "epoch": 0.7695924764890282, "grad_norm": 5.85331392288208, "learning_rate": 1.2275000000000001e-06, "loss": 7.8588, "step": 2455 }, { "epoch": 0.7699059561128526, "grad_norm": 7.215764999389648, "learning_rate": 1.2280000000000001e-06, "loss": 8.4039, "step": 2456 }, { "epoch": 0.7702194357366771, "grad_norm": 13.46800422668457, "learning_rate": 1.2285000000000002e-06, "loss": 11.4974, "step": 2457 }, { "epoch": 0.7705329153605016, "grad_norm": 7.22824764251709, "learning_rate": 1.229e-06, "loss": 8.1564, "step": 2458 }, { "epoch": 0.770846394984326, "grad_norm": 6.372166156768799, "learning_rate": 1.2295e-06, "loss": 6.8613, "step": 2459 }, { "epoch": 0.7711598746081505, "grad_norm": 8.008846282958984, "learning_rate": 1.23e-06, "loss": 8.2465, "step": 2460 }, { "epoch": 0.7714733542319749, "grad_norm": 10.947956085205078, "learning_rate": 1.2305000000000001e-06, "loss": 11.4562, "step": 2461 }, { "epoch": 0.7717868338557994, "grad_norm": 7.392462253570557, "learning_rate": 1.2310000000000002e-06, "loss": 9.5973, "step": 2462 }, { "epoch": 0.7721003134796238, "grad_norm": 6.9039082527160645, "learning_rate": 1.2315e-06, "loss": 7.2863, "step": 2463 }, { "epoch": 0.7724137931034483, "grad_norm": 6.517736911773682, "learning_rate": 1.2320000000000002e-06, "loss": 6.8215, "step": 2464 }, { "epoch": 0.7727272727272727, "grad_norm": 6.294345378875732, "learning_rate": 1.2325e-06, "loss": 6.9459, "step": 2465 }, { "epoch": 0.7730407523510971, "grad_norm": 7.055344581604004, "learning_rate": 1.233e-06, "loss": 7.9461, "step": 2466 }, { "epoch": 0.7733542319749216, "grad_norm": 8.949790954589844, "learning_rate": 1.2335000000000001e-06, "loss": 10.2358, "step": 2467 }, { "epoch": 0.773667711598746, "grad_norm": 12.344834327697754, "learning_rate": 1.234e-06, "loss": 11.9022, "step": 2468 }, { "epoch": 0.7739811912225706, "grad_norm": 13.830575942993164, "learning_rate": 1.2345000000000002e-06, "loss": 17.4536, "step": 2469 }, { "epoch": 0.774294670846395, "grad_norm": 10.294656753540039, "learning_rate": 1.235e-06, "loss": 7.7727, "step": 2470 }, { "epoch": 0.7746081504702195, "grad_norm": 6.9053874015808105, "learning_rate": 1.2355e-06, "loss": 8.134, "step": 2471 }, { "epoch": 0.7749216300940439, "grad_norm": 8.573596954345703, "learning_rate": 1.2360000000000001e-06, "loss": 9.5387, "step": 2472 }, { "epoch": 0.7752351097178684, "grad_norm": 8.290820121765137, "learning_rate": 1.2365000000000001e-06, "loss": 11.1582, "step": 2473 }, { "epoch": 0.7755485893416928, "grad_norm": 7.78951358795166, "learning_rate": 1.2370000000000002e-06, "loss": 9.6967, "step": 2474 }, { "epoch": 0.7758620689655172, "grad_norm": 9.113204956054688, "learning_rate": 1.2375e-06, "loss": 8.3096, "step": 2475 }, { "epoch": 0.7761755485893417, "grad_norm": 8.999139785766602, "learning_rate": 1.238e-06, "loss": 8.9482, "step": 2476 }, { "epoch": 0.7764890282131661, "grad_norm": 7.273940563201904, "learning_rate": 1.2385e-06, "loss": 7.3428, "step": 2477 }, { "epoch": 0.7768025078369906, "grad_norm": 9.114079475402832, "learning_rate": 1.2390000000000001e-06, "loss": 9.9958, "step": 2478 }, { "epoch": 0.777115987460815, "grad_norm": 9.299750328063965, "learning_rate": 1.2395000000000002e-06, "loss": 9.2364, "step": 2479 }, { "epoch": 0.7774294670846394, "grad_norm": 7.594107151031494, "learning_rate": 1.2400000000000002e-06, "loss": 8.9441, "step": 2480 }, { "epoch": 0.777742946708464, "grad_norm": 15.933536529541016, "learning_rate": 1.2405e-06, "loss": 15.7072, "step": 2481 }, { "epoch": 0.7780564263322884, "grad_norm": 6.593420028686523, "learning_rate": 1.241e-06, "loss": 7.2379, "step": 2482 }, { "epoch": 0.7783699059561129, "grad_norm": 10.504799842834473, "learning_rate": 1.2415e-06, "loss": 8.9226, "step": 2483 }, { "epoch": 0.7786833855799373, "grad_norm": 7.769631385803223, "learning_rate": 1.2420000000000001e-06, "loss": 7.9376, "step": 2484 }, { "epoch": 0.7789968652037618, "grad_norm": 9.674359321594238, "learning_rate": 1.2425000000000002e-06, "loss": 6.172, "step": 2485 }, { "epoch": 0.7793103448275862, "grad_norm": 10.36740493774414, "learning_rate": 1.243e-06, "loss": 10.3587, "step": 2486 }, { "epoch": 0.7796238244514107, "grad_norm": 10.151678085327148, "learning_rate": 1.2435e-06, "loss": 8.8859, "step": 2487 }, { "epoch": 0.7799373040752351, "grad_norm": 5.526301860809326, "learning_rate": 1.244e-06, "loss": 6.5573, "step": 2488 }, { "epoch": 0.7802507836990595, "grad_norm": 5.857382774353027, "learning_rate": 1.2445000000000001e-06, "loss": 6.264, "step": 2489 }, { "epoch": 0.780564263322884, "grad_norm": 7.313165187835693, "learning_rate": 1.2450000000000002e-06, "loss": 6.587, "step": 2490 }, { "epoch": 0.7808777429467084, "grad_norm": 10.243111610412598, "learning_rate": 1.2455e-06, "loss": 9.8566, "step": 2491 }, { "epoch": 0.7811912225705329, "grad_norm": 8.093012809753418, "learning_rate": 1.2460000000000002e-06, "loss": 7.5616, "step": 2492 }, { "epoch": 0.7815047021943574, "grad_norm": 9.839873313903809, "learning_rate": 1.2465e-06, "loss": 9.569, "step": 2493 }, { "epoch": 0.7818181818181819, "grad_norm": 9.529997825622559, "learning_rate": 1.247e-06, "loss": 8.8561, "step": 2494 }, { "epoch": 0.7821316614420063, "grad_norm": 7.144129276275635, "learning_rate": 1.2475000000000001e-06, "loss": 6.8548, "step": 2495 }, { "epoch": 0.7824451410658307, "grad_norm": 9.005447387695312, "learning_rate": 1.248e-06, "loss": 9.8978, "step": 2496 }, { "epoch": 0.7827586206896552, "grad_norm": 6.3512797355651855, "learning_rate": 1.2485000000000002e-06, "loss": 7.3197, "step": 2497 }, { "epoch": 0.7830721003134796, "grad_norm": 13.88167667388916, "learning_rate": 1.249e-06, "loss": 10.638, "step": 2498 }, { "epoch": 0.7833855799373041, "grad_norm": 8.67345905303955, "learning_rate": 1.2495e-06, "loss": 8.4422, "step": 2499 }, { "epoch": 0.7836990595611285, "grad_norm": 7.372941970825195, "learning_rate": 1.25e-06, "loss": 7.1321, "step": 2500 }, { "epoch": 0.784012539184953, "grad_norm": 7.833611488342285, "learning_rate": 1.2505000000000001e-06, "loss": 7.717, "step": 2501 }, { "epoch": 0.7843260188087774, "grad_norm": 7.675827980041504, "learning_rate": 1.251e-06, "loss": 6.8507, "step": 2502 }, { "epoch": 0.7846394984326018, "grad_norm": 7.610905170440674, "learning_rate": 1.2515000000000002e-06, "loss": 7.2601, "step": 2503 }, { "epoch": 0.7849529780564264, "grad_norm": 8.251374244689941, "learning_rate": 1.2520000000000003e-06, "loss": 7.8607, "step": 2504 }, { "epoch": 0.7852664576802508, "grad_norm": 8.651074409484863, "learning_rate": 1.2525e-06, "loss": 9.5066, "step": 2505 }, { "epoch": 0.7855799373040753, "grad_norm": 8.352408409118652, "learning_rate": 1.2530000000000001e-06, "loss": 6.5008, "step": 2506 }, { "epoch": 0.7858934169278997, "grad_norm": 8.274674415588379, "learning_rate": 1.2535e-06, "loss": 6.7467, "step": 2507 }, { "epoch": 0.7862068965517242, "grad_norm": 14.473176956176758, "learning_rate": 1.2540000000000002e-06, "loss": 14.1415, "step": 2508 }, { "epoch": 0.7865203761755486, "grad_norm": 7.974168300628662, "learning_rate": 1.2545000000000002e-06, "loss": 7.2298, "step": 2509 }, { "epoch": 0.786833855799373, "grad_norm": 7.596717357635498, "learning_rate": 1.255e-06, "loss": 7.7122, "step": 2510 }, { "epoch": 0.7871473354231975, "grad_norm": 11.445423126220703, "learning_rate": 1.2555e-06, "loss": 10.7785, "step": 2511 }, { "epoch": 0.7874608150470219, "grad_norm": 7.318839073181152, "learning_rate": 1.256e-06, "loss": 7.747, "step": 2512 }, { "epoch": 0.7877742946708464, "grad_norm": 8.10612964630127, "learning_rate": 1.2565000000000002e-06, "loss": 7.516, "step": 2513 }, { "epoch": 0.7880877742946708, "grad_norm": 6.949522972106934, "learning_rate": 1.2570000000000002e-06, "loss": 8.044, "step": 2514 }, { "epoch": 0.7884012539184952, "grad_norm": 10.29315185546875, "learning_rate": 1.2575e-06, "loss": 9.012, "step": 2515 }, { "epoch": 0.7887147335423198, "grad_norm": 8.57775592803955, "learning_rate": 1.258e-06, "loss": 6.2847, "step": 2516 }, { "epoch": 0.7890282131661442, "grad_norm": 11.213354110717773, "learning_rate": 1.2584999999999999e-06, "loss": 10.073, "step": 2517 }, { "epoch": 0.7893416927899687, "grad_norm": 12.789294242858887, "learning_rate": 1.2590000000000001e-06, "loss": 12.4063, "step": 2518 }, { "epoch": 0.7896551724137931, "grad_norm": 6.868427753448486, "learning_rate": 1.2595000000000002e-06, "loss": 5.4573, "step": 2519 }, { "epoch": 0.7899686520376176, "grad_norm": 6.975144863128662, "learning_rate": 1.26e-06, "loss": 7.0482, "step": 2520 }, { "epoch": 0.790282131661442, "grad_norm": 9.508519172668457, "learning_rate": 1.2605e-06, "loss": 9.9895, "step": 2521 }, { "epoch": 0.7905956112852665, "grad_norm": 12.231802940368652, "learning_rate": 1.261e-06, "loss": 10.782, "step": 2522 }, { "epoch": 0.7909090909090909, "grad_norm": 7.636473178863525, "learning_rate": 1.2615000000000001e-06, "loss": 7.4447, "step": 2523 }, { "epoch": 0.7912225705329153, "grad_norm": 14.089887619018555, "learning_rate": 1.2620000000000002e-06, "loss": 11.8825, "step": 2524 }, { "epoch": 0.7915360501567398, "grad_norm": 11.20523738861084, "learning_rate": 1.2625000000000002e-06, "loss": 10.0402, "step": 2525 }, { "epoch": 0.7918495297805642, "grad_norm": 10.182788848876953, "learning_rate": 1.263e-06, "loss": 7.4292, "step": 2526 }, { "epoch": 0.7921630094043887, "grad_norm": 8.2036771774292, "learning_rate": 1.2635e-06, "loss": 7.3361, "step": 2527 }, { "epoch": 0.7924764890282132, "grad_norm": 9.289824485778809, "learning_rate": 1.2640000000000003e-06, "loss": 7.6616, "step": 2528 }, { "epoch": 0.7927899686520377, "grad_norm": 10.154131889343262, "learning_rate": 1.2645000000000001e-06, "loss": 9.678, "step": 2529 }, { "epoch": 0.7931034482758621, "grad_norm": 8.044170379638672, "learning_rate": 1.2650000000000002e-06, "loss": 7.7272, "step": 2530 }, { "epoch": 0.7934169278996865, "grad_norm": 14.741924285888672, "learning_rate": 1.2655e-06, "loss": 11.6074, "step": 2531 }, { "epoch": 0.793730407523511, "grad_norm": 12.883913040161133, "learning_rate": 1.266e-06, "loss": 7.45, "step": 2532 }, { "epoch": 0.7940438871473354, "grad_norm": 14.291522026062012, "learning_rate": 1.2665000000000003e-06, "loss": 8.8233, "step": 2533 }, { "epoch": 0.7943573667711599, "grad_norm": 10.920492172241211, "learning_rate": 1.2670000000000001e-06, "loss": 8.2044, "step": 2534 }, { "epoch": 0.7946708463949843, "grad_norm": 9.552704811096191, "learning_rate": 1.2675000000000001e-06, "loss": 7.1354, "step": 2535 }, { "epoch": 0.7949843260188088, "grad_norm": 13.507481575012207, "learning_rate": 1.268e-06, "loss": 11.1275, "step": 2536 }, { "epoch": 0.7952978056426332, "grad_norm": 8.592397689819336, "learning_rate": 1.2685e-06, "loss": 6.2247, "step": 2537 }, { "epoch": 0.7956112852664576, "grad_norm": 8.421808242797852, "learning_rate": 1.2690000000000003e-06, "loss": 6.4027, "step": 2538 }, { "epoch": 0.7959247648902821, "grad_norm": 8.067540168762207, "learning_rate": 1.2695e-06, "loss": 8.3162, "step": 2539 }, { "epoch": 0.7962382445141066, "grad_norm": 8.634953498840332, "learning_rate": 1.2700000000000001e-06, "loss": 6.8064, "step": 2540 }, { "epoch": 0.7965517241379311, "grad_norm": 13.059370994567871, "learning_rate": 1.2705000000000002e-06, "loss": 10.4157, "step": 2541 }, { "epoch": 0.7968652037617555, "grad_norm": 7.6555304527282715, "learning_rate": 1.271e-06, "loss": 6.2527, "step": 2542 }, { "epoch": 0.79717868338558, "grad_norm": 9.67294692993164, "learning_rate": 1.2715000000000002e-06, "loss": 6.2133, "step": 2543 }, { "epoch": 0.7974921630094044, "grad_norm": 8.20207691192627, "learning_rate": 1.2720000000000003e-06, "loss": 7.4998, "step": 2544 }, { "epoch": 0.7978056426332288, "grad_norm": 9.94020938873291, "learning_rate": 1.2725e-06, "loss": 8.4282, "step": 2545 }, { "epoch": 0.7981191222570533, "grad_norm": 7.922036647796631, "learning_rate": 1.2730000000000001e-06, "loss": 6.9887, "step": 2546 }, { "epoch": 0.7984326018808777, "grad_norm": 11.113056182861328, "learning_rate": 1.2735e-06, "loss": 7.7985, "step": 2547 }, { "epoch": 0.7987460815047022, "grad_norm": 12.114452362060547, "learning_rate": 1.2740000000000002e-06, "loss": 8.92, "step": 2548 }, { "epoch": 0.7990595611285266, "grad_norm": 12.305609703063965, "learning_rate": 1.2745000000000002e-06, "loss": 10.3891, "step": 2549 }, { "epoch": 0.799373040752351, "grad_norm": 7.7950215339660645, "learning_rate": 1.275e-06, "loss": 6.9325, "step": 2550 }, { "epoch": 0.7996865203761756, "grad_norm": 12.936772346496582, "learning_rate": 1.2755000000000001e-06, "loss": 9.9608, "step": 2551 }, { "epoch": 0.8, "grad_norm": 9.42872428894043, "learning_rate": 1.276e-06, "loss": 7.2778, "step": 2552 }, { "epoch": 0.8003134796238245, "grad_norm": 12.10912036895752, "learning_rate": 1.2765000000000002e-06, "loss": 8.8379, "step": 2553 }, { "epoch": 0.8006269592476489, "grad_norm": 10.744391441345215, "learning_rate": 1.2770000000000002e-06, "loss": 8.0975, "step": 2554 }, { "epoch": 0.8009404388714734, "grad_norm": 9.00346565246582, "learning_rate": 1.2775e-06, "loss": 6.8766, "step": 2555 }, { "epoch": 0.8012539184952978, "grad_norm": 11.213250160217285, "learning_rate": 1.278e-06, "loss": 8.3033, "step": 2556 }, { "epoch": 0.8015673981191223, "grad_norm": 10.06607723236084, "learning_rate": 1.2785e-06, "loss": 6.8206, "step": 2557 }, { "epoch": 0.8018808777429467, "grad_norm": 16.033727645874023, "learning_rate": 1.2790000000000002e-06, "loss": 10.1014, "step": 2558 }, { "epoch": 0.8021943573667711, "grad_norm": 10.360695838928223, "learning_rate": 1.2795000000000002e-06, "loss": 7.3716, "step": 2559 }, { "epoch": 0.8025078369905956, "grad_norm": 16.054412841796875, "learning_rate": 1.28e-06, "loss": 9.5552, "step": 2560 }, { "epoch": 0.80282131661442, "grad_norm": 10.377695083618164, "learning_rate": 1.2805e-06, "loss": 7.2903, "step": 2561 }, { "epoch": 0.8031347962382445, "grad_norm": 10.614144325256348, "learning_rate": 1.281e-06, "loss": 7.9967, "step": 2562 }, { "epoch": 0.803448275862069, "grad_norm": 10.514567375183105, "learning_rate": 1.2815e-06, "loss": 6.7008, "step": 2563 }, { "epoch": 0.8037617554858935, "grad_norm": 8.607209205627441, "learning_rate": 1.2820000000000002e-06, "loss": 5.8722, "step": 2564 }, { "epoch": 0.8040752351097179, "grad_norm": 10.558499336242676, "learning_rate": 1.2825000000000002e-06, "loss": 7.2417, "step": 2565 }, { "epoch": 0.8043887147335423, "grad_norm": 17.161361694335938, "learning_rate": 1.283e-06, "loss": 13.5513, "step": 2566 }, { "epoch": 0.8047021943573668, "grad_norm": 11.054380416870117, "learning_rate": 1.2835e-06, "loss": 8.4185, "step": 2567 }, { "epoch": 0.8050156739811912, "grad_norm": 15.804526329040527, "learning_rate": 1.284e-06, "loss": 12.1668, "step": 2568 }, { "epoch": 0.8053291536050157, "grad_norm": 15.922082901000977, "learning_rate": 1.2845000000000002e-06, "loss": 11.7197, "step": 2569 }, { "epoch": 0.8056426332288401, "grad_norm": 15.542136192321777, "learning_rate": 1.2850000000000002e-06, "loss": 5.9863, "step": 2570 }, { "epoch": 0.8059561128526646, "grad_norm": 13.482390403747559, "learning_rate": 1.2855e-06, "loss": 8.9265, "step": 2571 }, { "epoch": 0.806269592476489, "grad_norm": 15.561339378356934, "learning_rate": 1.286e-06, "loss": 8.5574, "step": 2572 }, { "epoch": 0.8065830721003134, "grad_norm": 14.538595199584961, "learning_rate": 1.2864999999999999e-06, "loss": 11.4279, "step": 2573 }, { "epoch": 0.8068965517241379, "grad_norm": 14.449934005737305, "learning_rate": 1.2870000000000001e-06, "loss": 12.0358, "step": 2574 }, { "epoch": 0.8072100313479624, "grad_norm": 13.869661331176758, "learning_rate": 1.2875000000000002e-06, "loss": 10.5568, "step": 2575 }, { "epoch": 0.8075235109717869, "grad_norm": 10.084548950195312, "learning_rate": 1.288e-06, "loss": 6.7925, "step": 2576 }, { "epoch": 0.8078369905956113, "grad_norm": 8.329049110412598, "learning_rate": 1.2885e-06, "loss": 7.129, "step": 2577 }, { "epoch": 0.8081504702194358, "grad_norm": 15.03673267364502, "learning_rate": 1.289e-06, "loss": 10.9492, "step": 2578 }, { "epoch": 0.8084639498432602, "grad_norm": 11.203147888183594, "learning_rate": 1.2895e-06, "loss": 6.868, "step": 2579 }, { "epoch": 0.8087774294670846, "grad_norm": 14.008033752441406, "learning_rate": 1.2900000000000001e-06, "loss": 7.7931, "step": 2580 }, { "epoch": 0.8090909090909091, "grad_norm": 11.999140739440918, "learning_rate": 1.2905000000000002e-06, "loss": 5.7815, "step": 2581 }, { "epoch": 0.8094043887147335, "grad_norm": 14.67009449005127, "learning_rate": 1.291e-06, "loss": 10.9111, "step": 2582 }, { "epoch": 0.809717868338558, "grad_norm": 10.367982864379883, "learning_rate": 1.2915e-06, "loss": 7.1978, "step": 2583 }, { "epoch": 0.8100313479623824, "grad_norm": 15.708711624145508, "learning_rate": 1.2920000000000003e-06, "loss": 8.7541, "step": 2584 }, { "epoch": 0.8103448275862069, "grad_norm": 21.055742263793945, "learning_rate": 1.2925000000000001e-06, "loss": 9.4178, "step": 2585 }, { "epoch": 0.8106583072100313, "grad_norm": 11.601629257202148, "learning_rate": 1.2930000000000002e-06, "loss": 7.239, "step": 2586 }, { "epoch": 0.8109717868338558, "grad_norm": 19.32367515563965, "learning_rate": 1.2935e-06, "loss": 12.215, "step": 2587 }, { "epoch": 0.8112852664576803, "grad_norm": 12.446346282958984, "learning_rate": 1.294e-06, "loss": 8.1503, "step": 2588 }, { "epoch": 0.8115987460815047, "grad_norm": 15.01457691192627, "learning_rate": 1.2945000000000003e-06, "loss": 9.3474, "step": 2589 }, { "epoch": 0.8119122257053292, "grad_norm": 8.85611629486084, "learning_rate": 1.295e-06, "loss": 6.0037, "step": 2590 }, { "epoch": 0.8122257053291536, "grad_norm": 12.6907320022583, "learning_rate": 1.2955000000000001e-06, "loss": 6.035, "step": 2591 }, { "epoch": 0.812539184952978, "grad_norm": 29.200088500976562, "learning_rate": 1.296e-06, "loss": 15.6593, "step": 2592 }, { "epoch": 0.8128526645768025, "grad_norm": 13.449995994567871, "learning_rate": 1.2965e-06, "loss": 7.4237, "step": 2593 }, { "epoch": 0.8131661442006269, "grad_norm": 12.185378074645996, "learning_rate": 1.2970000000000002e-06, "loss": 7.6051, "step": 2594 }, { "epoch": 0.8134796238244514, "grad_norm": 13.01950454711914, "learning_rate": 1.2975e-06, "loss": 9.8676, "step": 2595 }, { "epoch": 0.8137931034482758, "grad_norm": 12.636890411376953, "learning_rate": 1.2980000000000001e-06, "loss": 7.8251, "step": 2596 }, { "epoch": 0.8141065830721003, "grad_norm": 14.959136962890625, "learning_rate": 1.2985e-06, "loss": 8.8174, "step": 2597 }, { "epoch": 0.8144200626959248, "grad_norm": 12.237345695495605, "learning_rate": 1.299e-06, "loss": 6.9888, "step": 2598 }, { "epoch": 0.8147335423197493, "grad_norm": 9.535272598266602, "learning_rate": 1.2995000000000002e-06, "loss": 5.9116, "step": 2599 }, { "epoch": 0.8150470219435737, "grad_norm": 15.491859436035156, "learning_rate": 1.3e-06, "loss": 7.9926, "step": 2600 }, { "epoch": 0.8153605015673981, "grad_norm": 11.194720268249512, "learning_rate": 1.3005e-06, "loss": 7.7533, "step": 2601 }, { "epoch": 0.8156739811912226, "grad_norm": 18.05571746826172, "learning_rate": 1.3010000000000001e-06, "loss": 9.5349, "step": 2602 }, { "epoch": 0.815987460815047, "grad_norm": 9.782732009887695, "learning_rate": 1.3015e-06, "loss": 6.5899, "step": 2603 }, { "epoch": 0.8163009404388715, "grad_norm": 15.516654968261719, "learning_rate": 1.3020000000000002e-06, "loss": 7.107, "step": 2604 }, { "epoch": 0.8166144200626959, "grad_norm": 13.18638801574707, "learning_rate": 1.3025000000000002e-06, "loss": 6.44, "step": 2605 }, { "epoch": 0.8169278996865204, "grad_norm": 14.614501953125, "learning_rate": 1.303e-06, "loss": 7.5293, "step": 2606 }, { "epoch": 0.8172413793103448, "grad_norm": 11.227118492126465, "learning_rate": 1.3035e-06, "loss": 7.8937, "step": 2607 }, { "epoch": 0.8175548589341692, "grad_norm": 15.58757495880127, "learning_rate": 1.304e-06, "loss": 8.0596, "step": 2608 }, { "epoch": 0.8178683385579937, "grad_norm": 22.042787551879883, "learning_rate": 1.3045000000000002e-06, "loss": 8.2819, "step": 2609 }, { "epoch": 0.8181818181818182, "grad_norm": 21.173282623291016, "learning_rate": 1.3050000000000002e-06, "loss": 10.4288, "step": 2610 }, { "epoch": 0.8184952978056427, "grad_norm": 18.32832145690918, "learning_rate": 1.3055e-06, "loss": 12.1147, "step": 2611 }, { "epoch": 0.8188087774294671, "grad_norm": 15.920726776123047, "learning_rate": 1.306e-06, "loss": 9.0858, "step": 2612 }, { "epoch": 0.8191222570532916, "grad_norm": 12.705510139465332, "learning_rate": 1.3065e-06, "loss": 8.1776, "step": 2613 }, { "epoch": 0.819435736677116, "grad_norm": 13.502321243286133, "learning_rate": 1.3070000000000001e-06, "loss": 7.3692, "step": 2614 }, { "epoch": 0.8197492163009404, "grad_norm": 15.334294319152832, "learning_rate": 1.3075000000000002e-06, "loss": 8.4849, "step": 2615 }, { "epoch": 0.8200626959247649, "grad_norm": 13.238600730895996, "learning_rate": 1.308e-06, "loss": 6.9999, "step": 2616 }, { "epoch": 0.8203761755485893, "grad_norm": 20.93524742126465, "learning_rate": 1.3085e-06, "loss": 7.2483, "step": 2617 }, { "epoch": 0.8206896551724138, "grad_norm": 13.544681549072266, "learning_rate": 1.309e-06, "loss": 7.8575, "step": 2618 }, { "epoch": 0.8210031347962382, "grad_norm": 14.38907241821289, "learning_rate": 1.3095000000000001e-06, "loss": 6.2083, "step": 2619 }, { "epoch": 0.8213166144200627, "grad_norm": 20.6708927154541, "learning_rate": 1.3100000000000002e-06, "loss": 9.0823, "step": 2620 }, { "epoch": 0.8216300940438871, "grad_norm": 12.318846702575684, "learning_rate": 1.3105000000000002e-06, "loss": 6.4326, "step": 2621 }, { "epoch": 0.8219435736677116, "grad_norm": 40.32315444946289, "learning_rate": 1.311e-06, "loss": 14.9634, "step": 2622 }, { "epoch": 0.8222570532915361, "grad_norm": 15.944631576538086, "learning_rate": 1.3115e-06, "loss": 7.8842, "step": 2623 }, { "epoch": 0.8225705329153605, "grad_norm": 12.112171173095703, "learning_rate": 1.3120000000000003e-06, "loss": 5.5937, "step": 2624 }, { "epoch": 0.822884012539185, "grad_norm": 13.884896278381348, "learning_rate": 1.3125000000000001e-06, "loss": 6.3687, "step": 2625 }, { "epoch": 0.8231974921630094, "grad_norm": 22.745121002197266, "learning_rate": 1.3130000000000002e-06, "loss": 12.1536, "step": 2626 }, { "epoch": 0.8235109717868339, "grad_norm": 18.648685455322266, "learning_rate": 1.3135e-06, "loss": 7.5739, "step": 2627 }, { "epoch": 0.8238244514106583, "grad_norm": 18.585302352905273, "learning_rate": 1.314e-06, "loss": 8.4919, "step": 2628 }, { "epoch": 0.8241379310344827, "grad_norm": 26.631200790405273, "learning_rate": 1.3145000000000003e-06, "loss": 8.0334, "step": 2629 }, { "epoch": 0.8244514106583072, "grad_norm": 22.677764892578125, "learning_rate": 1.3150000000000001e-06, "loss": 8.3661, "step": 2630 }, { "epoch": 0.8247648902821316, "grad_norm": 27.903852462768555, "learning_rate": 1.3155000000000002e-06, "loss": 10.0715, "step": 2631 }, { "epoch": 0.8250783699059561, "grad_norm": 26.193275451660156, "learning_rate": 1.316e-06, "loss": 9.49, "step": 2632 }, { "epoch": 0.8253918495297806, "grad_norm": 15.721944808959961, "learning_rate": 1.3165e-06, "loss": 7.3057, "step": 2633 }, { "epoch": 0.8257053291536051, "grad_norm": 16.153507232666016, "learning_rate": 1.3170000000000003e-06, "loss": 6.6323, "step": 2634 }, { "epoch": 0.8260188087774295, "grad_norm": 17.67839813232422, "learning_rate": 1.3175e-06, "loss": 7.1916, "step": 2635 }, { "epoch": 0.826332288401254, "grad_norm": 18.60005760192871, "learning_rate": 1.3180000000000001e-06, "loss": 7.4737, "step": 2636 }, { "epoch": 0.8266457680250784, "grad_norm": 16.918989181518555, "learning_rate": 1.3185e-06, "loss": 9.5991, "step": 2637 }, { "epoch": 0.8269592476489028, "grad_norm": 28.17279624938965, "learning_rate": 1.319e-06, "loss": 9.0958, "step": 2638 }, { "epoch": 0.8272727272727273, "grad_norm": 22.73778533935547, "learning_rate": 1.3195000000000002e-06, "loss": 6.3333, "step": 2639 }, { "epoch": 0.8275862068965517, "grad_norm": 17.459001541137695, "learning_rate": 1.32e-06, "loss": 6.0962, "step": 2640 }, { "epoch": 0.8278996865203762, "grad_norm": 13.217561721801758, "learning_rate": 1.3205e-06, "loss": 6.563, "step": 2641 }, { "epoch": 0.8282131661442006, "grad_norm": 15.55752182006836, "learning_rate": 1.3210000000000001e-06, "loss": 5.5877, "step": 2642 }, { "epoch": 0.828526645768025, "grad_norm": 56.43824005126953, "learning_rate": 1.3215e-06, "loss": 7.7625, "step": 2643 }, { "epoch": 0.8288401253918495, "grad_norm": 21.975671768188477, "learning_rate": 1.3220000000000002e-06, "loss": 9.7735, "step": 2644 }, { "epoch": 0.829153605015674, "grad_norm": 19.311349868774414, "learning_rate": 1.3225000000000003e-06, "loss": 8.4531, "step": 2645 }, { "epoch": 0.8294670846394985, "grad_norm": 39.477603912353516, "learning_rate": 1.323e-06, "loss": 14.1747, "step": 2646 }, { "epoch": 0.8297805642633229, "grad_norm": 23.62912368774414, "learning_rate": 1.3235000000000001e-06, "loss": 9.9994, "step": 2647 }, { "epoch": 0.8300940438871474, "grad_norm": 17.039447784423828, "learning_rate": 1.324e-06, "loss": 6.8518, "step": 2648 }, { "epoch": 0.8304075235109718, "grad_norm": 21.798898696899414, "learning_rate": 1.3245000000000002e-06, "loss": 8.8252, "step": 2649 }, { "epoch": 0.8307210031347962, "grad_norm": 31.107290267944336, "learning_rate": 1.3250000000000002e-06, "loss": 9.1876, "step": 2650 }, { "epoch": 0.8310344827586207, "grad_norm": 21.375978469848633, "learning_rate": 1.3255e-06, "loss": 6.6534, "step": 2651 }, { "epoch": 0.8313479623824451, "grad_norm": 16.57238006591797, "learning_rate": 1.326e-06, "loss": 5.7842, "step": 2652 }, { "epoch": 0.8316614420062696, "grad_norm": 25.424558639526367, "learning_rate": 1.3265e-06, "loss": 8.5218, "step": 2653 }, { "epoch": 0.831974921630094, "grad_norm": 32.927066802978516, "learning_rate": 1.3270000000000002e-06, "loss": 9.3753, "step": 2654 }, { "epoch": 0.8322884012539185, "grad_norm": 18.978023529052734, "learning_rate": 1.3275000000000002e-06, "loss": 6.1247, "step": 2655 }, { "epoch": 0.8326018808777429, "grad_norm": 38.223243713378906, "learning_rate": 1.328e-06, "loss": 11.0785, "step": 2656 }, { "epoch": 0.8329153605015674, "grad_norm": 16.340456008911133, "learning_rate": 1.3285e-06, "loss": 6.4915, "step": 2657 }, { "epoch": 0.8332288401253919, "grad_norm": 18.187742233276367, "learning_rate": 1.3290000000000001e-06, "loss": 8.5854, "step": 2658 }, { "epoch": 0.8335423197492163, "grad_norm": 17.474607467651367, "learning_rate": 1.3295000000000001e-06, "loss": 7.5377, "step": 2659 }, { "epoch": 0.8338557993730408, "grad_norm": 19.971267700195312, "learning_rate": 1.3300000000000002e-06, "loss": 7.2454, "step": 2660 }, { "epoch": 0.8341692789968652, "grad_norm": 14.669326782226562, "learning_rate": 1.3305000000000002e-06, "loss": 6.1117, "step": 2661 }, { "epoch": 0.8344827586206897, "grad_norm": 18.929859161376953, "learning_rate": 1.331e-06, "loss": 7.8862, "step": 2662 }, { "epoch": 0.8347962382445141, "grad_norm": 29.54069709777832, "learning_rate": 1.3315e-06, "loss": 7.5074, "step": 2663 }, { "epoch": 0.8351097178683385, "grad_norm": 21.967472076416016, "learning_rate": 1.3320000000000003e-06, "loss": 7.0764, "step": 2664 }, { "epoch": 0.835423197492163, "grad_norm": 45.06687927246094, "learning_rate": 1.3325000000000002e-06, "loss": 9.701, "step": 2665 }, { "epoch": 0.8357366771159874, "grad_norm": 24.894472122192383, "learning_rate": 1.3330000000000002e-06, "loss": 8.708, "step": 2666 }, { "epoch": 0.8360501567398119, "grad_norm": 20.155099868774414, "learning_rate": 1.3335e-06, "loss": 6.1855, "step": 2667 }, { "epoch": 0.8363636363636363, "grad_norm": 20.541330337524414, "learning_rate": 1.334e-06, "loss": 6.6755, "step": 2668 }, { "epoch": 0.8366771159874609, "grad_norm": 22.628978729248047, "learning_rate": 1.3345000000000003e-06, "loss": 8.2543, "step": 2669 }, { "epoch": 0.8369905956112853, "grad_norm": 22.389394760131836, "learning_rate": 1.3350000000000001e-06, "loss": 7.0468, "step": 2670 }, { "epoch": 0.8373040752351097, "grad_norm": 23.417722702026367, "learning_rate": 1.3355000000000002e-06, "loss": 6.792, "step": 2671 }, { "epoch": 0.8376175548589342, "grad_norm": 21.459794998168945, "learning_rate": 1.336e-06, "loss": 6.9663, "step": 2672 }, { "epoch": 0.8379310344827586, "grad_norm": 32.657989501953125, "learning_rate": 1.3365e-06, "loss": 5.9438, "step": 2673 }, { "epoch": 0.8382445141065831, "grad_norm": 47.682586669921875, "learning_rate": 1.3370000000000003e-06, "loss": 10.6549, "step": 2674 }, { "epoch": 0.8385579937304075, "grad_norm": 20.36661720275879, "learning_rate": 1.3375000000000001e-06, "loss": 6.8448, "step": 2675 }, { "epoch": 0.838871473354232, "grad_norm": 14.008045196533203, "learning_rate": 1.3380000000000001e-06, "loss": 6.8451, "step": 2676 }, { "epoch": 0.8391849529780564, "grad_norm": 17.530170440673828, "learning_rate": 1.3385e-06, "loss": 5.8831, "step": 2677 }, { "epoch": 0.8394984326018808, "grad_norm": 17.58102798461914, "learning_rate": 1.339e-06, "loss": 6.3528, "step": 2678 }, { "epoch": 0.8398119122257053, "grad_norm": 23.83525848388672, "learning_rate": 1.3395000000000003e-06, "loss": 7.7174, "step": 2679 }, { "epoch": 0.8401253918495298, "grad_norm": 27.301860809326172, "learning_rate": 1.34e-06, "loss": 5.6273, "step": 2680 }, { "epoch": 0.8404388714733543, "grad_norm": 29.820091247558594, "learning_rate": 1.3405000000000001e-06, "loss": 7.9886, "step": 2681 }, { "epoch": 0.8407523510971787, "grad_norm": 31.963327407836914, "learning_rate": 1.3410000000000002e-06, "loss": 6.5933, "step": 2682 }, { "epoch": 0.8410658307210032, "grad_norm": 23.809494018554688, "learning_rate": 1.3415e-06, "loss": 7.4239, "step": 2683 }, { "epoch": 0.8413793103448276, "grad_norm": 26.722423553466797, "learning_rate": 1.3420000000000002e-06, "loss": 6.8609, "step": 2684 }, { "epoch": 0.841692789968652, "grad_norm": 20.260234832763672, "learning_rate": 1.3425000000000003e-06, "loss": 7.6539, "step": 2685 }, { "epoch": 0.8420062695924765, "grad_norm": 29.476564407348633, "learning_rate": 1.343e-06, "loss": 8.3136, "step": 2686 }, { "epoch": 0.8423197492163009, "grad_norm": 17.281204223632812, "learning_rate": 1.3435000000000001e-06, "loss": 5.9271, "step": 2687 }, { "epoch": 0.8426332288401254, "grad_norm": 39.53363037109375, "learning_rate": 1.344e-06, "loss": 8.3828, "step": 2688 }, { "epoch": 0.8429467084639498, "grad_norm": 19.412582397460938, "learning_rate": 1.3445e-06, "loss": 6.6062, "step": 2689 }, { "epoch": 0.8432601880877743, "grad_norm": 31.106475830078125, "learning_rate": 1.3450000000000003e-06, "loss": 7.9131, "step": 2690 }, { "epoch": 0.8435736677115987, "grad_norm": 37.35240936279297, "learning_rate": 1.3455e-06, "loss": 8.8415, "step": 2691 }, { "epoch": 0.8438871473354232, "grad_norm": 26.39192771911621, "learning_rate": 1.3460000000000001e-06, "loss": 6.637, "step": 2692 }, { "epoch": 0.8442006269592477, "grad_norm": 21.605411529541016, "learning_rate": 1.3465e-06, "loss": 6.3293, "step": 2693 }, { "epoch": 0.8445141065830721, "grad_norm": 21.46399688720703, "learning_rate": 1.347e-06, "loss": 6.8912, "step": 2694 }, { "epoch": 0.8448275862068966, "grad_norm": 26.33460807800293, "learning_rate": 1.3475000000000002e-06, "loss": 8.7624, "step": 2695 }, { "epoch": 0.845141065830721, "grad_norm": 21.527713775634766, "learning_rate": 1.348e-06, "loss": 5.8286, "step": 2696 }, { "epoch": 0.8454545454545455, "grad_norm": 30.383634567260742, "learning_rate": 1.3485e-06, "loss": 9.5837, "step": 2697 }, { "epoch": 0.8457680250783699, "grad_norm": 33.206356048583984, "learning_rate": 1.3490000000000001e-06, "loss": 7.8532, "step": 2698 }, { "epoch": 0.8460815047021943, "grad_norm": 36.5329475402832, "learning_rate": 1.3495e-06, "loss": 5.4797, "step": 2699 }, { "epoch": 0.8463949843260188, "grad_norm": 30.74308204650879, "learning_rate": 1.3500000000000002e-06, "loss": 7.289, "step": 2700 }, { "epoch": 0.8467084639498432, "grad_norm": 38.16331481933594, "learning_rate": 1.3505000000000002e-06, "loss": 7.9117, "step": 2701 }, { "epoch": 0.8470219435736677, "grad_norm": 26.741302490234375, "learning_rate": 1.351e-06, "loss": 5.9725, "step": 2702 }, { "epoch": 0.8473354231974921, "grad_norm": 34.44478988647461, "learning_rate": 1.3515e-06, "loss": 7.5997, "step": 2703 }, { "epoch": 0.8476489028213167, "grad_norm": 23.709253311157227, "learning_rate": 1.352e-06, "loss": 6.8915, "step": 2704 }, { "epoch": 0.8479623824451411, "grad_norm": 24.891733169555664, "learning_rate": 1.3525000000000002e-06, "loss": 5.9107, "step": 2705 }, { "epoch": 0.8482758620689655, "grad_norm": 24.73875617980957, "learning_rate": 1.3530000000000002e-06, "loss": 6.2778, "step": 2706 }, { "epoch": 0.84858934169279, "grad_norm": 18.066926956176758, "learning_rate": 1.3535e-06, "loss": 6.1856, "step": 2707 }, { "epoch": 0.8489028213166144, "grad_norm": 40.347476959228516, "learning_rate": 1.354e-06, "loss": 7.0503, "step": 2708 }, { "epoch": 0.8492163009404389, "grad_norm": 31.80996322631836, "learning_rate": 1.3545e-06, "loss": 8.8319, "step": 2709 }, { "epoch": 0.8495297805642633, "grad_norm": 27.470136642456055, "learning_rate": 1.3550000000000002e-06, "loss": 7.3765, "step": 2710 }, { "epoch": 0.8498432601880878, "grad_norm": 83.07367706298828, "learning_rate": 1.3555000000000002e-06, "loss": 20.241, "step": 2711 }, { "epoch": 0.8501567398119122, "grad_norm": 30.60701560974121, "learning_rate": 1.356e-06, "loss": 6.9115, "step": 2712 }, { "epoch": 0.8504702194357366, "grad_norm": 31.232662200927734, "learning_rate": 1.3565e-06, "loss": 7.2203, "step": 2713 }, { "epoch": 0.8507836990595611, "grad_norm": 29.009035110473633, "learning_rate": 1.3569999999999999e-06, "loss": 7.3232, "step": 2714 }, { "epoch": 0.8510971786833855, "grad_norm": 20.113876342773438, "learning_rate": 1.3575000000000001e-06, "loss": 5.1684, "step": 2715 }, { "epoch": 0.8514106583072101, "grad_norm": 21.815345764160156, "learning_rate": 1.3580000000000002e-06, "loss": 6.1103, "step": 2716 }, { "epoch": 0.8517241379310345, "grad_norm": 28.0091552734375, "learning_rate": 1.3585e-06, "loss": 7.3426, "step": 2717 }, { "epoch": 0.852037617554859, "grad_norm": 30.732309341430664, "learning_rate": 1.359e-06, "loss": 6.6737, "step": 2718 }, { "epoch": 0.8523510971786834, "grad_norm": 24.529449462890625, "learning_rate": 1.3595e-06, "loss": 6.6836, "step": 2719 }, { "epoch": 0.8526645768025078, "grad_norm": 21.116939544677734, "learning_rate": 1.3600000000000001e-06, "loss": 6.1635, "step": 2720 }, { "epoch": 0.8529780564263323, "grad_norm": 27.107351303100586, "learning_rate": 1.3605000000000001e-06, "loss": 6.0729, "step": 2721 }, { "epoch": 0.8532915360501567, "grad_norm": 19.72542953491211, "learning_rate": 1.3610000000000002e-06, "loss": 4.9322, "step": 2722 }, { "epoch": 0.8536050156739812, "grad_norm": 20.744781494140625, "learning_rate": 1.3615e-06, "loss": 6.2112, "step": 2723 }, { "epoch": 0.8539184952978056, "grad_norm": 23.486122131347656, "learning_rate": 1.362e-06, "loss": 6.0354, "step": 2724 }, { "epoch": 0.85423197492163, "grad_norm": 37.22897720336914, "learning_rate": 1.3625000000000003e-06, "loss": 6.7026, "step": 2725 }, { "epoch": 0.8545454545454545, "grad_norm": 28.69023323059082, "learning_rate": 1.3630000000000001e-06, "loss": 6.349, "step": 2726 }, { "epoch": 0.854858934169279, "grad_norm": 31.381017684936523, "learning_rate": 1.3635000000000002e-06, "loss": 6.7903, "step": 2727 }, { "epoch": 0.8551724137931035, "grad_norm": 23.358259201049805, "learning_rate": 1.364e-06, "loss": 5.7168, "step": 2728 }, { "epoch": 0.8554858934169279, "grad_norm": 29.059661865234375, "learning_rate": 1.3645e-06, "loss": 6.2689, "step": 2729 }, { "epoch": 0.8557993730407524, "grad_norm": 28.51305389404297, "learning_rate": 1.3650000000000003e-06, "loss": 5.9654, "step": 2730 }, { "epoch": 0.8561128526645768, "grad_norm": 24.55581283569336, "learning_rate": 1.3655e-06, "loss": 5.0936, "step": 2731 }, { "epoch": 0.8564263322884013, "grad_norm": 28.580116271972656, "learning_rate": 1.3660000000000001e-06, "loss": 6.2048, "step": 2732 }, { "epoch": 0.8567398119122257, "grad_norm": 37.072593688964844, "learning_rate": 1.3665e-06, "loss": 6.9439, "step": 2733 }, { "epoch": 0.8570532915360501, "grad_norm": 24.373992919921875, "learning_rate": 1.367e-06, "loss": 6.8616, "step": 2734 }, { "epoch": 0.8573667711598746, "grad_norm": 37.73335647583008, "learning_rate": 1.3675000000000002e-06, "loss": 6.2993, "step": 2735 }, { "epoch": 0.857680250783699, "grad_norm": 80.02848815917969, "learning_rate": 1.368e-06, "loss": 14.2041, "step": 2736 }, { "epoch": 0.8579937304075235, "grad_norm": 52.7184944152832, "learning_rate": 1.3685000000000001e-06, "loss": 8.9868, "step": 2737 }, { "epoch": 0.8583072100313479, "grad_norm": 31.193477630615234, "learning_rate": 1.3690000000000001e-06, "loss": 7.0294, "step": 2738 }, { "epoch": 0.8586206896551725, "grad_norm": 30.320158004760742, "learning_rate": 1.3695e-06, "loss": 6.4768, "step": 2739 }, { "epoch": 0.8589341692789969, "grad_norm": 30.57577896118164, "learning_rate": 1.3700000000000002e-06, "loss": 5.7312, "step": 2740 }, { "epoch": 0.8592476489028213, "grad_norm": 25.569894790649414, "learning_rate": 1.3705000000000003e-06, "loss": 5.5384, "step": 2741 }, { "epoch": 0.8595611285266458, "grad_norm": 40.88070297241211, "learning_rate": 1.371e-06, "loss": 7.8681, "step": 2742 }, { "epoch": 0.8598746081504702, "grad_norm": 32.0083122253418, "learning_rate": 1.3715000000000001e-06, "loss": 6.5132, "step": 2743 }, { "epoch": 0.8601880877742947, "grad_norm": 33.715633392333984, "learning_rate": 1.372e-06, "loss": 6.0128, "step": 2744 }, { "epoch": 0.8605015673981191, "grad_norm": 43.099884033203125, "learning_rate": 1.3725000000000002e-06, "loss": 6.8138, "step": 2745 }, { "epoch": 0.8608150470219436, "grad_norm": 47.59553909301758, "learning_rate": 1.3730000000000002e-06, "loss": 8.2016, "step": 2746 }, { "epoch": 0.861128526645768, "grad_norm": 40.0462532043457, "learning_rate": 1.3735e-06, "loss": 6.6381, "step": 2747 }, { "epoch": 0.8614420062695924, "grad_norm": 35.96194076538086, "learning_rate": 1.374e-06, "loss": 7.399, "step": 2748 }, { "epoch": 0.8617554858934169, "grad_norm": 39.26766586303711, "learning_rate": 1.3745e-06, "loss": 6.7996, "step": 2749 }, { "epoch": 0.8620689655172413, "grad_norm": 51.32083511352539, "learning_rate": 1.3750000000000002e-06, "loss": 8.2172, "step": 2750 }, { "epoch": 0.8623824451410659, "grad_norm": 18.551576614379883, "learning_rate": 1.3755000000000002e-06, "loss": 6.0155, "step": 2751 }, { "epoch": 0.8626959247648903, "grad_norm": 20.321107864379883, "learning_rate": 1.376e-06, "loss": 5.1774, "step": 2752 }, { "epoch": 0.8630094043887148, "grad_norm": 25.40889549255371, "learning_rate": 1.3765e-06, "loss": 5.933, "step": 2753 }, { "epoch": 0.8633228840125392, "grad_norm": 40.312984466552734, "learning_rate": 1.377e-06, "loss": 6.7176, "step": 2754 }, { "epoch": 0.8636363636363636, "grad_norm": 38.97128677368164, "learning_rate": 1.3775000000000002e-06, "loss": 5.2988, "step": 2755 }, { "epoch": 0.8639498432601881, "grad_norm": 39.15882110595703, "learning_rate": 1.3780000000000002e-06, "loss": 7.5587, "step": 2756 }, { "epoch": 0.8642633228840125, "grad_norm": 37.12295913696289, "learning_rate": 1.3785e-06, "loss": 5.3687, "step": 2757 }, { "epoch": 0.864576802507837, "grad_norm": 69.34944152832031, "learning_rate": 1.379e-06, "loss": 10.4011, "step": 2758 }, { "epoch": 0.8648902821316614, "grad_norm": 46.22438430786133, "learning_rate": 1.3795e-06, "loss": 7.7162, "step": 2759 }, { "epoch": 0.8652037617554859, "grad_norm": 21.386926651000977, "learning_rate": 1.3800000000000001e-06, "loss": 4.8139, "step": 2760 }, { "epoch": 0.8655172413793103, "grad_norm": 41.118896484375, "learning_rate": 1.3805000000000002e-06, "loss": 6.995, "step": 2761 }, { "epoch": 0.8658307210031349, "grad_norm": 29.520702362060547, "learning_rate": 1.3810000000000002e-06, "loss": 5.1273, "step": 2762 }, { "epoch": 0.8661442006269593, "grad_norm": 36.20766830444336, "learning_rate": 1.3815e-06, "loss": 6.9384, "step": 2763 }, { "epoch": 0.8664576802507837, "grad_norm": 26.937095642089844, "learning_rate": 1.382e-06, "loss": 6.1545, "step": 2764 }, { "epoch": 0.8667711598746082, "grad_norm": 61.82893753051758, "learning_rate": 1.3825000000000003e-06, "loss": 8.894, "step": 2765 }, { "epoch": 0.8670846394984326, "grad_norm": 27.976224899291992, "learning_rate": 1.3830000000000001e-06, "loss": 4.7557, "step": 2766 }, { "epoch": 0.8673981191222571, "grad_norm": 79.30841827392578, "learning_rate": 1.3835000000000002e-06, "loss": 8.8842, "step": 2767 }, { "epoch": 0.8677115987460815, "grad_norm": 32.45915985107422, "learning_rate": 1.384e-06, "loss": 5.9744, "step": 2768 }, { "epoch": 0.868025078369906, "grad_norm": 27.075918197631836, "learning_rate": 1.3845e-06, "loss": 4.6374, "step": 2769 }, { "epoch": 0.8683385579937304, "grad_norm": 60.21990203857422, "learning_rate": 1.3850000000000003e-06, "loss": 6.1392, "step": 2770 }, { "epoch": 0.8686520376175548, "grad_norm": 38.180335998535156, "learning_rate": 1.3855000000000001e-06, "loss": 5.6472, "step": 2771 }, { "epoch": 0.8689655172413793, "grad_norm": 44.778709411621094, "learning_rate": 1.3860000000000002e-06, "loss": 6.1988, "step": 2772 }, { "epoch": 0.8692789968652037, "grad_norm": 44.12003707885742, "learning_rate": 1.3865e-06, "loss": 6.2107, "step": 2773 }, { "epoch": 0.8695924764890283, "grad_norm": 115.1012954711914, "learning_rate": 1.387e-06, "loss": 10.4227, "step": 2774 }, { "epoch": 0.8699059561128527, "grad_norm": 47.72286605834961, "learning_rate": 1.3875000000000003e-06, "loss": 5.7027, "step": 2775 }, { "epoch": 0.8702194357366771, "grad_norm": 42.00877380371094, "learning_rate": 1.388e-06, "loss": 7.9404, "step": 2776 }, { "epoch": 0.8705329153605016, "grad_norm": 30.263233184814453, "learning_rate": 1.3885000000000001e-06, "loss": 5.9391, "step": 2777 }, { "epoch": 0.870846394984326, "grad_norm": 32.171573638916016, "learning_rate": 1.3890000000000002e-06, "loss": 5.6703, "step": 2778 }, { "epoch": 0.8711598746081505, "grad_norm": 29.78256607055664, "learning_rate": 1.3895e-06, "loss": 5.0913, "step": 2779 }, { "epoch": 0.8714733542319749, "grad_norm": 60.958702087402344, "learning_rate": 1.3900000000000002e-06, "loss": 8.0207, "step": 2780 }, { "epoch": 0.8717868338557994, "grad_norm": 43.16213607788086, "learning_rate": 1.3905000000000003e-06, "loss": 6.1787, "step": 2781 }, { "epoch": 0.8721003134796238, "grad_norm": 46.14605712890625, "learning_rate": 1.3910000000000001e-06, "loss": 5.719, "step": 2782 }, { "epoch": 0.8724137931034482, "grad_norm": 43.69407272338867, "learning_rate": 1.3915000000000001e-06, "loss": 7.7847, "step": 2783 }, { "epoch": 0.8727272727272727, "grad_norm": 41.964820861816406, "learning_rate": 1.392e-06, "loss": 5.3784, "step": 2784 }, { "epoch": 0.8730407523510971, "grad_norm": 33.25990295410156, "learning_rate": 1.3925000000000002e-06, "loss": 5.3528, "step": 2785 }, { "epoch": 0.8733542319749217, "grad_norm": 37.829498291015625, "learning_rate": 1.3930000000000003e-06, "loss": 5.1219, "step": 2786 }, { "epoch": 0.8736677115987461, "grad_norm": 42.88932418823242, "learning_rate": 1.3935e-06, "loss": 6.6543, "step": 2787 }, { "epoch": 0.8739811912225706, "grad_norm": 63.86898422241211, "learning_rate": 1.3940000000000001e-06, "loss": 9.7226, "step": 2788 }, { "epoch": 0.874294670846395, "grad_norm": 45.50909423828125, "learning_rate": 1.3945e-06, "loss": 7.3516, "step": 2789 }, { "epoch": 0.8746081504702194, "grad_norm": 40.1636962890625, "learning_rate": 1.3950000000000002e-06, "loss": 5.1074, "step": 2790 }, { "epoch": 0.8749216300940439, "grad_norm": 37.66518020629883, "learning_rate": 1.3955000000000002e-06, "loss": 5.9449, "step": 2791 }, { "epoch": 0.8752351097178683, "grad_norm": 77.52034759521484, "learning_rate": 1.396e-06, "loss": 8.5409, "step": 2792 }, { "epoch": 0.8755485893416928, "grad_norm": 45.52507019042969, "learning_rate": 1.3965e-06, "loss": 6.171, "step": 2793 }, { "epoch": 0.8758620689655172, "grad_norm": 41.04618835449219, "learning_rate": 1.397e-06, "loss": 5.7984, "step": 2794 }, { "epoch": 0.8761755485893417, "grad_norm": 57.51970672607422, "learning_rate": 1.3975000000000002e-06, "loss": 6.6904, "step": 2795 }, { "epoch": 0.8764890282131661, "grad_norm": 59.90044403076172, "learning_rate": 1.3980000000000002e-06, "loss": 6.9548, "step": 2796 }, { "epoch": 0.8768025078369905, "grad_norm": 37.544612884521484, "learning_rate": 1.3985e-06, "loss": 6.0155, "step": 2797 }, { "epoch": 0.8771159874608151, "grad_norm": 43.11286926269531, "learning_rate": 1.399e-06, "loss": 5.8831, "step": 2798 }, { "epoch": 0.8774294670846395, "grad_norm": 22.86078453063965, "learning_rate": 1.3995000000000001e-06, "loss": 3.8691, "step": 2799 }, { "epoch": 0.877742946708464, "grad_norm": 24.594881057739258, "learning_rate": 1.4000000000000001e-06, "loss": 4.9944, "step": 2800 }, { "epoch": 0.8780564263322884, "grad_norm": 32.79366683959961, "learning_rate": 1.4005000000000002e-06, "loss": 4.5094, "step": 2801 }, { "epoch": 0.8783699059561129, "grad_norm": 40.043540954589844, "learning_rate": 1.4010000000000002e-06, "loss": 5.0449, "step": 2802 }, { "epoch": 0.8786833855799373, "grad_norm": 39.7636604309082, "learning_rate": 1.4015e-06, "loss": 4.8512, "step": 2803 }, { "epoch": 0.8789968652037617, "grad_norm": 38.312355041503906, "learning_rate": 1.402e-06, "loss": 4.9508, "step": 2804 }, { "epoch": 0.8793103448275862, "grad_norm": 40.00060272216797, "learning_rate": 1.4025000000000003e-06, "loss": 6.2614, "step": 2805 }, { "epoch": 0.8796238244514106, "grad_norm": 38.401275634765625, "learning_rate": 1.4030000000000002e-06, "loss": 5.2398, "step": 2806 }, { "epoch": 0.8799373040752351, "grad_norm": 40.27410125732422, "learning_rate": 1.4035000000000002e-06, "loss": 5.9456, "step": 2807 }, { "epoch": 0.8802507836990595, "grad_norm": 28.530675888061523, "learning_rate": 1.404e-06, "loss": 4.5628, "step": 2808 }, { "epoch": 0.8805642633228841, "grad_norm": 39.466033935546875, "learning_rate": 1.4045e-06, "loss": 5.2629, "step": 2809 }, { "epoch": 0.8808777429467085, "grad_norm": 29.432931900024414, "learning_rate": 1.4050000000000003e-06, "loss": 4.6265, "step": 2810 }, { "epoch": 0.881191222570533, "grad_norm": 48.174129486083984, "learning_rate": 1.4055000000000001e-06, "loss": 4.853, "step": 2811 }, { "epoch": 0.8815047021943574, "grad_norm": 53.55119323730469, "learning_rate": 1.4060000000000002e-06, "loss": 7.2377, "step": 2812 }, { "epoch": 0.8818181818181818, "grad_norm": 36.38772964477539, "learning_rate": 1.4065e-06, "loss": 5.2177, "step": 2813 }, { "epoch": 0.8821316614420063, "grad_norm": 34.217830657958984, "learning_rate": 1.407e-06, "loss": 5.3052, "step": 2814 }, { "epoch": 0.8824451410658307, "grad_norm": 42.84861373901367, "learning_rate": 1.4075e-06, "loss": 5.6933, "step": 2815 }, { "epoch": 0.8827586206896552, "grad_norm": 53.72264862060547, "learning_rate": 1.4080000000000001e-06, "loss": 5.3901, "step": 2816 }, { "epoch": 0.8830721003134796, "grad_norm": 53.93974685668945, "learning_rate": 1.4085000000000002e-06, "loss": 5.2435, "step": 2817 }, { "epoch": 0.883385579937304, "grad_norm": 28.91839027404785, "learning_rate": 1.4090000000000002e-06, "loss": 3.7364, "step": 2818 }, { "epoch": 0.8836990595611285, "grad_norm": 36.89215850830078, "learning_rate": 1.4095e-06, "loss": 4.5822, "step": 2819 }, { "epoch": 0.8840125391849529, "grad_norm": 33.66713333129883, "learning_rate": 1.41e-06, "loss": 4.6832, "step": 2820 }, { "epoch": 0.8843260188087775, "grad_norm": 44.25896453857422, "learning_rate": 1.4105000000000003e-06, "loss": 6.3235, "step": 2821 }, { "epoch": 0.8846394984326019, "grad_norm": 34.358551025390625, "learning_rate": 1.4110000000000001e-06, "loss": 5.9712, "step": 2822 }, { "epoch": 0.8849529780564264, "grad_norm": 36.53178787231445, "learning_rate": 1.4115000000000002e-06, "loss": 5.4185, "step": 2823 }, { "epoch": 0.8852664576802508, "grad_norm": 61.49184036254883, "learning_rate": 1.412e-06, "loss": 6.4534, "step": 2824 }, { "epoch": 0.8855799373040752, "grad_norm": 96.07270812988281, "learning_rate": 1.4125e-06, "loss": 9.2125, "step": 2825 }, { "epoch": 0.8858934169278997, "grad_norm": 44.58479690551758, "learning_rate": 1.4130000000000003e-06, "loss": 4.9576, "step": 2826 }, { "epoch": 0.8862068965517241, "grad_norm": 52.979209899902344, "learning_rate": 1.4135e-06, "loss": 5.3499, "step": 2827 }, { "epoch": 0.8865203761755486, "grad_norm": 37.45901107788086, "learning_rate": 1.4140000000000001e-06, "loss": 4.4372, "step": 2828 }, { "epoch": 0.886833855799373, "grad_norm": 44.921173095703125, "learning_rate": 1.4145e-06, "loss": 5.9111, "step": 2829 }, { "epoch": 0.8871473354231975, "grad_norm": 38.52549362182617, "learning_rate": 1.415e-06, "loss": 4.8988, "step": 2830 }, { "epoch": 0.8874608150470219, "grad_norm": 48.08058166503906, "learning_rate": 1.4155000000000003e-06, "loss": 4.7856, "step": 2831 }, { "epoch": 0.8877742946708463, "grad_norm": 32.446327209472656, "learning_rate": 1.416e-06, "loss": 4.5528, "step": 2832 }, { "epoch": 0.8880877742946709, "grad_norm": 86.36289978027344, "learning_rate": 1.4165000000000001e-06, "loss": 7.3174, "step": 2833 }, { "epoch": 0.8884012539184953, "grad_norm": 64.09017944335938, "learning_rate": 1.417e-06, "loss": 6.8142, "step": 2834 }, { "epoch": 0.8887147335423198, "grad_norm": 52.754486083984375, "learning_rate": 1.4175e-06, "loss": 6.6618, "step": 2835 }, { "epoch": 0.8890282131661442, "grad_norm": 43.92601013183594, "learning_rate": 1.4180000000000002e-06, "loss": 5.7932, "step": 2836 }, { "epoch": 0.8893416927899687, "grad_norm": 39.43013000488281, "learning_rate": 1.4185e-06, "loss": 5.5289, "step": 2837 }, { "epoch": 0.8896551724137931, "grad_norm": 42.97591018676758, "learning_rate": 1.419e-06, "loss": 4.7558, "step": 2838 }, { "epoch": 0.8899686520376175, "grad_norm": 43.5150032043457, "learning_rate": 1.4195000000000001e-06, "loss": 4.532, "step": 2839 }, { "epoch": 0.890282131661442, "grad_norm": 40.31664276123047, "learning_rate": 1.42e-06, "loss": 5.3421, "step": 2840 }, { "epoch": 0.8905956112852664, "grad_norm": 34.56455612182617, "learning_rate": 1.4205000000000002e-06, "loss": 4.0237, "step": 2841 }, { "epoch": 0.8909090909090909, "grad_norm": 30.62310218811035, "learning_rate": 1.4210000000000002e-06, "loss": 4.2576, "step": 2842 }, { "epoch": 0.8912225705329153, "grad_norm": 40.746028900146484, "learning_rate": 1.4215e-06, "loss": 4.587, "step": 2843 }, { "epoch": 0.8915360501567398, "grad_norm": 31.952014923095703, "learning_rate": 1.4220000000000001e-06, "loss": 3.8408, "step": 2844 }, { "epoch": 0.8918495297805643, "grad_norm": 26.450883865356445, "learning_rate": 1.4225e-06, "loss": 4.4055, "step": 2845 }, { "epoch": 0.8921630094043888, "grad_norm": 47.278404235839844, "learning_rate": 1.4230000000000002e-06, "loss": 4.9155, "step": 2846 }, { "epoch": 0.8924764890282132, "grad_norm": 55.63157272338867, "learning_rate": 1.4235000000000002e-06, "loss": 5.8403, "step": 2847 }, { "epoch": 0.8927899686520376, "grad_norm": 30.85458755493164, "learning_rate": 1.424e-06, "loss": 4.5499, "step": 2848 }, { "epoch": 0.8931034482758621, "grad_norm": 36.83681106567383, "learning_rate": 1.4245e-06, "loss": 5.0289, "step": 2849 }, { "epoch": 0.8934169278996865, "grad_norm": 32.328121185302734, "learning_rate": 1.425e-06, "loss": 4.2722, "step": 2850 }, { "epoch": 0.893730407523511, "grad_norm": 44.29472732543945, "learning_rate": 1.4255000000000002e-06, "loss": 4.4436, "step": 2851 }, { "epoch": 0.8940438871473354, "grad_norm": 46.61187744140625, "learning_rate": 1.4260000000000002e-06, "loss": 5.3737, "step": 2852 }, { "epoch": 0.8943573667711598, "grad_norm": 35.942100524902344, "learning_rate": 1.4265e-06, "loss": 4.4256, "step": 2853 }, { "epoch": 0.8946708463949843, "grad_norm": 32.366519927978516, "learning_rate": 1.427e-06, "loss": 3.8425, "step": 2854 }, { "epoch": 0.8949843260188087, "grad_norm": 35.66120529174805, "learning_rate": 1.4275e-06, "loss": 4.4804, "step": 2855 }, { "epoch": 0.8952978056426333, "grad_norm": 37.548492431640625, "learning_rate": 1.4280000000000001e-06, "loss": 5.4055, "step": 2856 }, { "epoch": 0.8956112852664577, "grad_norm": 41.95212936401367, "learning_rate": 1.4285000000000002e-06, "loss": 4.8837, "step": 2857 }, { "epoch": 0.8959247648902822, "grad_norm": 34.09566879272461, "learning_rate": 1.4290000000000002e-06, "loss": 4.8557, "step": 2858 }, { "epoch": 0.8962382445141066, "grad_norm": 52.16231918334961, "learning_rate": 1.4295e-06, "loss": 4.9953, "step": 2859 }, { "epoch": 0.896551724137931, "grad_norm": 38.20033264160156, "learning_rate": 1.43e-06, "loss": 4.2838, "step": 2860 }, { "epoch": 0.8968652037617555, "grad_norm": 52.53467559814453, "learning_rate": 1.4305000000000003e-06, "loss": 5.1196, "step": 2861 }, { "epoch": 0.8971786833855799, "grad_norm": 68.67937469482422, "learning_rate": 1.4310000000000001e-06, "loss": 5.4295, "step": 2862 }, { "epoch": 0.8974921630094044, "grad_norm": 39.02657699584961, "learning_rate": 1.4315000000000002e-06, "loss": 3.6579, "step": 2863 }, { "epoch": 0.8978056426332288, "grad_norm": 27.187744140625, "learning_rate": 1.432e-06, "loss": 3.7283, "step": 2864 }, { "epoch": 0.8981191222570533, "grad_norm": 42.428802490234375, "learning_rate": 1.4325e-06, "loss": 4.022, "step": 2865 }, { "epoch": 0.8984326018808777, "grad_norm": 29.769405364990234, "learning_rate": 1.4330000000000003e-06, "loss": 3.7161, "step": 2866 }, { "epoch": 0.8987460815047021, "grad_norm": 40.81663513183594, "learning_rate": 1.4335000000000001e-06, "loss": 4.9217, "step": 2867 }, { "epoch": 0.8990595611285267, "grad_norm": 48.31073760986328, "learning_rate": 1.4340000000000002e-06, "loss": 4.4489, "step": 2868 }, { "epoch": 0.8993730407523511, "grad_norm": 32.35671615600586, "learning_rate": 1.4345e-06, "loss": 4.1166, "step": 2869 }, { "epoch": 0.8996865203761756, "grad_norm": 46.619041442871094, "learning_rate": 1.435e-06, "loss": 6.4497, "step": 2870 }, { "epoch": 0.9, "grad_norm": 56.73939514160156, "learning_rate": 1.4355000000000003e-06, "loss": 5.7911, "step": 2871 }, { "epoch": 0.9003134796238245, "grad_norm": 31.99533462524414, "learning_rate": 1.436e-06, "loss": 3.6651, "step": 2872 }, { "epoch": 0.9006269592476489, "grad_norm": 37.6666374206543, "learning_rate": 1.4365000000000001e-06, "loss": 3.6502, "step": 2873 }, { "epoch": 0.9009404388714733, "grad_norm": 23.286277770996094, "learning_rate": 1.437e-06, "loss": 3.7086, "step": 2874 }, { "epoch": 0.9012539184952978, "grad_norm": 42.88998794555664, "learning_rate": 1.4375e-06, "loss": 5.5071, "step": 2875 }, { "epoch": 0.9015673981191222, "grad_norm": 37.3940315246582, "learning_rate": 1.4380000000000003e-06, "loss": 5.1036, "step": 2876 }, { "epoch": 0.9018808777429467, "grad_norm": 48.386104583740234, "learning_rate": 1.4385e-06, "loss": 4.5443, "step": 2877 }, { "epoch": 0.9021943573667711, "grad_norm": 63.84465408325195, "learning_rate": 1.4390000000000001e-06, "loss": 6.2741, "step": 2878 }, { "epoch": 0.9025078369905956, "grad_norm": 32.616859436035156, "learning_rate": 1.4395000000000002e-06, "loss": 4.203, "step": 2879 }, { "epoch": 0.9028213166144201, "grad_norm": 42.289005279541016, "learning_rate": 1.44e-06, "loss": 5.2399, "step": 2880 }, { "epoch": 0.9031347962382446, "grad_norm": 62.790042877197266, "learning_rate": 1.4405000000000002e-06, "loss": 4.648, "step": 2881 }, { "epoch": 0.903448275862069, "grad_norm": 51.915889739990234, "learning_rate": 1.4410000000000003e-06, "loss": 5.2985, "step": 2882 }, { "epoch": 0.9037617554858934, "grad_norm": 75.9524154663086, "learning_rate": 1.4415e-06, "loss": 6.2667, "step": 2883 }, { "epoch": 0.9040752351097179, "grad_norm": 28.834205627441406, "learning_rate": 1.4420000000000001e-06, "loss": 3.4066, "step": 2884 }, { "epoch": 0.9043887147335423, "grad_norm": 51.43450927734375, "learning_rate": 1.4425e-06, "loss": 6.0204, "step": 2885 }, { "epoch": 0.9047021943573668, "grad_norm": 76.39936065673828, "learning_rate": 1.4430000000000002e-06, "loss": 5.3644, "step": 2886 }, { "epoch": 0.9050156739811912, "grad_norm": 39.322025299072266, "learning_rate": 1.4435000000000002e-06, "loss": 4.0316, "step": 2887 }, { "epoch": 0.9053291536050156, "grad_norm": 30.268177032470703, "learning_rate": 1.444e-06, "loss": 4.4225, "step": 2888 }, { "epoch": 0.9056426332288401, "grad_norm": 32.31449508666992, "learning_rate": 1.4445e-06, "loss": 4.538, "step": 2889 }, { "epoch": 0.9059561128526645, "grad_norm": 26.894195556640625, "learning_rate": 1.445e-06, "loss": 3.7654, "step": 2890 }, { "epoch": 0.9062695924764891, "grad_norm": 67.31766510009766, "learning_rate": 1.4455000000000002e-06, "loss": 6.158, "step": 2891 }, { "epoch": 0.9065830721003135, "grad_norm": 24.468050003051758, "learning_rate": 1.4460000000000002e-06, "loss": 3.9294, "step": 2892 }, { "epoch": 0.906896551724138, "grad_norm": 47.52812576293945, "learning_rate": 1.4465e-06, "loss": 5.1586, "step": 2893 }, { "epoch": 0.9072100313479624, "grad_norm": 35.4903450012207, "learning_rate": 1.447e-06, "loss": 4.418, "step": 2894 }, { "epoch": 0.9075235109717869, "grad_norm": 31.2788028717041, "learning_rate": 1.4475000000000001e-06, "loss": 3.6287, "step": 2895 }, { "epoch": 0.9078369905956113, "grad_norm": 31.898345947265625, "learning_rate": 1.4480000000000002e-06, "loss": 3.8173, "step": 2896 }, { "epoch": 0.9081504702194357, "grad_norm": 59.66981506347656, "learning_rate": 1.4485000000000002e-06, "loss": 5.177, "step": 2897 }, { "epoch": 0.9084639498432602, "grad_norm": 50.7946891784668, "learning_rate": 1.4490000000000002e-06, "loss": 5.3483, "step": 2898 }, { "epoch": 0.9087774294670846, "grad_norm": 27.841205596923828, "learning_rate": 1.4495e-06, "loss": 3.2566, "step": 2899 }, { "epoch": 0.9090909090909091, "grad_norm": 29.948871612548828, "learning_rate": 1.45e-06, "loss": 4.0416, "step": 2900 }, { "epoch": 0.9094043887147335, "grad_norm": 23.60964584350586, "learning_rate": 1.4505000000000003e-06, "loss": 3.1585, "step": 2901 }, { "epoch": 0.909717868338558, "grad_norm": 33.11689376831055, "learning_rate": 1.4510000000000002e-06, "loss": 4.4712, "step": 2902 }, { "epoch": 0.9100313479623825, "grad_norm": 62.05201721191406, "learning_rate": 1.4515000000000002e-06, "loss": 5.4037, "step": 2903 }, { "epoch": 0.9103448275862069, "grad_norm": 105.85570526123047, "learning_rate": 1.452e-06, "loss": 5.0095, "step": 2904 }, { "epoch": 0.9106583072100314, "grad_norm": 45.89641189575195, "learning_rate": 1.4525e-06, "loss": 3.9751, "step": 2905 }, { "epoch": 0.9109717868338558, "grad_norm": 52.051944732666016, "learning_rate": 1.4530000000000003e-06, "loss": 4.7415, "step": 2906 }, { "epoch": 0.9112852664576803, "grad_norm": 67.7939682006836, "learning_rate": 1.4535000000000001e-06, "loss": 4.7054, "step": 2907 }, { "epoch": 0.9115987460815047, "grad_norm": 41.19660949707031, "learning_rate": 1.4540000000000002e-06, "loss": 5.2859, "step": 2908 }, { "epoch": 0.9119122257053291, "grad_norm": 46.74116516113281, "learning_rate": 1.4545e-06, "loss": 4.0541, "step": 2909 }, { "epoch": 0.9122257053291536, "grad_norm": 84.27909088134766, "learning_rate": 1.455e-06, "loss": 5.2779, "step": 2910 }, { "epoch": 0.912539184952978, "grad_norm": 53.99125289916992, "learning_rate": 1.4555000000000003e-06, "loss": 4.3379, "step": 2911 }, { "epoch": 0.9128526645768025, "grad_norm": 52.822410583496094, "learning_rate": 1.4560000000000001e-06, "loss": 5.3916, "step": 2912 }, { "epoch": 0.9131661442006269, "grad_norm": 53.39067840576172, "learning_rate": 1.4565000000000002e-06, "loss": 5.6752, "step": 2913 }, { "epoch": 0.9134796238244514, "grad_norm": 42.46992111206055, "learning_rate": 1.457e-06, "loss": 3.9782, "step": 2914 }, { "epoch": 0.9137931034482759, "grad_norm": 31.098447799682617, "learning_rate": 1.4575e-06, "loss": 3.5816, "step": 2915 }, { "epoch": 0.9141065830721004, "grad_norm": 49.15898513793945, "learning_rate": 1.4580000000000003e-06, "loss": 3.5684, "step": 2916 }, { "epoch": 0.9144200626959248, "grad_norm": 49.75162124633789, "learning_rate": 1.4585e-06, "loss": 4.8455, "step": 2917 }, { "epoch": 0.9147335423197492, "grad_norm": 51.76414108276367, "learning_rate": 1.4590000000000001e-06, "loss": 4.1124, "step": 2918 }, { "epoch": 0.9150470219435737, "grad_norm": 41.26679229736328, "learning_rate": 1.4595000000000002e-06, "loss": 4.2793, "step": 2919 }, { "epoch": 0.9153605015673981, "grad_norm": 43.92702865600586, "learning_rate": 1.46e-06, "loss": 4.6872, "step": 2920 }, { "epoch": 0.9156739811912226, "grad_norm": 34.27779769897461, "learning_rate": 1.4605000000000002e-06, "loss": 3.8484, "step": 2921 }, { "epoch": 0.915987460815047, "grad_norm": 31.705305099487305, "learning_rate": 1.4610000000000003e-06, "loss": 4.3887, "step": 2922 }, { "epoch": 0.9163009404388714, "grad_norm": 53.87886428833008, "learning_rate": 1.4615000000000001e-06, "loss": 4.4851, "step": 2923 }, { "epoch": 0.9166144200626959, "grad_norm": 43.97981262207031, "learning_rate": 1.4620000000000001e-06, "loss": 3.952, "step": 2924 }, { "epoch": 0.9169278996865203, "grad_norm": 36.8157958984375, "learning_rate": 1.4625e-06, "loss": 3.4762, "step": 2925 }, { "epoch": 0.9172413793103448, "grad_norm": 35.28675079345703, "learning_rate": 1.4630000000000002e-06, "loss": 4.2781, "step": 2926 }, { "epoch": 0.9175548589341693, "grad_norm": 37.354148864746094, "learning_rate": 1.4635000000000003e-06, "loss": 3.6935, "step": 2927 }, { "epoch": 0.9178683385579938, "grad_norm": 61.448665618896484, "learning_rate": 1.464e-06, "loss": 5.5435, "step": 2928 }, { "epoch": 0.9181818181818182, "grad_norm": 33.78268814086914, "learning_rate": 1.4645000000000001e-06, "loss": 4.5201, "step": 2929 }, { "epoch": 0.9184952978056427, "grad_norm": 37.02507019042969, "learning_rate": 1.465e-06, "loss": 3.5038, "step": 2930 }, { "epoch": 0.9188087774294671, "grad_norm": 29.58689308166504, "learning_rate": 1.4655000000000002e-06, "loss": 3.4179, "step": 2931 }, { "epoch": 0.9191222570532915, "grad_norm": 56.62324142456055, "learning_rate": 1.4660000000000002e-06, "loss": 3.9863, "step": 2932 }, { "epoch": 0.919435736677116, "grad_norm": 30.186080932617188, "learning_rate": 1.4665e-06, "loss": 3.6394, "step": 2933 }, { "epoch": 0.9197492163009404, "grad_norm": 91.04447937011719, "learning_rate": 1.467e-06, "loss": 5.1371, "step": 2934 }, { "epoch": 0.9200626959247649, "grad_norm": 94.30821990966797, "learning_rate": 1.4675000000000001e-06, "loss": 4.2498, "step": 2935 }, { "epoch": 0.9203761755485893, "grad_norm": 42.256168365478516, "learning_rate": 1.4680000000000002e-06, "loss": 4.5557, "step": 2936 }, { "epoch": 0.9206896551724137, "grad_norm": 38.17436599731445, "learning_rate": 1.4685000000000002e-06, "loss": 4.0215, "step": 2937 }, { "epoch": 0.9210031347962383, "grad_norm": 38.624698638916016, "learning_rate": 1.4690000000000003e-06, "loss": 3.5599, "step": 2938 }, { "epoch": 0.9213166144200627, "grad_norm": 38.51246643066406, "learning_rate": 1.4695e-06, "loss": 3.5211, "step": 2939 }, { "epoch": 0.9216300940438872, "grad_norm": 37.080074310302734, "learning_rate": 1.4700000000000001e-06, "loss": 5.7069, "step": 2940 }, { "epoch": 0.9219435736677116, "grad_norm": 37.63076400756836, "learning_rate": 1.4705e-06, "loss": 4.2385, "step": 2941 }, { "epoch": 0.9222570532915361, "grad_norm": 82.26602935791016, "learning_rate": 1.4710000000000002e-06, "loss": 3.6196, "step": 2942 }, { "epoch": 0.9225705329153605, "grad_norm": 44.46656036376953, "learning_rate": 1.4715000000000002e-06, "loss": 4.0862, "step": 2943 }, { "epoch": 0.922884012539185, "grad_norm": 34.88580322265625, "learning_rate": 1.472e-06, "loss": 3.5532, "step": 2944 }, { "epoch": 0.9231974921630094, "grad_norm": 41.69892501831055, "learning_rate": 1.4725e-06, "loss": 4.2038, "step": 2945 }, { "epoch": 0.9235109717868338, "grad_norm": 48.94770431518555, "learning_rate": 1.473e-06, "loss": 4.1548, "step": 2946 }, { "epoch": 0.9238244514106583, "grad_norm": 30.48263168334961, "learning_rate": 1.4735000000000002e-06, "loss": 3.5391, "step": 2947 }, { "epoch": 0.9241379310344827, "grad_norm": 37.1695671081543, "learning_rate": 1.4740000000000002e-06, "loss": 3.9293, "step": 2948 }, { "epoch": 0.9244514106583072, "grad_norm": 132.91793823242188, "learning_rate": 1.4745e-06, "loss": 4.7049, "step": 2949 }, { "epoch": 0.9247648902821317, "grad_norm": 52.81797790527344, "learning_rate": 1.475e-06, "loss": 4.619, "step": 2950 }, { "epoch": 0.9250783699059562, "grad_norm": 27.379749298095703, "learning_rate": 1.4754999999999999e-06, "loss": 3.5159, "step": 2951 }, { "epoch": 0.9253918495297806, "grad_norm": 37.0819206237793, "learning_rate": 1.4760000000000001e-06, "loss": 5.0518, "step": 2952 }, { "epoch": 0.925705329153605, "grad_norm": 60.48858642578125, "learning_rate": 1.4765000000000002e-06, "loss": 4.3286, "step": 2953 }, { "epoch": 0.9260188087774295, "grad_norm": 60.34362030029297, "learning_rate": 1.477e-06, "loss": 4.1523, "step": 2954 }, { "epoch": 0.9263322884012539, "grad_norm": 47.12152862548828, "learning_rate": 1.4775e-06, "loss": 3.5215, "step": 2955 }, { "epoch": 0.9266457680250784, "grad_norm": 33.692054748535156, "learning_rate": 1.478e-06, "loss": 3.7666, "step": 2956 }, { "epoch": 0.9269592476489028, "grad_norm": 32.28064727783203, "learning_rate": 1.4785000000000001e-06, "loss": 3.2797, "step": 2957 }, { "epoch": 0.9272727272727272, "grad_norm": 47.644588470458984, "learning_rate": 1.4790000000000002e-06, "loss": 4.0666, "step": 2958 }, { "epoch": 0.9275862068965517, "grad_norm": 37.37916564941406, "learning_rate": 1.4795000000000002e-06, "loss": 3.9653, "step": 2959 }, { "epoch": 0.9278996865203761, "grad_norm": 34.87968444824219, "learning_rate": 1.48e-06, "loss": 3.4317, "step": 2960 }, { "epoch": 0.9282131661442006, "grad_norm": 75.75358581542969, "learning_rate": 1.4805e-06, "loss": 5.1771, "step": 2961 }, { "epoch": 0.9285266457680251, "grad_norm": 39.766578674316406, "learning_rate": 1.4810000000000003e-06, "loss": 3.8669, "step": 2962 }, { "epoch": 0.9288401253918496, "grad_norm": 69.24491119384766, "learning_rate": 1.4815000000000001e-06, "loss": 4.7225, "step": 2963 }, { "epoch": 0.929153605015674, "grad_norm": 84.43096923828125, "learning_rate": 1.4820000000000002e-06, "loss": 4.2543, "step": 2964 }, { "epoch": 0.9294670846394985, "grad_norm": 41.02848434448242, "learning_rate": 1.4825e-06, "loss": 4.1251, "step": 2965 }, { "epoch": 0.9297805642633229, "grad_norm": 41.528892517089844, "learning_rate": 1.483e-06, "loss": 3.6634, "step": 2966 }, { "epoch": 0.9300940438871473, "grad_norm": 28.742746353149414, "learning_rate": 1.4835000000000003e-06, "loss": 4.2559, "step": 2967 }, { "epoch": 0.9304075235109718, "grad_norm": 52.95411682128906, "learning_rate": 1.4840000000000001e-06, "loss": 3.8098, "step": 2968 }, { "epoch": 0.9307210031347962, "grad_norm": 48.885982513427734, "learning_rate": 1.4845000000000001e-06, "loss": 4.4849, "step": 2969 }, { "epoch": 0.9310344827586207, "grad_norm": 36.8331298828125, "learning_rate": 1.485e-06, "loss": 4.2251, "step": 2970 }, { "epoch": 0.9313479623824451, "grad_norm": 33.98138427734375, "learning_rate": 1.4855e-06, "loss": 3.6462, "step": 2971 }, { "epoch": 0.9316614420062695, "grad_norm": 33.16653823852539, "learning_rate": 1.4860000000000003e-06, "loss": 3.4936, "step": 2972 }, { "epoch": 0.931974921630094, "grad_norm": 23.719505310058594, "learning_rate": 1.4865e-06, "loss": 3.3029, "step": 2973 }, { "epoch": 0.9322884012539185, "grad_norm": 34.3863410949707, "learning_rate": 1.4870000000000001e-06, "loss": 4.0512, "step": 2974 }, { "epoch": 0.932601880877743, "grad_norm": 31.7304744720459, "learning_rate": 1.4875000000000002e-06, "loss": 3.8679, "step": 2975 }, { "epoch": 0.9329153605015674, "grad_norm": 36.177642822265625, "learning_rate": 1.488e-06, "loss": 3.5125, "step": 2976 }, { "epoch": 0.9332288401253919, "grad_norm": 48.70751953125, "learning_rate": 1.4885000000000002e-06, "loss": 4.5736, "step": 2977 }, { "epoch": 0.9335423197492163, "grad_norm": 47.91338348388672, "learning_rate": 1.4890000000000003e-06, "loss": 4.3447, "step": 2978 }, { "epoch": 0.9338557993730408, "grad_norm": 38.861026763916016, "learning_rate": 1.4895e-06, "loss": 3.2831, "step": 2979 }, { "epoch": 0.9341692789968652, "grad_norm": 31.002511978149414, "learning_rate": 1.4900000000000001e-06, "loss": 3.899, "step": 2980 }, { "epoch": 0.9344827586206896, "grad_norm": 45.859554290771484, "learning_rate": 1.4905e-06, "loss": 3.303, "step": 2981 }, { "epoch": 0.9347962382445141, "grad_norm": 37.03654479980469, "learning_rate": 1.4910000000000002e-06, "loss": 3.4903, "step": 2982 }, { "epoch": 0.9351097178683385, "grad_norm": 51.43019485473633, "learning_rate": 1.4915000000000002e-06, "loss": 3.4536, "step": 2983 }, { "epoch": 0.935423197492163, "grad_norm": 24.4971866607666, "learning_rate": 1.492e-06, "loss": 3.258, "step": 2984 }, { "epoch": 0.9357366771159875, "grad_norm": 35.39488220214844, "learning_rate": 1.4925000000000001e-06, "loss": 3.8771, "step": 2985 }, { "epoch": 0.936050156739812, "grad_norm": 33.246829986572266, "learning_rate": 1.493e-06, "loss": 3.5061, "step": 2986 }, { "epoch": 0.9363636363636364, "grad_norm": 80.51388549804688, "learning_rate": 1.4935000000000002e-06, "loss": 3.0297, "step": 2987 }, { "epoch": 0.9366771159874608, "grad_norm": 32.59467315673828, "learning_rate": 1.4940000000000002e-06, "loss": 3.5332, "step": 2988 }, { "epoch": 0.9369905956112853, "grad_norm": 35.27291488647461, "learning_rate": 1.4945e-06, "loss": 3.4688, "step": 2989 }, { "epoch": 0.9373040752351097, "grad_norm": 34.252323150634766, "learning_rate": 1.495e-06, "loss": 4.57, "step": 2990 }, { "epoch": 0.9376175548589342, "grad_norm": 38.402591705322266, "learning_rate": 1.4955e-06, "loss": 3.7867, "step": 2991 }, { "epoch": 0.9379310344827586, "grad_norm": 60.38703918457031, "learning_rate": 1.4960000000000002e-06, "loss": 3.5518, "step": 2992 }, { "epoch": 0.938244514106583, "grad_norm": 49.44574737548828, "learning_rate": 1.4965000000000002e-06, "loss": 4.4428, "step": 2993 }, { "epoch": 0.9385579937304075, "grad_norm": 36.240272521972656, "learning_rate": 1.497e-06, "loss": 3.9223, "step": 2994 }, { "epoch": 0.9388714733542319, "grad_norm": 43.761146545410156, "learning_rate": 1.4975e-06, "loss": 4.0446, "step": 2995 }, { "epoch": 0.9391849529780564, "grad_norm": 37.836631774902344, "learning_rate": 1.498e-06, "loss": 4.5072, "step": 2996 }, { "epoch": 0.9394984326018809, "grad_norm": 43.07557678222656, "learning_rate": 1.4985000000000001e-06, "loss": 3.9492, "step": 2997 }, { "epoch": 0.9398119122257054, "grad_norm": 45.68537521362305, "learning_rate": 1.4990000000000002e-06, "loss": 3.9818, "step": 2998 }, { "epoch": 0.9401253918495298, "grad_norm": 39.437374114990234, "learning_rate": 1.4995000000000002e-06, "loss": 3.8931, "step": 2999 }, { "epoch": 0.9404388714733543, "grad_norm": 43.97903823852539, "learning_rate": 1.5e-06, "loss": 3.9224, "step": 3000 }, { "epoch": 0.9407523510971787, "grad_norm": 39.799964904785156, "learning_rate": 1.5005e-06, "loss": 4.0903, "step": 3001 }, { "epoch": 0.9410658307210031, "grad_norm": 31.609031677246094, "learning_rate": 1.5010000000000003e-06, "loss": 3.1316, "step": 3002 }, { "epoch": 0.9413793103448276, "grad_norm": 44.23056411743164, "learning_rate": 1.5015000000000002e-06, "loss": 3.8267, "step": 3003 }, { "epoch": 0.941692789968652, "grad_norm": 54.30805206298828, "learning_rate": 1.5020000000000002e-06, "loss": 4.5096, "step": 3004 }, { "epoch": 0.9420062695924765, "grad_norm": 35.796661376953125, "learning_rate": 1.5025e-06, "loss": 3.364, "step": 3005 }, { "epoch": 0.9423197492163009, "grad_norm": 38.125091552734375, "learning_rate": 1.503e-06, "loss": 3.3353, "step": 3006 }, { "epoch": 0.9426332288401253, "grad_norm": 55.130775451660156, "learning_rate": 1.5035000000000003e-06, "loss": 3.269, "step": 3007 }, { "epoch": 0.9429467084639498, "grad_norm": 33.1065788269043, "learning_rate": 1.5040000000000001e-06, "loss": 3.8043, "step": 3008 }, { "epoch": 0.9432601880877743, "grad_norm": 32.7559814453125, "learning_rate": 1.5045000000000002e-06, "loss": 3.3734, "step": 3009 }, { "epoch": 0.9435736677115988, "grad_norm": 54.44166946411133, "learning_rate": 1.505e-06, "loss": 3.8441, "step": 3010 }, { "epoch": 0.9438871473354232, "grad_norm": 42.8200569152832, "learning_rate": 1.5055e-06, "loss": 3.2862, "step": 3011 }, { "epoch": 0.9442006269592477, "grad_norm": 33.59670639038086, "learning_rate": 1.5060000000000003e-06, "loss": 3.8924, "step": 3012 }, { "epoch": 0.9445141065830721, "grad_norm": 39.061279296875, "learning_rate": 1.5065e-06, "loss": 3.7316, "step": 3013 }, { "epoch": 0.9448275862068966, "grad_norm": 37.6108283996582, "learning_rate": 1.5070000000000001e-06, "loss": 4.0276, "step": 3014 }, { "epoch": 0.945141065830721, "grad_norm": 31.421833038330078, "learning_rate": 1.5075000000000002e-06, "loss": 3.2679, "step": 3015 }, { "epoch": 0.9454545454545454, "grad_norm": 34.2425422668457, "learning_rate": 1.508e-06, "loss": 3.7938, "step": 3016 }, { "epoch": 0.9457680250783699, "grad_norm": 58.670650482177734, "learning_rate": 1.5085000000000003e-06, "loss": 4.0145, "step": 3017 }, { "epoch": 0.9460815047021943, "grad_norm": 20.024412155151367, "learning_rate": 1.5090000000000003e-06, "loss": 3.1288, "step": 3018 }, { "epoch": 0.9463949843260188, "grad_norm": 58.533390045166016, "learning_rate": 1.5095000000000001e-06, "loss": 3.3225, "step": 3019 }, { "epoch": 0.9467084639498433, "grad_norm": 54.33335876464844, "learning_rate": 1.5100000000000002e-06, "loss": 6.4205, "step": 3020 }, { "epoch": 0.9470219435736678, "grad_norm": 48.962615966796875, "learning_rate": 1.5105e-06, "loss": 4.808, "step": 3021 }, { "epoch": 0.9473354231974922, "grad_norm": 45.55655288696289, "learning_rate": 1.5110000000000002e-06, "loss": 4.7946, "step": 3022 }, { "epoch": 0.9476489028213166, "grad_norm": 50.928199768066406, "learning_rate": 1.5115000000000003e-06, "loss": 2.9686, "step": 3023 }, { "epoch": 0.9479623824451411, "grad_norm": 48.291133880615234, "learning_rate": 1.512e-06, "loss": 4.4833, "step": 3024 }, { "epoch": 0.9482758620689655, "grad_norm": 53.17333221435547, "learning_rate": 1.5125000000000001e-06, "loss": 3.8591, "step": 3025 }, { "epoch": 0.94858934169279, "grad_norm": 39.857887268066406, "learning_rate": 1.513e-06, "loss": 3.6312, "step": 3026 }, { "epoch": 0.9489028213166144, "grad_norm": 54.0508918762207, "learning_rate": 1.5135000000000002e-06, "loss": 3.8041, "step": 3027 }, { "epoch": 0.9492163009404389, "grad_norm": 80.95004272460938, "learning_rate": 1.5140000000000002e-06, "loss": 5.0221, "step": 3028 }, { "epoch": 0.9495297805642633, "grad_norm": 49.28989791870117, "learning_rate": 1.5145e-06, "loss": 3.7739, "step": 3029 }, { "epoch": 0.9498432601880877, "grad_norm": 47.68058395385742, "learning_rate": 1.5150000000000001e-06, "loss": 4.321, "step": 3030 }, { "epoch": 0.9501567398119122, "grad_norm": 38.63204574584961, "learning_rate": 1.5155e-06, "loss": 3.6415, "step": 3031 }, { "epoch": 0.9504702194357367, "grad_norm": 71.45962524414062, "learning_rate": 1.5160000000000002e-06, "loss": 4.3374, "step": 3032 }, { "epoch": 0.9507836990595612, "grad_norm": 33.07627868652344, "learning_rate": 1.5165000000000002e-06, "loss": 4.8849, "step": 3033 }, { "epoch": 0.9510971786833856, "grad_norm": 43.36479187011719, "learning_rate": 1.517e-06, "loss": 4.4194, "step": 3034 }, { "epoch": 0.95141065830721, "grad_norm": 39.327003479003906, "learning_rate": 1.5175e-06, "loss": 3.8361, "step": 3035 }, { "epoch": 0.9517241379310345, "grad_norm": 42.822105407714844, "learning_rate": 1.5180000000000001e-06, "loss": 3.613, "step": 3036 }, { "epoch": 0.9520376175548589, "grad_norm": 36.2182502746582, "learning_rate": 1.5185000000000002e-06, "loss": 2.9997, "step": 3037 }, { "epoch": 0.9523510971786834, "grad_norm": 53.4159049987793, "learning_rate": 1.5190000000000002e-06, "loss": 3.312, "step": 3038 }, { "epoch": 0.9526645768025078, "grad_norm": 31.3487491607666, "learning_rate": 1.5195000000000002e-06, "loss": 3.2757, "step": 3039 }, { "epoch": 0.9529780564263323, "grad_norm": 30.916259765625, "learning_rate": 1.52e-06, "loss": 3.5675, "step": 3040 }, { "epoch": 0.9532915360501567, "grad_norm": 38.4372673034668, "learning_rate": 1.5205e-06, "loss": 3.7267, "step": 3041 }, { "epoch": 0.9536050156739811, "grad_norm": 33.22764587402344, "learning_rate": 1.5210000000000003e-06, "loss": 4.0307, "step": 3042 }, { "epoch": 0.9539184952978056, "grad_norm": 44.29248809814453, "learning_rate": 1.5215000000000002e-06, "loss": 4.2784, "step": 3043 }, { "epoch": 0.9542319749216301, "grad_norm": 28.41790199279785, "learning_rate": 1.5220000000000002e-06, "loss": 4.0886, "step": 3044 }, { "epoch": 0.9545454545454546, "grad_norm": 27.84803009033203, "learning_rate": 1.5225e-06, "loss": 3.059, "step": 3045 }, { "epoch": 0.954858934169279, "grad_norm": 64.40411376953125, "learning_rate": 1.523e-06, "loss": 4.3402, "step": 3046 }, { "epoch": 0.9551724137931035, "grad_norm": 41.377769470214844, "learning_rate": 1.5235000000000003e-06, "loss": 4.1004, "step": 3047 }, { "epoch": 0.9554858934169279, "grad_norm": 28.825105667114258, "learning_rate": 1.5240000000000001e-06, "loss": 3.5855, "step": 3048 }, { "epoch": 0.9557993730407524, "grad_norm": 36.527618408203125, "learning_rate": 1.5245000000000002e-06, "loss": 3.7634, "step": 3049 }, { "epoch": 0.9561128526645768, "grad_norm": 48.31051254272461, "learning_rate": 1.525e-06, "loss": 3.2032, "step": 3050 }, { "epoch": 0.9564263322884012, "grad_norm": 49.092552185058594, "learning_rate": 1.5255e-06, "loss": 3.2192, "step": 3051 }, { "epoch": 0.9567398119122257, "grad_norm": 28.44539451599121, "learning_rate": 1.5260000000000003e-06, "loss": 3.1739, "step": 3052 }, { "epoch": 0.9570532915360501, "grad_norm": 48.597190856933594, "learning_rate": 1.5265000000000001e-06, "loss": 5.3544, "step": 3053 }, { "epoch": 0.9573667711598746, "grad_norm": 37.88104248046875, "learning_rate": 1.5270000000000002e-06, "loss": 3.1668, "step": 3054 }, { "epoch": 0.957680250783699, "grad_norm": 88.76260375976562, "learning_rate": 1.5275000000000002e-06, "loss": 4.3589, "step": 3055 }, { "epoch": 0.9579937304075236, "grad_norm": 31.21636962890625, "learning_rate": 1.528e-06, "loss": 3.3685, "step": 3056 }, { "epoch": 0.958307210031348, "grad_norm": 44.11945343017578, "learning_rate": 1.5285000000000003e-06, "loss": 3.3106, "step": 3057 }, { "epoch": 0.9586206896551724, "grad_norm": 46.66257095336914, "learning_rate": 1.5290000000000003e-06, "loss": 3.5772, "step": 3058 }, { "epoch": 0.9589341692789969, "grad_norm": 45.46147537231445, "learning_rate": 1.5295000000000001e-06, "loss": 6.1372, "step": 3059 }, { "epoch": 0.9592476489028213, "grad_norm": 37.89000701904297, "learning_rate": 1.5300000000000002e-06, "loss": 3.3852, "step": 3060 }, { "epoch": 0.9595611285266458, "grad_norm": 38.45157241821289, "learning_rate": 1.5305e-06, "loss": 3.4452, "step": 3061 }, { "epoch": 0.9598746081504702, "grad_norm": 29.158227920532227, "learning_rate": 1.5310000000000002e-06, "loss": 3.0486, "step": 3062 }, { "epoch": 0.9601880877742947, "grad_norm": 97.31979370117188, "learning_rate": 1.5315000000000003e-06, "loss": 4.3573, "step": 3063 }, { "epoch": 0.9605015673981191, "grad_norm": 63.818050384521484, "learning_rate": 1.5320000000000001e-06, "loss": 4.6664, "step": 3064 }, { "epoch": 0.9608150470219435, "grad_norm": 28.385637283325195, "learning_rate": 1.5325000000000002e-06, "loss": 3.4442, "step": 3065 }, { "epoch": 0.961128526645768, "grad_norm": 40.3345947265625, "learning_rate": 1.533e-06, "loss": 3.7686, "step": 3066 }, { "epoch": 0.9614420062695925, "grad_norm": 35.8496208190918, "learning_rate": 1.5335e-06, "loss": 3.5075, "step": 3067 }, { "epoch": 0.961755485893417, "grad_norm": 31.015871047973633, "learning_rate": 1.5340000000000003e-06, "loss": 3.5269, "step": 3068 }, { "epoch": 0.9620689655172414, "grad_norm": 26.467029571533203, "learning_rate": 1.5345e-06, "loss": 3.0488, "step": 3069 }, { "epoch": 0.9623824451410659, "grad_norm": 57.80925369262695, "learning_rate": 1.5350000000000001e-06, "loss": 4.5774, "step": 3070 }, { "epoch": 0.9626959247648903, "grad_norm": 68.83757019042969, "learning_rate": 1.5355e-06, "loss": 4.7216, "step": 3071 }, { "epoch": 0.9630094043887147, "grad_norm": 44.15302658081055, "learning_rate": 1.536e-06, "loss": 4.4129, "step": 3072 }, { "epoch": 0.9633228840125392, "grad_norm": 48.796966552734375, "learning_rate": 1.5365000000000002e-06, "loss": 3.7593, "step": 3073 }, { "epoch": 0.9636363636363636, "grad_norm": 52.55888748168945, "learning_rate": 1.537e-06, "loss": 3.7959, "step": 3074 }, { "epoch": 0.9639498432601881, "grad_norm": 31.702449798583984, "learning_rate": 1.5375e-06, "loss": 3.7677, "step": 3075 }, { "epoch": 0.9642633228840125, "grad_norm": 47.94102096557617, "learning_rate": 1.5380000000000001e-06, "loss": 3.5232, "step": 3076 }, { "epoch": 0.964576802507837, "grad_norm": 46.60993957519531, "learning_rate": 1.5385e-06, "loss": 4.0139, "step": 3077 }, { "epoch": 0.9648902821316614, "grad_norm": 28.9523983001709, "learning_rate": 1.5390000000000002e-06, "loss": 2.7694, "step": 3078 }, { "epoch": 0.965203761755486, "grad_norm": 44.399925231933594, "learning_rate": 1.5395000000000003e-06, "loss": 3.5384, "step": 3079 }, { "epoch": 0.9655172413793104, "grad_norm": 42.65142822265625, "learning_rate": 1.54e-06, "loss": 3.7671, "step": 3080 }, { "epoch": 0.9658307210031348, "grad_norm": 51.35871887207031, "learning_rate": 1.5405000000000001e-06, "loss": 3.4669, "step": 3081 }, { "epoch": 0.9661442006269593, "grad_norm": 41.16212844848633, "learning_rate": 1.541e-06, "loss": 3.4601, "step": 3082 }, { "epoch": 0.9664576802507837, "grad_norm": 36.21253204345703, "learning_rate": 1.5415000000000002e-06, "loss": 3.063, "step": 3083 }, { "epoch": 0.9667711598746082, "grad_norm": 29.120792388916016, "learning_rate": 1.5420000000000002e-06, "loss": 3.5839, "step": 3084 }, { "epoch": 0.9670846394984326, "grad_norm": 32.0816535949707, "learning_rate": 1.5425e-06, "loss": 3.7305, "step": 3085 }, { "epoch": 0.967398119122257, "grad_norm": 30.316551208496094, "learning_rate": 1.543e-06, "loss": 3.2055, "step": 3086 }, { "epoch": 0.9677115987460815, "grad_norm": 33.83134841918945, "learning_rate": 1.5435e-06, "loss": 3.3693, "step": 3087 }, { "epoch": 0.9680250783699059, "grad_norm": 47.991493225097656, "learning_rate": 1.5440000000000002e-06, "loss": 4.0049, "step": 3088 }, { "epoch": 0.9683385579937304, "grad_norm": 24.988670349121094, "learning_rate": 1.5445000000000002e-06, "loss": 3.81, "step": 3089 }, { "epoch": 0.9686520376175548, "grad_norm": 39.11276626586914, "learning_rate": 1.545e-06, "loss": 3.7266, "step": 3090 }, { "epoch": 0.9689655172413794, "grad_norm": 69.33980560302734, "learning_rate": 1.5455e-06, "loss": 4.3037, "step": 3091 }, { "epoch": 0.9692789968652038, "grad_norm": 43.25937271118164, "learning_rate": 1.546e-06, "loss": 3.6816, "step": 3092 }, { "epoch": 0.9695924764890282, "grad_norm": 56.0028190612793, "learning_rate": 1.5465000000000001e-06, "loss": 3.5904, "step": 3093 }, { "epoch": 0.9699059561128527, "grad_norm": 62.90016555786133, "learning_rate": 1.5470000000000002e-06, "loss": 4.1878, "step": 3094 }, { "epoch": 0.9702194357366771, "grad_norm": 27.32510757446289, "learning_rate": 1.5475000000000002e-06, "loss": 2.9088, "step": 3095 }, { "epoch": 0.9705329153605016, "grad_norm": 29.192907333374023, "learning_rate": 1.548e-06, "loss": 3.6891, "step": 3096 }, { "epoch": 0.970846394984326, "grad_norm": 29.538455963134766, "learning_rate": 1.5485e-06, "loss": 3.5678, "step": 3097 }, { "epoch": 0.9711598746081505, "grad_norm": 39.13743209838867, "learning_rate": 1.5490000000000003e-06, "loss": 3.5836, "step": 3098 }, { "epoch": 0.9714733542319749, "grad_norm": 35.45713424682617, "learning_rate": 1.5495000000000002e-06, "loss": 3.5822, "step": 3099 }, { "epoch": 0.9717868338557993, "grad_norm": 37.34980392456055, "learning_rate": 1.5500000000000002e-06, "loss": 3.4961, "step": 3100 }, { "epoch": 0.9721003134796238, "grad_norm": 52.05093765258789, "learning_rate": 1.5505e-06, "loss": 4.2753, "step": 3101 }, { "epoch": 0.9724137931034482, "grad_norm": 43.123226165771484, "learning_rate": 1.551e-06, "loss": 4.7245, "step": 3102 }, { "epoch": 0.9727272727272728, "grad_norm": 30.87041473388672, "learning_rate": 1.5515000000000003e-06, "loss": 3.187, "step": 3103 }, { "epoch": 0.9730407523510972, "grad_norm": 61.1220588684082, "learning_rate": 1.5520000000000001e-06, "loss": 3.7726, "step": 3104 }, { "epoch": 0.9733542319749217, "grad_norm": 24.472267150878906, "learning_rate": 1.5525000000000002e-06, "loss": 3.2109, "step": 3105 }, { "epoch": 0.9736677115987461, "grad_norm": 41.3580436706543, "learning_rate": 1.553e-06, "loss": 4.827, "step": 3106 }, { "epoch": 0.9739811912225705, "grad_norm": 23.06800079345703, "learning_rate": 1.5535e-06, "loss": 3.6523, "step": 3107 }, { "epoch": 0.974294670846395, "grad_norm": 45.498695373535156, "learning_rate": 1.5540000000000003e-06, "loss": 4.1928, "step": 3108 }, { "epoch": 0.9746081504702194, "grad_norm": 32.11305618286133, "learning_rate": 1.5545000000000001e-06, "loss": 3.395, "step": 3109 }, { "epoch": 0.9749216300940439, "grad_norm": 22.9804744720459, "learning_rate": 1.5550000000000001e-06, "loss": 2.9892, "step": 3110 }, { "epoch": 0.9752351097178683, "grad_norm": 33.303001403808594, "learning_rate": 1.5555e-06, "loss": 3.2164, "step": 3111 }, { "epoch": 0.9755485893416928, "grad_norm": 35.61023712158203, "learning_rate": 1.556e-06, "loss": 3.2413, "step": 3112 }, { "epoch": 0.9758620689655172, "grad_norm": 32.667476654052734, "learning_rate": 1.5565000000000003e-06, "loss": 4.2483, "step": 3113 }, { "epoch": 0.9761755485893417, "grad_norm": 30.902109146118164, "learning_rate": 1.557e-06, "loss": 3.5198, "step": 3114 }, { "epoch": 0.9764890282131662, "grad_norm": 29.37360191345215, "learning_rate": 1.5575000000000001e-06, "loss": 3.6189, "step": 3115 }, { "epoch": 0.9768025078369906, "grad_norm": 41.751956939697266, "learning_rate": 1.5580000000000002e-06, "loss": 3.3907, "step": 3116 }, { "epoch": 0.9771159874608151, "grad_norm": 48.0351676940918, "learning_rate": 1.5585e-06, "loss": 3.2474, "step": 3117 }, { "epoch": 0.9774294670846395, "grad_norm": 33.41267395019531, "learning_rate": 1.5590000000000002e-06, "loss": 3.4043, "step": 3118 }, { "epoch": 0.977742946708464, "grad_norm": 26.803760528564453, "learning_rate": 1.5595000000000003e-06, "loss": 3.016, "step": 3119 }, { "epoch": 0.9780564263322884, "grad_norm": 59.20711135864258, "learning_rate": 1.56e-06, "loss": 3.8982, "step": 3120 }, { "epoch": 0.9783699059561128, "grad_norm": 25.695276260375977, "learning_rate": 1.5605000000000001e-06, "loss": 2.929, "step": 3121 }, { "epoch": 0.9786833855799373, "grad_norm": 41.91707229614258, "learning_rate": 1.561e-06, "loss": 3.3785, "step": 3122 }, { "epoch": 0.9789968652037617, "grad_norm": 26.014942169189453, "learning_rate": 1.5615000000000002e-06, "loss": 4.3426, "step": 3123 }, { "epoch": 0.9793103448275862, "grad_norm": 43.355995178222656, "learning_rate": 1.5620000000000002e-06, "loss": 3.191, "step": 3124 }, { "epoch": 0.9796238244514106, "grad_norm": 52.087005615234375, "learning_rate": 1.5625e-06, "loss": 3.2585, "step": 3125 }, { "epoch": 0.9799373040752352, "grad_norm": 61.68992233276367, "learning_rate": 1.5630000000000001e-06, "loss": 3.7206, "step": 3126 }, { "epoch": 0.9802507836990596, "grad_norm": 34.2900390625, "learning_rate": 1.5635e-06, "loss": 3.2416, "step": 3127 }, { "epoch": 0.980564263322884, "grad_norm": 51.79576873779297, "learning_rate": 1.5640000000000002e-06, "loss": 5.8384, "step": 3128 }, { "epoch": 0.9808777429467085, "grad_norm": 26.661314010620117, "learning_rate": 1.5645000000000002e-06, "loss": 3.4689, "step": 3129 }, { "epoch": 0.9811912225705329, "grad_norm": 35.80608367919922, "learning_rate": 1.565e-06, "loss": 3.2023, "step": 3130 }, { "epoch": 0.9815047021943574, "grad_norm": 33.74555587768555, "learning_rate": 1.5655e-06, "loss": 3.3833, "step": 3131 }, { "epoch": 0.9818181818181818, "grad_norm": 40.17083740234375, "learning_rate": 1.566e-06, "loss": 3.5556, "step": 3132 }, { "epoch": 0.9821316614420063, "grad_norm": 39.58924865722656, "learning_rate": 1.5665000000000002e-06, "loss": 3.0832, "step": 3133 }, { "epoch": 0.9824451410658307, "grad_norm": 45.79573440551758, "learning_rate": 1.5670000000000002e-06, "loss": 3.5993, "step": 3134 }, { "epoch": 0.9827586206896551, "grad_norm": 30.124801635742188, "learning_rate": 1.5675e-06, "loss": 3.0965, "step": 3135 }, { "epoch": 0.9830721003134796, "grad_norm": 50.938419342041016, "learning_rate": 1.568e-06, "loss": 3.1747, "step": 3136 }, { "epoch": 0.983385579937304, "grad_norm": 37.803470611572266, "learning_rate": 1.5685e-06, "loss": 3.724, "step": 3137 }, { "epoch": 0.9836990595611286, "grad_norm": 45.604373931884766, "learning_rate": 1.5690000000000001e-06, "loss": 2.8654, "step": 3138 }, { "epoch": 0.984012539184953, "grad_norm": 29.251733779907227, "learning_rate": 1.5695000000000002e-06, "loss": 3.0197, "step": 3139 }, { "epoch": 0.9843260188087775, "grad_norm": 40.991294860839844, "learning_rate": 1.5700000000000002e-06, "loss": 3.0759, "step": 3140 }, { "epoch": 0.9846394984326019, "grad_norm": 39.70516586303711, "learning_rate": 1.5705e-06, "loss": 3.2381, "step": 3141 }, { "epoch": 0.9849529780564263, "grad_norm": 104.91124725341797, "learning_rate": 1.571e-06, "loss": 3.9789, "step": 3142 }, { "epoch": 0.9852664576802508, "grad_norm": 69.99864959716797, "learning_rate": 1.5715000000000003e-06, "loss": 5.4289, "step": 3143 }, { "epoch": 0.9855799373040752, "grad_norm": 47.31104278564453, "learning_rate": 1.5720000000000002e-06, "loss": 3.5801, "step": 3144 }, { "epoch": 0.9858934169278997, "grad_norm": 23.58669090270996, "learning_rate": 1.5725000000000002e-06, "loss": 3.2141, "step": 3145 }, { "epoch": 0.9862068965517241, "grad_norm": 60.75740432739258, "learning_rate": 1.573e-06, "loss": 3.1711, "step": 3146 }, { "epoch": 0.9865203761755486, "grad_norm": 35.55626678466797, "learning_rate": 1.5735e-06, "loss": 3.0043, "step": 3147 }, { "epoch": 0.986833855799373, "grad_norm": 56.99102783203125, "learning_rate": 1.5740000000000003e-06, "loss": 4.1672, "step": 3148 }, { "epoch": 0.9871473354231975, "grad_norm": 25.71376609802246, "learning_rate": 1.5745000000000001e-06, "loss": 2.92, "step": 3149 }, { "epoch": 0.987460815047022, "grad_norm": 31.668563842773438, "learning_rate": 1.5750000000000002e-06, "loss": 3.2671, "step": 3150 }, { "epoch": 0.9877742946708464, "grad_norm": 32.40145492553711, "learning_rate": 1.5755e-06, "loss": 3.1882, "step": 3151 }, { "epoch": 0.9880877742946709, "grad_norm": 33.56290054321289, "learning_rate": 1.576e-06, "loss": 2.97, "step": 3152 }, { "epoch": 0.9884012539184953, "grad_norm": 23.0459041595459, "learning_rate": 1.5765000000000003e-06, "loss": 2.8921, "step": 3153 }, { "epoch": 0.9887147335423198, "grad_norm": 31.948535919189453, "learning_rate": 1.577e-06, "loss": 3.348, "step": 3154 }, { "epoch": 0.9890282131661442, "grad_norm": 39.63069534301758, "learning_rate": 1.5775000000000001e-06, "loss": 2.9237, "step": 3155 }, { "epoch": 0.9893416927899686, "grad_norm": 32.34980392456055, "learning_rate": 1.5780000000000002e-06, "loss": 3.245, "step": 3156 }, { "epoch": 0.9896551724137931, "grad_norm": 41.89958953857422, "learning_rate": 1.5785e-06, "loss": 3.4284, "step": 3157 }, { "epoch": 0.9899686520376175, "grad_norm": 51.7044792175293, "learning_rate": 1.5790000000000003e-06, "loss": 3.0146, "step": 3158 }, { "epoch": 0.990282131661442, "grad_norm": 110.62555694580078, "learning_rate": 1.5795000000000003e-06, "loss": 3.7889, "step": 3159 }, { "epoch": 0.9905956112852664, "grad_norm": 42.60169982910156, "learning_rate": 1.5800000000000001e-06, "loss": 4.1996, "step": 3160 }, { "epoch": 0.990909090909091, "grad_norm": 45.62926483154297, "learning_rate": 1.5805000000000002e-06, "loss": 2.8792, "step": 3161 }, { "epoch": 0.9912225705329154, "grad_norm": 29.770870208740234, "learning_rate": 1.581e-06, "loss": 3.162, "step": 3162 }, { "epoch": 0.9915360501567398, "grad_norm": 40.25620651245117, "learning_rate": 1.5815000000000002e-06, "loss": 3.2687, "step": 3163 }, { "epoch": 0.9918495297805643, "grad_norm": 51.3452262878418, "learning_rate": 1.5820000000000003e-06, "loss": 3.6536, "step": 3164 }, { "epoch": 0.9921630094043887, "grad_norm": 42.561981201171875, "learning_rate": 1.5825e-06, "loss": 3.2372, "step": 3165 }, { "epoch": 0.9924764890282132, "grad_norm": 104.1058578491211, "learning_rate": 1.5830000000000001e-06, "loss": 3.6081, "step": 3166 }, { "epoch": 0.9927899686520376, "grad_norm": 33.43682098388672, "learning_rate": 1.5835e-06, "loss": 3.0076, "step": 3167 }, { "epoch": 0.993103448275862, "grad_norm": 31.840662002563477, "learning_rate": 1.5840000000000002e-06, "loss": 3.4222, "step": 3168 }, { "epoch": 0.9934169278996865, "grad_norm": 39.76411437988281, "learning_rate": 1.5845000000000002e-06, "loss": 3.4354, "step": 3169 }, { "epoch": 0.9937304075235109, "grad_norm": 45.577274322509766, "learning_rate": 1.585e-06, "loss": 4.2827, "step": 3170 }, { "epoch": 0.9940438871473354, "grad_norm": 40.2696647644043, "learning_rate": 1.5855000000000001e-06, "loss": 3.1225, "step": 3171 }, { "epoch": 0.9943573667711598, "grad_norm": 37.35997009277344, "learning_rate": 1.586e-06, "loss": 3.626, "step": 3172 }, { "epoch": 0.9946708463949844, "grad_norm": 29.363245010375977, "learning_rate": 1.5865000000000002e-06, "loss": 3.437, "step": 3173 }, { "epoch": 0.9949843260188088, "grad_norm": 30.479536056518555, "learning_rate": 1.5870000000000002e-06, "loss": 3.4304, "step": 3174 }, { "epoch": 0.9952978056426333, "grad_norm": 27.580543518066406, "learning_rate": 1.5875e-06, "loss": 3.2106, "step": 3175 }, { "epoch": 0.9956112852664577, "grad_norm": 39.4871711730957, "learning_rate": 1.588e-06, "loss": 3.465, "step": 3176 }, { "epoch": 0.9959247648902821, "grad_norm": 20.305816650390625, "learning_rate": 1.5885000000000001e-06, "loss": 2.8586, "step": 3177 }, { "epoch": 0.9962382445141066, "grad_norm": 54.459625244140625, "learning_rate": 1.5890000000000002e-06, "loss": 3.8607, "step": 3178 }, { "epoch": 0.996551724137931, "grad_norm": 45.21859359741211, "learning_rate": 1.5895000000000002e-06, "loss": 4.0433, "step": 3179 }, { "epoch": 0.9968652037617555, "grad_norm": 30.275541305541992, "learning_rate": 1.5900000000000002e-06, "loss": 3.1236, "step": 3180 }, { "epoch": 0.9971786833855799, "grad_norm": 24.845712661743164, "learning_rate": 1.5905e-06, "loss": 3.052, "step": 3181 }, { "epoch": 0.9974921630094044, "grad_norm": 47.068382263183594, "learning_rate": 1.591e-06, "loss": 4.1035, "step": 3182 }, { "epoch": 0.9978056426332288, "grad_norm": 86.9577865600586, "learning_rate": 1.5915000000000003e-06, "loss": 3.6548, "step": 3183 }, { "epoch": 0.9981191222570532, "grad_norm": 45.81187057495117, "learning_rate": 1.5920000000000002e-06, "loss": 4.0105, "step": 3184 }, { "epoch": 0.9984326018808778, "grad_norm": 54.567386627197266, "learning_rate": 1.5925000000000002e-06, "loss": 3.7583, "step": 3185 }, { "epoch": 0.9987460815047022, "grad_norm": 45.75572967529297, "learning_rate": 1.593e-06, "loss": 3.3029, "step": 3186 }, { "epoch": 0.9990595611285267, "grad_norm": 56.43427658081055, "learning_rate": 1.5935e-06, "loss": 2.8933, "step": 3187 }, { "epoch": 0.9993730407523511, "grad_norm": 41.16676712036133, "learning_rate": 1.594e-06, "loss": 3.0468, "step": 3188 }, { "epoch": 0.9996865203761756, "grad_norm": 38.499820709228516, "learning_rate": 1.5945000000000001e-06, "loss": 3.9996, "step": 3189 }, { "epoch": 1.0, "grad_norm": 20.5953426361084, "learning_rate": 1.5950000000000002e-06, "loss": 2.7745, "step": 3190 }, { "epoch": 1.0003134796238244, "grad_norm": 34.326446533203125, "learning_rate": 1.5955e-06, "loss": 3.4488, "step": 3191 }, { "epoch": 1.0006269592476489, "grad_norm": 25.23051643371582, "learning_rate": 1.596e-06, "loss": 2.7374, "step": 3192 }, { "epoch": 1.0006269592476489, "eval_loss": 5.7285990715026855, "eval_runtime": 21.0416, "eval_samples_per_second": 127.7, "eval_steps_per_second": 7.984, "step": 3192 }, { "epoch": 1.0009404388714733, "grad_norm": 27.973400115966797, "learning_rate": 1.5965e-06, "loss": 3.0561, "step": 3193 }, { "epoch": 1.0012539184952978, "grad_norm": 28.687578201293945, "learning_rate": 1.5970000000000001e-06, "loss": 3.1456, "step": 3194 }, { "epoch": 1.0015673981191222, "grad_norm": 54.51884078979492, "learning_rate": 1.5975000000000002e-06, "loss": 4.3574, "step": 3195 }, { "epoch": 1.0018808777429467, "grad_norm": 22.76805877685547, "learning_rate": 1.5980000000000002e-06, "loss": 3.5964, "step": 3196 }, { "epoch": 1.002194357366771, "grad_norm": 42.05739974975586, "learning_rate": 1.5985e-06, "loss": 2.9885, "step": 3197 }, { "epoch": 1.0025078369905955, "grad_norm": 55.0859260559082, "learning_rate": 1.599e-06, "loss": 3.6646, "step": 3198 }, { "epoch": 1.00282131661442, "grad_norm": 48.74159240722656, "learning_rate": 1.5995000000000003e-06, "loss": 3.5255, "step": 3199 }, { "epoch": 1.0031347962382444, "grad_norm": 28.012622833251953, "learning_rate": 1.6000000000000001e-06, "loss": 3.4958, "step": 3200 }, { "epoch": 1.0034482758620689, "grad_norm": 38.07377624511719, "learning_rate": 1.6005000000000002e-06, "loss": 3.3704, "step": 3201 }, { "epoch": 1.0037617554858935, "grad_norm": 28.49401092529297, "learning_rate": 1.601e-06, "loss": 4.2952, "step": 3202 }, { "epoch": 1.004075235109718, "grad_norm": 94.79960632324219, "learning_rate": 1.6015e-06, "loss": 5.6727, "step": 3203 }, { "epoch": 1.0043887147335424, "grad_norm": 21.47426414489746, "learning_rate": 1.6020000000000003e-06, "loss": 3.5125, "step": 3204 }, { "epoch": 1.0047021943573669, "grad_norm": 36.450836181640625, "learning_rate": 1.6025000000000001e-06, "loss": 2.8938, "step": 3205 }, { "epoch": 1.0050156739811913, "grad_norm": 38.39944076538086, "learning_rate": 1.6030000000000002e-06, "loss": 3.4479, "step": 3206 }, { "epoch": 1.0053291536050157, "grad_norm": 26.999618530273438, "learning_rate": 1.6035e-06, "loss": 3.1897, "step": 3207 }, { "epoch": 1.0056426332288402, "grad_norm": 38.76042938232422, "learning_rate": 1.604e-06, "loss": 3.3169, "step": 3208 }, { "epoch": 1.0059561128526646, "grad_norm": 34.40362548828125, "learning_rate": 1.6045000000000003e-06, "loss": 3.5853, "step": 3209 }, { "epoch": 1.006269592476489, "grad_norm": 29.437618255615234, "learning_rate": 1.605e-06, "loss": 2.737, "step": 3210 }, { "epoch": 1.0065830721003135, "grad_norm": 31.488759994506836, "learning_rate": 1.6055000000000001e-06, "loss": 4.0049, "step": 3211 }, { "epoch": 1.006896551724138, "grad_norm": 32.4052619934082, "learning_rate": 1.606e-06, "loss": 3.9704, "step": 3212 }, { "epoch": 1.0072100313479624, "grad_norm": 41.16117858886719, "learning_rate": 1.6065e-06, "loss": 3.4824, "step": 3213 }, { "epoch": 1.0075235109717868, "grad_norm": 39.67842102050781, "learning_rate": 1.6070000000000002e-06, "loss": 3.2149, "step": 3214 }, { "epoch": 1.0078369905956113, "grad_norm": 34.92125701904297, "learning_rate": 1.6075e-06, "loss": 3.4013, "step": 3215 }, { "epoch": 1.0081504702194357, "grad_norm": 35.19260787963867, "learning_rate": 1.608e-06, "loss": 3.0669, "step": 3216 }, { "epoch": 1.0084639498432602, "grad_norm": 30.965526580810547, "learning_rate": 1.6085000000000001e-06, "loss": 3.5108, "step": 3217 }, { "epoch": 1.0087774294670846, "grad_norm": 55.61679458618164, "learning_rate": 1.609e-06, "loss": 3.2511, "step": 3218 }, { "epoch": 1.009090909090909, "grad_norm": 52.900753021240234, "learning_rate": 1.6095000000000002e-06, "loss": 4.8786, "step": 3219 }, { "epoch": 1.0094043887147335, "grad_norm": 31.876253128051758, "learning_rate": 1.6100000000000003e-06, "loss": 3.5526, "step": 3220 }, { "epoch": 1.009717868338558, "grad_norm": 70.21955871582031, "learning_rate": 1.6105e-06, "loss": 3.2994, "step": 3221 }, { "epoch": 1.0100313479623824, "grad_norm": 57.42060470581055, "learning_rate": 1.6110000000000001e-06, "loss": 3.6681, "step": 3222 }, { "epoch": 1.0103448275862068, "grad_norm": 20.363557815551758, "learning_rate": 1.6115e-06, "loss": 2.969, "step": 3223 }, { "epoch": 1.0106583072100312, "grad_norm": 46.56084442138672, "learning_rate": 1.6120000000000002e-06, "loss": 3.2407, "step": 3224 }, { "epoch": 1.0109717868338557, "grad_norm": 26.832813262939453, "learning_rate": 1.6125000000000002e-06, "loss": 2.859, "step": 3225 }, { "epoch": 1.0112852664576804, "grad_norm": 39.61360549926758, "learning_rate": 1.613e-06, "loss": 2.8402, "step": 3226 }, { "epoch": 1.0115987460815048, "grad_norm": 59.345333099365234, "learning_rate": 1.6135e-06, "loss": 3.2726, "step": 3227 }, { "epoch": 1.0119122257053292, "grad_norm": 34.26584243774414, "learning_rate": 1.614e-06, "loss": 3.2933, "step": 3228 }, { "epoch": 1.0122257053291537, "grad_norm": 31.62544822692871, "learning_rate": 1.6145000000000002e-06, "loss": 2.8915, "step": 3229 }, { "epoch": 1.0125391849529781, "grad_norm": 24.0006160736084, "learning_rate": 1.6150000000000002e-06, "loss": 2.4861, "step": 3230 }, { "epoch": 1.0128526645768026, "grad_norm": 23.52138328552246, "learning_rate": 1.6155e-06, "loss": 3.1468, "step": 3231 }, { "epoch": 1.013166144200627, "grad_norm": 48.56825256347656, "learning_rate": 1.616e-06, "loss": 3.2097, "step": 3232 }, { "epoch": 1.0134796238244514, "grad_norm": 40.454193115234375, "learning_rate": 1.6165000000000001e-06, "loss": 3.646, "step": 3233 }, { "epoch": 1.013793103448276, "grad_norm": 27.93584632873535, "learning_rate": 1.6170000000000001e-06, "loss": 3.0358, "step": 3234 }, { "epoch": 1.0141065830721003, "grad_norm": 37.334739685058594, "learning_rate": 1.6175000000000002e-06, "loss": 3.0387, "step": 3235 }, { "epoch": 1.0144200626959248, "grad_norm": 29.4758358001709, "learning_rate": 1.6180000000000002e-06, "loss": 3.141, "step": 3236 }, { "epoch": 1.0147335423197492, "grad_norm": 19.978744506835938, "learning_rate": 1.6185e-06, "loss": 2.5931, "step": 3237 }, { "epoch": 1.0150470219435737, "grad_norm": 57.191749572753906, "learning_rate": 1.619e-06, "loss": 3.0585, "step": 3238 }, { "epoch": 1.015360501567398, "grad_norm": 36.548648834228516, "learning_rate": 1.6195000000000003e-06, "loss": 3.1298, "step": 3239 }, { "epoch": 1.0156739811912225, "grad_norm": 32.01398849487305, "learning_rate": 1.6200000000000002e-06, "loss": 3.2455, "step": 3240 }, { "epoch": 1.015987460815047, "grad_norm": 34.381431579589844, "learning_rate": 1.6205000000000002e-06, "loss": 3.4494, "step": 3241 }, { "epoch": 1.0163009404388714, "grad_norm": 59.548458099365234, "learning_rate": 1.621e-06, "loss": 4.1441, "step": 3242 }, { "epoch": 1.0166144200626959, "grad_norm": 59.294395446777344, "learning_rate": 1.6215e-06, "loss": 4.3494, "step": 3243 }, { "epoch": 1.0169278996865203, "grad_norm": 31.622285842895508, "learning_rate": 1.6220000000000003e-06, "loss": 3.1537, "step": 3244 }, { "epoch": 1.0172413793103448, "grad_norm": 33.20928955078125, "learning_rate": 1.6225000000000001e-06, "loss": 2.9726, "step": 3245 }, { "epoch": 1.0175548589341692, "grad_norm": 75.9956283569336, "learning_rate": 1.6230000000000002e-06, "loss": 3.3001, "step": 3246 }, { "epoch": 1.0178683385579936, "grad_norm": 44.67719650268555, "learning_rate": 1.6235e-06, "loss": 3.0272, "step": 3247 }, { "epoch": 1.018181818181818, "grad_norm": 27.121854782104492, "learning_rate": 1.624e-06, "loss": 3.3043, "step": 3248 }, { "epoch": 1.0184952978056427, "grad_norm": 34.486473083496094, "learning_rate": 1.6245000000000003e-06, "loss": 2.8611, "step": 3249 }, { "epoch": 1.0188087774294672, "grad_norm": 70.01057434082031, "learning_rate": 1.6250000000000001e-06, "loss": 3.9254, "step": 3250 }, { "epoch": 1.0191222570532916, "grad_norm": 35.10891342163086, "learning_rate": 1.6255000000000001e-06, "loss": 3.2911, "step": 3251 }, { "epoch": 1.019435736677116, "grad_norm": 45.17849349975586, "learning_rate": 1.626e-06, "loss": 3.5813, "step": 3252 }, { "epoch": 1.0197492163009405, "grad_norm": 36.457950592041016, "learning_rate": 1.6265e-06, "loss": 3.16, "step": 3253 }, { "epoch": 1.020062695924765, "grad_norm": 49.345062255859375, "learning_rate": 1.6270000000000003e-06, "loss": 4.0663, "step": 3254 }, { "epoch": 1.0203761755485894, "grad_norm": 22.22081756591797, "learning_rate": 1.6275e-06, "loss": 2.8179, "step": 3255 }, { "epoch": 1.0206896551724138, "grad_norm": 44.726043701171875, "learning_rate": 1.6280000000000001e-06, "loss": 3.1774, "step": 3256 }, { "epoch": 1.0210031347962383, "grad_norm": 63.41642379760742, "learning_rate": 1.6285000000000002e-06, "loss": 3.1283, "step": 3257 }, { "epoch": 1.0213166144200627, "grad_norm": 25.33333396911621, "learning_rate": 1.629e-06, "loss": 2.8745, "step": 3258 }, { "epoch": 1.0216300940438872, "grad_norm": 26.45008659362793, "learning_rate": 1.6295000000000002e-06, "loss": 3.3849, "step": 3259 }, { "epoch": 1.0219435736677116, "grad_norm": 29.69173812866211, "learning_rate": 1.6300000000000003e-06, "loss": 2.9018, "step": 3260 }, { "epoch": 1.022257053291536, "grad_norm": 29.18816375732422, "learning_rate": 1.6305e-06, "loss": 3.2486, "step": 3261 }, { "epoch": 1.0225705329153605, "grad_norm": 29.251667022705078, "learning_rate": 1.6310000000000001e-06, "loss": 2.9324, "step": 3262 }, { "epoch": 1.022884012539185, "grad_norm": 28.868215560913086, "learning_rate": 1.6315e-06, "loss": 2.872, "step": 3263 }, { "epoch": 1.0231974921630094, "grad_norm": 31.325496673583984, "learning_rate": 1.6320000000000002e-06, "loss": 3.2128, "step": 3264 }, { "epoch": 1.0235109717868338, "grad_norm": 42.07441711425781, "learning_rate": 1.6325000000000003e-06, "loss": 2.9007, "step": 3265 }, { "epoch": 1.0238244514106583, "grad_norm": 68.76583099365234, "learning_rate": 1.633e-06, "loss": 4.2267, "step": 3266 }, { "epoch": 1.0241379310344827, "grad_norm": 48.943477630615234, "learning_rate": 1.6335000000000001e-06, "loss": 3.0372, "step": 3267 }, { "epoch": 1.0244514106583071, "grad_norm": 40.075984954833984, "learning_rate": 1.634e-06, "loss": 3.9948, "step": 3268 }, { "epoch": 1.0247648902821316, "grad_norm": 21.297189712524414, "learning_rate": 1.6345000000000002e-06, "loss": 2.7858, "step": 3269 }, { "epoch": 1.025078369905956, "grad_norm": 40.05287551879883, "learning_rate": 1.6350000000000002e-06, "loss": 3.8012, "step": 3270 }, { "epoch": 1.0253918495297805, "grad_norm": 68.54923248291016, "learning_rate": 1.6355e-06, "loss": 3.6058, "step": 3271 }, { "epoch": 1.0257053291536051, "grad_norm": 37.00851058959961, "learning_rate": 1.636e-06, "loss": 3.347, "step": 3272 }, { "epoch": 1.0260188087774296, "grad_norm": 50.256134033203125, "learning_rate": 1.6365000000000001e-06, "loss": 3.2308, "step": 3273 }, { "epoch": 1.026332288401254, "grad_norm": 19.08588218688965, "learning_rate": 1.6370000000000002e-06, "loss": 2.9165, "step": 3274 }, { "epoch": 1.0266457680250785, "grad_norm": 26.314943313598633, "learning_rate": 1.6375000000000002e-06, "loss": 3.1059, "step": 3275 }, { "epoch": 1.026959247648903, "grad_norm": 74.63359069824219, "learning_rate": 1.6380000000000002e-06, "loss": 4.7147, "step": 3276 }, { "epoch": 1.0272727272727273, "grad_norm": 38.578792572021484, "learning_rate": 1.6385e-06, "loss": 3.2019, "step": 3277 }, { "epoch": 1.0275862068965518, "grad_norm": 40.135738372802734, "learning_rate": 1.639e-06, "loss": 2.7686, "step": 3278 }, { "epoch": 1.0278996865203762, "grad_norm": 23.660158157348633, "learning_rate": 1.6395000000000004e-06, "loss": 2.9382, "step": 3279 }, { "epoch": 1.0282131661442007, "grad_norm": 39.85804748535156, "learning_rate": 1.6400000000000002e-06, "loss": 2.8816, "step": 3280 }, { "epoch": 1.028526645768025, "grad_norm": 46.418426513671875, "learning_rate": 1.6405000000000002e-06, "loss": 3.0901, "step": 3281 }, { "epoch": 1.0288401253918495, "grad_norm": 190.2622528076172, "learning_rate": 1.641e-06, "loss": 2.9474, "step": 3282 }, { "epoch": 1.029153605015674, "grad_norm": 43.696956634521484, "learning_rate": 1.6415e-06, "loss": 3.5476, "step": 3283 }, { "epoch": 1.0294670846394984, "grad_norm": 22.224021911621094, "learning_rate": 1.6420000000000003e-06, "loss": 3.1479, "step": 3284 }, { "epoch": 1.0297805642633229, "grad_norm": 29.27781867980957, "learning_rate": 1.6425000000000002e-06, "loss": 2.6123, "step": 3285 }, { "epoch": 1.0300940438871473, "grad_norm": 24.40782356262207, "learning_rate": 1.6430000000000002e-06, "loss": 2.6373, "step": 3286 }, { "epoch": 1.0304075235109718, "grad_norm": 27.865283966064453, "learning_rate": 1.6435e-06, "loss": 3.657, "step": 3287 }, { "epoch": 1.0307210031347962, "grad_norm": 33.429283142089844, "learning_rate": 1.644e-06, "loss": 3.7763, "step": 3288 }, { "epoch": 1.0310344827586206, "grad_norm": 20.046314239501953, "learning_rate": 1.6445000000000003e-06, "loss": 3.4536, "step": 3289 }, { "epoch": 1.031347962382445, "grad_norm": 26.210359573364258, "learning_rate": 1.6450000000000001e-06, "loss": 2.8382, "step": 3290 }, { "epoch": 1.0316614420062695, "grad_norm": 27.057741165161133, "learning_rate": 1.6455000000000002e-06, "loss": 2.6236, "step": 3291 }, { "epoch": 1.031974921630094, "grad_norm": 60.8143310546875, "learning_rate": 1.646e-06, "loss": 4.561, "step": 3292 }, { "epoch": 1.0322884012539184, "grad_norm": 22.38360023498535, "learning_rate": 1.6465e-06, "loss": 2.4903, "step": 3293 }, { "epoch": 1.0326018808777429, "grad_norm": 32.65427780151367, "learning_rate": 1.6470000000000003e-06, "loss": 3.5099, "step": 3294 }, { "epoch": 1.0329153605015673, "grad_norm": 32.22104263305664, "learning_rate": 1.6475000000000001e-06, "loss": 2.9126, "step": 3295 }, { "epoch": 1.033228840125392, "grad_norm": 18.776548385620117, "learning_rate": 1.6480000000000001e-06, "loss": 2.7612, "step": 3296 }, { "epoch": 1.0335423197492164, "grad_norm": 25.23478126525879, "learning_rate": 1.6485000000000002e-06, "loss": 2.8399, "step": 3297 }, { "epoch": 1.0338557993730408, "grad_norm": 60.96337890625, "learning_rate": 1.649e-06, "loss": 2.9111, "step": 3298 }, { "epoch": 1.0341692789968653, "grad_norm": 26.42510414123535, "learning_rate": 1.6495000000000003e-06, "loss": 3.0101, "step": 3299 }, { "epoch": 1.0344827586206897, "grad_norm": 31.34357261657715, "learning_rate": 1.6500000000000003e-06, "loss": 2.8724, "step": 3300 }, { "epoch": 1.0347962382445142, "grad_norm": 64.86882781982422, "learning_rate": 1.6505000000000001e-06, "loss": 3.3274, "step": 3301 }, { "epoch": 1.0351097178683386, "grad_norm": 21.959516525268555, "learning_rate": 1.6510000000000002e-06, "loss": 3.0008, "step": 3302 }, { "epoch": 1.035423197492163, "grad_norm": 32.650943756103516, "learning_rate": 1.6515e-06, "loss": 3.1787, "step": 3303 }, { "epoch": 1.0357366771159875, "grad_norm": 36.1429328918457, "learning_rate": 1.6520000000000002e-06, "loss": 3.0022, "step": 3304 }, { "epoch": 1.036050156739812, "grad_norm": 48.24531936645508, "learning_rate": 1.6525000000000003e-06, "loss": 2.9519, "step": 3305 }, { "epoch": 1.0363636363636364, "grad_norm": 34.19054412841797, "learning_rate": 1.653e-06, "loss": 3.331, "step": 3306 }, { "epoch": 1.0366771159874608, "grad_norm": 18.031879425048828, "learning_rate": 1.6535000000000001e-06, "loss": 3.028, "step": 3307 }, { "epoch": 1.0369905956112853, "grad_norm": 29.253751754760742, "learning_rate": 1.654e-06, "loss": 2.7076, "step": 3308 }, { "epoch": 1.0373040752351097, "grad_norm": 43.910423278808594, "learning_rate": 1.6545000000000002e-06, "loss": 3.3512, "step": 3309 }, { "epoch": 1.0376175548589341, "grad_norm": 25.297222137451172, "learning_rate": 1.6550000000000002e-06, "loss": 3.3843, "step": 3310 }, { "epoch": 1.0379310344827586, "grad_norm": 27.861984252929688, "learning_rate": 1.6555e-06, "loss": 2.8931, "step": 3311 }, { "epoch": 1.038244514106583, "grad_norm": 29.09917640686035, "learning_rate": 1.6560000000000001e-06, "loss": 2.8308, "step": 3312 }, { "epoch": 1.0385579937304075, "grad_norm": 23.79581642150879, "learning_rate": 1.6565000000000001e-06, "loss": 2.7549, "step": 3313 }, { "epoch": 1.038871473354232, "grad_norm": 23.295839309692383, "learning_rate": 1.657e-06, "loss": 2.7089, "step": 3314 }, { "epoch": 1.0391849529780564, "grad_norm": 20.08068084716797, "learning_rate": 1.6575000000000002e-06, "loss": 2.7575, "step": 3315 }, { "epoch": 1.0394984326018808, "grad_norm": 28.336292266845703, "learning_rate": 1.6580000000000003e-06, "loss": 2.9939, "step": 3316 }, { "epoch": 1.0398119122257052, "grad_norm": 26.962081909179688, "learning_rate": 1.6585e-06, "loss": 2.9673, "step": 3317 }, { "epoch": 1.0401253918495297, "grad_norm": 71.06791687011719, "learning_rate": 1.6590000000000001e-06, "loss": 5.612, "step": 3318 }, { "epoch": 1.0404388714733543, "grad_norm": 22.833223342895508, "learning_rate": 1.6595e-06, "loss": 2.9429, "step": 3319 }, { "epoch": 1.0407523510971788, "grad_norm": 73.84259033203125, "learning_rate": 1.6600000000000002e-06, "loss": 4.2648, "step": 3320 }, { "epoch": 1.0410658307210032, "grad_norm": 38.24066925048828, "learning_rate": 1.6605000000000002e-06, "loss": 3.4204, "step": 3321 }, { "epoch": 1.0413793103448277, "grad_norm": 85.08003997802734, "learning_rate": 1.661e-06, "loss": 3.7887, "step": 3322 }, { "epoch": 1.041692789968652, "grad_norm": 36.931583404541016, "learning_rate": 1.6615e-06, "loss": 2.9921, "step": 3323 }, { "epoch": 1.0420062695924766, "grad_norm": 64.08250427246094, "learning_rate": 1.662e-06, "loss": 3.7738, "step": 3324 }, { "epoch": 1.042319749216301, "grad_norm": 32.0848388671875, "learning_rate": 1.6625000000000002e-06, "loss": 3.3746, "step": 3325 }, { "epoch": 1.0426332288401254, "grad_norm": 19.127561569213867, "learning_rate": 1.6630000000000002e-06, "loss": 2.7594, "step": 3326 }, { "epoch": 1.0429467084639499, "grad_norm": 97.25066375732422, "learning_rate": 1.6635e-06, "loss": 3.3868, "step": 3327 }, { "epoch": 1.0432601880877743, "grad_norm": 63.51421356201172, "learning_rate": 1.664e-06, "loss": 3.442, "step": 3328 }, { "epoch": 1.0435736677115988, "grad_norm": 38.333282470703125, "learning_rate": 1.6645e-06, "loss": 3.4402, "step": 3329 }, { "epoch": 1.0438871473354232, "grad_norm": 39.50896072387695, "learning_rate": 1.6650000000000002e-06, "loss": 3.0081, "step": 3330 }, { "epoch": 1.0442006269592476, "grad_norm": 43.86787033081055, "learning_rate": 1.6655000000000002e-06, "loss": 2.6378, "step": 3331 }, { "epoch": 1.044514106583072, "grad_norm": 25.9218692779541, "learning_rate": 1.666e-06, "loss": 2.7456, "step": 3332 }, { "epoch": 1.0448275862068965, "grad_norm": 41.19293212890625, "learning_rate": 1.6665e-06, "loss": 3.3948, "step": 3333 }, { "epoch": 1.045141065830721, "grad_norm": 27.02749252319336, "learning_rate": 1.667e-06, "loss": 3.3631, "step": 3334 }, { "epoch": 1.0454545454545454, "grad_norm": 48.45111083984375, "learning_rate": 1.6675000000000001e-06, "loss": 3.0751, "step": 3335 }, { "epoch": 1.0457680250783699, "grad_norm": 27.78561782836914, "learning_rate": 1.6680000000000002e-06, "loss": 3.616, "step": 3336 }, { "epoch": 1.0460815047021943, "grad_norm": 37.05852127075195, "learning_rate": 1.6685000000000002e-06, "loss": 2.9797, "step": 3337 }, { "epoch": 1.0463949843260187, "grad_norm": 33.00432205200195, "learning_rate": 1.669e-06, "loss": 3.8552, "step": 3338 }, { "epoch": 1.0467084639498432, "grad_norm": 55.03670120239258, "learning_rate": 1.6695e-06, "loss": 3.2324, "step": 3339 }, { "epoch": 1.0470219435736676, "grad_norm": 40.15440368652344, "learning_rate": 1.6700000000000003e-06, "loss": 3.3079, "step": 3340 }, { "epoch": 1.047335423197492, "grad_norm": 57.2756233215332, "learning_rate": 1.6705000000000001e-06, "loss": 3.9613, "step": 3341 }, { "epoch": 1.0476489028213165, "grad_norm": 44.24393844604492, "learning_rate": 1.6710000000000002e-06, "loss": 3.7156, "step": 3342 }, { "epoch": 1.0479623824451412, "grad_norm": 72.37336730957031, "learning_rate": 1.6715e-06, "loss": 2.5366, "step": 3343 }, { "epoch": 1.0482758620689656, "grad_norm": 44.30312728881836, "learning_rate": 1.672e-06, "loss": 3.3909, "step": 3344 }, { "epoch": 1.04858934169279, "grad_norm": 34.76454162597656, "learning_rate": 1.6725000000000003e-06, "loss": 3.5469, "step": 3345 }, { "epoch": 1.0489028213166145, "grad_norm": 50.73592758178711, "learning_rate": 1.6730000000000001e-06, "loss": 2.8253, "step": 3346 }, { "epoch": 1.049216300940439, "grad_norm": 33.41584396362305, "learning_rate": 1.6735000000000002e-06, "loss": 3.7759, "step": 3347 }, { "epoch": 1.0495297805642634, "grad_norm": 92.3950424194336, "learning_rate": 1.674e-06, "loss": 3.4374, "step": 3348 }, { "epoch": 1.0498432601880878, "grad_norm": 18.99493408203125, "learning_rate": 1.6745e-06, "loss": 2.7457, "step": 3349 }, { "epoch": 1.0501567398119123, "grad_norm": 65.04867553710938, "learning_rate": 1.6750000000000003e-06, "loss": 3.2615, "step": 3350 }, { "epoch": 1.0504702194357367, "grad_norm": 19.495983123779297, "learning_rate": 1.6755e-06, "loss": 2.5306, "step": 3351 }, { "epoch": 1.0507836990595611, "grad_norm": 70.9605712890625, "learning_rate": 1.6760000000000001e-06, "loss": 3.8023, "step": 3352 }, { "epoch": 1.0510971786833856, "grad_norm": 35.70189666748047, "learning_rate": 1.6765000000000002e-06, "loss": 2.8319, "step": 3353 }, { "epoch": 1.05141065830721, "grad_norm": 50.487388610839844, "learning_rate": 1.677e-06, "loss": 3.9747, "step": 3354 }, { "epoch": 1.0517241379310345, "grad_norm": 31.325843811035156, "learning_rate": 1.6775000000000002e-06, "loss": 3.1756, "step": 3355 }, { "epoch": 1.052037617554859, "grad_norm": 32.679195404052734, "learning_rate": 1.6780000000000003e-06, "loss": 2.9121, "step": 3356 }, { "epoch": 1.0523510971786834, "grad_norm": 40.90104675292969, "learning_rate": 1.6785000000000001e-06, "loss": 3.2628, "step": 3357 }, { "epoch": 1.0526645768025078, "grad_norm": 33.8775634765625, "learning_rate": 1.6790000000000001e-06, "loss": 2.3563, "step": 3358 }, { "epoch": 1.0529780564263322, "grad_norm": 25.625640869140625, "learning_rate": 1.6795e-06, "loss": 2.7023, "step": 3359 }, { "epoch": 1.0532915360501567, "grad_norm": 29.225643157958984, "learning_rate": 1.6800000000000002e-06, "loss": 2.8436, "step": 3360 }, { "epoch": 1.0536050156739811, "grad_norm": 24.310447692871094, "learning_rate": 1.6805000000000003e-06, "loss": 2.8803, "step": 3361 }, { "epoch": 1.0539184952978056, "grad_norm": 17.30785369873047, "learning_rate": 1.681e-06, "loss": 2.3881, "step": 3362 }, { "epoch": 1.05423197492163, "grad_norm": 21.964794158935547, "learning_rate": 1.6815000000000001e-06, "loss": 2.4327, "step": 3363 }, { "epoch": 1.0545454545454545, "grad_norm": 38.354225158691406, "learning_rate": 1.682e-06, "loss": 3.6111, "step": 3364 }, { "epoch": 1.054858934169279, "grad_norm": 27.03491973876953, "learning_rate": 1.6825000000000002e-06, "loss": 3.5078, "step": 3365 }, { "epoch": 1.0551724137931036, "grad_norm": 34.087646484375, "learning_rate": 1.6830000000000002e-06, "loss": 2.872, "step": 3366 }, { "epoch": 1.055485893416928, "grad_norm": 57.53993225097656, "learning_rate": 1.6835e-06, "loss": 3.4415, "step": 3367 }, { "epoch": 1.0557993730407524, "grad_norm": 119.94190216064453, "learning_rate": 1.684e-06, "loss": 3.5462, "step": 3368 }, { "epoch": 1.0561128526645769, "grad_norm": 46.50600051879883, "learning_rate": 1.6845e-06, "loss": 2.8154, "step": 3369 }, { "epoch": 1.0564263322884013, "grad_norm": 19.846092224121094, "learning_rate": 1.6850000000000002e-06, "loss": 2.6868, "step": 3370 }, { "epoch": 1.0567398119122258, "grad_norm": 42.88930892944336, "learning_rate": 1.6855000000000002e-06, "loss": 3.3179, "step": 3371 }, { "epoch": 1.0570532915360502, "grad_norm": 26.39129638671875, "learning_rate": 1.686e-06, "loss": 3.0197, "step": 3372 }, { "epoch": 1.0573667711598747, "grad_norm": 23.650426864624023, "learning_rate": 1.6865e-06, "loss": 3.1575, "step": 3373 }, { "epoch": 1.057680250783699, "grad_norm": 43.16667556762695, "learning_rate": 1.6870000000000001e-06, "loss": 3.9021, "step": 3374 }, { "epoch": 1.0579937304075235, "grad_norm": 27.793106079101562, "learning_rate": 1.6875000000000001e-06, "loss": 3.1221, "step": 3375 }, { "epoch": 1.058307210031348, "grad_norm": 29.778823852539062, "learning_rate": 1.6880000000000002e-06, "loss": 3.1776, "step": 3376 }, { "epoch": 1.0586206896551724, "grad_norm": 59.28425979614258, "learning_rate": 1.6885000000000002e-06, "loss": 3.1223, "step": 3377 }, { "epoch": 1.0589341692789969, "grad_norm": 26.89908218383789, "learning_rate": 1.689e-06, "loss": 2.7911, "step": 3378 }, { "epoch": 1.0592476489028213, "grad_norm": 33.18085479736328, "learning_rate": 1.6895e-06, "loss": 2.8164, "step": 3379 }, { "epoch": 1.0595611285266457, "grad_norm": 25.427278518676758, "learning_rate": 1.6900000000000003e-06, "loss": 2.6253, "step": 3380 }, { "epoch": 1.0598746081504702, "grad_norm": 67.79798126220703, "learning_rate": 1.6905000000000002e-06, "loss": 3.7484, "step": 3381 }, { "epoch": 1.0601880877742946, "grad_norm": 41.082977294921875, "learning_rate": 1.6910000000000002e-06, "loss": 3.2418, "step": 3382 }, { "epoch": 1.060501567398119, "grad_norm": 113.1292953491211, "learning_rate": 1.6915e-06, "loss": 3.2591, "step": 3383 }, { "epoch": 1.0608150470219435, "grad_norm": 29.729745864868164, "learning_rate": 1.692e-06, "loss": 3.3393, "step": 3384 }, { "epoch": 1.061128526645768, "grad_norm": 39.9203987121582, "learning_rate": 1.6925000000000003e-06, "loss": 3.3071, "step": 3385 }, { "epoch": 1.0614420062695924, "grad_norm": 51.5232048034668, "learning_rate": 1.6930000000000001e-06, "loss": 3.2022, "step": 3386 }, { "epoch": 1.0617554858934168, "grad_norm": 28.027128219604492, "learning_rate": 1.6935000000000002e-06, "loss": 2.9686, "step": 3387 }, { "epoch": 1.0620689655172413, "grad_norm": 17.445575714111328, "learning_rate": 1.694e-06, "loss": 2.3405, "step": 3388 }, { "epoch": 1.0623824451410657, "grad_norm": 31.220802307128906, "learning_rate": 1.6945e-06, "loss": 3.3004, "step": 3389 }, { "epoch": 1.0626959247648902, "grad_norm": 36.90548324584961, "learning_rate": 1.6950000000000003e-06, "loss": 3.3895, "step": 3390 }, { "epoch": 1.0630094043887148, "grad_norm": 19.605175018310547, "learning_rate": 1.6955000000000001e-06, "loss": 3.3256, "step": 3391 }, { "epoch": 1.0633228840125393, "grad_norm": 25.22102928161621, "learning_rate": 1.6960000000000002e-06, "loss": 3.0132, "step": 3392 }, { "epoch": 1.0636363636363637, "grad_norm": 31.76527976989746, "learning_rate": 1.6965000000000002e-06, "loss": 2.8584, "step": 3393 }, { "epoch": 1.0639498432601882, "grad_norm": 27.360450744628906, "learning_rate": 1.697e-06, "loss": 3.5555, "step": 3394 }, { "epoch": 1.0642633228840126, "grad_norm": 39.99030303955078, "learning_rate": 1.6975000000000003e-06, "loss": 3.9428, "step": 3395 }, { "epoch": 1.064576802507837, "grad_norm": 20.094100952148438, "learning_rate": 1.6980000000000003e-06, "loss": 2.9556, "step": 3396 }, { "epoch": 1.0648902821316615, "grad_norm": 27.564455032348633, "learning_rate": 1.6985000000000001e-06, "loss": 2.6065, "step": 3397 }, { "epoch": 1.065203761755486, "grad_norm": 23.814006805419922, "learning_rate": 1.6990000000000002e-06, "loss": 3.0498, "step": 3398 }, { "epoch": 1.0655172413793104, "grad_norm": 22.892719268798828, "learning_rate": 1.6995e-06, "loss": 3.1003, "step": 3399 }, { "epoch": 1.0658307210031348, "grad_norm": 37.36176681518555, "learning_rate": 1.7000000000000002e-06, "loss": 2.8643, "step": 3400 }, { "epoch": 1.0661442006269592, "grad_norm": 27.134132385253906, "learning_rate": 1.7005000000000003e-06, "loss": 3.4379, "step": 3401 }, { "epoch": 1.0664576802507837, "grad_norm": 45.84774398803711, "learning_rate": 1.701e-06, "loss": 3.2277, "step": 3402 }, { "epoch": 1.0667711598746081, "grad_norm": 40.72565841674805, "learning_rate": 1.7015000000000001e-06, "loss": 2.9327, "step": 3403 }, { "epoch": 1.0670846394984326, "grad_norm": 32.90715789794922, "learning_rate": 1.702e-06, "loss": 4.1643, "step": 3404 }, { "epoch": 1.067398119122257, "grad_norm": 29.4326229095459, "learning_rate": 1.7025000000000002e-06, "loss": 3.0146, "step": 3405 }, { "epoch": 1.0677115987460815, "grad_norm": 32.895835876464844, "learning_rate": 1.7030000000000003e-06, "loss": 2.61, "step": 3406 }, { "epoch": 1.068025078369906, "grad_norm": 26.282489776611328, "learning_rate": 1.7035e-06, "loss": 2.6263, "step": 3407 }, { "epoch": 1.0683385579937303, "grad_norm": 50.594940185546875, "learning_rate": 1.7040000000000001e-06, "loss": 2.5367, "step": 3408 }, { "epoch": 1.0686520376175548, "grad_norm": 23.735204696655273, "learning_rate": 1.7045e-06, "loss": 3.0807, "step": 3409 }, { "epoch": 1.0689655172413792, "grad_norm": 29.3944034576416, "learning_rate": 1.7050000000000002e-06, "loss": 2.8559, "step": 3410 }, { "epoch": 1.0692789968652037, "grad_norm": 101.75618743896484, "learning_rate": 1.7055000000000002e-06, "loss": 2.8017, "step": 3411 }, { "epoch": 1.069592476489028, "grad_norm": 32.16104507446289, "learning_rate": 1.706e-06, "loss": 3.0168, "step": 3412 }, { "epoch": 1.0699059561128528, "grad_norm": 33.96107864379883, "learning_rate": 1.7065e-06, "loss": 3.183, "step": 3413 }, { "epoch": 1.0702194357366772, "grad_norm": 24.979026794433594, "learning_rate": 1.7070000000000001e-06, "loss": 2.7527, "step": 3414 }, { "epoch": 1.0705329153605017, "grad_norm": 61.97425842285156, "learning_rate": 1.7075000000000002e-06, "loss": 3.9153, "step": 3415 }, { "epoch": 1.070846394984326, "grad_norm": 33.35090637207031, "learning_rate": 1.7080000000000002e-06, "loss": 3.1761, "step": 3416 }, { "epoch": 1.0711598746081505, "grad_norm": 24.40055274963379, "learning_rate": 1.7085000000000002e-06, "loss": 2.7182, "step": 3417 }, { "epoch": 1.071473354231975, "grad_norm": 36.82908630371094, "learning_rate": 1.709e-06, "loss": 3.2681, "step": 3418 }, { "epoch": 1.0717868338557994, "grad_norm": 75.78736114501953, "learning_rate": 1.7095000000000001e-06, "loss": 3.3688, "step": 3419 }, { "epoch": 1.0721003134796239, "grad_norm": 53.95949935913086, "learning_rate": 1.7100000000000004e-06, "loss": 3.3434, "step": 3420 }, { "epoch": 1.0724137931034483, "grad_norm": 38.91935348510742, "learning_rate": 1.7105000000000002e-06, "loss": 2.6255, "step": 3421 }, { "epoch": 1.0727272727272728, "grad_norm": 31.636262893676758, "learning_rate": 1.7110000000000002e-06, "loss": 3.2822, "step": 3422 }, { "epoch": 1.0730407523510972, "grad_norm": 49.99222183227539, "learning_rate": 1.7115e-06, "loss": 3.5524, "step": 3423 }, { "epoch": 1.0733542319749216, "grad_norm": 28.117033004760742, "learning_rate": 1.712e-06, "loss": 3.115, "step": 3424 }, { "epoch": 1.073667711598746, "grad_norm": 20.26447105407715, "learning_rate": 1.7125000000000003e-06, "loss": 2.6173, "step": 3425 }, { "epoch": 1.0739811912225705, "grad_norm": 17.405311584472656, "learning_rate": 1.7130000000000002e-06, "loss": 3.3576, "step": 3426 }, { "epoch": 1.074294670846395, "grad_norm": 42.87114715576172, "learning_rate": 1.7135000000000002e-06, "loss": 3.0969, "step": 3427 }, { "epoch": 1.0746081504702194, "grad_norm": 20.595788955688477, "learning_rate": 1.714e-06, "loss": 2.6237, "step": 3428 }, { "epoch": 1.0749216300940438, "grad_norm": 36.38020706176758, "learning_rate": 1.7145e-06, "loss": 3.5374, "step": 3429 }, { "epoch": 1.0752351097178683, "grad_norm": 31.2501163482666, "learning_rate": 1.7150000000000003e-06, "loss": 2.8522, "step": 3430 }, { "epoch": 1.0755485893416927, "grad_norm": 29.131214141845703, "learning_rate": 1.7155000000000001e-06, "loss": 2.7789, "step": 3431 }, { "epoch": 1.0758620689655172, "grad_norm": 26.90103530883789, "learning_rate": 1.7160000000000002e-06, "loss": 3.4374, "step": 3432 }, { "epoch": 1.0761755485893416, "grad_norm": 34.345088958740234, "learning_rate": 1.7165000000000002e-06, "loss": 2.7771, "step": 3433 }, { "epoch": 1.076489028213166, "grad_norm": 18.91362190246582, "learning_rate": 1.717e-06, "loss": 2.7293, "step": 3434 }, { "epoch": 1.0768025078369905, "grad_norm": 58.237918853759766, "learning_rate": 1.7175000000000003e-06, "loss": 2.9041, "step": 3435 }, { "epoch": 1.077115987460815, "grad_norm": 45.42815017700195, "learning_rate": 1.7180000000000003e-06, "loss": 3.1195, "step": 3436 }, { "epoch": 1.0774294670846394, "grad_norm": 20.961240768432617, "learning_rate": 1.7185000000000001e-06, "loss": 3.0985, "step": 3437 }, { "epoch": 1.077742946708464, "grad_norm": 31.146696090698242, "learning_rate": 1.7190000000000002e-06, "loss": 2.8074, "step": 3438 }, { "epoch": 1.0780564263322885, "grad_norm": 50.56280517578125, "learning_rate": 1.7195e-06, "loss": 3.9483, "step": 3439 }, { "epoch": 1.078369905956113, "grad_norm": 30.90845489501953, "learning_rate": 1.72e-06, "loss": 2.9276, "step": 3440 }, { "epoch": 1.0786833855799374, "grad_norm": 36.373931884765625, "learning_rate": 1.7205000000000003e-06, "loss": 2.7398, "step": 3441 }, { "epoch": 1.0789968652037618, "grad_norm": 38.69664764404297, "learning_rate": 1.7210000000000001e-06, "loss": 3.1982, "step": 3442 }, { "epoch": 1.0793103448275863, "grad_norm": 52.687503814697266, "learning_rate": 1.7215000000000002e-06, "loss": 2.7719, "step": 3443 }, { "epoch": 1.0796238244514107, "grad_norm": 28.561540603637695, "learning_rate": 1.722e-06, "loss": 2.7896, "step": 3444 }, { "epoch": 1.0799373040752351, "grad_norm": 104.28028869628906, "learning_rate": 1.7225e-06, "loss": 3.9382, "step": 3445 }, { "epoch": 1.0802507836990596, "grad_norm": 35.79922103881836, "learning_rate": 1.7230000000000003e-06, "loss": 3.2333, "step": 3446 }, { "epoch": 1.080564263322884, "grad_norm": 28.640287399291992, "learning_rate": 1.7235e-06, "loss": 2.6655, "step": 3447 }, { "epoch": 1.0808777429467085, "grad_norm": 28.324596405029297, "learning_rate": 1.7240000000000001e-06, "loss": 3.4944, "step": 3448 }, { "epoch": 1.081191222570533, "grad_norm": 35.781715393066406, "learning_rate": 1.7245e-06, "loss": 3.1321, "step": 3449 }, { "epoch": 1.0815047021943573, "grad_norm": 39.355499267578125, "learning_rate": 1.725e-06, "loss": 3.2125, "step": 3450 }, { "epoch": 1.0818181818181818, "grad_norm": 43.08556365966797, "learning_rate": 1.7255000000000003e-06, "loss": 3.4494, "step": 3451 }, { "epoch": 1.0821316614420062, "grad_norm": 49.84754943847656, "learning_rate": 1.726e-06, "loss": 3.232, "step": 3452 }, { "epoch": 1.0824451410658307, "grad_norm": 72.55428314208984, "learning_rate": 1.7265000000000001e-06, "loss": 3.0765, "step": 3453 }, { "epoch": 1.0827586206896551, "grad_norm": 43.84035110473633, "learning_rate": 1.7270000000000002e-06, "loss": 3.2667, "step": 3454 }, { "epoch": 1.0830721003134796, "grad_norm": 21.691329956054688, "learning_rate": 1.7275e-06, "loss": 2.6102, "step": 3455 }, { "epoch": 1.083385579937304, "grad_norm": 28.518117904663086, "learning_rate": 1.7280000000000002e-06, "loss": 3.7609, "step": 3456 }, { "epoch": 1.0836990595611284, "grad_norm": 20.661483764648438, "learning_rate": 1.7285000000000003e-06, "loss": 2.8076, "step": 3457 }, { "epoch": 1.0840125391849529, "grad_norm": 21.91556167602539, "learning_rate": 1.729e-06, "loss": 3.0376, "step": 3458 }, { "epoch": 1.0843260188087773, "grad_norm": 31.704559326171875, "learning_rate": 1.7295000000000001e-06, "loss": 2.8547, "step": 3459 }, { "epoch": 1.084639498432602, "grad_norm": 103.53755187988281, "learning_rate": 1.73e-06, "loss": 3.1362, "step": 3460 }, { "epoch": 1.0849529780564264, "grad_norm": 21.896881103515625, "learning_rate": 1.7305000000000002e-06, "loss": 2.8094, "step": 3461 }, { "epoch": 1.0852664576802509, "grad_norm": 26.890310287475586, "learning_rate": 1.7310000000000002e-06, "loss": 2.6375, "step": 3462 }, { "epoch": 1.0855799373040753, "grad_norm": 21.702211380004883, "learning_rate": 1.7315e-06, "loss": 2.839, "step": 3463 }, { "epoch": 1.0858934169278998, "grad_norm": 17.826766967773438, "learning_rate": 1.732e-06, "loss": 2.754, "step": 3464 }, { "epoch": 1.0862068965517242, "grad_norm": 27.036876678466797, "learning_rate": 1.7325e-06, "loss": 3.0524, "step": 3465 }, { "epoch": 1.0865203761755486, "grad_norm": 50.12677764892578, "learning_rate": 1.7330000000000002e-06, "loss": 4.3457, "step": 3466 }, { "epoch": 1.086833855799373, "grad_norm": 36.31730651855469, "learning_rate": 1.7335000000000002e-06, "loss": 2.7942, "step": 3467 }, { "epoch": 1.0871473354231975, "grad_norm": 29.601634979248047, "learning_rate": 1.734e-06, "loss": 2.8078, "step": 3468 }, { "epoch": 1.087460815047022, "grad_norm": 33.91842269897461, "learning_rate": 1.7345e-06, "loss": 2.7449, "step": 3469 }, { "epoch": 1.0877742946708464, "grad_norm": 27.274818420410156, "learning_rate": 1.7350000000000001e-06, "loss": 2.7981, "step": 3470 }, { "epoch": 1.0880877742946709, "grad_norm": 23.352020263671875, "learning_rate": 1.7355000000000002e-06, "loss": 2.7459, "step": 3471 }, { "epoch": 1.0884012539184953, "grad_norm": 39.16693115234375, "learning_rate": 1.7360000000000002e-06, "loss": 2.7875, "step": 3472 }, { "epoch": 1.0887147335423197, "grad_norm": 32.53330993652344, "learning_rate": 1.7365000000000002e-06, "loss": 3.6786, "step": 3473 }, { "epoch": 1.0890282131661442, "grad_norm": 124.53848266601562, "learning_rate": 1.737e-06, "loss": 3.1399, "step": 3474 }, { "epoch": 1.0893416927899686, "grad_norm": 42.1822395324707, "learning_rate": 1.7375e-06, "loss": 3.3776, "step": 3475 }, { "epoch": 1.089655172413793, "grad_norm": 19.54660987854004, "learning_rate": 1.7380000000000003e-06, "loss": 2.7341, "step": 3476 }, { "epoch": 1.0899686520376175, "grad_norm": 25.96428108215332, "learning_rate": 1.7385000000000002e-06, "loss": 2.8165, "step": 3477 }, { "epoch": 1.090282131661442, "grad_norm": 46.42351150512695, "learning_rate": 1.7390000000000002e-06, "loss": 2.9387, "step": 3478 }, { "epoch": 1.0905956112852664, "grad_norm": 16.924457550048828, "learning_rate": 1.7395e-06, "loss": 2.653, "step": 3479 }, { "epoch": 1.0909090909090908, "grad_norm": 25.124286651611328, "learning_rate": 1.74e-06, "loss": 2.5537, "step": 3480 }, { "epoch": 1.0912225705329153, "grad_norm": 64.10070037841797, "learning_rate": 1.7405000000000003e-06, "loss": 3.6643, "step": 3481 }, { "epoch": 1.0915360501567397, "grad_norm": 24.5253963470459, "learning_rate": 1.7410000000000001e-06, "loss": 2.676, "step": 3482 }, { "epoch": 1.0918495297805642, "grad_norm": 32.09825897216797, "learning_rate": 1.7415000000000002e-06, "loss": 3.3551, "step": 3483 }, { "epoch": 1.0921630094043888, "grad_norm": 35.84313201904297, "learning_rate": 1.742e-06, "loss": 3.4427, "step": 3484 }, { "epoch": 1.0924764890282133, "grad_norm": 51.0280647277832, "learning_rate": 1.7425e-06, "loss": 3.3035, "step": 3485 }, { "epoch": 1.0927899686520377, "grad_norm": 23.819229125976562, "learning_rate": 1.7430000000000003e-06, "loss": 2.9676, "step": 3486 }, { "epoch": 1.0931034482758621, "grad_norm": 38.78548812866211, "learning_rate": 1.7435000000000001e-06, "loss": 3.3918, "step": 3487 }, { "epoch": 1.0934169278996866, "grad_norm": 38.34880065917969, "learning_rate": 1.7440000000000002e-06, "loss": 3.0882, "step": 3488 }, { "epoch": 1.093730407523511, "grad_norm": 32.39979553222656, "learning_rate": 1.7445e-06, "loss": 2.9143, "step": 3489 }, { "epoch": 1.0940438871473355, "grad_norm": 51.53925704956055, "learning_rate": 1.745e-06, "loss": 2.7739, "step": 3490 }, { "epoch": 1.09435736677116, "grad_norm": 47.33768081665039, "learning_rate": 1.7455000000000003e-06, "loss": 3.2417, "step": 3491 }, { "epoch": 1.0946708463949844, "grad_norm": 23.84410285949707, "learning_rate": 1.746e-06, "loss": 2.9263, "step": 3492 }, { "epoch": 1.0949843260188088, "grad_norm": 27.154666900634766, "learning_rate": 1.7465000000000001e-06, "loss": 2.8824, "step": 3493 }, { "epoch": 1.0952978056426332, "grad_norm": 19.792095184326172, "learning_rate": 1.7470000000000002e-06, "loss": 2.5153, "step": 3494 }, { "epoch": 1.0956112852664577, "grad_norm": 29.83961296081543, "learning_rate": 1.7475e-06, "loss": 2.6682, "step": 3495 }, { "epoch": 1.0959247648902821, "grad_norm": 22.432218551635742, "learning_rate": 1.7480000000000002e-06, "loss": 2.6332, "step": 3496 }, { "epoch": 1.0962382445141066, "grad_norm": 34.51369857788086, "learning_rate": 1.7485000000000003e-06, "loss": 2.8702, "step": 3497 }, { "epoch": 1.096551724137931, "grad_norm": 22.819372177124023, "learning_rate": 1.7490000000000001e-06, "loss": 2.5255, "step": 3498 }, { "epoch": 1.0968652037617554, "grad_norm": 22.45680809020996, "learning_rate": 1.7495000000000001e-06, "loss": 3.092, "step": 3499 }, { "epoch": 1.09717868338558, "grad_norm": 47.499847412109375, "learning_rate": 1.75e-06, "loss": 4.2197, "step": 3500 }, { "epoch": 1.0974921630094043, "grad_norm": 75.1513442993164, "learning_rate": 1.7505000000000002e-06, "loss": 3.1316, "step": 3501 }, { "epoch": 1.0978056426332288, "grad_norm": 18.71675682067871, "learning_rate": 1.7510000000000003e-06, "loss": 2.5032, "step": 3502 }, { "epoch": 1.0981191222570532, "grad_norm": 51.24578094482422, "learning_rate": 1.7515e-06, "loss": 3.402, "step": 3503 }, { "epoch": 1.0984326018808777, "grad_norm": 16.40824317932129, "learning_rate": 1.7520000000000001e-06, "loss": 2.4963, "step": 3504 }, { "epoch": 1.098746081504702, "grad_norm": 87.81695556640625, "learning_rate": 1.7525e-06, "loss": 4.5785, "step": 3505 }, { "epoch": 1.0990595611285268, "grad_norm": 46.15744400024414, "learning_rate": 1.7530000000000002e-06, "loss": 3.1064, "step": 3506 }, { "epoch": 1.0993730407523512, "grad_norm": 34.15064239501953, "learning_rate": 1.7535000000000002e-06, "loss": 3.1321, "step": 3507 }, { "epoch": 1.0996865203761756, "grad_norm": 25.923816680908203, "learning_rate": 1.754e-06, "loss": 3.0893, "step": 3508 }, { "epoch": 1.1, "grad_norm": 32.922027587890625, "learning_rate": 1.7545e-06, "loss": 2.608, "step": 3509 }, { "epoch": 1.1003134796238245, "grad_norm": 37.616695404052734, "learning_rate": 1.7550000000000001e-06, "loss": 3.7035, "step": 3510 }, { "epoch": 1.100626959247649, "grad_norm": 35.45209884643555, "learning_rate": 1.7555000000000002e-06, "loss": 2.9419, "step": 3511 }, { "epoch": 1.1009404388714734, "grad_norm": 35.20869064331055, "learning_rate": 1.7560000000000002e-06, "loss": 3.0732, "step": 3512 }, { "epoch": 1.1012539184952979, "grad_norm": 44.86465835571289, "learning_rate": 1.7565000000000003e-06, "loss": 3.3239, "step": 3513 }, { "epoch": 1.1015673981191223, "grad_norm": 22.87040901184082, "learning_rate": 1.757e-06, "loss": 2.6935, "step": 3514 }, { "epoch": 1.1018808777429467, "grad_norm": 33.13298034667969, "learning_rate": 1.7575000000000001e-06, "loss": 2.7612, "step": 3515 }, { "epoch": 1.1021943573667712, "grad_norm": 18.739044189453125, "learning_rate": 1.7580000000000004e-06, "loss": 2.7834, "step": 3516 }, { "epoch": 1.1025078369905956, "grad_norm": 40.461647033691406, "learning_rate": 1.7585000000000002e-06, "loss": 3.0305, "step": 3517 }, { "epoch": 1.10282131661442, "grad_norm": 27.588546752929688, "learning_rate": 1.7590000000000002e-06, "loss": 2.4332, "step": 3518 }, { "epoch": 1.1031347962382445, "grad_norm": 26.194929122924805, "learning_rate": 1.7595e-06, "loss": 2.8379, "step": 3519 }, { "epoch": 1.103448275862069, "grad_norm": 34.12715530395508, "learning_rate": 1.76e-06, "loss": 2.7955, "step": 3520 }, { "epoch": 1.1037617554858934, "grad_norm": 24.855409622192383, "learning_rate": 1.7605000000000003e-06, "loss": 2.7204, "step": 3521 }, { "epoch": 1.1040752351097178, "grad_norm": 22.457441329956055, "learning_rate": 1.7610000000000002e-06, "loss": 3.2287, "step": 3522 }, { "epoch": 1.1043887147335423, "grad_norm": 21.55569839477539, "learning_rate": 1.7615000000000002e-06, "loss": 3.0252, "step": 3523 }, { "epoch": 1.1047021943573667, "grad_norm": 22.316848754882812, "learning_rate": 1.762e-06, "loss": 2.7789, "step": 3524 }, { "epoch": 1.1050156739811912, "grad_norm": 51.03260803222656, "learning_rate": 1.7625e-06, "loss": 3.6778, "step": 3525 }, { "epoch": 1.1053291536050156, "grad_norm": 20.623655319213867, "learning_rate": 1.7630000000000003e-06, "loss": 2.9086, "step": 3526 }, { "epoch": 1.10564263322884, "grad_norm": 22.799224853515625, "learning_rate": 1.7635000000000001e-06, "loss": 2.9854, "step": 3527 }, { "epoch": 1.1059561128526645, "grad_norm": 104.5829086303711, "learning_rate": 1.7640000000000002e-06, "loss": 3.4163, "step": 3528 }, { "epoch": 1.106269592476489, "grad_norm": 46.149879455566406, "learning_rate": 1.7645e-06, "loss": 2.5803, "step": 3529 }, { "epoch": 1.1065830721003134, "grad_norm": 21.742412567138672, "learning_rate": 1.765e-06, "loss": 2.3909, "step": 3530 }, { "epoch": 1.106896551724138, "grad_norm": 21.718233108520508, "learning_rate": 1.7655000000000003e-06, "loss": 2.7341, "step": 3531 }, { "epoch": 1.1072100313479625, "grad_norm": 19.40610122680664, "learning_rate": 1.7660000000000001e-06, "loss": 2.5342, "step": 3532 }, { "epoch": 1.107523510971787, "grad_norm": 29.3079891204834, "learning_rate": 1.7665000000000002e-06, "loss": 4.2854, "step": 3533 }, { "epoch": 1.1078369905956114, "grad_norm": 31.97117805480957, "learning_rate": 1.7670000000000002e-06, "loss": 2.8468, "step": 3534 }, { "epoch": 1.1081504702194358, "grad_norm": 32.589317321777344, "learning_rate": 1.7675e-06, "loss": 2.9813, "step": 3535 }, { "epoch": 1.1084639498432602, "grad_norm": 24.950849533081055, "learning_rate": 1.7680000000000003e-06, "loss": 2.7851, "step": 3536 }, { "epoch": 1.1087774294670847, "grad_norm": 29.49528694152832, "learning_rate": 1.7685000000000003e-06, "loss": 3.3423, "step": 3537 }, { "epoch": 1.1090909090909091, "grad_norm": 23.108978271484375, "learning_rate": 1.7690000000000001e-06, "loss": 2.9801, "step": 3538 }, { "epoch": 1.1094043887147336, "grad_norm": 26.269826889038086, "learning_rate": 1.7695000000000002e-06, "loss": 2.9439, "step": 3539 }, { "epoch": 1.109717868338558, "grad_norm": 25.8447322845459, "learning_rate": 1.77e-06, "loss": 2.7837, "step": 3540 }, { "epoch": 1.1100313479623825, "grad_norm": 25.49906349182129, "learning_rate": 1.7705000000000002e-06, "loss": 2.6496, "step": 3541 }, { "epoch": 1.110344827586207, "grad_norm": 31.025354385375977, "learning_rate": 1.7710000000000003e-06, "loss": 2.546, "step": 3542 }, { "epoch": 1.1106583072100313, "grad_norm": 34.72405242919922, "learning_rate": 1.7715000000000001e-06, "loss": 2.8026, "step": 3543 }, { "epoch": 1.1109717868338558, "grad_norm": 36.77132034301758, "learning_rate": 1.7720000000000001e-06, "loss": 3.3022, "step": 3544 }, { "epoch": 1.1112852664576802, "grad_norm": 32.44343566894531, "learning_rate": 1.7725e-06, "loss": 3.0009, "step": 3545 }, { "epoch": 1.1115987460815047, "grad_norm": 18.060049057006836, "learning_rate": 1.7730000000000002e-06, "loss": 2.5424, "step": 3546 }, { "epoch": 1.111912225705329, "grad_norm": 33.717308044433594, "learning_rate": 1.7735000000000003e-06, "loss": 3.7549, "step": 3547 }, { "epoch": 1.1122257053291535, "grad_norm": 18.197097778320312, "learning_rate": 1.774e-06, "loss": 2.7566, "step": 3548 }, { "epoch": 1.112539184952978, "grad_norm": 31.980981826782227, "learning_rate": 1.7745000000000001e-06, "loss": 3.3021, "step": 3549 }, { "epoch": 1.1128526645768024, "grad_norm": 47.19995880126953, "learning_rate": 1.7750000000000002e-06, "loss": 2.9785, "step": 3550 }, { "epoch": 1.1131661442006269, "grad_norm": 23.19911766052246, "learning_rate": 1.7755000000000002e-06, "loss": 3.0185, "step": 3551 }, { "epoch": 1.1134796238244513, "grad_norm": 35.693580627441406, "learning_rate": 1.7760000000000002e-06, "loss": 3.4506, "step": 3552 }, { "epoch": 1.113793103448276, "grad_norm": 36.20033645629883, "learning_rate": 1.7765000000000003e-06, "loss": 2.6537, "step": 3553 }, { "epoch": 1.1141065830721004, "grad_norm": 40.76347732543945, "learning_rate": 1.777e-06, "loss": 2.6187, "step": 3554 }, { "epoch": 1.1144200626959249, "grad_norm": 33.449562072753906, "learning_rate": 1.7775000000000001e-06, "loss": 3.0086, "step": 3555 }, { "epoch": 1.1147335423197493, "grad_norm": 70.72737884521484, "learning_rate": 1.7780000000000004e-06, "loss": 3.4293, "step": 3556 }, { "epoch": 1.1150470219435737, "grad_norm": 24.169923782348633, "learning_rate": 1.7785000000000002e-06, "loss": 2.5885, "step": 3557 }, { "epoch": 1.1153605015673982, "grad_norm": 22.239110946655273, "learning_rate": 1.7790000000000002e-06, "loss": 2.6941, "step": 3558 }, { "epoch": 1.1156739811912226, "grad_norm": 31.481800079345703, "learning_rate": 1.7795e-06, "loss": 2.9983, "step": 3559 }, { "epoch": 1.115987460815047, "grad_norm": 32.51259994506836, "learning_rate": 1.7800000000000001e-06, "loss": 2.5775, "step": 3560 }, { "epoch": 1.1163009404388715, "grad_norm": 23.58956527709961, "learning_rate": 1.7805000000000004e-06, "loss": 3.3801, "step": 3561 }, { "epoch": 1.116614420062696, "grad_norm": 22.67048454284668, "learning_rate": 1.7810000000000002e-06, "loss": 2.5815, "step": 3562 }, { "epoch": 1.1169278996865204, "grad_norm": 28.150209426879883, "learning_rate": 1.7815000000000002e-06, "loss": 2.6133, "step": 3563 }, { "epoch": 1.1172413793103448, "grad_norm": 63.03923416137695, "learning_rate": 1.782e-06, "loss": 2.9811, "step": 3564 }, { "epoch": 1.1175548589341693, "grad_norm": 101.25431060791016, "learning_rate": 1.7825e-06, "loss": 3.5238, "step": 3565 }, { "epoch": 1.1178683385579937, "grad_norm": 70.68958282470703, "learning_rate": 1.783e-06, "loss": 3.4413, "step": 3566 }, { "epoch": 1.1181818181818182, "grad_norm": 18.009140014648438, "learning_rate": 1.7835000000000002e-06, "loss": 2.4277, "step": 3567 }, { "epoch": 1.1184952978056426, "grad_norm": 41.71234893798828, "learning_rate": 1.7840000000000002e-06, "loss": 3.7431, "step": 3568 }, { "epoch": 1.118808777429467, "grad_norm": 62.48421096801758, "learning_rate": 1.7845e-06, "loss": 3.1696, "step": 3569 }, { "epoch": 1.1191222570532915, "grad_norm": 36.058467864990234, "learning_rate": 1.785e-06, "loss": 4.098, "step": 3570 }, { "epoch": 1.119435736677116, "grad_norm": 45.1207160949707, "learning_rate": 1.7855e-06, "loss": 2.8022, "step": 3571 }, { "epoch": 1.1197492163009404, "grad_norm": 200.1226348876953, "learning_rate": 1.7860000000000001e-06, "loss": 3.0398, "step": 3572 }, { "epoch": 1.1200626959247648, "grad_norm": 20.55527114868164, "learning_rate": 1.7865000000000002e-06, "loss": 2.7789, "step": 3573 }, { "epoch": 1.1203761755485893, "grad_norm": 42.2872428894043, "learning_rate": 1.7870000000000002e-06, "loss": 2.9237, "step": 3574 }, { "epoch": 1.1206896551724137, "grad_norm": 35.486915588378906, "learning_rate": 1.7875e-06, "loss": 3.1821, "step": 3575 }, { "epoch": 1.1210031347962381, "grad_norm": 40.49298095703125, "learning_rate": 1.788e-06, "loss": 3.2534, "step": 3576 }, { "epoch": 1.1213166144200626, "grad_norm": 24.555992126464844, "learning_rate": 1.7885000000000003e-06, "loss": 2.5797, "step": 3577 }, { "epoch": 1.1216300940438872, "grad_norm": 35.7850341796875, "learning_rate": 1.7890000000000002e-06, "loss": 3.2042, "step": 3578 }, { "epoch": 1.1219435736677117, "grad_norm": 47.83175277709961, "learning_rate": 1.7895000000000002e-06, "loss": 2.9337, "step": 3579 }, { "epoch": 1.1222570532915361, "grad_norm": 29.239980697631836, "learning_rate": 1.79e-06, "loss": 2.608, "step": 3580 }, { "epoch": 1.1225705329153606, "grad_norm": 22.09000015258789, "learning_rate": 1.7905e-06, "loss": 2.3577, "step": 3581 }, { "epoch": 1.122884012539185, "grad_norm": 33.30189514160156, "learning_rate": 1.7910000000000003e-06, "loss": 2.899, "step": 3582 }, { "epoch": 1.1231974921630095, "grad_norm": 18.214956283569336, "learning_rate": 1.7915000000000001e-06, "loss": 2.6793, "step": 3583 }, { "epoch": 1.123510971786834, "grad_norm": 29.422252655029297, "learning_rate": 1.7920000000000002e-06, "loss": 2.7343, "step": 3584 }, { "epoch": 1.1238244514106583, "grad_norm": 54.86771011352539, "learning_rate": 1.7925e-06, "loss": 2.9922, "step": 3585 }, { "epoch": 1.1241379310344828, "grad_norm": 33.0071907043457, "learning_rate": 1.793e-06, "loss": 2.7256, "step": 3586 }, { "epoch": 1.1244514106583072, "grad_norm": 38.235755920410156, "learning_rate": 1.7935000000000003e-06, "loss": 3.1273, "step": 3587 }, { "epoch": 1.1247648902821317, "grad_norm": 34.35319900512695, "learning_rate": 1.794e-06, "loss": 2.7182, "step": 3588 }, { "epoch": 1.125078369905956, "grad_norm": 24.03645896911621, "learning_rate": 1.7945000000000001e-06, "loss": 3.0134, "step": 3589 }, { "epoch": 1.1253918495297806, "grad_norm": 22.386899948120117, "learning_rate": 1.7950000000000002e-06, "loss": 2.56, "step": 3590 }, { "epoch": 1.125705329153605, "grad_norm": 21.036725997924805, "learning_rate": 1.7955e-06, "loss": 3.0022, "step": 3591 }, { "epoch": 1.1260188087774294, "grad_norm": 37.754215240478516, "learning_rate": 1.7960000000000003e-06, "loss": 2.8063, "step": 3592 }, { "epoch": 1.1263322884012539, "grad_norm": 34.971317291259766, "learning_rate": 1.7965000000000003e-06, "loss": 2.4451, "step": 3593 }, { "epoch": 1.1266457680250783, "grad_norm": 27.26007080078125, "learning_rate": 1.7970000000000001e-06, "loss": 2.6376, "step": 3594 }, { "epoch": 1.1269592476489028, "grad_norm": 31.517719268798828, "learning_rate": 1.7975000000000002e-06, "loss": 2.7497, "step": 3595 }, { "epoch": 1.1272727272727272, "grad_norm": 32.8683967590332, "learning_rate": 1.798e-06, "loss": 2.6048, "step": 3596 }, { "epoch": 1.1275862068965516, "grad_norm": 25.371549606323242, "learning_rate": 1.7985000000000002e-06, "loss": 2.5012, "step": 3597 }, { "epoch": 1.127899686520376, "grad_norm": 41.98431396484375, "learning_rate": 1.7990000000000003e-06, "loss": 3.0868, "step": 3598 }, { "epoch": 1.1282131661442005, "grad_norm": 32.28398513793945, "learning_rate": 1.7995e-06, "loss": 2.5616, "step": 3599 }, { "epoch": 1.1285266457680252, "grad_norm": 41.10795974731445, "learning_rate": 1.8000000000000001e-06, "loss": 2.9102, "step": 3600 }, { "epoch": 1.1288401253918496, "grad_norm": 40.71234130859375, "learning_rate": 1.8005e-06, "loss": 3.9757, "step": 3601 }, { "epoch": 1.129153605015674, "grad_norm": 29.76866340637207, "learning_rate": 1.8010000000000002e-06, "loss": 3.0691, "step": 3602 }, { "epoch": 1.1294670846394985, "grad_norm": 43.9595832824707, "learning_rate": 1.8015000000000002e-06, "loss": 2.918, "step": 3603 }, { "epoch": 1.129780564263323, "grad_norm": 34.53847122192383, "learning_rate": 1.802e-06, "loss": 3.1386, "step": 3604 }, { "epoch": 1.1300940438871474, "grad_norm": 29.814184188842773, "learning_rate": 1.8025000000000001e-06, "loss": 3.0358, "step": 3605 }, { "epoch": 1.1304075235109718, "grad_norm": 36.51237106323242, "learning_rate": 1.803e-06, "loss": 3.2394, "step": 3606 }, { "epoch": 1.1307210031347963, "grad_norm": 21.537824630737305, "learning_rate": 1.8035000000000002e-06, "loss": 2.8301, "step": 3607 }, { "epoch": 1.1310344827586207, "grad_norm": 53.89097213745117, "learning_rate": 1.8040000000000002e-06, "loss": 3.1312, "step": 3608 }, { "epoch": 1.1313479623824452, "grad_norm": 25.06736946105957, "learning_rate": 1.8045e-06, "loss": 2.9153, "step": 3609 }, { "epoch": 1.1316614420062696, "grad_norm": 21.754596710205078, "learning_rate": 1.805e-06, "loss": 3.1249, "step": 3610 }, { "epoch": 1.131974921630094, "grad_norm": 71.8990707397461, "learning_rate": 1.8055000000000001e-06, "loss": 2.648, "step": 3611 }, { "epoch": 1.1322884012539185, "grad_norm": 29.600143432617188, "learning_rate": 1.8060000000000002e-06, "loss": 2.9847, "step": 3612 }, { "epoch": 1.132601880877743, "grad_norm": 20.959943771362305, "learning_rate": 1.8065000000000002e-06, "loss": 2.438, "step": 3613 }, { "epoch": 1.1329153605015674, "grad_norm": 31.19559669494629, "learning_rate": 1.8070000000000002e-06, "loss": 3.2214, "step": 3614 }, { "epoch": 1.1332288401253918, "grad_norm": 42.91053009033203, "learning_rate": 1.8075e-06, "loss": 2.8965, "step": 3615 }, { "epoch": 1.1335423197492163, "grad_norm": 82.3578872680664, "learning_rate": 1.808e-06, "loss": 3.4786, "step": 3616 }, { "epoch": 1.1338557993730407, "grad_norm": 33.798587799072266, "learning_rate": 1.8085000000000003e-06, "loss": 3.9894, "step": 3617 }, { "epoch": 1.1341692789968651, "grad_norm": 23.663040161132812, "learning_rate": 1.8090000000000002e-06, "loss": 2.8058, "step": 3618 }, { "epoch": 1.1344827586206896, "grad_norm": 25.969789505004883, "learning_rate": 1.8095000000000002e-06, "loss": 2.981, "step": 3619 }, { "epoch": 1.134796238244514, "grad_norm": 34.255165100097656, "learning_rate": 1.81e-06, "loss": 3.3658, "step": 3620 }, { "epoch": 1.1351097178683385, "grad_norm": 25.960176467895508, "learning_rate": 1.8105e-06, "loss": 2.7904, "step": 3621 }, { "epoch": 1.135423197492163, "grad_norm": 23.85843849182129, "learning_rate": 1.8110000000000003e-06, "loss": 3.1468, "step": 3622 }, { "epoch": 1.1357366771159874, "grad_norm": 40.975677490234375, "learning_rate": 1.8115000000000001e-06, "loss": 2.6138, "step": 3623 }, { "epoch": 1.1360501567398118, "grad_norm": 53.12677001953125, "learning_rate": 1.8120000000000002e-06, "loss": 3.0403, "step": 3624 }, { "epoch": 1.1363636363636362, "grad_norm": 28.290157318115234, "learning_rate": 1.8125e-06, "loss": 2.918, "step": 3625 }, { "epoch": 1.136677115987461, "grad_norm": 25.967588424682617, "learning_rate": 1.813e-06, "loss": 3.2999, "step": 3626 }, { "epoch": 1.1369905956112853, "grad_norm": 28.83611297607422, "learning_rate": 1.8135000000000003e-06, "loss": 2.4556, "step": 3627 }, { "epoch": 1.1373040752351098, "grad_norm": 33.06849670410156, "learning_rate": 1.8140000000000001e-06, "loss": 2.8766, "step": 3628 }, { "epoch": 1.1376175548589342, "grad_norm": 109.74273681640625, "learning_rate": 1.8145000000000002e-06, "loss": 2.9064, "step": 3629 }, { "epoch": 1.1379310344827587, "grad_norm": 29.229536056518555, "learning_rate": 1.8150000000000002e-06, "loss": 2.6592, "step": 3630 }, { "epoch": 1.1382445141065831, "grad_norm": 55.26179885864258, "learning_rate": 1.8155e-06, "loss": 2.8912, "step": 3631 }, { "epoch": 1.1385579937304076, "grad_norm": 23.847789764404297, "learning_rate": 1.8160000000000003e-06, "loss": 2.6741, "step": 3632 }, { "epoch": 1.138871473354232, "grad_norm": 93.24641418457031, "learning_rate": 1.8165000000000003e-06, "loss": 3.6647, "step": 3633 }, { "epoch": 1.1391849529780564, "grad_norm": 24.152320861816406, "learning_rate": 1.8170000000000001e-06, "loss": 2.5677, "step": 3634 }, { "epoch": 1.1394984326018809, "grad_norm": 37.18745040893555, "learning_rate": 1.8175000000000002e-06, "loss": 2.7119, "step": 3635 }, { "epoch": 1.1398119122257053, "grad_norm": 18.85127067565918, "learning_rate": 1.818e-06, "loss": 2.6048, "step": 3636 }, { "epoch": 1.1401253918495298, "grad_norm": 64.41525268554688, "learning_rate": 1.8185000000000002e-06, "loss": 3.0811, "step": 3637 }, { "epoch": 1.1404388714733542, "grad_norm": 47.21731185913086, "learning_rate": 1.8190000000000003e-06, "loss": 2.8125, "step": 3638 }, { "epoch": 1.1407523510971787, "grad_norm": 87.95037078857422, "learning_rate": 1.8195000000000001e-06, "loss": 3.9969, "step": 3639 }, { "epoch": 1.141065830721003, "grad_norm": 69.32103729248047, "learning_rate": 1.8200000000000002e-06, "loss": 3.752, "step": 3640 }, { "epoch": 1.1413793103448275, "grad_norm": 42.07493591308594, "learning_rate": 1.8205e-06, "loss": 3.4211, "step": 3641 }, { "epoch": 1.141692789968652, "grad_norm": 27.793582916259766, "learning_rate": 1.8210000000000002e-06, "loss": 2.6521, "step": 3642 }, { "epoch": 1.1420062695924764, "grad_norm": 25.326906204223633, "learning_rate": 1.8215000000000003e-06, "loss": 2.6708, "step": 3643 }, { "epoch": 1.1423197492163009, "grad_norm": 52.35178756713867, "learning_rate": 1.822e-06, "loss": 3.6085, "step": 3644 }, { "epoch": 1.1426332288401253, "grad_norm": 26.3311824798584, "learning_rate": 1.8225000000000001e-06, "loss": 3.0224, "step": 3645 }, { "epoch": 1.14294670846395, "grad_norm": 23.436771392822266, "learning_rate": 1.823e-06, "loss": 3.0373, "step": 3646 }, { "epoch": 1.1432601880877744, "grad_norm": 21.673803329467773, "learning_rate": 1.8235000000000002e-06, "loss": 2.7107, "step": 3647 }, { "epoch": 1.1435736677115989, "grad_norm": 18.09606170654297, "learning_rate": 1.8240000000000002e-06, "loss": 2.8579, "step": 3648 }, { "epoch": 1.1438871473354233, "grad_norm": 15.521315574645996, "learning_rate": 1.8245e-06, "loss": 2.6568, "step": 3649 }, { "epoch": 1.1442006269592477, "grad_norm": 22.498708724975586, "learning_rate": 1.825e-06, "loss": 2.571, "step": 3650 }, { "epoch": 1.1445141065830722, "grad_norm": 40.23341369628906, "learning_rate": 1.8255000000000001e-06, "loss": 2.7546, "step": 3651 }, { "epoch": 1.1448275862068966, "grad_norm": 20.80241584777832, "learning_rate": 1.8260000000000002e-06, "loss": 2.4507, "step": 3652 }, { "epoch": 1.145141065830721, "grad_norm": 16.20147705078125, "learning_rate": 1.8265000000000002e-06, "loss": 2.4255, "step": 3653 }, { "epoch": 1.1454545454545455, "grad_norm": 25.723865509033203, "learning_rate": 1.8270000000000003e-06, "loss": 2.5392, "step": 3654 }, { "epoch": 1.14576802507837, "grad_norm": 35.9820671081543, "learning_rate": 1.8275e-06, "loss": 3.0196, "step": 3655 }, { "epoch": 1.1460815047021944, "grad_norm": 22.180850982666016, "learning_rate": 1.8280000000000001e-06, "loss": 2.7761, "step": 3656 }, { "epoch": 1.1463949843260188, "grad_norm": 29.314672470092773, "learning_rate": 1.8285000000000004e-06, "loss": 3.6485, "step": 3657 }, { "epoch": 1.1467084639498433, "grad_norm": 48.099334716796875, "learning_rate": 1.8290000000000002e-06, "loss": 2.9849, "step": 3658 }, { "epoch": 1.1470219435736677, "grad_norm": 28.2821102142334, "learning_rate": 1.8295000000000002e-06, "loss": 2.592, "step": 3659 }, { "epoch": 1.1473354231974922, "grad_norm": 25.157413482666016, "learning_rate": 1.83e-06, "loss": 2.9226, "step": 3660 }, { "epoch": 1.1476489028213166, "grad_norm": 57.659461975097656, "learning_rate": 1.8305e-06, "loss": 2.345, "step": 3661 }, { "epoch": 1.147962382445141, "grad_norm": 38.85525894165039, "learning_rate": 1.8310000000000003e-06, "loss": 3.045, "step": 3662 }, { "epoch": 1.1482758620689655, "grad_norm": 60.336917877197266, "learning_rate": 1.8315000000000002e-06, "loss": 3.0008, "step": 3663 }, { "epoch": 1.14858934169279, "grad_norm": 30.43072509765625, "learning_rate": 1.8320000000000002e-06, "loss": 3.0494, "step": 3664 }, { "epoch": 1.1489028213166144, "grad_norm": 70.96421813964844, "learning_rate": 1.8325e-06, "loss": 2.9899, "step": 3665 }, { "epoch": 1.1492163009404388, "grad_norm": 24.374061584472656, "learning_rate": 1.833e-06, "loss": 2.9697, "step": 3666 }, { "epoch": 1.1495297805642632, "grad_norm": 23.166545867919922, "learning_rate": 1.8335000000000003e-06, "loss": 2.4767, "step": 3667 }, { "epoch": 1.1498432601880877, "grad_norm": 23.386566162109375, "learning_rate": 1.8340000000000001e-06, "loss": 2.6403, "step": 3668 }, { "epoch": 1.1501567398119121, "grad_norm": 46.93557357788086, "learning_rate": 1.8345000000000002e-06, "loss": 3.3012, "step": 3669 }, { "epoch": 1.1504702194357366, "grad_norm": 73.08238220214844, "learning_rate": 1.8350000000000002e-06, "loss": 3.0202, "step": 3670 }, { "epoch": 1.150783699059561, "grad_norm": 28.248533248901367, "learning_rate": 1.8355e-06, "loss": 3.7443, "step": 3671 }, { "epoch": 1.1510971786833855, "grad_norm": 27.702686309814453, "learning_rate": 1.8360000000000003e-06, "loss": 3.2577, "step": 3672 }, { "epoch": 1.1514106583072101, "grad_norm": 27.311302185058594, "learning_rate": 1.8365000000000003e-06, "loss": 2.8966, "step": 3673 }, { "epoch": 1.1517241379310346, "grad_norm": 51.159976959228516, "learning_rate": 1.8370000000000002e-06, "loss": 3.3169, "step": 3674 }, { "epoch": 1.152037617554859, "grad_norm": 32.074832916259766, "learning_rate": 1.8375000000000002e-06, "loss": 2.2569, "step": 3675 }, { "epoch": 1.1523510971786834, "grad_norm": 35.85694885253906, "learning_rate": 1.838e-06, "loss": 2.7525, "step": 3676 }, { "epoch": 1.152664576802508, "grad_norm": 24.524778366088867, "learning_rate": 1.8385000000000003e-06, "loss": 3.047, "step": 3677 }, { "epoch": 1.1529780564263323, "grad_norm": 38.84641647338867, "learning_rate": 1.8390000000000003e-06, "loss": 3.4556, "step": 3678 }, { "epoch": 1.1532915360501568, "grad_norm": 22.57017707824707, "learning_rate": 1.8395000000000001e-06, "loss": 2.5072, "step": 3679 }, { "epoch": 1.1536050156739812, "grad_norm": 62.06796646118164, "learning_rate": 1.8400000000000002e-06, "loss": 3.7502, "step": 3680 }, { "epoch": 1.1539184952978057, "grad_norm": 19.97160530090332, "learning_rate": 1.8405e-06, "loss": 2.7028, "step": 3681 }, { "epoch": 1.15423197492163, "grad_norm": 34.08852005004883, "learning_rate": 1.8410000000000002e-06, "loss": 2.8553, "step": 3682 }, { "epoch": 1.1545454545454545, "grad_norm": 24.2499942779541, "learning_rate": 1.8415000000000003e-06, "loss": 3.4699, "step": 3683 }, { "epoch": 1.154858934169279, "grad_norm": 26.6270694732666, "learning_rate": 1.8420000000000001e-06, "loss": 2.8316, "step": 3684 }, { "epoch": 1.1551724137931034, "grad_norm": 18.375274658203125, "learning_rate": 1.8425000000000001e-06, "loss": 2.711, "step": 3685 }, { "epoch": 1.1554858934169279, "grad_norm": 18.933826446533203, "learning_rate": 1.843e-06, "loss": 2.9843, "step": 3686 }, { "epoch": 1.1557993730407523, "grad_norm": 19.643722534179688, "learning_rate": 1.8435000000000002e-06, "loss": 2.8432, "step": 3687 }, { "epoch": 1.1561128526645768, "grad_norm": 103.77330017089844, "learning_rate": 1.8440000000000003e-06, "loss": 2.9164, "step": 3688 }, { "epoch": 1.1564263322884012, "grad_norm": 103.93336486816406, "learning_rate": 1.8445e-06, "loss": 2.6389, "step": 3689 }, { "epoch": 1.1567398119122256, "grad_norm": 16.40753746032715, "learning_rate": 1.8450000000000001e-06, "loss": 2.5037, "step": 3690 }, { "epoch": 1.15705329153605, "grad_norm": 31.478145599365234, "learning_rate": 1.8455000000000002e-06, "loss": 2.0934, "step": 3691 }, { "epoch": 1.1573667711598745, "grad_norm": 17.865964889526367, "learning_rate": 1.846e-06, "loss": 2.6648, "step": 3692 }, { "epoch": 1.1576802507836992, "grad_norm": 22.187891006469727, "learning_rate": 1.8465000000000002e-06, "loss": 2.4895, "step": 3693 }, { "epoch": 1.1579937304075236, "grad_norm": 22.099071502685547, "learning_rate": 1.8470000000000003e-06, "loss": 2.8543, "step": 3694 }, { "epoch": 1.158307210031348, "grad_norm": 67.45195007324219, "learning_rate": 1.8475e-06, "loss": 2.8602, "step": 3695 }, { "epoch": 1.1586206896551725, "grad_norm": 29.56635284423828, "learning_rate": 1.8480000000000001e-06, "loss": 2.5955, "step": 3696 }, { "epoch": 1.158934169278997, "grad_norm": 23.719898223876953, "learning_rate": 1.8485e-06, "loss": 2.8614, "step": 3697 }, { "epoch": 1.1592476489028214, "grad_norm": 27.395971298217773, "learning_rate": 1.8490000000000002e-06, "loss": 2.8527, "step": 3698 }, { "epoch": 1.1595611285266458, "grad_norm": 18.854639053344727, "learning_rate": 1.8495000000000002e-06, "loss": 2.5318, "step": 3699 }, { "epoch": 1.1598746081504703, "grad_norm": 17.002197265625, "learning_rate": 1.85e-06, "loss": 2.5792, "step": 3700 }, { "epoch": 1.1601880877742947, "grad_norm": 22.917043685913086, "learning_rate": 1.8505000000000001e-06, "loss": 2.4931, "step": 3701 }, { "epoch": 1.1605015673981192, "grad_norm": 45.23145294189453, "learning_rate": 1.851e-06, "loss": 2.9302, "step": 3702 }, { "epoch": 1.1608150470219436, "grad_norm": 34.68376922607422, "learning_rate": 1.8515000000000002e-06, "loss": 2.8426, "step": 3703 }, { "epoch": 1.161128526645768, "grad_norm": 30.177770614624023, "learning_rate": 1.8520000000000002e-06, "loss": 2.4237, "step": 3704 }, { "epoch": 1.1614420062695925, "grad_norm": 62.9578971862793, "learning_rate": 1.8525e-06, "loss": 4.0864, "step": 3705 }, { "epoch": 1.161755485893417, "grad_norm": 28.61890411376953, "learning_rate": 1.853e-06, "loss": 2.6907, "step": 3706 }, { "epoch": 1.1620689655172414, "grad_norm": 32.607295989990234, "learning_rate": 1.8535000000000001e-06, "loss": 2.6239, "step": 3707 }, { "epoch": 1.1623824451410658, "grad_norm": 32.68761444091797, "learning_rate": 1.8540000000000002e-06, "loss": 2.8229, "step": 3708 }, { "epoch": 1.1626959247648903, "grad_norm": 26.804277420043945, "learning_rate": 1.8545000000000002e-06, "loss": 2.6434, "step": 3709 }, { "epoch": 1.1630094043887147, "grad_norm": 30.005191802978516, "learning_rate": 1.8550000000000002e-06, "loss": 3.8045, "step": 3710 }, { "epoch": 1.1633228840125391, "grad_norm": 91.55377960205078, "learning_rate": 1.8555e-06, "loss": 2.93, "step": 3711 }, { "epoch": 1.1636363636363636, "grad_norm": 40.16615295410156, "learning_rate": 1.856e-06, "loss": 2.7624, "step": 3712 }, { "epoch": 1.163949843260188, "grad_norm": 29.836450576782227, "learning_rate": 1.8565000000000004e-06, "loss": 2.6629, "step": 3713 }, { "epoch": 1.1642633228840125, "grad_norm": 46.64818572998047, "learning_rate": 1.8570000000000002e-06, "loss": 3.0849, "step": 3714 }, { "epoch": 1.164576802507837, "grad_norm": 27.5681095123291, "learning_rate": 1.8575000000000002e-06, "loss": 2.522, "step": 3715 }, { "epoch": 1.1648902821316613, "grad_norm": 35.278045654296875, "learning_rate": 1.858e-06, "loss": 2.6238, "step": 3716 }, { "epoch": 1.1652037617554858, "grad_norm": 18.06938934326172, "learning_rate": 1.8585e-06, "loss": 2.4448, "step": 3717 }, { "epoch": 1.1655172413793102, "grad_norm": 25.87566566467285, "learning_rate": 1.8590000000000003e-06, "loss": 2.6019, "step": 3718 }, { "epoch": 1.1658307210031347, "grad_norm": 36.46488952636719, "learning_rate": 1.8595000000000002e-06, "loss": 3.7691, "step": 3719 }, { "epoch": 1.1661442006269593, "grad_norm": 31.678668975830078, "learning_rate": 1.8600000000000002e-06, "loss": 2.6131, "step": 3720 }, { "epoch": 1.1664576802507838, "grad_norm": 26.029878616333008, "learning_rate": 1.8605e-06, "loss": 3.1041, "step": 3721 }, { "epoch": 1.1667711598746082, "grad_norm": 31.820192337036133, "learning_rate": 1.861e-06, "loss": 2.592, "step": 3722 }, { "epoch": 1.1670846394984327, "grad_norm": 39.595542907714844, "learning_rate": 1.8615000000000003e-06, "loss": 2.8279, "step": 3723 }, { "epoch": 1.167398119122257, "grad_norm": 42.314151763916016, "learning_rate": 1.8620000000000001e-06, "loss": 3.0745, "step": 3724 }, { "epoch": 1.1677115987460815, "grad_norm": 38.720455169677734, "learning_rate": 1.8625000000000002e-06, "loss": 2.8476, "step": 3725 }, { "epoch": 1.168025078369906, "grad_norm": 65.40015411376953, "learning_rate": 1.863e-06, "loss": 4.3198, "step": 3726 }, { "epoch": 1.1683385579937304, "grad_norm": 139.53359985351562, "learning_rate": 1.8635e-06, "loss": 2.9609, "step": 3727 }, { "epoch": 1.1686520376175549, "grad_norm": 28.293638229370117, "learning_rate": 1.8640000000000003e-06, "loss": 2.9129, "step": 3728 }, { "epoch": 1.1689655172413793, "grad_norm": 27.988492965698242, "learning_rate": 1.8645e-06, "loss": 2.3799, "step": 3729 }, { "epoch": 1.1692789968652038, "grad_norm": 14.3427095413208, "learning_rate": 1.8650000000000001e-06, "loss": 2.2311, "step": 3730 }, { "epoch": 1.1695924764890282, "grad_norm": 40.683189392089844, "learning_rate": 1.8655000000000002e-06, "loss": 2.8328, "step": 3731 }, { "epoch": 1.1699059561128526, "grad_norm": 39.50870132446289, "learning_rate": 1.866e-06, "loss": 2.3849, "step": 3732 }, { "epoch": 1.170219435736677, "grad_norm": 39.7574462890625, "learning_rate": 1.8665000000000003e-06, "loss": 2.6584, "step": 3733 }, { "epoch": 1.1705329153605015, "grad_norm": 31.841501235961914, "learning_rate": 1.8670000000000003e-06, "loss": 3.718, "step": 3734 }, { "epoch": 1.170846394984326, "grad_norm": 57.89259338378906, "learning_rate": 1.8675000000000001e-06, "loss": 2.6181, "step": 3735 }, { "epoch": 1.1711598746081504, "grad_norm": 44.258182525634766, "learning_rate": 1.8680000000000002e-06, "loss": 2.5498, "step": 3736 }, { "epoch": 1.1714733542319749, "grad_norm": 50.148067474365234, "learning_rate": 1.8685e-06, "loss": 2.8004, "step": 3737 }, { "epoch": 1.1717868338557993, "grad_norm": 38.96084976196289, "learning_rate": 1.8690000000000002e-06, "loss": 3.2472, "step": 3738 }, { "epoch": 1.1721003134796237, "grad_norm": 13.656538963317871, "learning_rate": 1.8695000000000003e-06, "loss": 2.5985, "step": 3739 }, { "epoch": 1.1724137931034484, "grad_norm": 32.31047821044922, "learning_rate": 1.87e-06, "loss": 2.8887, "step": 3740 }, { "epoch": 1.1727272727272728, "grad_norm": 14.074935913085938, "learning_rate": 1.8705000000000001e-06, "loss": 2.8068, "step": 3741 }, { "epoch": 1.1730407523510973, "grad_norm": 23.464031219482422, "learning_rate": 1.871e-06, "loss": 3.1187, "step": 3742 }, { "epoch": 1.1733542319749217, "grad_norm": 26.261005401611328, "learning_rate": 1.8715000000000002e-06, "loss": 2.7901, "step": 3743 }, { "epoch": 1.1736677115987462, "grad_norm": 28.60563850402832, "learning_rate": 1.8720000000000002e-06, "loss": 3.1748, "step": 3744 }, { "epoch": 1.1739811912225706, "grad_norm": 38.07712936401367, "learning_rate": 1.8725e-06, "loss": 2.8271, "step": 3745 }, { "epoch": 1.174294670846395, "grad_norm": 114.45543670654297, "learning_rate": 1.8730000000000001e-06, "loss": 2.6053, "step": 3746 }, { "epoch": 1.1746081504702195, "grad_norm": 22.89457893371582, "learning_rate": 1.8735000000000001e-06, "loss": 2.4069, "step": 3747 }, { "epoch": 1.174921630094044, "grad_norm": 38.383277893066406, "learning_rate": 1.8740000000000002e-06, "loss": 3.9307, "step": 3748 }, { "epoch": 1.1752351097178684, "grad_norm": 42.46405029296875, "learning_rate": 1.8745000000000002e-06, "loss": 3.3047, "step": 3749 }, { "epoch": 1.1755485893416928, "grad_norm": 17.158899307250977, "learning_rate": 1.8750000000000003e-06, "loss": 2.7314, "step": 3750 }, { "epoch": 1.1758620689655173, "grad_norm": 28.514450073242188, "learning_rate": 1.8755e-06, "loss": 3.8771, "step": 3751 }, { "epoch": 1.1761755485893417, "grad_norm": 33.93141174316406, "learning_rate": 1.8760000000000001e-06, "loss": 2.4683, "step": 3752 }, { "epoch": 1.1764890282131661, "grad_norm": 17.947032928466797, "learning_rate": 1.8765000000000002e-06, "loss": 2.5114, "step": 3753 }, { "epoch": 1.1768025078369906, "grad_norm": 39.59684371948242, "learning_rate": 1.8770000000000002e-06, "loss": 2.6302, "step": 3754 }, { "epoch": 1.177115987460815, "grad_norm": 41.95842361450195, "learning_rate": 1.8775000000000002e-06, "loss": 3.1639, "step": 3755 }, { "epoch": 1.1774294670846395, "grad_norm": 36.54000473022461, "learning_rate": 1.878e-06, "loss": 3.0288, "step": 3756 }, { "epoch": 1.177742946708464, "grad_norm": 33.35289764404297, "learning_rate": 1.8785e-06, "loss": 2.664, "step": 3757 }, { "epoch": 1.1780564263322884, "grad_norm": 24.627397537231445, "learning_rate": 1.8790000000000003e-06, "loss": 2.712, "step": 3758 }, { "epoch": 1.1783699059561128, "grad_norm": 28.08515167236328, "learning_rate": 1.8795000000000002e-06, "loss": 2.8725, "step": 3759 }, { "epoch": 1.1786833855799372, "grad_norm": 61.68031311035156, "learning_rate": 1.8800000000000002e-06, "loss": 2.816, "step": 3760 }, { "epoch": 1.1789968652037617, "grad_norm": 87.2820053100586, "learning_rate": 1.8805e-06, "loss": 3.6346, "step": 3761 }, { "epoch": 1.1793103448275861, "grad_norm": 71.9816665649414, "learning_rate": 1.881e-06, "loss": 3.4425, "step": 3762 }, { "epoch": 1.1796238244514106, "grad_norm": 20.74867057800293, "learning_rate": 1.8815000000000003e-06, "loss": 2.4513, "step": 3763 }, { "epoch": 1.179937304075235, "grad_norm": 35.601165771484375, "learning_rate": 1.8820000000000001e-06, "loss": 3.3946, "step": 3764 }, { "epoch": 1.1802507836990594, "grad_norm": 55.310768127441406, "learning_rate": 1.8825000000000002e-06, "loss": 2.894, "step": 3765 }, { "epoch": 1.1805642633228839, "grad_norm": 41.345428466796875, "learning_rate": 1.883e-06, "loss": 2.7336, "step": 3766 }, { "epoch": 1.1808777429467086, "grad_norm": 23.19663429260254, "learning_rate": 1.8835e-06, "loss": 2.5383, "step": 3767 }, { "epoch": 1.181191222570533, "grad_norm": 35.46568298339844, "learning_rate": 1.8840000000000003e-06, "loss": 3.1715, "step": 3768 }, { "epoch": 1.1815047021943574, "grad_norm": 20.419200897216797, "learning_rate": 1.8845000000000001e-06, "loss": 2.8013, "step": 3769 }, { "epoch": 1.1818181818181819, "grad_norm": 19.463045120239258, "learning_rate": 1.8850000000000002e-06, "loss": 2.4863, "step": 3770 }, { "epoch": 1.1821316614420063, "grad_norm": 34.91453552246094, "learning_rate": 1.8855000000000002e-06, "loss": 2.868, "step": 3771 }, { "epoch": 1.1824451410658308, "grad_norm": 16.39441680908203, "learning_rate": 1.886e-06, "loss": 2.2358, "step": 3772 }, { "epoch": 1.1827586206896552, "grad_norm": 29.26861000061035, "learning_rate": 1.8865000000000003e-06, "loss": 2.2603, "step": 3773 }, { "epoch": 1.1830721003134796, "grad_norm": 38.201534271240234, "learning_rate": 1.8870000000000003e-06, "loss": 3.9458, "step": 3774 }, { "epoch": 1.183385579937304, "grad_norm": 37.52971649169922, "learning_rate": 1.8875000000000001e-06, "loss": 2.4286, "step": 3775 }, { "epoch": 1.1836990595611285, "grad_norm": 36.337181091308594, "learning_rate": 1.8880000000000002e-06, "loss": 3.3092, "step": 3776 }, { "epoch": 1.184012539184953, "grad_norm": 44.544219970703125, "learning_rate": 1.8885e-06, "loss": 2.5326, "step": 3777 }, { "epoch": 1.1843260188087774, "grad_norm": 60.01102066040039, "learning_rate": 1.8890000000000003e-06, "loss": 2.7836, "step": 3778 }, { "epoch": 1.1846394984326019, "grad_norm": 26.756061553955078, "learning_rate": 1.8895000000000003e-06, "loss": 2.7295, "step": 3779 }, { "epoch": 1.1849529780564263, "grad_norm": 51.751670837402344, "learning_rate": 1.8900000000000001e-06, "loss": 2.6031, "step": 3780 }, { "epoch": 1.1852664576802507, "grad_norm": 37.55747985839844, "learning_rate": 1.8905000000000002e-06, "loss": 3.4953, "step": 3781 }, { "epoch": 1.1855799373040752, "grad_norm": 17.792362213134766, "learning_rate": 1.891e-06, "loss": 2.438, "step": 3782 }, { "epoch": 1.1858934169278996, "grad_norm": 42.68111801147461, "learning_rate": 1.8915000000000002e-06, "loss": 2.534, "step": 3783 }, { "epoch": 1.186206896551724, "grad_norm": 20.284029006958008, "learning_rate": 1.8920000000000003e-06, "loss": 2.8496, "step": 3784 }, { "epoch": 1.1865203761755485, "grad_norm": 22.83557891845703, "learning_rate": 1.8925e-06, "loss": 2.8779, "step": 3785 }, { "epoch": 1.186833855799373, "grad_norm": 20.988258361816406, "learning_rate": 1.8930000000000001e-06, "loss": 3.2298, "step": 3786 }, { "epoch": 1.1871473354231976, "grad_norm": 52.38770294189453, "learning_rate": 1.8935e-06, "loss": 2.6163, "step": 3787 }, { "epoch": 1.187460815047022, "grad_norm": 65.69499969482422, "learning_rate": 1.8940000000000002e-06, "loss": 3.5213, "step": 3788 }, { "epoch": 1.1877742946708465, "grad_norm": 30.15526008605957, "learning_rate": 1.8945000000000002e-06, "loss": 2.7002, "step": 3789 }, { "epoch": 1.188087774294671, "grad_norm": 52.1644401550293, "learning_rate": 1.895e-06, "loss": 3.1846, "step": 3790 }, { "epoch": 1.1884012539184954, "grad_norm": 24.36962890625, "learning_rate": 1.8955e-06, "loss": 2.4365, "step": 3791 }, { "epoch": 1.1887147335423198, "grad_norm": 19.295175552368164, "learning_rate": 1.8960000000000001e-06, "loss": 2.8495, "step": 3792 }, { "epoch": 1.1890282131661443, "grad_norm": 26.62535285949707, "learning_rate": 1.8965000000000002e-06, "loss": 2.7075, "step": 3793 }, { "epoch": 1.1893416927899687, "grad_norm": 40.80854034423828, "learning_rate": 1.8970000000000002e-06, "loss": 2.7312, "step": 3794 }, { "epoch": 1.1896551724137931, "grad_norm": 57.89705276489258, "learning_rate": 1.8975000000000003e-06, "loss": 2.5595, "step": 3795 }, { "epoch": 1.1899686520376176, "grad_norm": 26.473291397094727, "learning_rate": 1.898e-06, "loss": 2.5633, "step": 3796 }, { "epoch": 1.190282131661442, "grad_norm": 44.334529876708984, "learning_rate": 1.8985000000000001e-06, "loss": 2.4269, "step": 3797 }, { "epoch": 1.1905956112852665, "grad_norm": 27.930118560791016, "learning_rate": 1.8990000000000004e-06, "loss": 2.9725, "step": 3798 }, { "epoch": 1.190909090909091, "grad_norm": 45.654109954833984, "learning_rate": 1.8995000000000002e-06, "loss": 3.2554, "step": 3799 }, { "epoch": 1.1912225705329154, "grad_norm": 16.584321975708008, "learning_rate": 1.9000000000000002e-06, "loss": 2.5039, "step": 3800 }, { "epoch": 1.1915360501567398, "grad_norm": 30.959943771362305, "learning_rate": 1.9005e-06, "loss": 2.9912, "step": 3801 }, { "epoch": 1.1918495297805642, "grad_norm": 31.72456169128418, "learning_rate": 1.901e-06, "loss": 2.6606, "step": 3802 }, { "epoch": 1.1921630094043887, "grad_norm": 23.852758407592773, "learning_rate": 1.9015000000000003e-06, "loss": 2.865, "step": 3803 }, { "epoch": 1.1924764890282131, "grad_norm": 34.01231002807617, "learning_rate": 1.9020000000000002e-06, "loss": 2.3327, "step": 3804 }, { "epoch": 1.1927899686520376, "grad_norm": 57.21983337402344, "learning_rate": 1.9025000000000002e-06, "loss": 2.9505, "step": 3805 }, { "epoch": 1.193103448275862, "grad_norm": 129.81326293945312, "learning_rate": 1.903e-06, "loss": 2.698, "step": 3806 }, { "epoch": 1.1934169278996865, "grad_norm": 51.74315643310547, "learning_rate": 1.9035e-06, "loss": 3.5234, "step": 3807 }, { "epoch": 1.193730407523511, "grad_norm": 22.179800033569336, "learning_rate": 1.9040000000000003e-06, "loss": 2.6641, "step": 3808 }, { "epoch": 1.1940438871473353, "grad_norm": 15.864103317260742, "learning_rate": 1.9045000000000001e-06, "loss": 2.7566, "step": 3809 }, { "epoch": 1.1943573667711598, "grad_norm": 29.442760467529297, "learning_rate": 1.9050000000000002e-06, "loss": 3.1718, "step": 3810 }, { "epoch": 1.1946708463949842, "grad_norm": 67.95341491699219, "learning_rate": 1.9055000000000002e-06, "loss": 3.1144, "step": 3811 }, { "epoch": 1.1949843260188087, "grad_norm": 27.336172103881836, "learning_rate": 1.906e-06, "loss": 2.7236, "step": 3812 }, { "epoch": 1.195297805642633, "grad_norm": 76.92893981933594, "learning_rate": 1.9065e-06, "loss": 2.2788, "step": 3813 }, { "epoch": 1.1956112852664578, "grad_norm": 28.78301239013672, "learning_rate": 1.9070000000000003e-06, "loss": 2.6817, "step": 3814 }, { "epoch": 1.1959247648902822, "grad_norm": 57.42724609375, "learning_rate": 1.9075000000000004e-06, "loss": 3.3968, "step": 3815 }, { "epoch": 1.1962382445141067, "grad_norm": 25.262216567993164, "learning_rate": 1.908e-06, "loss": 2.6335, "step": 3816 }, { "epoch": 1.196551724137931, "grad_norm": 19.571935653686523, "learning_rate": 1.9085e-06, "loss": 2.5299, "step": 3817 }, { "epoch": 1.1968652037617555, "grad_norm": 23.045181274414062, "learning_rate": 1.909e-06, "loss": 2.6344, "step": 3818 }, { "epoch": 1.19717868338558, "grad_norm": 53.79544448852539, "learning_rate": 1.9095e-06, "loss": 3.178, "step": 3819 }, { "epoch": 1.1974921630094044, "grad_norm": 25.499238967895508, "learning_rate": 1.9100000000000003e-06, "loss": 3.3268, "step": 3820 }, { "epoch": 1.1978056426332289, "grad_norm": 44.6558837890625, "learning_rate": 1.9105e-06, "loss": 3.7756, "step": 3821 }, { "epoch": 1.1981191222570533, "grad_norm": 33.91236114501953, "learning_rate": 1.911e-06, "loss": 2.9224, "step": 3822 }, { "epoch": 1.1984326018808777, "grad_norm": 40.67926788330078, "learning_rate": 1.9115e-06, "loss": 2.646, "step": 3823 }, { "epoch": 1.1987460815047022, "grad_norm": 28.280128479003906, "learning_rate": 1.912e-06, "loss": 2.5554, "step": 3824 }, { "epoch": 1.1990595611285266, "grad_norm": 31.250951766967773, "learning_rate": 1.9125000000000003e-06, "loss": 2.6868, "step": 3825 }, { "epoch": 1.199373040752351, "grad_norm": 53.51694107055664, "learning_rate": 1.913e-06, "loss": 3.5194, "step": 3826 }, { "epoch": 1.1996865203761755, "grad_norm": 44.22115707397461, "learning_rate": 1.9135e-06, "loss": 3.2924, "step": 3827 }, { "epoch": 1.2, "grad_norm": 27.570470809936523, "learning_rate": 1.9140000000000002e-06, "loss": 2.7071, "step": 3828 }, { "epoch": 1.2003134796238244, "grad_norm": 29.043636322021484, "learning_rate": 1.9145e-06, "loss": 2.9722, "step": 3829 }, { "epoch": 1.2006269592476488, "grad_norm": 48.598636627197266, "learning_rate": 1.9150000000000003e-06, "loss": 2.6605, "step": 3830 }, { "epoch": 1.2009404388714733, "grad_norm": 25.80156898498535, "learning_rate": 1.9155e-06, "loss": 2.7321, "step": 3831 }, { "epoch": 1.2012539184952977, "grad_norm": 36.27217483520508, "learning_rate": 1.916e-06, "loss": 2.3933, "step": 3832 }, { "epoch": 1.2015673981191222, "grad_norm": 50.53173065185547, "learning_rate": 1.9165e-06, "loss": 2.8351, "step": 3833 }, { "epoch": 1.2018808777429468, "grad_norm": 33.122230529785156, "learning_rate": 1.9170000000000005e-06, "loss": 3.0608, "step": 3834 }, { "epoch": 1.2021943573667713, "grad_norm": 66.05673217773438, "learning_rate": 1.9175000000000003e-06, "loss": 3.3519, "step": 3835 }, { "epoch": 1.2025078369905957, "grad_norm": 27.911020278930664, "learning_rate": 1.918e-06, "loss": 2.6034, "step": 3836 }, { "epoch": 1.2028213166144202, "grad_norm": 43.20354080200195, "learning_rate": 1.9185e-06, "loss": 2.8307, "step": 3837 }, { "epoch": 1.2031347962382446, "grad_norm": 23.005508422851562, "learning_rate": 1.919e-06, "loss": 2.6978, "step": 3838 }, { "epoch": 1.203448275862069, "grad_norm": 67.49423217773438, "learning_rate": 1.9195000000000004e-06, "loss": 3.424, "step": 3839 }, { "epoch": 1.2037617554858935, "grad_norm": 30.004865646362305, "learning_rate": 1.9200000000000003e-06, "loss": 2.5142, "step": 3840 }, { "epoch": 1.204075235109718, "grad_norm": 33.494834899902344, "learning_rate": 1.9205e-06, "loss": 3.2469, "step": 3841 }, { "epoch": 1.2043887147335424, "grad_norm": 27.38806915283203, "learning_rate": 1.921e-06, "loss": 2.7793, "step": 3842 }, { "epoch": 1.2047021943573668, "grad_norm": 52.8104133605957, "learning_rate": 1.9215e-06, "loss": 2.6053, "step": 3843 }, { "epoch": 1.2050156739811912, "grad_norm": 18.836341857910156, "learning_rate": 1.9220000000000004e-06, "loss": 2.3581, "step": 3844 }, { "epoch": 1.2053291536050157, "grad_norm": 28.024110794067383, "learning_rate": 1.9225000000000002e-06, "loss": 2.6733, "step": 3845 }, { "epoch": 1.2056426332288401, "grad_norm": 13.282366752624512, "learning_rate": 1.923e-06, "loss": 2.4797, "step": 3846 }, { "epoch": 1.2059561128526646, "grad_norm": 16.323963165283203, "learning_rate": 1.9235e-06, "loss": 2.3768, "step": 3847 }, { "epoch": 1.206269592476489, "grad_norm": 27.969438552856445, "learning_rate": 1.924e-06, "loss": 2.2672, "step": 3848 }, { "epoch": 1.2065830721003135, "grad_norm": 27.688213348388672, "learning_rate": 1.9245000000000004e-06, "loss": 2.5915, "step": 3849 }, { "epoch": 1.206896551724138, "grad_norm": 26.980260848999023, "learning_rate": 1.925e-06, "loss": 2.6703, "step": 3850 }, { "epoch": 1.2072100313479623, "grad_norm": 25.653573989868164, "learning_rate": 1.9255e-06, "loss": 2.3697, "step": 3851 }, { "epoch": 1.2075235109717868, "grad_norm": 16.955244064331055, "learning_rate": 1.9260000000000003e-06, "loss": 2.5476, "step": 3852 }, { "epoch": 1.2078369905956112, "grad_norm": 51.82334518432617, "learning_rate": 1.9265e-06, "loss": 2.7162, "step": 3853 }, { "epoch": 1.2081504702194357, "grad_norm": 19.893192291259766, "learning_rate": 1.9270000000000004e-06, "loss": 2.3514, "step": 3854 }, { "epoch": 1.20846394984326, "grad_norm": 19.577396392822266, "learning_rate": 1.9275e-06, "loss": 2.6409, "step": 3855 }, { "epoch": 1.2087774294670846, "grad_norm": 15.481810569763184, "learning_rate": 1.928e-06, "loss": 2.5522, "step": 3856 }, { "epoch": 1.209090909090909, "grad_norm": 20.681058883666992, "learning_rate": 1.9285000000000003e-06, "loss": 2.7032, "step": 3857 }, { "epoch": 1.2094043887147334, "grad_norm": 36.33036804199219, "learning_rate": 1.929e-06, "loss": 3.2028, "step": 3858 }, { "epoch": 1.2097178683385579, "grad_norm": 20.78473472595215, "learning_rate": 1.9295000000000003e-06, "loss": 2.487, "step": 3859 }, { "epoch": 1.2100313479623825, "grad_norm": 89.45281982421875, "learning_rate": 1.93e-06, "loss": 5.2864, "step": 3860 }, { "epoch": 1.210344827586207, "grad_norm": 32.773216247558594, "learning_rate": 1.9305e-06, "loss": 2.6761, "step": 3861 }, { "epoch": 1.2106583072100314, "grad_norm": 17.46109962463379, "learning_rate": 1.9310000000000002e-06, "loss": 2.4264, "step": 3862 }, { "epoch": 1.2109717868338559, "grad_norm": 28.945659637451172, "learning_rate": 1.9315e-06, "loss": 2.73, "step": 3863 }, { "epoch": 1.2112852664576803, "grad_norm": 150.18414306640625, "learning_rate": 1.9320000000000003e-06, "loss": 2.8617, "step": 3864 }, { "epoch": 1.2115987460815048, "grad_norm": 27.77288055419922, "learning_rate": 1.9325e-06, "loss": 2.7205, "step": 3865 }, { "epoch": 1.2119122257053292, "grad_norm": 23.874618530273438, "learning_rate": 1.933e-06, "loss": 2.5498, "step": 3866 }, { "epoch": 1.2122257053291536, "grad_norm": 118.2887191772461, "learning_rate": 1.9335e-06, "loss": 2.8452, "step": 3867 }, { "epoch": 1.212539184952978, "grad_norm": 44.203243255615234, "learning_rate": 1.934e-06, "loss": 3.3568, "step": 3868 }, { "epoch": 1.2128526645768025, "grad_norm": 22.49334144592285, "learning_rate": 1.9345000000000003e-06, "loss": 3.0688, "step": 3869 }, { "epoch": 1.213166144200627, "grad_norm": 20.258432388305664, "learning_rate": 1.935e-06, "loss": 2.781, "step": 3870 }, { "epoch": 1.2134796238244514, "grad_norm": 26.8780517578125, "learning_rate": 1.9355000000000004e-06, "loss": 3.2032, "step": 3871 }, { "epoch": 1.2137931034482758, "grad_norm": 27.263999938964844, "learning_rate": 1.936e-06, "loss": 2.3119, "step": 3872 }, { "epoch": 1.2141065830721003, "grad_norm": 29.607118606567383, "learning_rate": 1.9365e-06, "loss": 2.9999, "step": 3873 }, { "epoch": 1.2144200626959247, "grad_norm": 36.53831481933594, "learning_rate": 1.9370000000000003e-06, "loss": 3.1496, "step": 3874 }, { "epoch": 1.2147335423197492, "grad_norm": 18.061384201049805, "learning_rate": 1.9375e-06, "loss": 2.596, "step": 3875 }, { "epoch": 1.2150470219435736, "grad_norm": 41.549800872802734, "learning_rate": 1.9380000000000003e-06, "loss": 3.3622, "step": 3876 }, { "epoch": 1.215360501567398, "grad_norm": 26.330778121948242, "learning_rate": 1.9385e-06, "loss": 2.5022, "step": 3877 }, { "epoch": 1.2156739811912225, "grad_norm": 41.321956634521484, "learning_rate": 1.939e-06, "loss": 3.0533, "step": 3878 }, { "epoch": 1.215987460815047, "grad_norm": 16.474641799926758, "learning_rate": 1.9395000000000002e-06, "loss": 2.7315, "step": 3879 }, { "epoch": 1.2163009404388714, "grad_norm": 46.212467193603516, "learning_rate": 1.94e-06, "loss": 2.9196, "step": 3880 }, { "epoch": 1.216614420062696, "grad_norm": 31.109722137451172, "learning_rate": 1.9405000000000003e-06, "loss": 2.6945, "step": 3881 }, { "epoch": 1.2169278996865205, "grad_norm": 18.186649322509766, "learning_rate": 1.941e-06, "loss": 2.4777, "step": 3882 }, { "epoch": 1.217241379310345, "grad_norm": 28.388580322265625, "learning_rate": 1.9415e-06, "loss": 2.8091, "step": 3883 }, { "epoch": 1.2175548589341694, "grad_norm": 16.219085693359375, "learning_rate": 1.942e-06, "loss": 2.3407, "step": 3884 }, { "epoch": 1.2178683385579938, "grad_norm": 25.237590789794922, "learning_rate": 1.9425e-06, "loss": 2.5794, "step": 3885 }, { "epoch": 1.2181818181818183, "grad_norm": 17.25516700744629, "learning_rate": 1.9430000000000003e-06, "loss": 2.3214, "step": 3886 }, { "epoch": 1.2184952978056427, "grad_norm": 20.798234939575195, "learning_rate": 1.9435e-06, "loss": 2.4373, "step": 3887 }, { "epoch": 1.2188087774294671, "grad_norm": 18.155517578125, "learning_rate": 1.944e-06, "loss": 2.3363, "step": 3888 }, { "epoch": 1.2191222570532916, "grad_norm": 27.072912216186523, "learning_rate": 1.9445e-06, "loss": 2.6745, "step": 3889 }, { "epoch": 1.219435736677116, "grad_norm": 34.35905075073242, "learning_rate": 1.945e-06, "loss": 2.7749, "step": 3890 }, { "epoch": 1.2197492163009405, "grad_norm": 19.70022964477539, "learning_rate": 1.9455000000000003e-06, "loss": 2.7745, "step": 3891 }, { "epoch": 1.220062695924765, "grad_norm": 13.153515815734863, "learning_rate": 1.946e-06, "loss": 2.3028, "step": 3892 }, { "epoch": 1.2203761755485893, "grad_norm": 42.72065734863281, "learning_rate": 1.9465e-06, "loss": 2.7218, "step": 3893 }, { "epoch": 1.2206896551724138, "grad_norm": 38.443115234375, "learning_rate": 1.947e-06, "loss": 2.8484, "step": 3894 }, { "epoch": 1.2210031347962382, "grad_norm": 20.11655044555664, "learning_rate": 1.9475000000000004e-06, "loss": 2.6273, "step": 3895 }, { "epoch": 1.2213166144200627, "grad_norm": 46.73358917236328, "learning_rate": 1.9480000000000002e-06, "loss": 2.4925, "step": 3896 }, { "epoch": 1.2216300940438871, "grad_norm": 82.11383056640625, "learning_rate": 1.9485e-06, "loss": 3.4626, "step": 3897 }, { "epoch": 1.2219435736677116, "grad_norm": 22.496259689331055, "learning_rate": 1.949e-06, "loss": 2.4596, "step": 3898 }, { "epoch": 1.222257053291536, "grad_norm": 21.611501693725586, "learning_rate": 1.9495e-06, "loss": 2.3298, "step": 3899 }, { "epoch": 1.2225705329153604, "grad_norm": 16.648698806762695, "learning_rate": 1.9500000000000004e-06, "loss": 2.5577, "step": 3900 }, { "epoch": 1.2228840125391849, "grad_norm": 17.495756149291992, "learning_rate": 1.9505000000000002e-06, "loss": 2.3889, "step": 3901 }, { "epoch": 1.2231974921630093, "grad_norm": 25.892221450805664, "learning_rate": 1.951e-06, "loss": 2.4928, "step": 3902 }, { "epoch": 1.2235109717868338, "grad_norm": 27.336132049560547, "learning_rate": 1.9515e-06, "loss": 2.8628, "step": 3903 }, { "epoch": 1.2238244514106582, "grad_norm": 21.546873092651367, "learning_rate": 1.952e-06, "loss": 2.6961, "step": 3904 }, { "epoch": 1.2241379310344827, "grad_norm": 24.31972312927246, "learning_rate": 1.9525000000000004e-06, "loss": 2.6013, "step": 3905 }, { "epoch": 1.224451410658307, "grad_norm": 39.16084671020508, "learning_rate": 1.953e-06, "loss": 2.5389, "step": 3906 }, { "epoch": 1.2247648902821318, "grad_norm": 218.3396453857422, "learning_rate": 1.9535e-06, "loss": 3.3436, "step": 3907 }, { "epoch": 1.2250783699059562, "grad_norm": 27.832767486572266, "learning_rate": 1.9540000000000003e-06, "loss": 2.7819, "step": 3908 }, { "epoch": 1.2253918495297806, "grad_norm": 13.001688003540039, "learning_rate": 1.9545e-06, "loss": 2.5069, "step": 3909 }, { "epoch": 1.225705329153605, "grad_norm": 14.492598533630371, "learning_rate": 1.9550000000000003e-06, "loss": 2.1693, "step": 3910 }, { "epoch": 1.2260188087774295, "grad_norm": 60.928680419921875, "learning_rate": 1.9555e-06, "loss": 2.4301, "step": 3911 }, { "epoch": 1.226332288401254, "grad_norm": 64.4627914428711, "learning_rate": 1.956e-06, "loss": 3.0376, "step": 3912 }, { "epoch": 1.2266457680250784, "grad_norm": 61.340904235839844, "learning_rate": 1.9565000000000002e-06, "loss": 2.5944, "step": 3913 }, { "epoch": 1.2269592476489029, "grad_norm": 33.53535842895508, "learning_rate": 1.9570000000000005e-06, "loss": 3.2278, "step": 3914 }, { "epoch": 1.2272727272727273, "grad_norm": 59.37514114379883, "learning_rate": 1.9575000000000003e-06, "loss": 2.9019, "step": 3915 }, { "epoch": 1.2275862068965517, "grad_norm": 18.784873962402344, "learning_rate": 1.958e-06, "loss": 2.5877, "step": 3916 }, { "epoch": 1.2278996865203762, "grad_norm": 30.17689323425293, "learning_rate": 1.9585e-06, "loss": 2.6973, "step": 3917 }, { "epoch": 1.2282131661442006, "grad_norm": 29.341289520263672, "learning_rate": 1.9590000000000002e-06, "loss": 2.8073, "step": 3918 }, { "epoch": 1.228526645768025, "grad_norm": 46.65104293823242, "learning_rate": 1.9595000000000005e-06, "loss": 2.9391, "step": 3919 }, { "epoch": 1.2288401253918495, "grad_norm": 31.895326614379883, "learning_rate": 1.9600000000000003e-06, "loss": 2.5958, "step": 3920 }, { "epoch": 1.229153605015674, "grad_norm": 48.45664596557617, "learning_rate": 1.9605e-06, "loss": 2.5412, "step": 3921 }, { "epoch": 1.2294670846394984, "grad_norm": 13.672981262207031, "learning_rate": 1.961e-06, "loss": 2.3096, "step": 3922 }, { "epoch": 1.2297805642633228, "grad_norm": 30.28753662109375, "learning_rate": 1.9615e-06, "loss": 2.4116, "step": 3923 }, { "epoch": 1.2300940438871473, "grad_norm": 39.5670280456543, "learning_rate": 1.9620000000000004e-06, "loss": 3.0972, "step": 3924 }, { "epoch": 1.2304075235109717, "grad_norm": 17.293338775634766, "learning_rate": 1.9625000000000003e-06, "loss": 2.3708, "step": 3925 }, { "epoch": 1.2307210031347962, "grad_norm": 25.217784881591797, "learning_rate": 1.963e-06, "loss": 2.3573, "step": 3926 }, { "epoch": 1.2310344827586206, "grad_norm": 29.8195743560791, "learning_rate": 1.9635e-06, "loss": 2.9449, "step": 3927 }, { "epoch": 1.2313479623824453, "grad_norm": 55.18169403076172, "learning_rate": 1.964e-06, "loss": 2.5012, "step": 3928 }, { "epoch": 1.2316614420062697, "grad_norm": 28.454805374145508, "learning_rate": 1.9645000000000004e-06, "loss": 2.8373, "step": 3929 }, { "epoch": 1.2319749216300941, "grad_norm": 24.07086944580078, "learning_rate": 1.9650000000000002e-06, "loss": 2.5167, "step": 3930 }, { "epoch": 1.2322884012539186, "grad_norm": 26.33777618408203, "learning_rate": 1.9655e-06, "loss": 3.0033, "step": 3931 }, { "epoch": 1.232601880877743, "grad_norm": 26.97733497619629, "learning_rate": 1.9660000000000003e-06, "loss": 2.8334, "step": 3932 }, { "epoch": 1.2329153605015675, "grad_norm": 38.315311431884766, "learning_rate": 1.9665e-06, "loss": 3.6005, "step": 3933 }, { "epoch": 1.233228840125392, "grad_norm": 37.27223587036133, "learning_rate": 1.9670000000000004e-06, "loss": 3.188, "step": 3934 }, { "epoch": 1.2335423197492164, "grad_norm": 45.02766036987305, "learning_rate": 1.9675000000000002e-06, "loss": 3.2214, "step": 3935 }, { "epoch": 1.2338557993730408, "grad_norm": 24.59618377685547, "learning_rate": 1.968e-06, "loss": 2.6329, "step": 3936 }, { "epoch": 1.2341692789968652, "grad_norm": 30.170249938964844, "learning_rate": 1.9685000000000003e-06, "loss": 2.9202, "step": 3937 }, { "epoch": 1.2344827586206897, "grad_norm": 20.308732986450195, "learning_rate": 1.969e-06, "loss": 2.4924, "step": 3938 }, { "epoch": 1.2347962382445141, "grad_norm": 35.08584213256836, "learning_rate": 1.9695e-06, "loss": 2.1855, "step": 3939 }, { "epoch": 1.2351097178683386, "grad_norm": 24.181821823120117, "learning_rate": 1.97e-06, "loss": 2.4697, "step": 3940 }, { "epoch": 1.235423197492163, "grad_norm": 16.770347595214844, "learning_rate": 1.9705e-06, "loss": 2.365, "step": 3941 }, { "epoch": 1.2357366771159874, "grad_norm": 23.461381912231445, "learning_rate": 1.9710000000000003e-06, "loss": 2.9199, "step": 3942 }, { "epoch": 1.236050156739812, "grad_norm": 20.003427505493164, "learning_rate": 1.9715e-06, "loss": 2.3127, "step": 3943 }, { "epoch": 1.2363636363636363, "grad_norm": 33.976295471191406, "learning_rate": 1.972e-06, "loss": 2.6393, "step": 3944 }, { "epoch": 1.2366771159874608, "grad_norm": 61.14518737792969, "learning_rate": 1.9725e-06, "loss": 2.6706, "step": 3945 }, { "epoch": 1.2369905956112852, "grad_norm": 56.58893585205078, "learning_rate": 1.973e-06, "loss": 2.5611, "step": 3946 }, { "epoch": 1.2373040752351097, "grad_norm": 51.902347564697266, "learning_rate": 1.9735000000000002e-06, "loss": 2.9941, "step": 3947 }, { "epoch": 1.237617554858934, "grad_norm": 21.652944564819336, "learning_rate": 1.974e-06, "loss": 2.7382, "step": 3948 }, { "epoch": 1.2379310344827585, "grad_norm": 44.5388298034668, "learning_rate": 1.9745e-06, "loss": 3.0546, "step": 3949 }, { "epoch": 1.238244514106583, "grad_norm": 24.932762145996094, "learning_rate": 1.975e-06, "loss": 2.3333, "step": 3950 }, { "epoch": 1.2385579937304074, "grad_norm": 27.233287811279297, "learning_rate": 1.9755000000000004e-06, "loss": 2.8792, "step": 3951 }, { "epoch": 1.2388714733542319, "grad_norm": 24.123865127563477, "learning_rate": 1.9760000000000002e-06, "loss": 2.5114, "step": 3952 }, { "epoch": 1.2391849529780563, "grad_norm": 25.245868682861328, "learning_rate": 1.9765e-06, "loss": 2.279, "step": 3953 }, { "epoch": 1.239498432601881, "grad_norm": 38.38374710083008, "learning_rate": 1.977e-06, "loss": 2.7466, "step": 3954 }, { "epoch": 1.2398119122257054, "grad_norm": 18.390085220336914, "learning_rate": 1.9775e-06, "loss": 2.734, "step": 3955 }, { "epoch": 1.2401253918495299, "grad_norm": 36.78880310058594, "learning_rate": 1.9780000000000004e-06, "loss": 2.7813, "step": 3956 }, { "epoch": 1.2404388714733543, "grad_norm": 66.53557586669922, "learning_rate": 1.9785e-06, "loss": 2.5124, "step": 3957 }, { "epoch": 1.2407523510971787, "grad_norm": 103.23516082763672, "learning_rate": 1.979e-06, "loss": 2.4264, "step": 3958 }, { "epoch": 1.2410658307210032, "grad_norm": 49.23095703125, "learning_rate": 1.9795e-06, "loss": 2.6846, "step": 3959 }, { "epoch": 1.2413793103448276, "grad_norm": 15.24638557434082, "learning_rate": 1.98e-06, "loss": 2.175, "step": 3960 }, { "epoch": 1.241692789968652, "grad_norm": 39.227378845214844, "learning_rate": 1.9805000000000004e-06, "loss": 3.1209, "step": 3961 }, { "epoch": 1.2420062695924765, "grad_norm": 32.122440338134766, "learning_rate": 1.981e-06, "loss": 2.5471, "step": 3962 }, { "epoch": 1.242319749216301, "grad_norm": 22.874025344848633, "learning_rate": 1.9815e-06, "loss": 3.025, "step": 3963 }, { "epoch": 1.2426332288401254, "grad_norm": 25.451332092285156, "learning_rate": 1.982e-06, "loss": 2.7846, "step": 3964 }, { "epoch": 1.2429467084639498, "grad_norm": 65.10897064208984, "learning_rate": 1.9825e-06, "loss": 2.6536, "step": 3965 }, { "epoch": 1.2432601880877743, "grad_norm": 41.65166091918945, "learning_rate": 1.9830000000000003e-06, "loss": 3.1561, "step": 3966 }, { "epoch": 1.2435736677115987, "grad_norm": 60.49795150756836, "learning_rate": 1.9835e-06, "loss": 3.6986, "step": 3967 }, { "epoch": 1.2438871473354232, "grad_norm": 21.87533187866211, "learning_rate": 1.984e-06, "loss": 2.6943, "step": 3968 }, { "epoch": 1.2442006269592476, "grad_norm": 14.572546005249023, "learning_rate": 1.9845000000000002e-06, "loss": 2.5391, "step": 3969 }, { "epoch": 1.244514106583072, "grad_norm": 24.31571388244629, "learning_rate": 1.985e-06, "loss": 2.9149, "step": 3970 }, { "epoch": 1.2448275862068965, "grad_norm": 24.90671730041504, "learning_rate": 1.9855000000000003e-06, "loss": 2.8762, "step": 3971 }, { "epoch": 1.245141065830721, "grad_norm": 24.37377166748047, "learning_rate": 1.986e-06, "loss": 2.6314, "step": 3972 }, { "epoch": 1.2454545454545454, "grad_norm": 19.115659713745117, "learning_rate": 1.9865e-06, "loss": 2.46, "step": 3973 }, { "epoch": 1.2457680250783698, "grad_norm": 21.906410217285156, "learning_rate": 1.987e-06, "loss": 2.7101, "step": 3974 }, { "epoch": 1.2460815047021945, "grad_norm": 30.39508819580078, "learning_rate": 1.9875000000000005e-06, "loss": 2.9468, "step": 3975 }, { "epoch": 1.246394984326019, "grad_norm": 40.24324417114258, "learning_rate": 1.9880000000000003e-06, "loss": 2.492, "step": 3976 }, { "epoch": 1.2467084639498434, "grad_norm": 18.688854217529297, "learning_rate": 1.9885e-06, "loss": 2.4316, "step": 3977 }, { "epoch": 1.2470219435736678, "grad_norm": 114.26399993896484, "learning_rate": 1.989e-06, "loss": 3.3366, "step": 3978 }, { "epoch": 1.2473354231974922, "grad_norm": 13.459341049194336, "learning_rate": 1.9895e-06, "loss": 2.2597, "step": 3979 }, { "epoch": 1.2476489028213167, "grad_norm": 24.026880264282227, "learning_rate": 1.9900000000000004e-06, "loss": 2.5099, "step": 3980 }, { "epoch": 1.2479623824451411, "grad_norm": 30.318620681762695, "learning_rate": 1.9905000000000003e-06, "loss": 2.4107, "step": 3981 }, { "epoch": 1.2482758620689656, "grad_norm": 14.148162841796875, "learning_rate": 1.991e-06, "loss": 2.4989, "step": 3982 }, { "epoch": 1.24858934169279, "grad_norm": 34.825653076171875, "learning_rate": 1.9915e-06, "loss": 3.2823, "step": 3983 }, { "epoch": 1.2489028213166145, "grad_norm": 23.375911712646484, "learning_rate": 1.992e-06, "loss": 2.8601, "step": 3984 }, { "epoch": 1.249216300940439, "grad_norm": 75.13655853271484, "learning_rate": 1.9925000000000004e-06, "loss": 2.4779, "step": 3985 }, { "epoch": 1.2495297805642633, "grad_norm": 12.734591484069824, "learning_rate": 1.9930000000000002e-06, "loss": 2.1784, "step": 3986 }, { "epoch": 1.2498432601880878, "grad_norm": 34.51366424560547, "learning_rate": 1.9935e-06, "loss": 2.566, "step": 3987 }, { "epoch": 1.2501567398119122, "grad_norm": 16.85667610168457, "learning_rate": 1.9940000000000003e-06, "loss": 2.4532, "step": 3988 }, { "epoch": 1.2504702194357367, "grad_norm": 66.03433227539062, "learning_rate": 1.9945e-06, "loss": 2.7693, "step": 3989 }, { "epoch": 1.250783699059561, "grad_norm": 49.91973114013672, "learning_rate": 1.9950000000000004e-06, "loss": 3.3225, "step": 3990 }, { "epoch": 1.250783699059561, "eval_loss": 3.8327951431274414, "eval_runtime": 21.0158, "eval_samples_per_second": 127.856, "eval_steps_per_second": 7.994, "step": 3990 }, { "epoch": 1.2510971786833855, "grad_norm": 28.950902938842773, "learning_rate": 1.9955e-06, "loss": 2.7085, "step": 3991 }, { "epoch": 1.25141065830721, "grad_norm": 48.79054641723633, "learning_rate": 1.996e-06, "loss": 2.7858, "step": 3992 }, { "epoch": 1.2517241379310344, "grad_norm": 25.618690490722656, "learning_rate": 1.9965000000000003e-06, "loss": 2.6399, "step": 3993 }, { "epoch": 1.2520376175548589, "grad_norm": 13.908455848693848, "learning_rate": 1.997e-06, "loss": 2.2832, "step": 3994 }, { "epoch": 1.2523510971786833, "grad_norm": 16.477970123291016, "learning_rate": 1.9975000000000004e-06, "loss": 2.6561, "step": 3995 }, { "epoch": 1.2526645768025078, "grad_norm": 39.627376556396484, "learning_rate": 1.998e-06, "loss": 2.8566, "step": 3996 }, { "epoch": 1.2529780564263322, "grad_norm": 15.530925750732422, "learning_rate": 1.9985e-06, "loss": 2.701, "step": 3997 }, { "epoch": 1.2532915360501566, "grad_norm": 71.57562255859375, "learning_rate": 1.9990000000000003e-06, "loss": 3.97, "step": 3998 }, { "epoch": 1.253605015673981, "grad_norm": 50.008113861083984, "learning_rate": 1.9995e-06, "loss": 2.8347, "step": 3999 }, { "epoch": 1.2539184952978055, "grad_norm": 58.43327331542969, "learning_rate": 2.0000000000000003e-06, "loss": 3.1059, "step": 4000 }, { "epoch": 1.25423197492163, "grad_norm": 27.29250144958496, "learning_rate": 2.0005e-06, "loss": 2.5445, "step": 4001 }, { "epoch": 1.2545454545454544, "grad_norm": 14.431353569030762, "learning_rate": 2.001e-06, "loss": 2.4125, "step": 4002 }, { "epoch": 1.254858934169279, "grad_norm": 34.21233367919922, "learning_rate": 2.0015000000000002e-06, "loss": 3.0746, "step": 4003 }, { "epoch": 1.2551724137931035, "grad_norm": 35.75080871582031, "learning_rate": 2.002e-06, "loss": 2.6628, "step": 4004 }, { "epoch": 1.255485893416928, "grad_norm": 39.31830978393555, "learning_rate": 2.0025000000000003e-06, "loss": 2.8678, "step": 4005 }, { "epoch": 1.2557993730407524, "grad_norm": 25.93036651611328, "learning_rate": 2.003e-06, "loss": 2.7643, "step": 4006 }, { "epoch": 1.2561128526645768, "grad_norm": 16.806570053100586, "learning_rate": 2.0035e-06, "loss": 2.3947, "step": 4007 }, { "epoch": 1.2564263322884013, "grad_norm": 101.331787109375, "learning_rate": 2.004e-06, "loss": 3.1663, "step": 4008 }, { "epoch": 1.2567398119122257, "grad_norm": 141.9791717529297, "learning_rate": 2.0045e-06, "loss": 3.1777, "step": 4009 }, { "epoch": 1.2570532915360502, "grad_norm": 18.34634017944336, "learning_rate": 2.0050000000000003e-06, "loss": 2.7777, "step": 4010 }, { "epoch": 1.2573667711598746, "grad_norm": 19.01302719116211, "learning_rate": 2.0055e-06, "loss": 2.5083, "step": 4011 }, { "epoch": 1.257680250783699, "grad_norm": 39.72523880004883, "learning_rate": 2.0060000000000004e-06, "loss": 2.4291, "step": 4012 }, { "epoch": 1.2579937304075235, "grad_norm": 24.59177017211914, "learning_rate": 2.0065e-06, "loss": 2.7645, "step": 4013 }, { "epoch": 1.258307210031348, "grad_norm": 28.052520751953125, "learning_rate": 2.007e-06, "loss": 2.4558, "step": 4014 }, { "epoch": 1.2586206896551724, "grad_norm": 18.298568725585938, "learning_rate": 2.0075000000000003e-06, "loss": 2.3395, "step": 4015 }, { "epoch": 1.2589341692789968, "grad_norm": 19.31409454345703, "learning_rate": 2.008e-06, "loss": 2.6477, "step": 4016 }, { "epoch": 1.2592476489028213, "grad_norm": 61.653419494628906, "learning_rate": 2.0085000000000003e-06, "loss": 3.8697, "step": 4017 }, { "epoch": 1.2595611285266457, "grad_norm": 33.65704345703125, "learning_rate": 2.009e-06, "loss": 2.7283, "step": 4018 }, { "epoch": 1.2598746081504701, "grad_norm": 49.89202117919922, "learning_rate": 2.0095e-06, "loss": 2.9251, "step": 4019 }, { "epoch": 1.2601880877742948, "grad_norm": 25.42795181274414, "learning_rate": 2.0100000000000002e-06, "loss": 2.4845, "step": 4020 }, { "epoch": 1.2605015673981192, "grad_norm": 19.237651824951172, "learning_rate": 2.0105e-06, "loss": 2.5299, "step": 4021 }, { "epoch": 1.2608150470219437, "grad_norm": 20.69957160949707, "learning_rate": 2.0110000000000003e-06, "loss": 2.7474, "step": 4022 }, { "epoch": 1.2611285266457681, "grad_norm": 47.299766540527344, "learning_rate": 2.0115e-06, "loss": 2.7008, "step": 4023 }, { "epoch": 1.2614420062695926, "grad_norm": 26.17997169494629, "learning_rate": 2.012e-06, "loss": 2.9662, "step": 4024 }, { "epoch": 1.261755485893417, "grad_norm": 14.28285026550293, "learning_rate": 2.0125000000000002e-06, "loss": 2.4881, "step": 4025 }, { "epoch": 1.2620689655172415, "grad_norm": 27.487199783325195, "learning_rate": 2.013e-06, "loss": 2.4464, "step": 4026 }, { "epoch": 1.262382445141066, "grad_norm": 20.70392608642578, "learning_rate": 2.0135000000000003e-06, "loss": 2.4101, "step": 4027 }, { "epoch": 1.2626959247648903, "grad_norm": 19.474803924560547, "learning_rate": 2.014e-06, "loss": 2.0501, "step": 4028 }, { "epoch": 1.2630094043887148, "grad_norm": 39.99258804321289, "learning_rate": 2.0145e-06, "loss": 2.5471, "step": 4029 }, { "epoch": 1.2633228840125392, "grad_norm": 26.0811767578125, "learning_rate": 2.015e-06, "loss": 2.3593, "step": 4030 }, { "epoch": 1.2636363636363637, "grad_norm": 19.96682357788086, "learning_rate": 2.0155000000000004e-06, "loss": 2.6898, "step": 4031 }, { "epoch": 1.263949843260188, "grad_norm": 16.560874938964844, "learning_rate": 2.0160000000000003e-06, "loss": 2.4135, "step": 4032 }, { "epoch": 1.2642633228840126, "grad_norm": 44.8605842590332, "learning_rate": 2.0165e-06, "loss": 3.0805, "step": 4033 }, { "epoch": 1.264576802507837, "grad_norm": 14.197938919067383, "learning_rate": 2.017e-06, "loss": 2.8498, "step": 4034 }, { "epoch": 1.2648902821316614, "grad_norm": 28.6807861328125, "learning_rate": 2.0175e-06, "loss": 2.584, "step": 4035 }, { "epoch": 1.2652037617554859, "grad_norm": 69.7178726196289, "learning_rate": 2.0180000000000004e-06, "loss": 2.8437, "step": 4036 }, { "epoch": 1.2655172413793103, "grad_norm": 19.459928512573242, "learning_rate": 2.0185000000000002e-06, "loss": 2.813, "step": 4037 }, { "epoch": 1.2658307210031348, "grad_norm": 12.748106956481934, "learning_rate": 2.019e-06, "loss": 2.5893, "step": 4038 }, { "epoch": 1.2661442006269592, "grad_norm": 58.045040130615234, "learning_rate": 2.0195e-06, "loss": 2.9038, "step": 4039 }, { "epoch": 1.2664576802507836, "grad_norm": 23.788843154907227, "learning_rate": 2.02e-06, "loss": 2.1679, "step": 4040 }, { "epoch": 1.266771159874608, "grad_norm": 29.40966033935547, "learning_rate": 2.0205000000000004e-06, "loss": 2.2847, "step": 4041 }, { "epoch": 1.2670846394984325, "grad_norm": 19.985910415649414, "learning_rate": 2.0210000000000002e-06, "loss": 2.9521, "step": 4042 }, { "epoch": 1.267398119122257, "grad_norm": 91.1566390991211, "learning_rate": 2.0215e-06, "loss": 3.4109, "step": 4043 }, { "epoch": 1.2677115987460814, "grad_norm": 34.98796463012695, "learning_rate": 2.022e-06, "loss": 2.7613, "step": 4044 }, { "epoch": 1.2680250783699059, "grad_norm": 24.939620971679688, "learning_rate": 2.0225e-06, "loss": 2.2957, "step": 4045 }, { "epoch": 1.2683385579937303, "grad_norm": 66.59795379638672, "learning_rate": 2.0230000000000004e-06, "loss": 2.8673, "step": 4046 }, { "epoch": 1.2686520376175547, "grad_norm": 14.904571533203125, "learning_rate": 2.0235e-06, "loss": 2.5483, "step": 4047 }, { "epoch": 1.2689655172413792, "grad_norm": 24.920495986938477, "learning_rate": 2.024e-06, "loss": 2.8943, "step": 4048 }, { "epoch": 1.2692789968652036, "grad_norm": 26.274133682250977, "learning_rate": 2.0245000000000003e-06, "loss": 2.5536, "step": 4049 }, { "epoch": 1.2695924764890283, "grad_norm": 30.83713722229004, "learning_rate": 2.025e-06, "loss": 2.3649, "step": 4050 }, { "epoch": 1.2699059561128527, "grad_norm": 25.9537296295166, "learning_rate": 2.0255000000000003e-06, "loss": 2.3121, "step": 4051 }, { "epoch": 1.2702194357366772, "grad_norm": 16.14348602294922, "learning_rate": 2.026e-06, "loss": 2.421, "step": 4052 }, { "epoch": 1.2705329153605016, "grad_norm": 54.47565841674805, "learning_rate": 2.0265e-06, "loss": 3.325, "step": 4053 }, { "epoch": 1.270846394984326, "grad_norm": 42.51778793334961, "learning_rate": 2.0270000000000002e-06, "loss": 3.1028, "step": 4054 }, { "epoch": 1.2711598746081505, "grad_norm": 28.14337921142578, "learning_rate": 2.0275000000000005e-06, "loss": 2.5186, "step": 4055 }, { "epoch": 1.271473354231975, "grad_norm": 14.081459999084473, "learning_rate": 2.0280000000000003e-06, "loss": 2.3879, "step": 4056 }, { "epoch": 1.2717868338557994, "grad_norm": 32.00901412963867, "learning_rate": 2.0285e-06, "loss": 2.5596, "step": 4057 }, { "epoch": 1.2721003134796238, "grad_norm": 15.976505279541016, "learning_rate": 2.029e-06, "loss": 2.8214, "step": 4058 }, { "epoch": 1.2724137931034483, "grad_norm": 16.56611442565918, "learning_rate": 2.0295000000000002e-06, "loss": 2.4008, "step": 4059 }, { "epoch": 1.2727272727272727, "grad_norm": 37.53776931762695, "learning_rate": 2.0300000000000005e-06, "loss": 2.5785, "step": 4060 }, { "epoch": 1.2730407523510971, "grad_norm": 27.125139236450195, "learning_rate": 2.0305000000000003e-06, "loss": 2.687, "step": 4061 }, { "epoch": 1.2733542319749216, "grad_norm": 21.245471954345703, "learning_rate": 2.031e-06, "loss": 2.6162, "step": 4062 }, { "epoch": 1.273667711598746, "grad_norm": 20.019229888916016, "learning_rate": 2.0315e-06, "loss": 2.2253, "step": 4063 }, { "epoch": 1.2739811912225705, "grad_norm": 14.50991153717041, "learning_rate": 2.032e-06, "loss": 2.3581, "step": 4064 }, { "epoch": 1.274294670846395, "grad_norm": 19.146757125854492, "learning_rate": 2.0325e-06, "loss": 3.0703, "step": 4065 }, { "epoch": 1.2746081504702194, "grad_norm": 55.10849380493164, "learning_rate": 2.0330000000000003e-06, "loss": 2.585, "step": 4066 }, { "epoch": 1.274921630094044, "grad_norm": 15.288493156433105, "learning_rate": 2.0335e-06, "loss": 2.4462, "step": 4067 }, { "epoch": 1.2752351097178685, "grad_norm": 27.375669479370117, "learning_rate": 2.0340000000000003e-06, "loss": 2.4656, "step": 4068 }, { "epoch": 1.275548589341693, "grad_norm": 21.695545196533203, "learning_rate": 2.0345e-06, "loss": 2.659, "step": 4069 }, { "epoch": 1.2758620689655173, "grad_norm": 16.515893936157227, "learning_rate": 2.035e-06, "loss": 2.5209, "step": 4070 }, { "epoch": 1.2761755485893418, "grad_norm": 22.51224708557129, "learning_rate": 2.0355000000000002e-06, "loss": 2.46, "step": 4071 }, { "epoch": 1.2764890282131662, "grad_norm": 19.671100616455078, "learning_rate": 2.036e-06, "loss": 2.6473, "step": 4072 }, { "epoch": 1.2768025078369907, "grad_norm": 21.77365493774414, "learning_rate": 2.0365000000000003e-06, "loss": 2.7338, "step": 4073 }, { "epoch": 1.2771159874608151, "grad_norm": 19.614160537719727, "learning_rate": 2.037e-06, "loss": 2.583, "step": 4074 }, { "epoch": 1.2774294670846396, "grad_norm": 28.54057502746582, "learning_rate": 2.0375e-06, "loss": 3.0976, "step": 4075 }, { "epoch": 1.277742946708464, "grad_norm": 16.560401916503906, "learning_rate": 2.0380000000000002e-06, "loss": 2.4488, "step": 4076 }, { "epoch": 1.2780564263322884, "grad_norm": 22.638416290283203, "learning_rate": 2.0385e-06, "loss": 2.3756, "step": 4077 }, { "epoch": 1.2783699059561129, "grad_norm": 29.103925704956055, "learning_rate": 2.0390000000000003e-06, "loss": 2.5986, "step": 4078 }, { "epoch": 1.2786833855799373, "grad_norm": 30.395748138427734, "learning_rate": 2.0395e-06, "loss": 3.4387, "step": 4079 }, { "epoch": 1.2789968652037618, "grad_norm": 33.306697845458984, "learning_rate": 2.04e-06, "loss": 2.4712, "step": 4080 }, { "epoch": 1.2793103448275862, "grad_norm": 29.530553817749023, "learning_rate": 2.0405e-06, "loss": 2.4654, "step": 4081 }, { "epoch": 1.2796238244514107, "grad_norm": 21.296916961669922, "learning_rate": 2.041e-06, "loss": 2.4743, "step": 4082 }, { "epoch": 1.279937304075235, "grad_norm": 25.111255645751953, "learning_rate": 2.0415000000000003e-06, "loss": 2.4727, "step": 4083 }, { "epoch": 1.2802507836990595, "grad_norm": 22.278154373168945, "learning_rate": 2.042e-06, "loss": 2.2619, "step": 4084 }, { "epoch": 1.280564263322884, "grad_norm": 33.86940383911133, "learning_rate": 2.0425e-06, "loss": 2.5844, "step": 4085 }, { "epoch": 1.2808777429467084, "grad_norm": 22.564645767211914, "learning_rate": 2.043e-06, "loss": 2.6423, "step": 4086 }, { "epoch": 1.2811912225705329, "grad_norm": 24.984695434570312, "learning_rate": 2.0435e-06, "loss": 2.3474, "step": 4087 }, { "epoch": 1.2815047021943573, "grad_norm": 73.06382751464844, "learning_rate": 2.0440000000000003e-06, "loss": 2.3279, "step": 4088 }, { "epoch": 1.2818181818181817, "grad_norm": 24.00354766845703, "learning_rate": 2.0445e-06, "loss": 2.4972, "step": 4089 }, { "epoch": 1.2821316614420062, "grad_norm": 28.285858154296875, "learning_rate": 2.045e-06, "loss": 2.4998, "step": 4090 }, { "epoch": 1.2824451410658306, "grad_norm": 52.50784683227539, "learning_rate": 2.0455e-06, "loss": 3.6619, "step": 4091 }, { "epoch": 1.282758620689655, "grad_norm": 26.97268295288086, "learning_rate": 2.0460000000000004e-06, "loss": 2.7269, "step": 4092 }, { "epoch": 1.2830721003134795, "grad_norm": 26.74242401123047, "learning_rate": 2.0465000000000002e-06, "loss": 2.3844, "step": 4093 }, { "epoch": 1.283385579937304, "grad_norm": 35.4241943359375, "learning_rate": 2.047e-06, "loss": 2.8735, "step": 4094 }, { "epoch": 1.2836990595611284, "grad_norm": 26.71773910522461, "learning_rate": 2.0475e-06, "loss": 2.6532, "step": 4095 }, { "epoch": 1.284012539184953, "grad_norm": 41.195072174072266, "learning_rate": 2.048e-06, "loss": 2.6725, "step": 4096 }, { "epoch": 1.2843260188087775, "grad_norm": 63.87744903564453, "learning_rate": 2.0485000000000004e-06, "loss": 3.0468, "step": 4097 }, { "epoch": 1.284639498432602, "grad_norm": 26.91925048828125, "learning_rate": 2.049e-06, "loss": 2.7028, "step": 4098 }, { "epoch": 1.2849529780564264, "grad_norm": 36.127159118652344, "learning_rate": 2.0495e-06, "loss": 2.801, "step": 4099 }, { "epoch": 1.2852664576802508, "grad_norm": 45.2885856628418, "learning_rate": 2.05e-06, "loss": 2.5757, "step": 4100 }, { "epoch": 1.2855799373040753, "grad_norm": 21.48421859741211, "learning_rate": 2.0505e-06, "loss": 2.5982, "step": 4101 }, { "epoch": 1.2858934169278997, "grad_norm": 21.25076675415039, "learning_rate": 2.0510000000000004e-06, "loss": 2.4623, "step": 4102 }, { "epoch": 1.2862068965517242, "grad_norm": 22.256900787353516, "learning_rate": 2.0515e-06, "loss": 2.4937, "step": 4103 }, { "epoch": 1.2865203761755486, "grad_norm": 35.315425872802734, "learning_rate": 2.052e-06, "loss": 2.5487, "step": 4104 }, { "epoch": 1.286833855799373, "grad_norm": 21.41267204284668, "learning_rate": 2.0525000000000003e-06, "loss": 2.6247, "step": 4105 }, { "epoch": 1.2871473354231975, "grad_norm": 23.571125030517578, "learning_rate": 2.053e-06, "loss": 2.4657, "step": 4106 }, { "epoch": 1.287460815047022, "grad_norm": 20.08584976196289, "learning_rate": 2.0535000000000003e-06, "loss": 2.6455, "step": 4107 }, { "epoch": 1.2877742946708464, "grad_norm": 26.432762145996094, "learning_rate": 2.054e-06, "loss": 2.3619, "step": 4108 }, { "epoch": 1.2880877742946708, "grad_norm": 28.0866641998291, "learning_rate": 2.0545e-06, "loss": 2.5186, "step": 4109 }, { "epoch": 1.2884012539184952, "grad_norm": 58.62767028808594, "learning_rate": 2.0550000000000002e-06, "loss": 2.4484, "step": 4110 }, { "epoch": 1.2887147335423197, "grad_norm": 19.969892501831055, "learning_rate": 2.0555000000000005e-06, "loss": 2.268, "step": 4111 }, { "epoch": 1.2890282131661441, "grad_norm": 20.90323257446289, "learning_rate": 2.0560000000000003e-06, "loss": 2.4765, "step": 4112 }, { "epoch": 1.2893416927899686, "grad_norm": 33.036624908447266, "learning_rate": 2.0565e-06, "loss": 2.888, "step": 4113 }, { "epoch": 1.2896551724137932, "grad_norm": 24.629642486572266, "learning_rate": 2.057e-06, "loss": 2.5645, "step": 4114 }, { "epoch": 1.2899686520376177, "grad_norm": 12.752878189086914, "learning_rate": 2.0575e-06, "loss": 2.5795, "step": 4115 }, { "epoch": 1.2902821316614421, "grad_norm": 33.927825927734375, "learning_rate": 2.0580000000000005e-06, "loss": 2.7409, "step": 4116 }, { "epoch": 1.2905956112852666, "grad_norm": 12.693510055541992, "learning_rate": 2.0585000000000003e-06, "loss": 2.3876, "step": 4117 }, { "epoch": 1.290909090909091, "grad_norm": 37.0070915222168, "learning_rate": 2.059e-06, "loss": 2.8024, "step": 4118 }, { "epoch": 1.2912225705329154, "grad_norm": 24.57074546813965, "learning_rate": 2.0595e-06, "loss": 2.6066, "step": 4119 }, { "epoch": 1.29153605015674, "grad_norm": 28.160541534423828, "learning_rate": 2.06e-06, "loss": 2.6939, "step": 4120 }, { "epoch": 1.2918495297805643, "grad_norm": 37.937007904052734, "learning_rate": 2.0605000000000004e-06, "loss": 2.9812, "step": 4121 }, { "epoch": 1.2921630094043888, "grad_norm": 16.546600341796875, "learning_rate": 2.0610000000000003e-06, "loss": 2.44, "step": 4122 }, { "epoch": 1.2924764890282132, "grad_norm": 71.4891128540039, "learning_rate": 2.0615e-06, "loss": 2.7788, "step": 4123 }, { "epoch": 1.2927899686520377, "grad_norm": 33.58329772949219, "learning_rate": 2.062e-06, "loss": 2.6808, "step": 4124 }, { "epoch": 1.293103448275862, "grad_norm": 46.48095703125, "learning_rate": 2.0625e-06, "loss": 2.614, "step": 4125 }, { "epoch": 1.2934169278996865, "grad_norm": 26.5600643157959, "learning_rate": 2.0630000000000004e-06, "loss": 2.6591, "step": 4126 }, { "epoch": 1.293730407523511, "grad_norm": 26.45882225036621, "learning_rate": 2.0635000000000002e-06, "loss": 2.9518, "step": 4127 }, { "epoch": 1.2940438871473354, "grad_norm": 34.29823684692383, "learning_rate": 2.064e-06, "loss": 2.7566, "step": 4128 }, { "epoch": 1.2943573667711599, "grad_norm": 20.29995346069336, "learning_rate": 2.0645000000000003e-06, "loss": 2.8827, "step": 4129 }, { "epoch": 1.2946708463949843, "grad_norm": 14.124770164489746, "learning_rate": 2.065e-06, "loss": 2.2836, "step": 4130 }, { "epoch": 1.2949843260188088, "grad_norm": 20.10064697265625, "learning_rate": 2.0655000000000004e-06, "loss": 3.0029, "step": 4131 }, { "epoch": 1.2952978056426332, "grad_norm": 41.575111389160156, "learning_rate": 2.066e-06, "loss": 2.2947, "step": 4132 }, { "epoch": 1.2956112852664576, "grad_norm": 15.761857032775879, "learning_rate": 2.0665e-06, "loss": 2.3627, "step": 4133 }, { "epoch": 1.295924764890282, "grad_norm": 15.230303764343262, "learning_rate": 2.0670000000000003e-06, "loss": 2.9609, "step": 4134 }, { "epoch": 1.2962382445141065, "grad_norm": 25.583053588867188, "learning_rate": 2.0675e-06, "loss": 2.3677, "step": 4135 }, { "epoch": 1.296551724137931, "grad_norm": 22.213247299194336, "learning_rate": 2.0680000000000004e-06, "loss": 2.5241, "step": 4136 }, { "epoch": 1.2968652037617554, "grad_norm": 25.54466438293457, "learning_rate": 2.0685e-06, "loss": 2.6116, "step": 4137 }, { "epoch": 1.2971786833855798, "grad_norm": 23.40324592590332, "learning_rate": 2.069e-06, "loss": 2.124, "step": 4138 }, { "epoch": 1.2974921630094043, "grad_norm": 41.573062896728516, "learning_rate": 2.0695000000000003e-06, "loss": 3.7863, "step": 4139 }, { "epoch": 1.2978056426332287, "grad_norm": 13.942071914672852, "learning_rate": 2.07e-06, "loss": 2.3136, "step": 4140 }, { "epoch": 1.2981191222570532, "grad_norm": 15.588929176330566, "learning_rate": 2.0705000000000003e-06, "loss": 2.3499, "step": 4141 }, { "epoch": 1.2984326018808776, "grad_norm": 117.79044342041016, "learning_rate": 2.071e-06, "loss": 2.8299, "step": 4142 }, { "epoch": 1.2987460815047023, "grad_norm": 33.719791412353516, "learning_rate": 2.0715e-06, "loss": 2.3125, "step": 4143 }, { "epoch": 1.2990595611285267, "grad_norm": 53.90446472167969, "learning_rate": 2.0720000000000002e-06, "loss": 2.7984, "step": 4144 }, { "epoch": 1.2993730407523512, "grad_norm": 61.47092819213867, "learning_rate": 2.0725e-06, "loss": 3.0994, "step": 4145 }, { "epoch": 1.2996865203761756, "grad_norm": 33.92884063720703, "learning_rate": 2.0730000000000003e-06, "loss": 3.1014, "step": 4146 }, { "epoch": 1.3, "grad_norm": 29.12774085998535, "learning_rate": 2.0735e-06, "loss": 2.4252, "step": 4147 }, { "epoch": 1.3003134796238245, "grad_norm": 20.700414657592773, "learning_rate": 2.0740000000000004e-06, "loss": 3.0268, "step": 4148 }, { "epoch": 1.300626959247649, "grad_norm": 48.68048095703125, "learning_rate": 2.0745000000000002e-06, "loss": 2.9331, "step": 4149 }, { "epoch": 1.3009404388714734, "grad_norm": 17.82050323486328, "learning_rate": 2.075e-06, "loss": 2.3952, "step": 4150 }, { "epoch": 1.3012539184952978, "grad_norm": 20.30829429626465, "learning_rate": 2.0755000000000003e-06, "loss": 2.4276, "step": 4151 }, { "epoch": 1.3015673981191223, "grad_norm": 28.34781265258789, "learning_rate": 2.076e-06, "loss": 2.5247, "step": 4152 }, { "epoch": 1.3018808777429467, "grad_norm": 15.627279281616211, "learning_rate": 2.0765000000000004e-06, "loss": 2.5008, "step": 4153 }, { "epoch": 1.3021943573667711, "grad_norm": 24.798925399780273, "learning_rate": 2.077e-06, "loss": 2.7212, "step": 4154 }, { "epoch": 1.3025078369905956, "grad_norm": 27.888092041015625, "learning_rate": 2.0775e-06, "loss": 2.6418, "step": 4155 }, { "epoch": 1.30282131661442, "grad_norm": 15.778170585632324, "learning_rate": 2.0780000000000003e-06, "loss": 2.2588, "step": 4156 }, { "epoch": 1.3031347962382445, "grad_norm": 21.949968338012695, "learning_rate": 2.0785e-06, "loss": 2.2574, "step": 4157 }, { "epoch": 1.303448275862069, "grad_norm": 25.079078674316406, "learning_rate": 2.0790000000000003e-06, "loss": 2.8032, "step": 4158 }, { "epoch": 1.3037617554858933, "grad_norm": 25.341110229492188, "learning_rate": 2.0795e-06, "loss": 2.5002, "step": 4159 }, { "epoch": 1.3040752351097178, "grad_norm": 44.27704620361328, "learning_rate": 2.08e-06, "loss": 3.2607, "step": 4160 }, { "epoch": 1.3043887147335425, "grad_norm": 47.27009582519531, "learning_rate": 2.0805000000000002e-06, "loss": 2.3683, "step": 4161 }, { "epoch": 1.304702194357367, "grad_norm": 29.553010940551758, "learning_rate": 2.081e-06, "loss": 2.5803, "step": 4162 }, { "epoch": 1.3050156739811913, "grad_norm": 25.054651260375977, "learning_rate": 2.0815000000000003e-06, "loss": 2.265, "step": 4163 }, { "epoch": 1.3053291536050158, "grad_norm": 83.10655212402344, "learning_rate": 2.082e-06, "loss": 2.4293, "step": 4164 }, { "epoch": 1.3056426332288402, "grad_norm": 20.534961700439453, "learning_rate": 2.0825e-06, "loss": 2.5412, "step": 4165 }, { "epoch": 1.3059561128526647, "grad_norm": 27.032581329345703, "learning_rate": 2.0830000000000002e-06, "loss": 2.8988, "step": 4166 }, { "epoch": 1.306269592476489, "grad_norm": 19.306673049926758, "learning_rate": 2.0835e-06, "loss": 2.5186, "step": 4167 }, { "epoch": 1.3065830721003135, "grad_norm": 34.469871520996094, "learning_rate": 2.0840000000000003e-06, "loss": 2.4477, "step": 4168 }, { "epoch": 1.306896551724138, "grad_norm": 23.936790466308594, "learning_rate": 2.0845e-06, "loss": 2.4572, "step": 4169 }, { "epoch": 1.3072100313479624, "grad_norm": 26.452882766723633, "learning_rate": 2.085e-06, "loss": 2.33, "step": 4170 }, { "epoch": 1.3075235109717869, "grad_norm": 36.12260437011719, "learning_rate": 2.0855e-06, "loss": 2.9178, "step": 4171 }, { "epoch": 1.3078369905956113, "grad_norm": 19.62932777404785, "learning_rate": 2.0860000000000004e-06, "loss": 2.5279, "step": 4172 }, { "epoch": 1.3081504702194358, "grad_norm": 21.609783172607422, "learning_rate": 2.0865000000000003e-06, "loss": 2.7524, "step": 4173 }, { "epoch": 1.3084639498432602, "grad_norm": 62.739112854003906, "learning_rate": 2.087e-06, "loss": 2.985, "step": 4174 }, { "epoch": 1.3087774294670846, "grad_norm": 34.94454574584961, "learning_rate": 2.0875e-06, "loss": 3.0326, "step": 4175 }, { "epoch": 1.309090909090909, "grad_norm": 54.120731353759766, "learning_rate": 2.088e-06, "loss": 3.2535, "step": 4176 }, { "epoch": 1.3094043887147335, "grad_norm": 35.5368537902832, "learning_rate": 2.0885000000000004e-06, "loss": 2.9125, "step": 4177 }, { "epoch": 1.309717868338558, "grad_norm": 38.99037170410156, "learning_rate": 2.0890000000000002e-06, "loss": 2.7989, "step": 4178 }, { "epoch": 1.3100313479623824, "grad_norm": 14.405024528503418, "learning_rate": 2.0895e-06, "loss": 2.4821, "step": 4179 }, { "epoch": 1.3103448275862069, "grad_norm": 25.834924697875977, "learning_rate": 2.09e-06, "loss": 2.5991, "step": 4180 }, { "epoch": 1.3106583072100313, "grad_norm": 23.313091278076172, "learning_rate": 2.0905e-06, "loss": 2.4887, "step": 4181 }, { "epoch": 1.3109717868338557, "grad_norm": 23.875391006469727, "learning_rate": 2.0910000000000004e-06, "loss": 2.8547, "step": 4182 }, { "epoch": 1.3112852664576802, "grad_norm": 22.093515396118164, "learning_rate": 2.0915000000000002e-06, "loss": 2.6691, "step": 4183 }, { "epoch": 1.3115987460815046, "grad_norm": 14.514626502990723, "learning_rate": 2.092e-06, "loss": 2.4244, "step": 4184 }, { "epoch": 1.311912225705329, "grad_norm": 50.77438735961914, "learning_rate": 2.0925000000000003e-06, "loss": 2.8881, "step": 4185 }, { "epoch": 1.3122257053291535, "grad_norm": 34.015419006347656, "learning_rate": 2.093e-06, "loss": 2.848, "step": 4186 }, { "epoch": 1.312539184952978, "grad_norm": 32.5399284362793, "learning_rate": 2.0935000000000004e-06, "loss": 2.7736, "step": 4187 }, { "epoch": 1.3128526645768024, "grad_norm": 22.101634979248047, "learning_rate": 2.094e-06, "loss": 2.662, "step": 4188 }, { "epoch": 1.3131661442006268, "grad_norm": 22.13960838317871, "learning_rate": 2.0945e-06, "loss": 2.4422, "step": 4189 }, { "epoch": 1.3134796238244515, "grad_norm": 34.24423599243164, "learning_rate": 2.0950000000000003e-06, "loss": 2.9224, "step": 4190 }, { "epoch": 1.313793103448276, "grad_norm": 26.192934036254883, "learning_rate": 2.0955e-06, "loss": 2.556, "step": 4191 }, { "epoch": 1.3141065830721004, "grad_norm": 22.4783992767334, "learning_rate": 2.0960000000000003e-06, "loss": 3.2724, "step": 4192 }, { "epoch": 1.3144200626959248, "grad_norm": 30.03993034362793, "learning_rate": 2.0965e-06, "loss": 2.2862, "step": 4193 }, { "epoch": 1.3147335423197493, "grad_norm": 44.098018646240234, "learning_rate": 2.097e-06, "loss": 2.8181, "step": 4194 }, { "epoch": 1.3150470219435737, "grad_norm": 34.14385223388672, "learning_rate": 2.0975000000000002e-06, "loss": 2.5418, "step": 4195 }, { "epoch": 1.3153605015673981, "grad_norm": 25.11463165283203, "learning_rate": 2.098e-06, "loss": 2.5286, "step": 4196 }, { "epoch": 1.3156739811912226, "grad_norm": 28.286510467529297, "learning_rate": 2.0985000000000003e-06, "loss": 2.2723, "step": 4197 }, { "epoch": 1.315987460815047, "grad_norm": 29.696142196655273, "learning_rate": 2.099e-06, "loss": 2.5886, "step": 4198 }, { "epoch": 1.3163009404388715, "grad_norm": 16.753244400024414, "learning_rate": 2.0995e-06, "loss": 2.6589, "step": 4199 }, { "epoch": 1.316614420062696, "grad_norm": 23.720779418945312, "learning_rate": 2.1000000000000002e-06, "loss": 2.5827, "step": 4200 }, { "epoch": 1.3169278996865204, "grad_norm": 15.066795349121094, "learning_rate": 2.1005e-06, "loss": 2.4665, "step": 4201 }, { "epoch": 1.3172413793103448, "grad_norm": 26.113821029663086, "learning_rate": 2.1010000000000003e-06, "loss": 3.0411, "step": 4202 }, { "epoch": 1.3175548589341692, "grad_norm": 43.554443359375, "learning_rate": 2.1015e-06, "loss": 2.5904, "step": 4203 }, { "epoch": 1.3178683385579937, "grad_norm": 20.175739288330078, "learning_rate": 2.102e-06, "loss": 2.5329, "step": 4204 }, { "epoch": 1.3181818181818181, "grad_norm": 15.298727035522461, "learning_rate": 2.1025e-06, "loss": 2.4321, "step": 4205 }, { "epoch": 1.3184952978056426, "grad_norm": 48.11604309082031, "learning_rate": 2.103e-06, "loss": 2.5175, "step": 4206 }, { "epoch": 1.318808777429467, "grad_norm": 73.75093841552734, "learning_rate": 2.1035000000000003e-06, "loss": 2.6158, "step": 4207 }, { "epoch": 1.3191222570532917, "grad_norm": 49.25511932373047, "learning_rate": 2.104e-06, "loss": 2.5209, "step": 4208 }, { "epoch": 1.3194357366771161, "grad_norm": 19.584362030029297, "learning_rate": 2.1045000000000003e-06, "loss": 2.7566, "step": 4209 }, { "epoch": 1.3197492163009406, "grad_norm": 31.413415908813477, "learning_rate": 2.105e-06, "loss": 2.6506, "step": 4210 }, { "epoch": 1.320062695924765, "grad_norm": 24.068866729736328, "learning_rate": 2.1055e-06, "loss": 2.6964, "step": 4211 }, { "epoch": 1.3203761755485894, "grad_norm": 18.091983795166016, "learning_rate": 2.1060000000000002e-06, "loss": 2.4337, "step": 4212 }, { "epoch": 1.3206896551724139, "grad_norm": 21.102827072143555, "learning_rate": 2.1065e-06, "loss": 2.5642, "step": 4213 }, { "epoch": 1.3210031347962383, "grad_norm": 36.744781494140625, "learning_rate": 2.1070000000000003e-06, "loss": 2.3418, "step": 4214 }, { "epoch": 1.3213166144200628, "grad_norm": 113.0008316040039, "learning_rate": 2.1075e-06, "loss": 3.1146, "step": 4215 }, { "epoch": 1.3216300940438872, "grad_norm": 28.350322723388672, "learning_rate": 2.108e-06, "loss": 2.5306, "step": 4216 }, { "epoch": 1.3219435736677116, "grad_norm": 15.95138931274414, "learning_rate": 2.1085000000000002e-06, "loss": 2.3577, "step": 4217 }, { "epoch": 1.322257053291536, "grad_norm": 54.72137451171875, "learning_rate": 2.109e-06, "loss": 2.5567, "step": 4218 }, { "epoch": 1.3225705329153605, "grad_norm": 14.051758766174316, "learning_rate": 2.1095000000000003e-06, "loss": 2.2394, "step": 4219 }, { "epoch": 1.322884012539185, "grad_norm": 27.43135643005371, "learning_rate": 2.11e-06, "loss": 2.2638, "step": 4220 }, { "epoch": 1.3231974921630094, "grad_norm": 19.857425689697266, "learning_rate": 2.1105e-06, "loss": 2.7347, "step": 4221 }, { "epoch": 1.3235109717868339, "grad_norm": 16.808263778686523, "learning_rate": 2.111e-06, "loss": 2.3323, "step": 4222 }, { "epoch": 1.3238244514106583, "grad_norm": 31.731477737426758, "learning_rate": 2.1115e-06, "loss": 2.594, "step": 4223 }, { "epoch": 1.3241379310344827, "grad_norm": 42.76530075073242, "learning_rate": 2.1120000000000003e-06, "loss": 2.4346, "step": 4224 }, { "epoch": 1.3244514106583072, "grad_norm": 13.989803314208984, "learning_rate": 2.1125e-06, "loss": 2.5349, "step": 4225 }, { "epoch": 1.3247648902821316, "grad_norm": 18.851530075073242, "learning_rate": 2.113e-06, "loss": 2.4989, "step": 4226 }, { "epoch": 1.325078369905956, "grad_norm": 57.27273178100586, "learning_rate": 2.1135e-06, "loss": 2.3327, "step": 4227 }, { "epoch": 1.3253918495297805, "grad_norm": 21.600278854370117, "learning_rate": 2.1140000000000004e-06, "loss": 2.6823, "step": 4228 }, { "epoch": 1.325705329153605, "grad_norm": 18.557979583740234, "learning_rate": 2.1145000000000003e-06, "loss": 2.8352, "step": 4229 }, { "epoch": 1.3260188087774294, "grad_norm": 10.887141227722168, "learning_rate": 2.115e-06, "loss": 2.4158, "step": 4230 }, { "epoch": 1.3263322884012538, "grad_norm": 29.551910400390625, "learning_rate": 2.1155e-06, "loss": 2.6161, "step": 4231 }, { "epoch": 1.3266457680250783, "grad_norm": 46.62107849121094, "learning_rate": 2.116e-06, "loss": 3.0583, "step": 4232 }, { "epoch": 1.3269592476489027, "grad_norm": 22.17835235595703, "learning_rate": 2.1165000000000004e-06, "loss": 2.5605, "step": 4233 }, { "epoch": 1.3272727272727272, "grad_norm": 15.31783390045166, "learning_rate": 2.1170000000000002e-06, "loss": 2.3408, "step": 4234 }, { "epoch": 1.3275862068965516, "grad_norm": 32.893497467041016, "learning_rate": 2.1175e-06, "loss": 2.7138, "step": 4235 }, { "epoch": 1.327899686520376, "grad_norm": 37.14299774169922, "learning_rate": 2.118e-06, "loss": 2.4061, "step": 4236 }, { "epoch": 1.3282131661442007, "grad_norm": 45.42509841918945, "learning_rate": 2.1185e-06, "loss": 3.1221, "step": 4237 }, { "epoch": 1.3285266457680251, "grad_norm": 34.94936752319336, "learning_rate": 2.1190000000000004e-06, "loss": 2.9343, "step": 4238 }, { "epoch": 1.3288401253918496, "grad_norm": 41.42539978027344, "learning_rate": 2.1195e-06, "loss": 2.8523, "step": 4239 }, { "epoch": 1.329153605015674, "grad_norm": 25.758621215820312, "learning_rate": 2.12e-06, "loss": 2.2589, "step": 4240 }, { "epoch": 1.3294670846394985, "grad_norm": 37.20207595825195, "learning_rate": 2.1205e-06, "loss": 2.4667, "step": 4241 }, { "epoch": 1.329780564263323, "grad_norm": 14.378911018371582, "learning_rate": 2.121e-06, "loss": 2.6144, "step": 4242 }, { "epoch": 1.3300940438871474, "grad_norm": 30.05830955505371, "learning_rate": 2.1215000000000004e-06, "loss": 2.7819, "step": 4243 }, { "epoch": 1.3304075235109718, "grad_norm": 24.931983947753906, "learning_rate": 2.122e-06, "loss": 3.0302, "step": 4244 }, { "epoch": 1.3307210031347962, "grad_norm": 34.86129379272461, "learning_rate": 2.1225e-06, "loss": 3.0939, "step": 4245 }, { "epoch": 1.3310344827586207, "grad_norm": 30.962717056274414, "learning_rate": 2.1230000000000003e-06, "loss": 2.6051, "step": 4246 }, { "epoch": 1.3313479623824451, "grad_norm": 23.41077423095703, "learning_rate": 2.1235e-06, "loss": 2.5413, "step": 4247 }, { "epoch": 1.3316614420062696, "grad_norm": 16.523805618286133, "learning_rate": 2.1240000000000003e-06, "loss": 2.4509, "step": 4248 }, { "epoch": 1.331974921630094, "grad_norm": 25.446996688842773, "learning_rate": 2.1245e-06, "loss": 2.3192, "step": 4249 }, { "epoch": 1.3322884012539185, "grad_norm": 50.319541931152344, "learning_rate": 2.125e-06, "loss": 2.6523, "step": 4250 }, { "epoch": 1.332601880877743, "grad_norm": 30.700286865234375, "learning_rate": 2.1255000000000002e-06, "loss": 2.7076, "step": 4251 }, { "epoch": 1.3329153605015673, "grad_norm": 25.385419845581055, "learning_rate": 2.1260000000000005e-06, "loss": 2.541, "step": 4252 }, { "epoch": 1.3332288401253918, "grad_norm": 21.421798706054688, "learning_rate": 2.1265000000000003e-06, "loss": 3.12, "step": 4253 }, { "epoch": 1.3335423197492162, "grad_norm": 33.206661224365234, "learning_rate": 2.127e-06, "loss": 2.5614, "step": 4254 }, { "epoch": 1.3338557993730409, "grad_norm": 31.233118057250977, "learning_rate": 2.1275e-06, "loss": 2.4797, "step": 4255 }, { "epoch": 1.3341692789968653, "grad_norm": 21.808687210083008, "learning_rate": 2.128e-06, "loss": 2.3788, "step": 4256 }, { "epoch": 1.3344827586206898, "grad_norm": 17.938234329223633, "learning_rate": 2.1285000000000005e-06, "loss": 2.4703, "step": 4257 }, { "epoch": 1.3347962382445142, "grad_norm": 43.056495666503906, "learning_rate": 2.1290000000000003e-06, "loss": 2.7721, "step": 4258 }, { "epoch": 1.3351097178683387, "grad_norm": 27.85002899169922, "learning_rate": 2.1295e-06, "loss": 2.2889, "step": 4259 }, { "epoch": 1.335423197492163, "grad_norm": 14.552286148071289, "learning_rate": 2.13e-06, "loss": 2.3216, "step": 4260 }, { "epoch": 1.3357366771159875, "grad_norm": 53.797794342041016, "learning_rate": 2.1305e-06, "loss": 2.9491, "step": 4261 }, { "epoch": 1.336050156739812, "grad_norm": 28.714412689208984, "learning_rate": 2.1310000000000004e-06, "loss": 2.4772, "step": 4262 }, { "epoch": 1.3363636363636364, "grad_norm": 21.757843017578125, "learning_rate": 2.1315000000000003e-06, "loss": 2.6578, "step": 4263 }, { "epoch": 1.3366771159874609, "grad_norm": 18.476587295532227, "learning_rate": 2.132e-06, "loss": 2.4202, "step": 4264 }, { "epoch": 1.3369905956112853, "grad_norm": 13.806644439697266, "learning_rate": 2.1325000000000003e-06, "loss": 2.2175, "step": 4265 }, { "epoch": 1.3373040752351097, "grad_norm": 18.443056106567383, "learning_rate": 2.133e-06, "loss": 2.6481, "step": 4266 }, { "epoch": 1.3376175548589342, "grad_norm": 50.59961700439453, "learning_rate": 2.1335000000000004e-06, "loss": 3.1419, "step": 4267 }, { "epoch": 1.3379310344827586, "grad_norm": 94.30250549316406, "learning_rate": 2.1340000000000002e-06, "loss": 3.4595, "step": 4268 }, { "epoch": 1.338244514106583, "grad_norm": 28.462663650512695, "learning_rate": 2.1345e-06, "loss": 2.2085, "step": 4269 }, { "epoch": 1.3385579937304075, "grad_norm": 43.41497802734375, "learning_rate": 2.1350000000000003e-06, "loss": 2.5531, "step": 4270 }, { "epoch": 1.338871473354232, "grad_norm": 14.940423965454102, "learning_rate": 2.1355e-06, "loss": 2.3722, "step": 4271 }, { "epoch": 1.3391849529780564, "grad_norm": 18.490428924560547, "learning_rate": 2.1360000000000004e-06, "loss": 2.3419, "step": 4272 }, { "epoch": 1.3394984326018808, "grad_norm": 30.41261100769043, "learning_rate": 2.1365000000000002e-06, "loss": 2.837, "step": 4273 }, { "epoch": 1.3398119122257053, "grad_norm": 31.712522506713867, "learning_rate": 2.137e-06, "loss": 3.1905, "step": 4274 }, { "epoch": 1.3401253918495297, "grad_norm": 16.07421875, "learning_rate": 2.1375000000000003e-06, "loss": 2.1409, "step": 4275 }, { "epoch": 1.3404388714733542, "grad_norm": 21.069381713867188, "learning_rate": 2.138e-06, "loss": 2.3961, "step": 4276 }, { "epoch": 1.3407523510971786, "grad_norm": 36.013614654541016, "learning_rate": 2.1385000000000004e-06, "loss": 2.3364, "step": 4277 }, { "epoch": 1.341065830721003, "grad_norm": 17.197298049926758, "learning_rate": 2.139e-06, "loss": 2.506, "step": 4278 }, { "epoch": 1.3413793103448275, "grad_norm": 17.188724517822266, "learning_rate": 2.1395e-06, "loss": 2.6044, "step": 4279 }, { "epoch": 1.341692789968652, "grad_norm": 25.85085105895996, "learning_rate": 2.1400000000000003e-06, "loss": 3.0611, "step": 4280 }, { "epoch": 1.3420062695924764, "grad_norm": 18.796472549438477, "learning_rate": 2.1405e-06, "loss": 2.241, "step": 4281 }, { "epoch": 1.3423197492163008, "grad_norm": 16.755332946777344, "learning_rate": 2.1410000000000003e-06, "loss": 2.2393, "step": 4282 }, { "epoch": 1.3426332288401253, "grad_norm": 19.352550506591797, "learning_rate": 2.1415e-06, "loss": 2.61, "step": 4283 }, { "epoch": 1.34294670846395, "grad_norm": 29.244054794311523, "learning_rate": 2.142e-06, "loss": 2.7959, "step": 4284 }, { "epoch": 1.3432601880877744, "grad_norm": 40.46537780761719, "learning_rate": 2.1425000000000002e-06, "loss": 3.0749, "step": 4285 }, { "epoch": 1.3435736677115988, "grad_norm": 17.840002059936523, "learning_rate": 2.143e-06, "loss": 2.266, "step": 4286 }, { "epoch": 1.3438871473354232, "grad_norm": 47.583038330078125, "learning_rate": 2.1435000000000003e-06, "loss": 2.668, "step": 4287 }, { "epoch": 1.3442006269592477, "grad_norm": 23.055110931396484, "learning_rate": 2.144e-06, "loss": 2.6395, "step": 4288 }, { "epoch": 1.3445141065830721, "grad_norm": 38.746795654296875, "learning_rate": 2.1445000000000004e-06, "loss": 3.8204, "step": 4289 }, { "epoch": 1.3448275862068966, "grad_norm": 19.897480010986328, "learning_rate": 2.1450000000000002e-06, "loss": 2.3241, "step": 4290 }, { "epoch": 1.345141065830721, "grad_norm": 257.05810546875, "learning_rate": 2.1455e-06, "loss": 3.1263, "step": 4291 }, { "epoch": 1.3454545454545455, "grad_norm": 61.294918060302734, "learning_rate": 2.1460000000000003e-06, "loss": 2.7713, "step": 4292 }, { "epoch": 1.34576802507837, "grad_norm": 21.1854305267334, "learning_rate": 2.1465e-06, "loss": 2.2955, "step": 4293 }, { "epoch": 1.3460815047021943, "grad_norm": 21.163742065429688, "learning_rate": 2.1470000000000004e-06, "loss": 2.4182, "step": 4294 }, { "epoch": 1.3463949843260188, "grad_norm": 18.480819702148438, "learning_rate": 2.1475e-06, "loss": 2.6374, "step": 4295 }, { "epoch": 1.3467084639498432, "grad_norm": 13.95515251159668, "learning_rate": 2.148e-06, "loss": 2.2971, "step": 4296 }, { "epoch": 1.3470219435736677, "grad_norm": 40.44957733154297, "learning_rate": 2.1485000000000003e-06, "loss": 2.5437, "step": 4297 }, { "epoch": 1.347335423197492, "grad_norm": 63.295997619628906, "learning_rate": 2.149e-06, "loss": 2.7892, "step": 4298 }, { "epoch": 1.3476489028213166, "grad_norm": 33.74834060668945, "learning_rate": 2.1495000000000003e-06, "loss": 2.737, "step": 4299 }, { "epoch": 1.347962382445141, "grad_norm": 28.50554656982422, "learning_rate": 2.15e-06, "loss": 2.3759, "step": 4300 }, { "epoch": 1.3482758620689654, "grad_norm": 22.857704162597656, "learning_rate": 2.1505e-06, "loss": 2.9179, "step": 4301 }, { "epoch": 1.34858934169279, "grad_norm": 15.061685562133789, "learning_rate": 2.1510000000000002e-06, "loss": 2.2522, "step": 4302 }, { "epoch": 1.3489028213166145, "grad_norm": 19.26007080078125, "learning_rate": 2.1515e-06, "loss": 2.4863, "step": 4303 }, { "epoch": 1.349216300940439, "grad_norm": 62.05062484741211, "learning_rate": 2.1520000000000003e-06, "loss": 3.1491, "step": 4304 }, { "epoch": 1.3495297805642634, "grad_norm": 21.983184814453125, "learning_rate": 2.1525e-06, "loss": 2.4744, "step": 4305 }, { "epoch": 1.3498432601880879, "grad_norm": 60.55722427368164, "learning_rate": 2.153e-06, "loss": 2.5645, "step": 4306 }, { "epoch": 1.3501567398119123, "grad_norm": 14.09309196472168, "learning_rate": 2.1535000000000002e-06, "loss": 2.71, "step": 4307 }, { "epoch": 1.3504702194357368, "grad_norm": 43.35044860839844, "learning_rate": 2.1540000000000005e-06, "loss": 2.8857, "step": 4308 }, { "epoch": 1.3507836990595612, "grad_norm": 23.708900451660156, "learning_rate": 2.1545000000000003e-06, "loss": 2.5606, "step": 4309 }, { "epoch": 1.3510971786833856, "grad_norm": 20.863204956054688, "learning_rate": 2.155e-06, "loss": 2.6164, "step": 4310 }, { "epoch": 1.35141065830721, "grad_norm": 16.67975616455078, "learning_rate": 2.1555e-06, "loss": 2.3722, "step": 4311 }, { "epoch": 1.3517241379310345, "grad_norm": 16.18927001953125, "learning_rate": 2.156e-06, "loss": 2.2388, "step": 4312 }, { "epoch": 1.352037617554859, "grad_norm": 21.1960506439209, "learning_rate": 2.1565000000000004e-06, "loss": 2.4443, "step": 4313 }, { "epoch": 1.3523510971786834, "grad_norm": 50.9581184387207, "learning_rate": 2.1570000000000003e-06, "loss": 2.742, "step": 4314 }, { "epoch": 1.3526645768025078, "grad_norm": 29.12017250061035, "learning_rate": 2.1575e-06, "loss": 2.7771, "step": 4315 }, { "epoch": 1.3529780564263323, "grad_norm": 16.846023559570312, "learning_rate": 2.158e-06, "loss": 2.5052, "step": 4316 }, { "epoch": 1.3532915360501567, "grad_norm": 50.6660270690918, "learning_rate": 2.1585e-06, "loss": 2.8763, "step": 4317 }, { "epoch": 1.3536050156739812, "grad_norm": 15.816275596618652, "learning_rate": 2.1590000000000004e-06, "loss": 2.5491, "step": 4318 }, { "epoch": 1.3539184952978056, "grad_norm": 52.377525329589844, "learning_rate": 2.1595000000000002e-06, "loss": 2.6742, "step": 4319 }, { "epoch": 1.35423197492163, "grad_norm": 370.2469482421875, "learning_rate": 2.16e-06, "loss": 2.6247, "step": 4320 }, { "epoch": 1.3545454545454545, "grad_norm": 46.37929153442383, "learning_rate": 2.1605e-06, "loss": 3.0429, "step": 4321 }, { "epoch": 1.354858934169279, "grad_norm": 44.83418655395508, "learning_rate": 2.161e-06, "loss": 2.7189, "step": 4322 }, { "epoch": 1.3551724137931034, "grad_norm": 12.778990745544434, "learning_rate": 2.1615000000000004e-06, "loss": 2.3526, "step": 4323 }, { "epoch": 1.3554858934169278, "grad_norm": 21.996225357055664, "learning_rate": 2.1620000000000002e-06, "loss": 2.9304, "step": 4324 }, { "epoch": 1.3557993730407523, "grad_norm": 16.98282814025879, "learning_rate": 2.1625e-06, "loss": 2.3532, "step": 4325 }, { "epoch": 1.3561128526645767, "grad_norm": 21.168886184692383, "learning_rate": 2.1630000000000003e-06, "loss": 2.7953, "step": 4326 }, { "epoch": 1.3564263322884011, "grad_norm": 23.252321243286133, "learning_rate": 2.1635e-06, "loss": 2.3939, "step": 4327 }, { "epoch": 1.3567398119122256, "grad_norm": 24.73003578186035, "learning_rate": 2.1640000000000004e-06, "loss": 2.3674, "step": 4328 }, { "epoch": 1.35705329153605, "grad_norm": 40.02139663696289, "learning_rate": 2.1645e-06, "loss": 2.5032, "step": 4329 }, { "epoch": 1.3573667711598745, "grad_norm": 19.943395614624023, "learning_rate": 2.165e-06, "loss": 2.5056, "step": 4330 }, { "epoch": 1.3576802507836991, "grad_norm": 18.23369789123535, "learning_rate": 2.1655000000000003e-06, "loss": 2.3793, "step": 4331 }, { "epoch": 1.3579937304075236, "grad_norm": 222.7714385986328, "learning_rate": 2.166e-06, "loss": 2.5105, "step": 4332 }, { "epoch": 1.358307210031348, "grad_norm": 18.434926986694336, "learning_rate": 2.1665000000000003e-06, "loss": 2.3635, "step": 4333 }, { "epoch": 1.3586206896551725, "grad_norm": 48.5857048034668, "learning_rate": 2.167e-06, "loss": 2.5115, "step": 4334 }, { "epoch": 1.358934169278997, "grad_norm": 24.345428466796875, "learning_rate": 2.1675e-06, "loss": 3.1558, "step": 4335 }, { "epoch": 1.3592476489028213, "grad_norm": 24.408594131469727, "learning_rate": 2.1680000000000002e-06, "loss": 2.7029, "step": 4336 }, { "epoch": 1.3595611285266458, "grad_norm": 52.189109802246094, "learning_rate": 2.1685e-06, "loss": 2.9846, "step": 4337 }, { "epoch": 1.3598746081504702, "grad_norm": 58.49778366088867, "learning_rate": 2.1690000000000003e-06, "loss": 2.9688, "step": 4338 }, { "epoch": 1.3601880877742947, "grad_norm": 49.80043029785156, "learning_rate": 2.1695e-06, "loss": 2.7229, "step": 4339 }, { "epoch": 1.3605015673981191, "grad_norm": 29.659666061401367, "learning_rate": 2.17e-06, "loss": 3.1569, "step": 4340 }, { "epoch": 1.3608150470219436, "grad_norm": 16.015514373779297, "learning_rate": 2.1705000000000002e-06, "loss": 2.3724, "step": 4341 }, { "epoch": 1.361128526645768, "grad_norm": 26.102127075195312, "learning_rate": 2.171e-06, "loss": 2.4503, "step": 4342 }, { "epoch": 1.3614420062695924, "grad_norm": 17.50678062438965, "learning_rate": 2.1715000000000003e-06, "loss": 2.0142, "step": 4343 }, { "epoch": 1.3617554858934169, "grad_norm": 19.81712532043457, "learning_rate": 2.172e-06, "loss": 2.5277, "step": 4344 }, { "epoch": 1.3620689655172413, "grad_norm": 40.197269439697266, "learning_rate": 2.1725000000000004e-06, "loss": 2.5226, "step": 4345 }, { "epoch": 1.3623824451410658, "grad_norm": 83.39940643310547, "learning_rate": 2.173e-06, "loss": 3.6011, "step": 4346 }, { "epoch": 1.3626959247648902, "grad_norm": 27.205394744873047, "learning_rate": 2.1735e-06, "loss": 2.4416, "step": 4347 }, { "epoch": 1.3630094043887147, "grad_norm": 80.09081268310547, "learning_rate": 2.1740000000000003e-06, "loss": 2.7729, "step": 4348 }, { "epoch": 1.3633228840125393, "grad_norm": 11.565630912780762, "learning_rate": 2.1745e-06, "loss": 2.2525, "step": 4349 }, { "epoch": 1.3636363636363638, "grad_norm": 18.49755859375, "learning_rate": 2.1750000000000004e-06, "loss": 2.3291, "step": 4350 }, { "epoch": 1.3639498432601882, "grad_norm": 27.100292205810547, "learning_rate": 2.1755e-06, "loss": 2.4243, "step": 4351 }, { "epoch": 1.3642633228840126, "grad_norm": 18.38869285583496, "learning_rate": 2.176e-06, "loss": 2.7035, "step": 4352 }, { "epoch": 1.364576802507837, "grad_norm": 37.79602813720703, "learning_rate": 2.1765000000000003e-06, "loss": 2.428, "step": 4353 }, { "epoch": 1.3648902821316615, "grad_norm": 12.350529670715332, "learning_rate": 2.177e-06, "loss": 2.4816, "step": 4354 }, { "epoch": 1.365203761755486, "grad_norm": 33.481956481933594, "learning_rate": 2.1775000000000003e-06, "loss": 2.5288, "step": 4355 }, { "epoch": 1.3655172413793104, "grad_norm": 54.41046905517578, "learning_rate": 2.178e-06, "loss": 2.6805, "step": 4356 }, { "epoch": 1.3658307210031349, "grad_norm": 32.40873336791992, "learning_rate": 2.1785e-06, "loss": 2.4722, "step": 4357 }, { "epoch": 1.3661442006269593, "grad_norm": 13.064126014709473, "learning_rate": 2.1790000000000002e-06, "loss": 2.1733, "step": 4358 }, { "epoch": 1.3664576802507837, "grad_norm": 26.094358444213867, "learning_rate": 2.1795e-06, "loss": 2.9938, "step": 4359 }, { "epoch": 1.3667711598746082, "grad_norm": 15.258416175842285, "learning_rate": 2.1800000000000003e-06, "loss": 2.3782, "step": 4360 }, { "epoch": 1.3670846394984326, "grad_norm": 21.28451919555664, "learning_rate": 2.1805e-06, "loss": 2.8561, "step": 4361 }, { "epoch": 1.367398119122257, "grad_norm": 32.01693344116211, "learning_rate": 2.181e-06, "loss": 2.6646, "step": 4362 }, { "epoch": 1.3677115987460815, "grad_norm": 22.742752075195312, "learning_rate": 2.1815e-06, "loss": 2.3205, "step": 4363 }, { "epoch": 1.368025078369906, "grad_norm": 17.927749633789062, "learning_rate": 2.182e-06, "loss": 2.4711, "step": 4364 }, { "epoch": 1.3683385579937304, "grad_norm": 11.641703605651855, "learning_rate": 2.1825000000000003e-06, "loss": 2.9432, "step": 4365 }, { "epoch": 1.3686520376175548, "grad_norm": 19.79996109008789, "learning_rate": 2.183e-06, "loss": 2.511, "step": 4366 }, { "epoch": 1.3689655172413793, "grad_norm": 15.667484283447266, "learning_rate": 2.1835e-06, "loss": 2.2901, "step": 4367 }, { "epoch": 1.3692789968652037, "grad_norm": 20.27992820739746, "learning_rate": 2.184e-06, "loss": 2.6813, "step": 4368 }, { "epoch": 1.3695924764890282, "grad_norm": 13.770139694213867, "learning_rate": 2.1845000000000004e-06, "loss": 2.275, "step": 4369 }, { "epoch": 1.3699059561128526, "grad_norm": 50.38872146606445, "learning_rate": 2.1850000000000003e-06, "loss": 3.0291, "step": 4370 }, { "epoch": 1.370219435736677, "grad_norm": 23.927005767822266, "learning_rate": 2.1855e-06, "loss": 2.3998, "step": 4371 }, { "epoch": 1.3705329153605015, "grad_norm": 20.26904296875, "learning_rate": 2.186e-06, "loss": 2.5815, "step": 4372 }, { "epoch": 1.370846394984326, "grad_norm": 16.98995018005371, "learning_rate": 2.1865e-06, "loss": 2.415, "step": 4373 }, { "epoch": 1.3711598746081504, "grad_norm": 52.3552131652832, "learning_rate": 2.1870000000000004e-06, "loss": 2.8687, "step": 4374 }, { "epoch": 1.3714733542319748, "grad_norm": 40.32173538208008, "learning_rate": 2.1875000000000002e-06, "loss": 2.2187, "step": 4375 }, { "epoch": 1.3717868338557992, "grad_norm": 25.02348518371582, "learning_rate": 2.188e-06, "loss": 2.166, "step": 4376 }, { "epoch": 1.3721003134796237, "grad_norm": 25.96660614013672, "learning_rate": 2.1885e-06, "loss": 2.5841, "step": 4377 }, { "epoch": 1.3724137931034484, "grad_norm": 37.6526985168457, "learning_rate": 2.189e-06, "loss": 2.7103, "step": 4378 }, { "epoch": 1.3727272727272728, "grad_norm": 21.919849395751953, "learning_rate": 2.1895000000000004e-06, "loss": 2.9718, "step": 4379 }, { "epoch": 1.3730407523510972, "grad_norm": 29.23079490661621, "learning_rate": 2.19e-06, "loss": 2.2765, "step": 4380 }, { "epoch": 1.3733542319749217, "grad_norm": 37.13560104370117, "learning_rate": 2.1905e-06, "loss": 2.5976, "step": 4381 }, { "epoch": 1.3736677115987461, "grad_norm": 18.35036277770996, "learning_rate": 2.191e-06, "loss": 2.4564, "step": 4382 }, { "epoch": 1.3739811912225706, "grad_norm": 25.74932861328125, "learning_rate": 2.1915e-06, "loss": 2.43, "step": 4383 }, { "epoch": 1.374294670846395, "grad_norm": 22.727750778198242, "learning_rate": 2.1920000000000004e-06, "loss": 1.9894, "step": 4384 }, { "epoch": 1.3746081504702194, "grad_norm": 32.51811599731445, "learning_rate": 2.1925e-06, "loss": 2.3064, "step": 4385 }, { "epoch": 1.374921630094044, "grad_norm": 60.5367317199707, "learning_rate": 2.193e-06, "loss": 2.453, "step": 4386 }, { "epoch": 1.3752351097178683, "grad_norm": 16.200775146484375, "learning_rate": 2.1935000000000003e-06, "loss": 2.3594, "step": 4387 }, { "epoch": 1.3755485893416928, "grad_norm": 27.528079986572266, "learning_rate": 2.194e-06, "loss": 3.6015, "step": 4388 }, { "epoch": 1.3758620689655172, "grad_norm": 32.33339309692383, "learning_rate": 2.1945000000000003e-06, "loss": 2.6115, "step": 4389 }, { "epoch": 1.3761755485893417, "grad_norm": 52.41084289550781, "learning_rate": 2.195e-06, "loss": 3.0907, "step": 4390 }, { "epoch": 1.376489028213166, "grad_norm": 15.238964080810547, "learning_rate": 2.1955e-06, "loss": 2.2765, "step": 4391 }, { "epoch": 1.3768025078369905, "grad_norm": 42.972206115722656, "learning_rate": 2.1960000000000002e-06, "loss": 3.603, "step": 4392 }, { "epoch": 1.377115987460815, "grad_norm": 13.6290922164917, "learning_rate": 2.1965000000000005e-06, "loss": 2.302, "step": 4393 }, { "epoch": 1.3774294670846394, "grad_norm": 11.237236022949219, "learning_rate": 2.1970000000000003e-06, "loss": 2.3066, "step": 4394 }, { "epoch": 1.3777429467084639, "grad_norm": 18.652162551879883, "learning_rate": 2.1975e-06, "loss": 2.8458, "step": 4395 }, { "epoch": 1.3780564263322885, "grad_norm": 20.605178833007812, "learning_rate": 2.198e-06, "loss": 2.1595, "step": 4396 }, { "epoch": 1.378369905956113, "grad_norm": 24.695146560668945, "learning_rate": 2.1985e-06, "loss": 2.3621, "step": 4397 }, { "epoch": 1.3786833855799374, "grad_norm": 19.619272232055664, "learning_rate": 2.1990000000000005e-06, "loss": 2.431, "step": 4398 }, { "epoch": 1.3789968652037619, "grad_norm": 22.981626510620117, "learning_rate": 2.1995000000000003e-06, "loss": 2.3365, "step": 4399 }, { "epoch": 1.3793103448275863, "grad_norm": 19.587995529174805, "learning_rate": 2.2e-06, "loss": 2.4834, "step": 4400 }, { "epoch": 1.3796238244514107, "grad_norm": 25.917057037353516, "learning_rate": 2.2005e-06, "loss": 2.5633, "step": 4401 }, { "epoch": 1.3799373040752352, "grad_norm": 46.26815414428711, "learning_rate": 2.201e-06, "loss": 2.897, "step": 4402 }, { "epoch": 1.3802507836990596, "grad_norm": 91.74821472167969, "learning_rate": 2.2015000000000004e-06, "loss": 2.8106, "step": 4403 }, { "epoch": 1.380564263322884, "grad_norm": 26.963945388793945, "learning_rate": 2.2020000000000003e-06, "loss": 2.4099, "step": 4404 }, { "epoch": 1.3808777429467085, "grad_norm": 29.442838668823242, "learning_rate": 2.2025e-06, "loss": 2.529, "step": 4405 }, { "epoch": 1.381191222570533, "grad_norm": 23.21820831298828, "learning_rate": 2.2030000000000003e-06, "loss": 2.5125, "step": 4406 }, { "epoch": 1.3815047021943574, "grad_norm": 38.09263610839844, "learning_rate": 2.2035e-06, "loss": 2.4609, "step": 4407 }, { "epoch": 1.3818181818181818, "grad_norm": 17.69852638244629, "learning_rate": 2.2040000000000004e-06, "loss": 3.0537, "step": 4408 }, { "epoch": 1.3821316614420063, "grad_norm": 12.266457557678223, "learning_rate": 2.2045000000000002e-06, "loss": 2.3415, "step": 4409 }, { "epoch": 1.3824451410658307, "grad_norm": 19.07096290588379, "learning_rate": 2.205e-06, "loss": 2.213, "step": 4410 }, { "epoch": 1.3827586206896552, "grad_norm": 19.074338912963867, "learning_rate": 2.2055000000000003e-06, "loss": 2.6777, "step": 4411 }, { "epoch": 1.3830721003134796, "grad_norm": 36.32673263549805, "learning_rate": 2.206e-06, "loss": 2.8512, "step": 4412 }, { "epoch": 1.383385579937304, "grad_norm": 11.210381507873535, "learning_rate": 2.2065000000000004e-06, "loss": 2.3242, "step": 4413 }, { "epoch": 1.3836990595611285, "grad_norm": 21.469053268432617, "learning_rate": 2.2070000000000002e-06, "loss": 2.3212, "step": 4414 }, { "epoch": 1.384012539184953, "grad_norm": 31.558183670043945, "learning_rate": 2.2075e-06, "loss": 2.4386, "step": 4415 }, { "epoch": 1.3843260188087774, "grad_norm": 13.501238822937012, "learning_rate": 2.2080000000000003e-06, "loss": 2.2854, "step": 4416 }, { "epoch": 1.3846394984326018, "grad_norm": 22.445682525634766, "learning_rate": 2.2085e-06, "loss": 2.6778, "step": 4417 }, { "epoch": 1.3849529780564263, "grad_norm": 28.20890235900879, "learning_rate": 2.2090000000000004e-06, "loss": 2.5873, "step": 4418 }, { "epoch": 1.3852664576802507, "grad_norm": 19.3082275390625, "learning_rate": 2.2095e-06, "loss": 2.2456, "step": 4419 }, { "epoch": 1.3855799373040751, "grad_norm": 19.401933670043945, "learning_rate": 2.21e-06, "loss": 2.4951, "step": 4420 }, { "epoch": 1.3858934169278996, "grad_norm": 38.82876968383789, "learning_rate": 2.2105000000000003e-06, "loss": 2.6545, "step": 4421 }, { "epoch": 1.386206896551724, "grad_norm": 28.123075485229492, "learning_rate": 2.211e-06, "loss": 2.5914, "step": 4422 }, { "epoch": 1.3865203761755485, "grad_norm": 29.51925277709961, "learning_rate": 2.2115000000000003e-06, "loss": 2.3974, "step": 4423 }, { "epoch": 1.386833855799373, "grad_norm": 19.382112503051758, "learning_rate": 2.212e-06, "loss": 2.4665, "step": 4424 }, { "epoch": 1.3871473354231976, "grad_norm": 43.75849533081055, "learning_rate": 2.2125e-06, "loss": 2.5744, "step": 4425 }, { "epoch": 1.387460815047022, "grad_norm": 58.4948616027832, "learning_rate": 2.2130000000000002e-06, "loss": 2.1157, "step": 4426 }, { "epoch": 1.3877742946708465, "grad_norm": 46.00054168701172, "learning_rate": 2.2135e-06, "loss": 2.3246, "step": 4427 }, { "epoch": 1.388087774294671, "grad_norm": 26.322551727294922, "learning_rate": 2.2140000000000003e-06, "loss": 2.5217, "step": 4428 }, { "epoch": 1.3884012539184953, "grad_norm": 163.227783203125, "learning_rate": 2.2145e-06, "loss": 3.6124, "step": 4429 }, { "epoch": 1.3887147335423198, "grad_norm": 47.88997268676758, "learning_rate": 2.2150000000000004e-06, "loss": 2.552, "step": 4430 }, { "epoch": 1.3890282131661442, "grad_norm": 26.8756103515625, "learning_rate": 2.2155000000000002e-06, "loss": 2.6725, "step": 4431 }, { "epoch": 1.3893416927899687, "grad_norm": 17.241147994995117, "learning_rate": 2.216e-06, "loss": 2.2438, "step": 4432 }, { "epoch": 1.389655172413793, "grad_norm": 40.995697021484375, "learning_rate": 2.2165000000000003e-06, "loss": 2.6698, "step": 4433 }, { "epoch": 1.3899686520376175, "grad_norm": 17.262331008911133, "learning_rate": 2.217e-06, "loss": 2.5749, "step": 4434 }, { "epoch": 1.390282131661442, "grad_norm": 18.621685028076172, "learning_rate": 2.2175000000000004e-06, "loss": 2.4253, "step": 4435 }, { "epoch": 1.3905956112852664, "grad_norm": 30.165283203125, "learning_rate": 2.218e-06, "loss": 2.3773, "step": 4436 }, { "epoch": 1.3909090909090909, "grad_norm": 17.12919807434082, "learning_rate": 2.2185e-06, "loss": 2.3988, "step": 4437 }, { "epoch": 1.3912225705329153, "grad_norm": 63.46233367919922, "learning_rate": 2.219e-06, "loss": 2.7529, "step": 4438 }, { "epoch": 1.3915360501567398, "grad_norm": 30.933561325073242, "learning_rate": 2.2195e-06, "loss": 2.4569, "step": 4439 }, { "epoch": 1.3918495297805642, "grad_norm": 12.137598991394043, "learning_rate": 2.2200000000000003e-06, "loss": 2.2861, "step": 4440 }, { "epoch": 1.3921630094043886, "grad_norm": 38.616188049316406, "learning_rate": 2.2205e-06, "loss": 2.258, "step": 4441 }, { "epoch": 1.3924764890282133, "grad_norm": 23.51974105834961, "learning_rate": 2.221e-06, "loss": 2.4517, "step": 4442 }, { "epoch": 1.3927899686520377, "grad_norm": 15.76685619354248, "learning_rate": 2.2215000000000002e-06, "loss": 2.4993, "step": 4443 }, { "epoch": 1.3931034482758622, "grad_norm": 18.977575302124023, "learning_rate": 2.222e-06, "loss": 2.4115, "step": 4444 }, { "epoch": 1.3934169278996866, "grad_norm": 32.84709548950195, "learning_rate": 2.2225000000000003e-06, "loss": 2.6831, "step": 4445 }, { "epoch": 1.393730407523511, "grad_norm": 41.960670471191406, "learning_rate": 2.223e-06, "loss": 2.5226, "step": 4446 }, { "epoch": 1.3940438871473355, "grad_norm": 14.326679229736328, "learning_rate": 2.2235e-06, "loss": 2.8367, "step": 4447 }, { "epoch": 1.39435736677116, "grad_norm": 17.317045211791992, "learning_rate": 2.2240000000000002e-06, "loss": 2.5747, "step": 4448 }, { "epoch": 1.3946708463949844, "grad_norm": 27.15635108947754, "learning_rate": 2.2245000000000005e-06, "loss": 2.6648, "step": 4449 }, { "epoch": 1.3949843260188088, "grad_norm": 22.993974685668945, "learning_rate": 2.2250000000000003e-06, "loss": 2.3917, "step": 4450 }, { "epoch": 1.3952978056426333, "grad_norm": 15.782670021057129, "learning_rate": 2.2255e-06, "loss": 2.229, "step": 4451 }, { "epoch": 1.3956112852664577, "grad_norm": 20.578027725219727, "learning_rate": 2.226e-06, "loss": 2.5032, "step": 4452 }, { "epoch": 1.3959247648902822, "grad_norm": 26.298973083496094, "learning_rate": 2.2265e-06, "loss": 2.4797, "step": 4453 }, { "epoch": 1.3962382445141066, "grad_norm": 17.40005111694336, "learning_rate": 2.2270000000000004e-06, "loss": 2.354, "step": 4454 }, { "epoch": 1.396551724137931, "grad_norm": 22.421411514282227, "learning_rate": 2.2275000000000003e-06, "loss": 2.2626, "step": 4455 }, { "epoch": 1.3968652037617555, "grad_norm": 35.079627990722656, "learning_rate": 2.228e-06, "loss": 2.4148, "step": 4456 }, { "epoch": 1.39717868338558, "grad_norm": 22.324115753173828, "learning_rate": 2.2285e-06, "loss": 2.5046, "step": 4457 }, { "epoch": 1.3974921630094044, "grad_norm": 47.19550323486328, "learning_rate": 2.229e-06, "loss": 2.8975, "step": 4458 }, { "epoch": 1.3978056426332288, "grad_norm": 13.401718139648438, "learning_rate": 2.2295000000000004e-06, "loss": 2.3232, "step": 4459 }, { "epoch": 1.3981191222570533, "grad_norm": 32.29413986206055, "learning_rate": 2.2300000000000002e-06, "loss": 2.6065, "step": 4460 }, { "epoch": 1.3984326018808777, "grad_norm": 28.262371063232422, "learning_rate": 2.2305e-06, "loss": 2.1882, "step": 4461 }, { "epoch": 1.3987460815047021, "grad_norm": 17.66546058654785, "learning_rate": 2.231e-06, "loss": 2.3207, "step": 4462 }, { "epoch": 1.3990595611285266, "grad_norm": 32.400699615478516, "learning_rate": 2.2315e-06, "loss": 2.1721, "step": 4463 }, { "epoch": 1.399373040752351, "grad_norm": 18.58637237548828, "learning_rate": 2.2320000000000004e-06, "loss": 2.2821, "step": 4464 }, { "epoch": 1.3996865203761755, "grad_norm": 19.35753059387207, "learning_rate": 2.2325000000000002e-06, "loss": 3.3655, "step": 4465 }, { "epoch": 1.4, "grad_norm": 28.945085525512695, "learning_rate": 2.233e-06, "loss": 3.0592, "step": 4466 }, { "epoch": 1.4003134796238244, "grad_norm": 10.396805763244629, "learning_rate": 2.2335000000000003e-06, "loss": 2.4453, "step": 4467 }, { "epoch": 1.4006269592476488, "grad_norm": 49.911705017089844, "learning_rate": 2.234e-06, "loss": 2.4485, "step": 4468 }, { "epoch": 1.4009404388714732, "grad_norm": 79.0918960571289, "learning_rate": 2.2345000000000004e-06, "loss": 1.9819, "step": 4469 }, { "epoch": 1.4012539184952977, "grad_norm": 22.501100540161133, "learning_rate": 2.235e-06, "loss": 2.4589, "step": 4470 }, { "epoch": 1.4015673981191221, "grad_norm": 36.03977966308594, "learning_rate": 2.2355e-06, "loss": 2.4418, "step": 4471 }, { "epoch": 1.4018808777429468, "grad_norm": 16.972667694091797, "learning_rate": 2.2360000000000003e-06, "loss": 2.3223, "step": 4472 }, { "epoch": 1.4021943573667712, "grad_norm": 14.537110328674316, "learning_rate": 2.2365e-06, "loss": 2.3733, "step": 4473 }, { "epoch": 1.4025078369905957, "grad_norm": 53.65971755981445, "learning_rate": 2.2370000000000004e-06, "loss": 2.5804, "step": 4474 }, { "epoch": 1.40282131661442, "grad_norm": 19.41697883605957, "learning_rate": 2.2375e-06, "loss": 2.681, "step": 4475 }, { "epoch": 1.4031347962382446, "grad_norm": 15.1221923828125, "learning_rate": 2.238e-06, "loss": 2.1564, "step": 4476 }, { "epoch": 1.403448275862069, "grad_norm": 19.52741241455078, "learning_rate": 2.2385000000000003e-06, "loss": 2.1644, "step": 4477 }, { "epoch": 1.4037617554858934, "grad_norm": 48.018367767333984, "learning_rate": 2.239e-06, "loss": 2.6797, "step": 4478 }, { "epoch": 1.4040752351097179, "grad_norm": 41.445098876953125, "learning_rate": 2.2395000000000003e-06, "loss": 2.8707, "step": 4479 }, { "epoch": 1.4043887147335423, "grad_norm": 53.0460319519043, "learning_rate": 2.24e-06, "loss": 3.1522, "step": 4480 }, { "epoch": 1.4047021943573668, "grad_norm": 17.2395076751709, "learning_rate": 2.2405e-06, "loss": 2.2166, "step": 4481 }, { "epoch": 1.4050156739811912, "grad_norm": 22.412208557128906, "learning_rate": 2.2410000000000002e-06, "loss": 2.4682, "step": 4482 }, { "epoch": 1.4053291536050156, "grad_norm": 54.72515106201172, "learning_rate": 2.2415e-06, "loss": 2.521, "step": 4483 }, { "epoch": 1.40564263322884, "grad_norm": 26.304935455322266, "learning_rate": 2.2420000000000003e-06, "loss": 2.0913, "step": 4484 }, { "epoch": 1.4059561128526645, "grad_norm": 16.559471130371094, "learning_rate": 2.2425e-06, "loss": 2.1316, "step": 4485 }, { "epoch": 1.406269592476489, "grad_norm": 24.916353225708008, "learning_rate": 2.2430000000000004e-06, "loss": 2.594, "step": 4486 }, { "epoch": 1.4065830721003134, "grad_norm": 31.964988708496094, "learning_rate": 2.2435e-06, "loss": 2.5372, "step": 4487 }, { "epoch": 1.4068965517241379, "grad_norm": 60.68818283081055, "learning_rate": 2.244e-06, "loss": 2.9889, "step": 4488 }, { "epoch": 1.4072100313479625, "grad_norm": 16.230152130126953, "learning_rate": 2.2445000000000003e-06, "loss": 2.4079, "step": 4489 }, { "epoch": 1.407523510971787, "grad_norm": 19.387493133544922, "learning_rate": 2.245e-06, "loss": 2.5174, "step": 4490 }, { "epoch": 1.4078369905956114, "grad_norm": 14.607769012451172, "learning_rate": 2.2455000000000004e-06, "loss": 2.1988, "step": 4491 }, { "epoch": 1.4081504702194358, "grad_norm": 23.40793228149414, "learning_rate": 2.246e-06, "loss": 2.7084, "step": 4492 }, { "epoch": 1.4084639498432603, "grad_norm": 75.68866729736328, "learning_rate": 2.2465e-06, "loss": 2.672, "step": 4493 }, { "epoch": 1.4087774294670847, "grad_norm": 66.80253601074219, "learning_rate": 2.2470000000000003e-06, "loss": 2.8853, "step": 4494 }, { "epoch": 1.4090909090909092, "grad_norm": 35.71215057373047, "learning_rate": 2.2475e-06, "loss": 2.5797, "step": 4495 }, { "epoch": 1.4094043887147336, "grad_norm": 119.49893188476562, "learning_rate": 2.2480000000000003e-06, "loss": 2.1623, "step": 4496 }, { "epoch": 1.409717868338558, "grad_norm": 20.179807662963867, "learning_rate": 2.2485e-06, "loss": 2.3686, "step": 4497 }, { "epoch": 1.4100313479623825, "grad_norm": 33.06385040283203, "learning_rate": 2.249e-06, "loss": 3.224, "step": 4498 }, { "epoch": 1.410344827586207, "grad_norm": 28.012863159179688, "learning_rate": 2.2495000000000002e-06, "loss": 2.5953, "step": 4499 }, { "epoch": 1.4106583072100314, "grad_norm": 23.963817596435547, "learning_rate": 2.25e-06, "loss": 2.0826, "step": 4500 }, { "epoch": 1.4109717868338558, "grad_norm": 18.82379913330078, "learning_rate": 2.2505000000000003e-06, "loss": 2.5676, "step": 4501 }, { "epoch": 1.4112852664576803, "grad_norm": 29.09153175354004, "learning_rate": 2.251e-06, "loss": 2.5424, "step": 4502 }, { "epoch": 1.4115987460815047, "grad_norm": 24.60426902770996, "learning_rate": 2.2515e-06, "loss": 2.4428, "step": 4503 }, { "epoch": 1.4119122257053291, "grad_norm": 12.828309059143066, "learning_rate": 2.252e-06, "loss": 2.4397, "step": 4504 }, { "epoch": 1.4122257053291536, "grad_norm": 10.570173263549805, "learning_rate": 2.2525e-06, "loss": 2.289, "step": 4505 }, { "epoch": 1.412539184952978, "grad_norm": 11.049269676208496, "learning_rate": 2.2530000000000003e-06, "loss": 2.133, "step": 4506 }, { "epoch": 1.4128526645768025, "grad_norm": 77.35066986083984, "learning_rate": 2.2535e-06, "loss": 2.4723, "step": 4507 }, { "epoch": 1.413166144200627, "grad_norm": 56.341064453125, "learning_rate": 2.254e-06, "loss": 3.4043, "step": 4508 }, { "epoch": 1.4134796238244514, "grad_norm": 18.001556396484375, "learning_rate": 2.2545e-06, "loss": 2.5181, "step": 4509 }, { "epoch": 1.4137931034482758, "grad_norm": 11.483774185180664, "learning_rate": 2.2550000000000004e-06, "loss": 2.4675, "step": 4510 }, { "epoch": 1.4141065830721002, "grad_norm": 92.04100799560547, "learning_rate": 2.2555000000000003e-06, "loss": 3.0268, "step": 4511 }, { "epoch": 1.4144200626959247, "grad_norm": 25.606027603149414, "learning_rate": 2.256e-06, "loss": 2.5514, "step": 4512 }, { "epoch": 1.4147335423197491, "grad_norm": 25.288551330566406, "learning_rate": 2.2565e-06, "loss": 2.4726, "step": 4513 }, { "epoch": 1.4150470219435736, "grad_norm": 30.17818260192871, "learning_rate": 2.257e-06, "loss": 2.3311, "step": 4514 }, { "epoch": 1.415360501567398, "grad_norm": 42.18940734863281, "learning_rate": 2.2575000000000004e-06, "loss": 2.6688, "step": 4515 }, { "epoch": 1.4156739811912225, "grad_norm": 18.35502815246582, "learning_rate": 2.2580000000000002e-06, "loss": 2.3161, "step": 4516 }, { "epoch": 1.415987460815047, "grad_norm": 20.00831413269043, "learning_rate": 2.2585e-06, "loss": 2.3968, "step": 4517 }, { "epoch": 1.4163009404388713, "grad_norm": 21.05620574951172, "learning_rate": 2.259e-06, "loss": 2.4714, "step": 4518 }, { "epoch": 1.416614420062696, "grad_norm": 21.366777420043945, "learning_rate": 2.2595e-06, "loss": 2.7164, "step": 4519 }, { "epoch": 1.4169278996865204, "grad_norm": 21.714614868164062, "learning_rate": 2.2600000000000004e-06, "loss": 2.5011, "step": 4520 }, { "epoch": 1.4172413793103449, "grad_norm": 29.17438316345215, "learning_rate": 2.2605e-06, "loss": 2.6016, "step": 4521 }, { "epoch": 1.4175548589341693, "grad_norm": 21.83228874206543, "learning_rate": 2.261e-06, "loss": 2.3649, "step": 4522 }, { "epoch": 1.4178683385579938, "grad_norm": 36.300941467285156, "learning_rate": 2.2615000000000003e-06, "loss": 3.4691, "step": 4523 }, { "epoch": 1.4181818181818182, "grad_norm": 52.76793670654297, "learning_rate": 2.262e-06, "loss": 2.8731, "step": 4524 }, { "epoch": 1.4184952978056427, "grad_norm": 44.001461029052734, "learning_rate": 2.2625000000000004e-06, "loss": 2.4197, "step": 4525 }, { "epoch": 1.418808777429467, "grad_norm": 14.982115745544434, "learning_rate": 2.263e-06, "loss": 2.2569, "step": 4526 }, { "epoch": 1.4191222570532915, "grad_norm": 14.590595245361328, "learning_rate": 2.2635e-06, "loss": 2.3871, "step": 4527 }, { "epoch": 1.419435736677116, "grad_norm": 23.147096633911133, "learning_rate": 2.2640000000000003e-06, "loss": 2.2429, "step": 4528 }, { "epoch": 1.4197492163009404, "grad_norm": 25.500978469848633, "learning_rate": 2.2645000000000005e-06, "loss": 2.4662, "step": 4529 }, { "epoch": 1.4200626959247649, "grad_norm": 56.13792037963867, "learning_rate": 2.2650000000000003e-06, "loss": 3.5036, "step": 4530 }, { "epoch": 1.4203761755485893, "grad_norm": 27.729318618774414, "learning_rate": 2.2655e-06, "loss": 2.3546, "step": 4531 }, { "epoch": 1.4206896551724137, "grad_norm": 19.810440063476562, "learning_rate": 2.266e-06, "loss": 2.6285, "step": 4532 }, { "epoch": 1.4210031347962382, "grad_norm": 19.486492156982422, "learning_rate": 2.2665000000000002e-06, "loss": 2.3256, "step": 4533 }, { "epoch": 1.4213166144200626, "grad_norm": 32.0871696472168, "learning_rate": 2.2670000000000005e-06, "loss": 2.2786, "step": 4534 }, { "epoch": 1.421630094043887, "grad_norm": 10.243577003479004, "learning_rate": 2.2675000000000003e-06, "loss": 2.8419, "step": 4535 }, { "epoch": 1.4219435736677117, "grad_norm": 15.97329044342041, "learning_rate": 2.268e-06, "loss": 2.5527, "step": 4536 }, { "epoch": 1.4222570532915362, "grad_norm": 24.04090118408203, "learning_rate": 2.2685e-06, "loss": 2.4916, "step": 4537 }, { "epoch": 1.4225705329153606, "grad_norm": 33.413352966308594, "learning_rate": 2.2690000000000002e-06, "loss": 2.8125, "step": 4538 }, { "epoch": 1.422884012539185, "grad_norm": 42.97076416015625, "learning_rate": 2.2695000000000005e-06, "loss": 3.0929, "step": 4539 }, { "epoch": 1.4231974921630095, "grad_norm": 50.45700454711914, "learning_rate": 2.2700000000000003e-06, "loss": 2.6702, "step": 4540 }, { "epoch": 1.423510971786834, "grad_norm": 13.318603515625, "learning_rate": 2.2705e-06, "loss": 2.333, "step": 4541 }, { "epoch": 1.4238244514106584, "grad_norm": 30.421119689941406, "learning_rate": 2.271e-06, "loss": 2.4485, "step": 4542 }, { "epoch": 1.4241379310344828, "grad_norm": 31.187414169311523, "learning_rate": 2.2715e-06, "loss": 2.3791, "step": 4543 }, { "epoch": 1.4244514106583073, "grad_norm": 14.010133743286133, "learning_rate": 2.2720000000000004e-06, "loss": 2.2676, "step": 4544 }, { "epoch": 1.4247648902821317, "grad_norm": 18.06087875366211, "learning_rate": 2.2725000000000003e-06, "loss": 2.3697, "step": 4545 }, { "epoch": 1.4250783699059562, "grad_norm": 28.26776123046875, "learning_rate": 2.273e-06, "loss": 2.2602, "step": 4546 }, { "epoch": 1.4253918495297806, "grad_norm": 78.11388397216797, "learning_rate": 2.2735000000000003e-06, "loss": 3.5192, "step": 4547 }, { "epoch": 1.425705329153605, "grad_norm": 30.041149139404297, "learning_rate": 2.274e-06, "loss": 2.425, "step": 4548 }, { "epoch": 1.4260188087774295, "grad_norm": 35.09251403808594, "learning_rate": 2.2745000000000004e-06, "loss": 2.3432, "step": 4549 }, { "epoch": 1.426332288401254, "grad_norm": 20.941200256347656, "learning_rate": 2.2750000000000002e-06, "loss": 2.6254, "step": 4550 }, { "epoch": 1.4266457680250784, "grad_norm": 25.721513748168945, "learning_rate": 2.2755e-06, "loss": 3.0662, "step": 4551 }, { "epoch": 1.4269592476489028, "grad_norm": 19.826330184936523, "learning_rate": 2.2760000000000003e-06, "loss": 2.2957, "step": 4552 }, { "epoch": 1.4272727272727272, "grad_norm": 27.101993560791016, "learning_rate": 2.2765e-06, "loss": 2.7257, "step": 4553 }, { "epoch": 1.4275862068965517, "grad_norm": 30.94487190246582, "learning_rate": 2.2770000000000004e-06, "loss": 2.3578, "step": 4554 }, { "epoch": 1.4278996865203761, "grad_norm": 20.627241134643555, "learning_rate": 2.2775000000000002e-06, "loss": 2.4182, "step": 4555 }, { "epoch": 1.4282131661442006, "grad_norm": 41.094146728515625, "learning_rate": 2.278e-06, "loss": 2.4636, "step": 4556 }, { "epoch": 1.428526645768025, "grad_norm": 22.273983001708984, "learning_rate": 2.2785000000000003e-06, "loss": 2.4823, "step": 4557 }, { "epoch": 1.4288401253918495, "grad_norm": 15.536066055297852, "learning_rate": 2.279e-06, "loss": 2.3666, "step": 4558 }, { "epoch": 1.429153605015674, "grad_norm": 19.999664306640625, "learning_rate": 2.2795000000000004e-06, "loss": 2.3434, "step": 4559 }, { "epoch": 1.4294670846394983, "grad_norm": 29.738088607788086, "learning_rate": 2.28e-06, "loss": 3.1658, "step": 4560 }, { "epoch": 1.4297805642633228, "grad_norm": 28.52236557006836, "learning_rate": 2.2805e-06, "loss": 2.5009, "step": 4561 }, { "epoch": 1.4300940438871472, "grad_norm": 58.76250457763672, "learning_rate": 2.2810000000000003e-06, "loss": 2.8935, "step": 4562 }, { "epoch": 1.4304075235109717, "grad_norm": 13.827445030212402, "learning_rate": 2.2815e-06, "loss": 2.5779, "step": 4563 }, { "epoch": 1.430721003134796, "grad_norm": 24.913976669311523, "learning_rate": 2.282e-06, "loss": 2.5309, "step": 4564 }, { "epoch": 1.4310344827586206, "grad_norm": 14.8967924118042, "learning_rate": 2.2825e-06, "loss": 2.3282, "step": 4565 }, { "epoch": 1.4313479623824452, "grad_norm": 27.908395767211914, "learning_rate": 2.2830000000000004e-06, "loss": 2.1359, "step": 4566 }, { "epoch": 1.4316614420062697, "grad_norm": 44.49998474121094, "learning_rate": 2.2835000000000002e-06, "loss": 2.3364, "step": 4567 }, { "epoch": 1.431974921630094, "grad_norm": 78.90461730957031, "learning_rate": 2.284e-06, "loss": 2.8442, "step": 4568 }, { "epoch": 1.4322884012539185, "grad_norm": 61.37258529663086, "learning_rate": 2.2845e-06, "loss": 2.2926, "step": 4569 }, { "epoch": 1.432601880877743, "grad_norm": 17.646041870117188, "learning_rate": 2.285e-06, "loss": 2.453, "step": 4570 }, { "epoch": 1.4329153605015674, "grad_norm": 26.07595443725586, "learning_rate": 2.2855000000000004e-06, "loss": 2.5075, "step": 4571 }, { "epoch": 1.4332288401253919, "grad_norm": 30.174352645874023, "learning_rate": 2.2860000000000002e-06, "loss": 2.6651, "step": 4572 }, { "epoch": 1.4335423197492163, "grad_norm": 12.280838012695312, "learning_rate": 2.2865e-06, "loss": 2.3135, "step": 4573 }, { "epoch": 1.4338557993730408, "grad_norm": 17.326416015625, "learning_rate": 2.287e-06, "loss": 2.3963, "step": 4574 }, { "epoch": 1.4341692789968652, "grad_norm": 48.17243957519531, "learning_rate": 2.2875e-06, "loss": 2.5341, "step": 4575 }, { "epoch": 1.4344827586206896, "grad_norm": 75.73432922363281, "learning_rate": 2.2880000000000004e-06, "loss": 2.8398, "step": 4576 }, { "epoch": 1.434796238244514, "grad_norm": 280.7005920410156, "learning_rate": 2.2885e-06, "loss": 3.0092, "step": 4577 }, { "epoch": 1.4351097178683385, "grad_norm": 15.257850646972656, "learning_rate": 2.289e-06, "loss": 2.2889, "step": 4578 }, { "epoch": 1.435423197492163, "grad_norm": 31.832490921020508, "learning_rate": 2.2895e-06, "loss": 2.4703, "step": 4579 }, { "epoch": 1.4357366771159874, "grad_norm": 24.959474563598633, "learning_rate": 2.29e-06, "loss": 1.9712, "step": 4580 }, { "epoch": 1.4360501567398118, "grad_norm": 30.74005126953125, "learning_rate": 2.2905000000000003e-06, "loss": 2.8657, "step": 4581 }, { "epoch": 1.4363636363636363, "grad_norm": 36.41959762573242, "learning_rate": 2.291e-06, "loss": 2.5452, "step": 4582 }, { "epoch": 1.436677115987461, "grad_norm": 28.88225746154785, "learning_rate": 2.2915e-06, "loss": 2.789, "step": 4583 }, { "epoch": 1.4369905956112854, "grad_norm": 22.670028686523438, "learning_rate": 2.2920000000000002e-06, "loss": 2.5401, "step": 4584 }, { "epoch": 1.4373040752351098, "grad_norm": 17.49390411376953, "learning_rate": 2.2925e-06, "loss": 2.2482, "step": 4585 }, { "epoch": 1.4376175548589343, "grad_norm": 43.223487854003906, "learning_rate": 2.2930000000000003e-06, "loss": 2.6491, "step": 4586 }, { "epoch": 1.4379310344827587, "grad_norm": 28.049163818359375, "learning_rate": 2.2935e-06, "loss": 2.4151, "step": 4587 }, { "epoch": 1.4382445141065832, "grad_norm": 23.431224822998047, "learning_rate": 2.294e-06, "loss": 2.3691, "step": 4588 }, { "epoch": 1.4385579937304076, "grad_norm": 12.071837425231934, "learning_rate": 2.2945000000000002e-06, "loss": 2.3541, "step": 4589 }, { "epoch": 1.438871473354232, "grad_norm": 28.91056251525879, "learning_rate": 2.2950000000000005e-06, "loss": 2.4894, "step": 4590 }, { "epoch": 1.4391849529780565, "grad_norm": 22.10399055480957, "learning_rate": 2.2955000000000003e-06, "loss": 2.3852, "step": 4591 }, { "epoch": 1.439498432601881, "grad_norm": 23.183076858520508, "learning_rate": 2.296e-06, "loss": 2.2229, "step": 4592 }, { "epoch": 1.4398119122257054, "grad_norm": 15.805237770080566, "learning_rate": 2.2965e-06, "loss": 2.3327, "step": 4593 }, { "epoch": 1.4401253918495298, "grad_norm": 24.32073211669922, "learning_rate": 2.297e-06, "loss": 2.4777, "step": 4594 }, { "epoch": 1.4404388714733543, "grad_norm": 29.6636905670166, "learning_rate": 2.2975000000000004e-06, "loss": 2.4994, "step": 4595 }, { "epoch": 1.4407523510971787, "grad_norm": 21.071327209472656, "learning_rate": 2.2980000000000003e-06, "loss": 2.5011, "step": 4596 }, { "epoch": 1.4410658307210031, "grad_norm": 30.68376922607422, "learning_rate": 2.2985e-06, "loss": 2.6823, "step": 4597 }, { "epoch": 1.4413793103448276, "grad_norm": 49.39833068847656, "learning_rate": 2.299e-06, "loss": 3.1115, "step": 4598 }, { "epoch": 1.441692789968652, "grad_norm": 13.447183609008789, "learning_rate": 2.2995e-06, "loss": 2.4562, "step": 4599 }, { "epoch": 1.4420062695924765, "grad_norm": 27.433025360107422, "learning_rate": 2.3000000000000004e-06, "loss": 2.4388, "step": 4600 }, { "epoch": 1.442319749216301, "grad_norm": 26.882158279418945, "learning_rate": 2.3005000000000003e-06, "loss": 2.3435, "step": 4601 }, { "epoch": 1.4426332288401253, "grad_norm": 24.261220932006836, "learning_rate": 2.301e-06, "loss": 2.383, "step": 4602 }, { "epoch": 1.4429467084639498, "grad_norm": 18.171157836914062, "learning_rate": 2.3015000000000003e-06, "loss": 2.3532, "step": 4603 }, { "epoch": 1.4432601880877742, "grad_norm": 27.184085845947266, "learning_rate": 2.302e-06, "loss": 2.3824, "step": 4604 }, { "epoch": 1.4435736677115987, "grad_norm": 25.665733337402344, "learning_rate": 2.3025000000000004e-06, "loss": 2.2047, "step": 4605 }, { "epoch": 1.4438871473354231, "grad_norm": 29.451372146606445, "learning_rate": 2.3030000000000002e-06, "loss": 2.244, "step": 4606 }, { "epoch": 1.4442006269592476, "grad_norm": 16.427772521972656, "learning_rate": 2.3035e-06, "loss": 2.5831, "step": 4607 }, { "epoch": 1.444514106583072, "grad_norm": 30.14790916442871, "learning_rate": 2.3040000000000003e-06, "loss": 2.8414, "step": 4608 }, { "epoch": 1.4448275862068964, "grad_norm": 16.03703498840332, "learning_rate": 2.3045e-06, "loss": 2.26, "step": 4609 }, { "epoch": 1.4451410658307209, "grad_norm": 126.84292602539062, "learning_rate": 2.3050000000000004e-06, "loss": 2.7126, "step": 4610 }, { "epoch": 1.4454545454545453, "grad_norm": 18.969295501708984, "learning_rate": 2.3055e-06, "loss": 2.0886, "step": 4611 }, { "epoch": 1.4457680250783698, "grad_norm": 42.71048355102539, "learning_rate": 2.306e-06, "loss": 3.2491, "step": 4612 }, { "epoch": 1.4460815047021944, "grad_norm": 18.34791374206543, "learning_rate": 2.3065000000000003e-06, "loss": 2.4025, "step": 4613 }, { "epoch": 1.4463949843260189, "grad_norm": 23.1646728515625, "learning_rate": 2.307e-06, "loss": 2.9714, "step": 4614 }, { "epoch": 1.4467084639498433, "grad_norm": 39.93225860595703, "learning_rate": 2.3075000000000004e-06, "loss": 2.9443, "step": 4615 }, { "epoch": 1.4470219435736678, "grad_norm": 17.18674087524414, "learning_rate": 2.308e-06, "loss": 2.4679, "step": 4616 }, { "epoch": 1.4473354231974922, "grad_norm": 40.53770065307617, "learning_rate": 2.3085e-06, "loss": 2.764, "step": 4617 }, { "epoch": 1.4476489028213166, "grad_norm": 36.780208587646484, "learning_rate": 2.3090000000000003e-06, "loss": 2.8383, "step": 4618 }, { "epoch": 1.447962382445141, "grad_norm": 14.106476783752441, "learning_rate": 2.3095e-06, "loss": 2.404, "step": 4619 }, { "epoch": 1.4482758620689655, "grad_norm": 13.64006233215332, "learning_rate": 2.3100000000000003e-06, "loss": 2.2731, "step": 4620 }, { "epoch": 1.44858934169279, "grad_norm": 26.457382202148438, "learning_rate": 2.3105e-06, "loss": 2.7478, "step": 4621 }, { "epoch": 1.4489028213166144, "grad_norm": 38.400936126708984, "learning_rate": 2.311e-06, "loss": 2.5102, "step": 4622 }, { "epoch": 1.4492163009404389, "grad_norm": 41.096771240234375, "learning_rate": 2.3115000000000002e-06, "loss": 2.3325, "step": 4623 }, { "epoch": 1.4495297805642633, "grad_norm": 37.16647720336914, "learning_rate": 2.312e-06, "loss": 2.8084, "step": 4624 }, { "epoch": 1.4498432601880877, "grad_norm": 12.608713150024414, "learning_rate": 2.3125000000000003e-06, "loss": 2.474, "step": 4625 }, { "epoch": 1.4501567398119122, "grad_norm": 11.61148452758789, "learning_rate": 2.313e-06, "loss": 2.4906, "step": 4626 }, { "epoch": 1.4504702194357366, "grad_norm": 152.3636474609375, "learning_rate": 2.3135000000000004e-06, "loss": 2.8927, "step": 4627 }, { "epoch": 1.450783699059561, "grad_norm": 22.41962242126465, "learning_rate": 2.314e-06, "loss": 2.3821, "step": 4628 }, { "epoch": 1.4510971786833855, "grad_norm": 12.883655548095703, "learning_rate": 2.3145e-06, "loss": 2.2177, "step": 4629 }, { "epoch": 1.4514106583072102, "grad_norm": 85.4560317993164, "learning_rate": 2.3150000000000003e-06, "loss": 3.2228, "step": 4630 }, { "epoch": 1.4517241379310346, "grad_norm": 26.63489532470703, "learning_rate": 2.3155e-06, "loss": 2.3603, "step": 4631 }, { "epoch": 1.452037617554859, "grad_norm": 32.63912582397461, "learning_rate": 2.3160000000000004e-06, "loss": 2.5887, "step": 4632 }, { "epoch": 1.4523510971786835, "grad_norm": 27.428691864013672, "learning_rate": 2.3165e-06, "loss": 2.6201, "step": 4633 }, { "epoch": 1.452664576802508, "grad_norm": 12.853342056274414, "learning_rate": 2.317e-06, "loss": 2.4724, "step": 4634 }, { "epoch": 1.4529780564263324, "grad_norm": 31.954179763793945, "learning_rate": 2.3175000000000003e-06, "loss": 2.8328, "step": 4635 }, { "epoch": 1.4532915360501568, "grad_norm": 17.225984573364258, "learning_rate": 2.318e-06, "loss": 2.4483, "step": 4636 }, { "epoch": 1.4536050156739813, "grad_norm": 39.024723052978516, "learning_rate": 2.3185000000000003e-06, "loss": 2.6955, "step": 4637 }, { "epoch": 1.4539184952978057, "grad_norm": 13.877724647521973, "learning_rate": 2.319e-06, "loss": 2.5151, "step": 4638 }, { "epoch": 1.4542319749216301, "grad_norm": 20.05229377746582, "learning_rate": 2.3195e-06, "loss": 2.4511, "step": 4639 }, { "epoch": 1.4545454545454546, "grad_norm": 40.79631805419922, "learning_rate": 2.3200000000000002e-06, "loss": 3.2781, "step": 4640 }, { "epoch": 1.454858934169279, "grad_norm": 26.610153198242188, "learning_rate": 2.3205e-06, "loss": 2.237, "step": 4641 }, { "epoch": 1.4551724137931035, "grad_norm": 43.11115264892578, "learning_rate": 2.3210000000000003e-06, "loss": 1.9168, "step": 4642 }, { "epoch": 1.455485893416928, "grad_norm": 31.909210205078125, "learning_rate": 2.3215e-06, "loss": 2.3807, "step": 4643 }, { "epoch": 1.4557993730407524, "grad_norm": 34.96866989135742, "learning_rate": 2.322e-06, "loss": 2.5393, "step": 4644 }, { "epoch": 1.4561128526645768, "grad_norm": 16.582456588745117, "learning_rate": 2.3225e-06, "loss": 2.4451, "step": 4645 }, { "epoch": 1.4564263322884012, "grad_norm": 13.87610912322998, "learning_rate": 2.3230000000000005e-06, "loss": 2.4433, "step": 4646 }, { "epoch": 1.4567398119122257, "grad_norm": 23.799659729003906, "learning_rate": 2.3235000000000003e-06, "loss": 2.5115, "step": 4647 }, { "epoch": 1.4570532915360501, "grad_norm": 27.039108276367188, "learning_rate": 2.324e-06, "loss": 2.2724, "step": 4648 }, { "epoch": 1.4573667711598746, "grad_norm": 21.01291275024414, "learning_rate": 2.3245e-06, "loss": 2.5922, "step": 4649 }, { "epoch": 1.457680250783699, "grad_norm": 25.368284225463867, "learning_rate": 2.325e-06, "loss": 3.2784, "step": 4650 }, { "epoch": 1.4579937304075234, "grad_norm": 21.040912628173828, "learning_rate": 2.3255000000000004e-06, "loss": 2.2421, "step": 4651 }, { "epoch": 1.458307210031348, "grad_norm": 27.654041290283203, "learning_rate": 2.3260000000000003e-06, "loss": 2.5671, "step": 4652 }, { "epoch": 1.4586206896551723, "grad_norm": 14.569658279418945, "learning_rate": 2.3265e-06, "loss": 2.3534, "step": 4653 }, { "epoch": 1.4589341692789968, "grad_norm": 27.91167640686035, "learning_rate": 2.327e-06, "loss": 2.7539, "step": 4654 }, { "epoch": 1.4592476489028212, "grad_norm": 17.407148361206055, "learning_rate": 2.3275e-06, "loss": 2.5176, "step": 4655 }, { "epoch": 1.4595611285266457, "grad_norm": 20.95456886291504, "learning_rate": 2.3280000000000004e-06, "loss": 2.2847, "step": 4656 }, { "epoch": 1.45987460815047, "grad_norm": 15.59076976776123, "learning_rate": 2.3285000000000002e-06, "loss": 2.2954, "step": 4657 }, { "epoch": 1.4601880877742945, "grad_norm": 18.56783676147461, "learning_rate": 2.329e-06, "loss": 2.4062, "step": 4658 }, { "epoch": 1.4605015673981192, "grad_norm": 18.09410858154297, "learning_rate": 2.3295e-06, "loss": 2.2867, "step": 4659 }, { "epoch": 1.4608150470219436, "grad_norm": 10.945845603942871, "learning_rate": 2.33e-06, "loss": 2.2486, "step": 4660 }, { "epoch": 1.461128526645768, "grad_norm": 19.199756622314453, "learning_rate": 2.3305000000000004e-06, "loss": 2.6845, "step": 4661 }, { "epoch": 1.4614420062695925, "grad_norm": 15.891563415527344, "learning_rate": 2.3310000000000002e-06, "loss": 2.6765, "step": 4662 }, { "epoch": 1.461755485893417, "grad_norm": 31.088024139404297, "learning_rate": 2.3315e-06, "loss": 2.3015, "step": 4663 }, { "epoch": 1.4620689655172414, "grad_norm": 72.7515869140625, "learning_rate": 2.3320000000000003e-06, "loss": 2.8665, "step": 4664 }, { "epoch": 1.4623824451410659, "grad_norm": 22.520038604736328, "learning_rate": 2.3325e-06, "loss": 2.517, "step": 4665 }, { "epoch": 1.4626959247648903, "grad_norm": 19.68512725830078, "learning_rate": 2.3330000000000004e-06, "loss": 2.5447, "step": 4666 }, { "epoch": 1.4630094043887147, "grad_norm": 55.96286392211914, "learning_rate": 2.3335e-06, "loss": 2.6517, "step": 4667 }, { "epoch": 1.4633228840125392, "grad_norm": 38.48992919921875, "learning_rate": 2.334e-06, "loss": 2.9361, "step": 4668 }, { "epoch": 1.4636363636363636, "grad_norm": 25.144210815429688, "learning_rate": 2.3345000000000003e-06, "loss": 2.3563, "step": 4669 }, { "epoch": 1.463949843260188, "grad_norm": 16.994159698486328, "learning_rate": 2.3350000000000005e-06, "loss": 2.4947, "step": 4670 }, { "epoch": 1.4642633228840125, "grad_norm": 30.933141708374023, "learning_rate": 2.3355000000000003e-06, "loss": 2.469, "step": 4671 }, { "epoch": 1.464576802507837, "grad_norm": 33.76378631591797, "learning_rate": 2.336e-06, "loss": 2.8162, "step": 4672 }, { "epoch": 1.4648902821316614, "grad_norm": 13.684147834777832, "learning_rate": 2.3365e-06, "loss": 2.3385, "step": 4673 }, { "epoch": 1.4652037617554858, "grad_norm": 33.54969787597656, "learning_rate": 2.3370000000000002e-06, "loss": 2.4732, "step": 4674 }, { "epoch": 1.4655172413793103, "grad_norm": 21.35070037841797, "learning_rate": 2.3375000000000005e-06, "loss": 2.8555, "step": 4675 }, { "epoch": 1.4658307210031347, "grad_norm": 20.555835723876953, "learning_rate": 2.3380000000000003e-06, "loss": 2.4961, "step": 4676 }, { "epoch": 1.4661442006269594, "grad_norm": 26.707595825195312, "learning_rate": 2.3385e-06, "loss": 2.4832, "step": 4677 }, { "epoch": 1.4664576802507838, "grad_norm": 18.20785140991211, "learning_rate": 2.339e-06, "loss": 2.9194, "step": 4678 }, { "epoch": 1.4667711598746083, "grad_norm": 94.4541244506836, "learning_rate": 2.3395000000000002e-06, "loss": 2.3439, "step": 4679 }, { "epoch": 1.4670846394984327, "grad_norm": 25.442766189575195, "learning_rate": 2.3400000000000005e-06, "loss": 2.6061, "step": 4680 }, { "epoch": 1.4673981191222571, "grad_norm": 33.864078521728516, "learning_rate": 2.3405000000000003e-06, "loss": 2.5939, "step": 4681 }, { "epoch": 1.4677115987460816, "grad_norm": 41.32418441772461, "learning_rate": 2.341e-06, "loss": 3.0855, "step": 4682 }, { "epoch": 1.468025078369906, "grad_norm": 20.600738525390625, "learning_rate": 2.3415000000000004e-06, "loss": 2.4121, "step": 4683 }, { "epoch": 1.4683385579937305, "grad_norm": 21.15034294128418, "learning_rate": 2.342e-06, "loss": 2.4212, "step": 4684 }, { "epoch": 1.468652037617555, "grad_norm": 16.126068115234375, "learning_rate": 2.3425000000000004e-06, "loss": 2.5768, "step": 4685 }, { "epoch": 1.4689655172413794, "grad_norm": 18.89388084411621, "learning_rate": 2.3430000000000003e-06, "loss": 2.1556, "step": 4686 }, { "epoch": 1.4692789968652038, "grad_norm": 49.33967208862305, "learning_rate": 2.3435e-06, "loss": 2.4046, "step": 4687 }, { "epoch": 1.4695924764890282, "grad_norm": 18.668262481689453, "learning_rate": 2.3440000000000003e-06, "loss": 2.4608, "step": 4688 }, { "epoch": 1.4699059561128527, "grad_norm": 15.122355461120605, "learning_rate": 2.3445e-06, "loss": 2.412, "step": 4689 }, { "epoch": 1.4702194357366771, "grad_norm": 9.459912300109863, "learning_rate": 2.345e-06, "loss": 2.3784, "step": 4690 }, { "epoch": 1.4705329153605016, "grad_norm": 13.91557788848877, "learning_rate": 2.3455000000000002e-06, "loss": 2.6558, "step": 4691 }, { "epoch": 1.470846394984326, "grad_norm": 53.33099365234375, "learning_rate": 2.346e-06, "loss": 2.5759, "step": 4692 }, { "epoch": 1.4711598746081505, "grad_norm": 26.81603240966797, "learning_rate": 2.3465000000000003e-06, "loss": 2.4183, "step": 4693 }, { "epoch": 1.471473354231975, "grad_norm": 15.679391860961914, "learning_rate": 2.347e-06, "loss": 2.8987, "step": 4694 }, { "epoch": 1.4717868338557993, "grad_norm": 16.509136199951172, "learning_rate": 2.3475e-06, "loss": 2.4195, "step": 4695 }, { "epoch": 1.4721003134796238, "grad_norm": 30.074337005615234, "learning_rate": 2.3480000000000002e-06, "loss": 2.5757, "step": 4696 }, { "epoch": 1.4724137931034482, "grad_norm": 19.880718231201172, "learning_rate": 2.3485e-06, "loss": 2.3326, "step": 4697 }, { "epoch": 1.4727272727272727, "grad_norm": 37.2962646484375, "learning_rate": 2.3490000000000003e-06, "loss": 2.1726, "step": 4698 }, { "epoch": 1.473040752351097, "grad_norm": 29.554244995117188, "learning_rate": 2.3495e-06, "loss": 2.377, "step": 4699 }, { "epoch": 1.4733542319749215, "grad_norm": 24.98297691345215, "learning_rate": 2.35e-06, "loss": 2.597, "step": 4700 }, { "epoch": 1.473667711598746, "grad_norm": 13.754597663879395, "learning_rate": 2.3505e-06, "loss": 2.4814, "step": 4701 }, { "epoch": 1.4739811912225704, "grad_norm": 54.246803283691406, "learning_rate": 2.351e-06, "loss": 2.4151, "step": 4702 }, { "epoch": 1.4742946708463949, "grad_norm": 9.66529369354248, "learning_rate": 2.3515000000000003e-06, "loss": 2.3911, "step": 4703 }, { "epoch": 1.4746081504702193, "grad_norm": 16.990436553955078, "learning_rate": 2.352e-06, "loss": 2.7343, "step": 4704 }, { "epoch": 1.4749216300940438, "grad_norm": 18.184396743774414, "learning_rate": 2.3525e-06, "loss": 2.3241, "step": 4705 }, { "epoch": 1.4752351097178684, "grad_norm": 16.607677459716797, "learning_rate": 2.353e-06, "loss": 2.5718, "step": 4706 }, { "epoch": 1.4755485893416929, "grad_norm": 25.31608772277832, "learning_rate": 2.3535000000000004e-06, "loss": 2.612, "step": 4707 }, { "epoch": 1.4758620689655173, "grad_norm": 9.628955841064453, "learning_rate": 2.3540000000000002e-06, "loss": 2.3875, "step": 4708 }, { "epoch": 1.4761755485893417, "grad_norm": 20.810176849365234, "learning_rate": 2.3545e-06, "loss": 2.4388, "step": 4709 }, { "epoch": 1.4764890282131662, "grad_norm": 47.8430061340332, "learning_rate": 2.355e-06, "loss": 2.5104, "step": 4710 }, { "epoch": 1.4768025078369906, "grad_norm": 44.49211502075195, "learning_rate": 2.3555e-06, "loss": 2.1969, "step": 4711 }, { "epoch": 1.477115987460815, "grad_norm": 16.758508682250977, "learning_rate": 2.3560000000000004e-06, "loss": 2.4783, "step": 4712 }, { "epoch": 1.4774294670846395, "grad_norm": 20.397388458251953, "learning_rate": 2.3565000000000002e-06, "loss": 2.361, "step": 4713 }, { "epoch": 1.477742946708464, "grad_norm": 26.805633544921875, "learning_rate": 2.357e-06, "loss": 2.5757, "step": 4714 }, { "epoch": 1.4780564263322884, "grad_norm": 44.96598815917969, "learning_rate": 2.3575e-06, "loss": 2.8678, "step": 4715 }, { "epoch": 1.4783699059561128, "grad_norm": 23.84420394897461, "learning_rate": 2.358e-06, "loss": 2.7157, "step": 4716 }, { "epoch": 1.4786833855799373, "grad_norm": 14.02918529510498, "learning_rate": 2.3585000000000004e-06, "loss": 2.2836, "step": 4717 }, { "epoch": 1.4789968652037617, "grad_norm": 16.57590675354004, "learning_rate": 2.359e-06, "loss": 2.1943, "step": 4718 }, { "epoch": 1.4793103448275862, "grad_norm": 23.227548599243164, "learning_rate": 2.3595e-06, "loss": 3.5659, "step": 4719 }, { "epoch": 1.4796238244514106, "grad_norm": 12.566490173339844, "learning_rate": 2.3600000000000003e-06, "loss": 2.4076, "step": 4720 }, { "epoch": 1.479937304075235, "grad_norm": 30.77694320678711, "learning_rate": 2.3605e-06, "loss": 2.5922, "step": 4721 }, { "epoch": 1.4802507836990595, "grad_norm": 15.837915420532227, "learning_rate": 2.3610000000000003e-06, "loss": 2.3494, "step": 4722 }, { "epoch": 1.480564263322884, "grad_norm": 18.752384185791016, "learning_rate": 2.3615e-06, "loss": 2.6885, "step": 4723 }, { "epoch": 1.4808777429467086, "grad_norm": 21.009431838989258, "learning_rate": 2.362e-06, "loss": 2.3239, "step": 4724 }, { "epoch": 1.481191222570533, "grad_norm": 51.4163703918457, "learning_rate": 2.3625000000000003e-06, "loss": 2.4585, "step": 4725 }, { "epoch": 1.4815047021943575, "grad_norm": 10.593263626098633, "learning_rate": 2.3630000000000005e-06, "loss": 2.2508, "step": 4726 }, { "epoch": 1.481818181818182, "grad_norm": 27.31687355041504, "learning_rate": 2.3635000000000003e-06, "loss": 2.5703, "step": 4727 }, { "epoch": 1.4821316614420064, "grad_norm": 27.05815315246582, "learning_rate": 2.364e-06, "loss": 2.7473, "step": 4728 }, { "epoch": 1.4824451410658308, "grad_norm": 18.825220108032227, "learning_rate": 2.3645e-06, "loss": 2.3339, "step": 4729 }, { "epoch": 1.4827586206896552, "grad_norm": 45.73810958862305, "learning_rate": 2.3650000000000002e-06, "loss": 2.526, "step": 4730 }, { "epoch": 1.4830721003134797, "grad_norm": 17.639097213745117, "learning_rate": 2.3655000000000005e-06, "loss": 2.6365, "step": 4731 }, { "epoch": 1.4833855799373041, "grad_norm": 20.23157501220703, "learning_rate": 2.3660000000000003e-06, "loss": 2.3239, "step": 4732 }, { "epoch": 1.4836990595611286, "grad_norm": 13.595239639282227, "learning_rate": 2.3665e-06, "loss": 2.2712, "step": 4733 }, { "epoch": 1.484012539184953, "grad_norm": 90.09672546386719, "learning_rate": 2.367e-06, "loss": 3.2016, "step": 4734 }, { "epoch": 1.4843260188087775, "grad_norm": 96.69264221191406, "learning_rate": 2.3675e-06, "loss": 2.354, "step": 4735 }, { "epoch": 1.484639498432602, "grad_norm": 32.80999755859375, "learning_rate": 2.3680000000000005e-06, "loss": 3.1623, "step": 4736 }, { "epoch": 1.4849529780564263, "grad_norm": 29.02104377746582, "learning_rate": 2.3685000000000003e-06, "loss": 2.4772, "step": 4737 }, { "epoch": 1.4852664576802508, "grad_norm": 19.033187866210938, "learning_rate": 2.369e-06, "loss": 2.3753, "step": 4738 }, { "epoch": 1.4855799373040752, "grad_norm": 20.396085739135742, "learning_rate": 2.3695e-06, "loss": 2.3814, "step": 4739 }, { "epoch": 1.4858934169278997, "grad_norm": 43.776119232177734, "learning_rate": 2.37e-06, "loss": 2.4351, "step": 4740 }, { "epoch": 1.486206896551724, "grad_norm": 16.356319427490234, "learning_rate": 2.3705000000000004e-06, "loss": 2.385, "step": 4741 }, { "epoch": 1.4865203761755486, "grad_norm": 27.0319766998291, "learning_rate": 2.3710000000000003e-06, "loss": 2.6299, "step": 4742 }, { "epoch": 1.486833855799373, "grad_norm": 28.409624099731445, "learning_rate": 2.3715e-06, "loss": 2.5679, "step": 4743 }, { "epoch": 1.4871473354231974, "grad_norm": 15.29839038848877, "learning_rate": 2.3720000000000003e-06, "loss": 2.2037, "step": 4744 }, { "epoch": 1.4874608150470219, "grad_norm": 44.7656364440918, "learning_rate": 2.3725e-06, "loss": 2.4191, "step": 4745 }, { "epoch": 1.4877742946708463, "grad_norm": 33.07748031616211, "learning_rate": 2.3730000000000004e-06, "loss": 2.9531, "step": 4746 }, { "epoch": 1.4880877742946708, "grad_norm": 15.354388236999512, "learning_rate": 2.3735000000000002e-06, "loss": 2.264, "step": 4747 }, { "epoch": 1.4884012539184952, "grad_norm": 52.4686164855957, "learning_rate": 2.374e-06, "loss": 2.2487, "step": 4748 }, { "epoch": 1.4887147335423196, "grad_norm": 13.158205032348633, "learning_rate": 2.3745000000000003e-06, "loss": 2.4897, "step": 4749 }, { "epoch": 1.489028213166144, "grad_norm": 32.53001403808594, "learning_rate": 2.375e-06, "loss": 2.7913, "step": 4750 }, { "epoch": 1.4893416927899685, "grad_norm": 15.134367942810059, "learning_rate": 2.3755000000000004e-06, "loss": 2.4354, "step": 4751 }, { "epoch": 1.489655172413793, "grad_norm": 14.022037506103516, "learning_rate": 2.376e-06, "loss": 2.8385, "step": 4752 }, { "epoch": 1.4899686520376176, "grad_norm": 12.180167198181152, "learning_rate": 2.3765e-06, "loss": 2.355, "step": 4753 }, { "epoch": 1.490282131661442, "grad_norm": 24.225900650024414, "learning_rate": 2.3770000000000003e-06, "loss": 2.7134, "step": 4754 }, { "epoch": 1.4905956112852665, "grad_norm": 14.048494338989258, "learning_rate": 2.3775e-06, "loss": 2.3146, "step": 4755 }, { "epoch": 1.490909090909091, "grad_norm": 52.83781433105469, "learning_rate": 2.3780000000000004e-06, "loss": 2.6985, "step": 4756 }, { "epoch": 1.4912225705329154, "grad_norm": 16.474990844726562, "learning_rate": 2.3785e-06, "loss": 2.3649, "step": 4757 }, { "epoch": 1.4915360501567398, "grad_norm": 71.45556640625, "learning_rate": 2.379e-06, "loss": 2.3475, "step": 4758 }, { "epoch": 1.4918495297805643, "grad_norm": 115.68441009521484, "learning_rate": 2.3795000000000003e-06, "loss": 3.3305, "step": 4759 }, { "epoch": 1.4921630094043887, "grad_norm": 33.583370208740234, "learning_rate": 2.38e-06, "loss": 2.6491, "step": 4760 }, { "epoch": 1.4924764890282132, "grad_norm": 67.39979553222656, "learning_rate": 2.3805000000000003e-06, "loss": 3.0342, "step": 4761 }, { "epoch": 1.4927899686520376, "grad_norm": 18.05873680114746, "learning_rate": 2.381e-06, "loss": 2.2855, "step": 4762 }, { "epoch": 1.493103448275862, "grad_norm": 19.0296573638916, "learning_rate": 2.3815000000000004e-06, "loss": 2.3671, "step": 4763 }, { "epoch": 1.4934169278996865, "grad_norm": 20.803741455078125, "learning_rate": 2.3820000000000002e-06, "loss": 2.1159, "step": 4764 }, { "epoch": 1.493730407523511, "grad_norm": 21.71002197265625, "learning_rate": 2.3825e-06, "loss": 2.1941, "step": 4765 }, { "epoch": 1.4940438871473354, "grad_norm": 27.901458740234375, "learning_rate": 2.3830000000000003e-06, "loss": 2.3726, "step": 4766 }, { "epoch": 1.4943573667711598, "grad_norm": 22.505205154418945, "learning_rate": 2.3835e-06, "loss": 2.2415, "step": 4767 }, { "epoch": 1.4946708463949843, "grad_norm": 59.68888473510742, "learning_rate": 2.3840000000000004e-06, "loss": 2.935, "step": 4768 }, { "epoch": 1.4949843260188087, "grad_norm": 24.86325454711914, "learning_rate": 2.3845e-06, "loss": 2.3544, "step": 4769 }, { "epoch": 1.4952978056426331, "grad_norm": 19.455236434936523, "learning_rate": 2.385e-06, "loss": 2.2056, "step": 4770 }, { "epoch": 1.4956112852664578, "grad_norm": 12.818082809448242, "learning_rate": 2.3855000000000003e-06, "loss": 2.2469, "step": 4771 }, { "epoch": 1.4959247648902823, "grad_norm": 41.96872329711914, "learning_rate": 2.386e-06, "loss": 2.7509, "step": 4772 }, { "epoch": 1.4962382445141067, "grad_norm": 19.305591583251953, "learning_rate": 2.3865000000000004e-06, "loss": 2.315, "step": 4773 }, { "epoch": 1.4965517241379311, "grad_norm": 37.54866027832031, "learning_rate": 2.387e-06, "loss": 2.1162, "step": 4774 }, { "epoch": 1.4968652037617556, "grad_norm": 21.865880966186523, "learning_rate": 2.3875e-06, "loss": 2.6231, "step": 4775 }, { "epoch": 1.49717868338558, "grad_norm": 16.861860275268555, "learning_rate": 2.3880000000000003e-06, "loss": 2.2176, "step": 4776 }, { "epoch": 1.4974921630094045, "grad_norm": 31.959991455078125, "learning_rate": 2.3885e-06, "loss": 2.1846, "step": 4777 }, { "epoch": 1.497805642633229, "grad_norm": 32.17866897583008, "learning_rate": 2.3890000000000003e-06, "loss": 2.3376, "step": 4778 }, { "epoch": 1.4981191222570533, "grad_norm": 22.17281723022461, "learning_rate": 2.3895e-06, "loss": 2.3265, "step": 4779 }, { "epoch": 1.4984326018808778, "grad_norm": 19.636741638183594, "learning_rate": 2.39e-06, "loss": 2.3045, "step": 4780 }, { "epoch": 1.4987460815047022, "grad_norm": 13.541577339172363, "learning_rate": 2.3905000000000002e-06, "loss": 2.8675, "step": 4781 }, { "epoch": 1.4990595611285267, "grad_norm": 14.708398818969727, "learning_rate": 2.391e-06, "loss": 2.4103, "step": 4782 }, { "epoch": 1.4993730407523511, "grad_norm": 20.451107025146484, "learning_rate": 2.3915000000000003e-06, "loss": 2.6472, "step": 4783 }, { "epoch": 1.4996865203761756, "grad_norm": 46.59673309326172, "learning_rate": 2.392e-06, "loss": 2.3958, "step": 4784 }, { "epoch": 1.5, "grad_norm": 22.707223892211914, "learning_rate": 2.3925e-06, "loss": 2.0873, "step": 4785 }, { "epoch": 1.5003134796238244, "grad_norm": 51.4575080871582, "learning_rate": 2.3930000000000002e-06, "loss": 2.4381, "step": 4786 }, { "epoch": 1.5006269592476489, "grad_norm": 17.58340072631836, "learning_rate": 2.3935000000000005e-06, "loss": 2.2001, "step": 4787 }, { "epoch": 1.5009404388714733, "grad_norm": 94.2140884399414, "learning_rate": 2.3940000000000003e-06, "loss": 2.8093, "step": 4788 }, { "epoch": 1.5009404388714733, "eval_loss": 3.0960214138031006, "eval_runtime": 21.033, "eval_samples_per_second": 127.752, "eval_steps_per_second": 7.987, "step": 4788 }, { "epoch": 1.5012539184952978, "grad_norm": 27.263410568237305, "learning_rate": 2.3945e-06, "loss": 2.5208, "step": 4789 }, { "epoch": 1.5015673981191222, "grad_norm": 46.74039077758789, "learning_rate": 2.395e-06, "loss": 2.3394, "step": 4790 }, { "epoch": 1.5018808777429467, "grad_norm": 29.4154109954834, "learning_rate": 2.3955e-06, "loss": 2.3911, "step": 4791 }, { "epoch": 1.502194357366771, "grad_norm": 21.53428840637207, "learning_rate": 2.3960000000000004e-06, "loss": 2.2403, "step": 4792 }, { "epoch": 1.5025078369905955, "grad_norm": 12.628145217895508, "learning_rate": 2.3965000000000003e-06, "loss": 2.8572, "step": 4793 }, { "epoch": 1.50282131661442, "grad_norm": 44.54032516479492, "learning_rate": 2.397e-06, "loss": 2.4828, "step": 4794 }, { "epoch": 1.5031347962382444, "grad_norm": 16.90373420715332, "learning_rate": 2.3975e-06, "loss": 2.3013, "step": 4795 }, { "epoch": 1.5034482758620689, "grad_norm": 17.77326774597168, "learning_rate": 2.398e-06, "loss": 2.414, "step": 4796 }, { "epoch": 1.5037617554858933, "grad_norm": 19.9771728515625, "learning_rate": 2.3985000000000004e-06, "loss": 2.236, "step": 4797 }, { "epoch": 1.5040752351097177, "grad_norm": 16.979183197021484, "learning_rate": 2.3990000000000002e-06, "loss": 2.2729, "step": 4798 }, { "epoch": 1.5043887147335422, "grad_norm": 28.2900333404541, "learning_rate": 2.3995e-06, "loss": 2.4509, "step": 4799 }, { "epoch": 1.5047021943573666, "grad_norm": 16.051536560058594, "learning_rate": 2.4000000000000003e-06, "loss": 2.4083, "step": 4800 }, { "epoch": 1.505015673981191, "grad_norm": 43.82846450805664, "learning_rate": 2.4005e-06, "loss": 2.3903, "step": 4801 }, { "epoch": 1.5053291536050155, "grad_norm": 24.0675106048584, "learning_rate": 2.4010000000000004e-06, "loss": 3.2002, "step": 4802 }, { "epoch": 1.5056426332288402, "grad_norm": 16.599925994873047, "learning_rate": 2.4015000000000002e-06, "loss": 2.2859, "step": 4803 }, { "epoch": 1.5059561128526646, "grad_norm": 15.7088623046875, "learning_rate": 2.402e-06, "loss": 2.1144, "step": 4804 }, { "epoch": 1.506269592476489, "grad_norm": 17.363964080810547, "learning_rate": 2.4025000000000003e-06, "loss": 2.4393, "step": 4805 }, { "epoch": 1.5065830721003135, "grad_norm": 40.69404602050781, "learning_rate": 2.4030000000000005e-06, "loss": 2.4038, "step": 4806 }, { "epoch": 1.506896551724138, "grad_norm": 11.931968688964844, "learning_rate": 2.4035000000000004e-06, "loss": 2.2415, "step": 4807 }, { "epoch": 1.5072100313479624, "grad_norm": 15.336716651916504, "learning_rate": 2.404e-06, "loss": 2.3751, "step": 4808 }, { "epoch": 1.5075235109717868, "grad_norm": 26.773588180541992, "learning_rate": 2.4045e-06, "loss": 2.4447, "step": 4809 }, { "epoch": 1.5078369905956113, "grad_norm": 10.982498168945312, "learning_rate": 2.4050000000000003e-06, "loss": 2.2266, "step": 4810 }, { "epoch": 1.5081504702194357, "grad_norm": 18.58058738708496, "learning_rate": 2.4055000000000005e-06, "loss": 2.4837, "step": 4811 }, { "epoch": 1.5084639498432602, "grad_norm": 25.92458152770996, "learning_rate": 2.4060000000000003e-06, "loss": 2.3429, "step": 4812 }, { "epoch": 1.5087774294670846, "grad_norm": 22.95157241821289, "learning_rate": 2.4065e-06, "loss": 2.3551, "step": 4813 }, { "epoch": 1.509090909090909, "grad_norm": 12.792777061462402, "learning_rate": 2.407e-06, "loss": 2.1755, "step": 4814 }, { "epoch": 1.5094043887147337, "grad_norm": 20.182924270629883, "learning_rate": 2.4075000000000002e-06, "loss": 2.0571, "step": 4815 }, { "epoch": 1.5097178683385581, "grad_norm": 16.23738670349121, "learning_rate": 2.408e-06, "loss": 2.681, "step": 4816 }, { "epoch": 1.5100313479623826, "grad_norm": 12.84005355834961, "learning_rate": 2.4085000000000003e-06, "loss": 2.9673, "step": 4817 }, { "epoch": 1.510344827586207, "grad_norm": 19.747169494628906, "learning_rate": 2.409e-06, "loss": 2.3013, "step": 4818 }, { "epoch": 1.5106583072100315, "grad_norm": 27.003398895263672, "learning_rate": 2.4095e-06, "loss": 3.092, "step": 4819 }, { "epoch": 1.510971786833856, "grad_norm": 26.648151397705078, "learning_rate": 2.4100000000000002e-06, "loss": 2.2164, "step": 4820 }, { "epoch": 1.5112852664576804, "grad_norm": 12.34875202178955, "learning_rate": 2.4105e-06, "loss": 2.6219, "step": 4821 }, { "epoch": 1.5115987460815048, "grad_norm": 27.272109985351562, "learning_rate": 2.4110000000000003e-06, "loss": 2.3687, "step": 4822 }, { "epoch": 1.5119122257053292, "grad_norm": 33.198692321777344, "learning_rate": 2.4115e-06, "loss": 2.7016, "step": 4823 }, { "epoch": 1.5122257053291537, "grad_norm": 15.936403274536133, "learning_rate": 2.4120000000000004e-06, "loss": 2.6341, "step": 4824 }, { "epoch": 1.5125391849529781, "grad_norm": 37.71348190307617, "learning_rate": 2.4125e-06, "loss": 2.3681, "step": 4825 }, { "epoch": 1.5128526645768026, "grad_norm": 13.023741722106934, "learning_rate": 2.413e-06, "loss": 2.473, "step": 4826 }, { "epoch": 1.513166144200627, "grad_norm": 13.259977340698242, "learning_rate": 2.4135000000000003e-06, "loss": 2.703, "step": 4827 }, { "epoch": 1.5134796238244514, "grad_norm": 59.40862274169922, "learning_rate": 2.414e-06, "loss": 3.2762, "step": 4828 }, { "epoch": 1.513793103448276, "grad_norm": 24.48963737487793, "learning_rate": 2.4145000000000003e-06, "loss": 2.906, "step": 4829 }, { "epoch": 1.5141065830721003, "grad_norm": 13.027275085449219, "learning_rate": 2.415e-06, "loss": 2.5957, "step": 4830 }, { "epoch": 1.5144200626959248, "grad_norm": 13.824039459228516, "learning_rate": 2.4155e-06, "loss": 2.2586, "step": 4831 }, { "epoch": 1.5147335423197492, "grad_norm": 22.69841194152832, "learning_rate": 2.4160000000000002e-06, "loss": 2.7997, "step": 4832 }, { "epoch": 1.5150470219435737, "grad_norm": 28.543188095092773, "learning_rate": 2.4165e-06, "loss": 2.6812, "step": 4833 }, { "epoch": 1.515360501567398, "grad_norm": 66.29344940185547, "learning_rate": 2.4170000000000003e-06, "loss": 3.7326, "step": 4834 }, { "epoch": 1.5156739811912225, "grad_norm": 35.89635467529297, "learning_rate": 2.4175e-06, "loss": 2.3554, "step": 4835 }, { "epoch": 1.515987460815047, "grad_norm": 21.20792579650879, "learning_rate": 2.418e-06, "loss": 2.4513, "step": 4836 }, { "epoch": 1.5163009404388714, "grad_norm": 24.268932342529297, "learning_rate": 2.4185000000000002e-06, "loss": 2.4245, "step": 4837 }, { "epoch": 1.5166144200626959, "grad_norm": 464.6249694824219, "learning_rate": 2.419e-06, "loss": 2.7334, "step": 4838 }, { "epoch": 1.5169278996865203, "grad_norm": 17.084293365478516, "learning_rate": 2.4195000000000003e-06, "loss": 2.1875, "step": 4839 }, { "epoch": 1.5172413793103448, "grad_norm": 47.49144744873047, "learning_rate": 2.42e-06, "loss": 2.5519, "step": 4840 }, { "epoch": 1.5175548589341692, "grad_norm": 19.089134216308594, "learning_rate": 2.4205e-06, "loss": 3.1702, "step": 4841 }, { "epoch": 1.5178683385579936, "grad_norm": 21.179597854614258, "learning_rate": 2.421e-06, "loss": 2.2937, "step": 4842 }, { "epoch": 1.518181818181818, "grad_norm": 12.67155933380127, "learning_rate": 2.4215000000000004e-06, "loss": 2.2309, "step": 4843 }, { "epoch": 1.5184952978056425, "grad_norm": 15.11013126373291, "learning_rate": 2.4220000000000003e-06, "loss": 2.2038, "step": 4844 }, { "epoch": 1.518808777429467, "grad_norm": 13.126248359680176, "learning_rate": 2.4225e-06, "loss": 2.4207, "step": 4845 }, { "epoch": 1.5191222570532914, "grad_norm": 55.65666198730469, "learning_rate": 2.423e-06, "loss": 2.9468, "step": 4846 }, { "epoch": 1.5194357366771158, "grad_norm": 12.58440113067627, "learning_rate": 2.4235e-06, "loss": 2.3262, "step": 4847 }, { "epoch": 1.5197492163009403, "grad_norm": 37.81396484375, "learning_rate": 2.4240000000000004e-06, "loss": 2.6327, "step": 4848 }, { "epoch": 1.5200626959247647, "grad_norm": 16.613920211791992, "learning_rate": 2.4245000000000002e-06, "loss": 2.4177, "step": 4849 }, { "epoch": 1.5203761755485894, "grad_norm": 25.54189682006836, "learning_rate": 2.425e-06, "loss": 2.4132, "step": 4850 }, { "epoch": 1.5206896551724138, "grad_norm": 11.360184669494629, "learning_rate": 2.4255e-06, "loss": 2.3275, "step": 4851 }, { "epoch": 1.5210031347962383, "grad_norm": 22.406675338745117, "learning_rate": 2.426e-06, "loss": 2.4512, "step": 4852 }, { "epoch": 1.5213166144200627, "grad_norm": 15.627935409545898, "learning_rate": 2.4265000000000004e-06, "loss": 2.3716, "step": 4853 }, { "epoch": 1.5216300940438872, "grad_norm": 42.328453063964844, "learning_rate": 2.4270000000000002e-06, "loss": 2.3246, "step": 4854 }, { "epoch": 1.5219435736677116, "grad_norm": 16.802934646606445, "learning_rate": 2.4275e-06, "loss": 2.1961, "step": 4855 }, { "epoch": 1.522257053291536, "grad_norm": 17.857563018798828, "learning_rate": 2.428e-06, "loss": 2.2421, "step": 4856 }, { "epoch": 1.5225705329153605, "grad_norm": 12.803881645202637, "learning_rate": 2.4285e-06, "loss": 2.4404, "step": 4857 }, { "epoch": 1.522884012539185, "grad_norm": 18.709604263305664, "learning_rate": 2.4290000000000004e-06, "loss": 2.6333, "step": 4858 }, { "epoch": 1.5231974921630094, "grad_norm": 25.00420379638672, "learning_rate": 2.4295e-06, "loss": 2.4124, "step": 4859 }, { "epoch": 1.5235109717868338, "grad_norm": 13.070844650268555, "learning_rate": 2.43e-06, "loss": 2.4096, "step": 4860 }, { "epoch": 1.5238244514106583, "grad_norm": 20.735050201416016, "learning_rate": 2.4305000000000003e-06, "loss": 2.0841, "step": 4861 }, { "epoch": 1.524137931034483, "grad_norm": 30.139467239379883, "learning_rate": 2.431e-06, "loss": 2.527, "step": 4862 }, { "epoch": 1.5244514106583074, "grad_norm": 31.38504409790039, "learning_rate": 2.4315000000000004e-06, "loss": 2.6405, "step": 4863 }, { "epoch": 1.5247648902821318, "grad_norm": 14.734469413757324, "learning_rate": 2.432e-06, "loss": 2.147, "step": 4864 }, { "epoch": 1.5250783699059562, "grad_norm": 21.783294677734375, "learning_rate": 2.4325e-06, "loss": 2.2365, "step": 4865 }, { "epoch": 1.5253918495297807, "grad_norm": 20.588727951049805, "learning_rate": 2.4330000000000003e-06, "loss": 2.3363, "step": 4866 }, { "epoch": 1.5257053291536051, "grad_norm": 17.136247634887695, "learning_rate": 2.4335000000000005e-06, "loss": 2.4411, "step": 4867 }, { "epoch": 1.5260188087774296, "grad_norm": 11.98731517791748, "learning_rate": 2.4340000000000003e-06, "loss": 2.3687, "step": 4868 }, { "epoch": 1.526332288401254, "grad_norm": 18.351879119873047, "learning_rate": 2.4345e-06, "loss": 2.4216, "step": 4869 }, { "epoch": 1.5266457680250785, "grad_norm": 15.112627029418945, "learning_rate": 2.435e-06, "loss": 2.2907, "step": 4870 }, { "epoch": 1.526959247648903, "grad_norm": 33.07174301147461, "learning_rate": 2.4355000000000002e-06, "loss": 2.5456, "step": 4871 }, { "epoch": 1.5272727272727273, "grad_norm": 13.481283187866211, "learning_rate": 2.4360000000000005e-06, "loss": 2.4042, "step": 4872 }, { "epoch": 1.5275862068965518, "grad_norm": 16.563289642333984, "learning_rate": 2.4365000000000003e-06, "loss": 2.2126, "step": 4873 }, { "epoch": 1.5278996865203762, "grad_norm": 15.1494779586792, "learning_rate": 2.437e-06, "loss": 2.4135, "step": 4874 }, { "epoch": 1.5282131661442007, "grad_norm": 27.80925941467285, "learning_rate": 2.4375e-06, "loss": 2.4168, "step": 4875 }, { "epoch": 1.528526645768025, "grad_norm": 13.605854034423828, "learning_rate": 2.438e-06, "loss": 2.4283, "step": 4876 }, { "epoch": 1.5288401253918495, "grad_norm": 24.888622283935547, "learning_rate": 2.4385000000000005e-06, "loss": 2.242, "step": 4877 }, { "epoch": 1.529153605015674, "grad_norm": 23.643573760986328, "learning_rate": 2.4390000000000003e-06, "loss": 2.5237, "step": 4878 }, { "epoch": 1.5294670846394984, "grad_norm": 14.883064270019531, "learning_rate": 2.4395e-06, "loss": 2.3608, "step": 4879 }, { "epoch": 1.5297805642633229, "grad_norm": 38.991031646728516, "learning_rate": 2.4400000000000004e-06, "loss": 2.3225, "step": 4880 }, { "epoch": 1.5300940438871473, "grad_norm": 112.9505386352539, "learning_rate": 2.4405e-06, "loss": 2.8802, "step": 4881 }, { "epoch": 1.5304075235109718, "grad_norm": 8.81878662109375, "learning_rate": 2.4410000000000004e-06, "loss": 2.2658, "step": 4882 }, { "epoch": 1.5307210031347962, "grad_norm": 27.886585235595703, "learning_rate": 2.4415000000000003e-06, "loss": 2.8646, "step": 4883 }, { "epoch": 1.5310344827586206, "grad_norm": 17.828750610351562, "learning_rate": 2.442e-06, "loss": 2.484, "step": 4884 }, { "epoch": 1.531347962382445, "grad_norm": 16.510709762573242, "learning_rate": 2.4425000000000003e-06, "loss": 2.273, "step": 4885 }, { "epoch": 1.5316614420062695, "grad_norm": 17.356098175048828, "learning_rate": 2.443e-06, "loss": 2.3664, "step": 4886 }, { "epoch": 1.531974921630094, "grad_norm": 22.031627655029297, "learning_rate": 2.4435000000000004e-06, "loss": 2.6871, "step": 4887 }, { "epoch": 1.5322884012539184, "grad_norm": 15.652859687805176, "learning_rate": 2.4440000000000002e-06, "loss": 2.469, "step": 4888 }, { "epoch": 1.5326018808777429, "grad_norm": 18.956735610961914, "learning_rate": 2.4445e-06, "loss": 2.2246, "step": 4889 }, { "epoch": 1.5329153605015673, "grad_norm": 16.19228744506836, "learning_rate": 2.4450000000000003e-06, "loss": 2.7894, "step": 4890 }, { "epoch": 1.5332288401253917, "grad_norm": 36.512962341308594, "learning_rate": 2.4455e-06, "loss": 3.2316, "step": 4891 }, { "epoch": 1.5335423197492162, "grad_norm": 194.37013244628906, "learning_rate": 2.4460000000000004e-06, "loss": 3.3238, "step": 4892 }, { "epoch": 1.5338557993730406, "grad_norm": 27.753740310668945, "learning_rate": 2.4465e-06, "loss": 2.5502, "step": 4893 }, { "epoch": 1.534169278996865, "grad_norm": 18.22922706604004, "learning_rate": 2.447e-06, "loss": 2.3534, "step": 4894 }, { "epoch": 1.5344827586206895, "grad_norm": 14.241109848022461, "learning_rate": 2.4475000000000003e-06, "loss": 2.2446, "step": 4895 }, { "epoch": 1.534796238244514, "grad_norm": 38.08067321777344, "learning_rate": 2.448e-06, "loss": 2.9727, "step": 4896 }, { "epoch": 1.5351097178683386, "grad_norm": 52.77937698364258, "learning_rate": 2.4485000000000004e-06, "loss": 2.7533, "step": 4897 }, { "epoch": 1.535423197492163, "grad_norm": 19.61460304260254, "learning_rate": 2.449e-06, "loss": 2.3417, "step": 4898 }, { "epoch": 1.5357366771159875, "grad_norm": 38.986412048339844, "learning_rate": 2.4495e-06, "loss": 2.6992, "step": 4899 }, { "epoch": 1.536050156739812, "grad_norm": 33.46477508544922, "learning_rate": 2.4500000000000003e-06, "loss": 2.2636, "step": 4900 }, { "epoch": 1.5363636363636364, "grad_norm": 59.14734649658203, "learning_rate": 2.4505e-06, "loss": 2.4793, "step": 4901 }, { "epoch": 1.5366771159874608, "grad_norm": 16.077693939208984, "learning_rate": 2.4510000000000003e-06, "loss": 2.2937, "step": 4902 }, { "epoch": 1.5369905956112853, "grad_norm": 25.643287658691406, "learning_rate": 2.4515e-06, "loss": 2.6153, "step": 4903 }, { "epoch": 1.5373040752351097, "grad_norm": 25.702070236206055, "learning_rate": 2.4520000000000004e-06, "loss": 2.3666, "step": 4904 }, { "epoch": 1.5376175548589341, "grad_norm": 44.60346603393555, "learning_rate": 2.4525000000000002e-06, "loss": 2.481, "step": 4905 }, { "epoch": 1.5379310344827586, "grad_norm": 28.352134704589844, "learning_rate": 2.453e-06, "loss": 2.6935, "step": 4906 }, { "epoch": 1.538244514106583, "grad_norm": 15.22026252746582, "learning_rate": 2.4535000000000003e-06, "loss": 2.3636, "step": 4907 }, { "epoch": 1.5385579937304075, "grad_norm": 29.389680862426758, "learning_rate": 2.454e-06, "loss": 2.2899, "step": 4908 }, { "epoch": 1.5388714733542321, "grad_norm": 16.111085891723633, "learning_rate": 2.4545000000000004e-06, "loss": 2.5567, "step": 4909 }, { "epoch": 1.5391849529780566, "grad_norm": 17.4854679107666, "learning_rate": 2.4550000000000002e-06, "loss": 2.4468, "step": 4910 }, { "epoch": 1.539498432601881, "grad_norm": 26.088729858398438, "learning_rate": 2.4555e-06, "loss": 2.4288, "step": 4911 }, { "epoch": 1.5398119122257055, "grad_norm": 18.86445426940918, "learning_rate": 2.4560000000000003e-06, "loss": 2.2143, "step": 4912 }, { "epoch": 1.54012539184953, "grad_norm": 11.737680435180664, "learning_rate": 2.4565e-06, "loss": 2.4803, "step": 4913 }, { "epoch": 1.5404388714733543, "grad_norm": 13.126084327697754, "learning_rate": 2.4570000000000004e-06, "loss": 2.6635, "step": 4914 }, { "epoch": 1.5407523510971788, "grad_norm": 14.669913291931152, "learning_rate": 2.4575e-06, "loss": 2.1352, "step": 4915 }, { "epoch": 1.5410658307210032, "grad_norm": 8.558469772338867, "learning_rate": 2.458e-06, "loss": 2.4981, "step": 4916 }, { "epoch": 1.5413793103448277, "grad_norm": 37.0605583190918, "learning_rate": 2.4585000000000003e-06, "loss": 2.7061, "step": 4917 }, { "epoch": 1.541692789968652, "grad_norm": 25.720998764038086, "learning_rate": 2.459e-06, "loss": 2.3515, "step": 4918 }, { "epoch": 1.5420062695924766, "grad_norm": 17.545047760009766, "learning_rate": 2.4595000000000003e-06, "loss": 2.2531, "step": 4919 }, { "epoch": 1.542319749216301, "grad_norm": 14.890344619750977, "learning_rate": 2.46e-06, "loss": 2.5966, "step": 4920 }, { "epoch": 1.5426332288401254, "grad_norm": 15.457768440246582, "learning_rate": 2.4605e-06, "loss": 2.1787, "step": 4921 }, { "epoch": 1.5429467084639499, "grad_norm": 10.848100662231445, "learning_rate": 2.4610000000000002e-06, "loss": 2.1762, "step": 4922 }, { "epoch": 1.5432601880877743, "grad_norm": 18.135581970214844, "learning_rate": 2.4615000000000005e-06, "loss": 2.0205, "step": 4923 }, { "epoch": 1.5435736677115988, "grad_norm": 16.26011848449707, "learning_rate": 2.4620000000000003e-06, "loss": 2.4316, "step": 4924 }, { "epoch": 1.5438871473354232, "grad_norm": 19.314348220825195, "learning_rate": 2.4625e-06, "loss": 2.4738, "step": 4925 }, { "epoch": 1.5442006269592476, "grad_norm": 21.843786239624023, "learning_rate": 2.463e-06, "loss": 2.2117, "step": 4926 }, { "epoch": 1.544514106583072, "grad_norm": 13.87489128112793, "learning_rate": 2.4635000000000002e-06, "loss": 2.2631, "step": 4927 }, { "epoch": 1.5448275862068965, "grad_norm": 19.082761764526367, "learning_rate": 2.4640000000000005e-06, "loss": 2.4595, "step": 4928 }, { "epoch": 1.545141065830721, "grad_norm": 49.517860412597656, "learning_rate": 2.4645000000000003e-06, "loss": 2.4809, "step": 4929 }, { "epoch": 1.5454545454545454, "grad_norm": 12.239781379699707, "learning_rate": 2.465e-06, "loss": 2.3829, "step": 4930 }, { "epoch": 1.5457680250783699, "grad_norm": 17.61081886291504, "learning_rate": 2.4655e-06, "loss": 2.2738, "step": 4931 }, { "epoch": 1.5460815047021943, "grad_norm": 9.764055252075195, "learning_rate": 2.466e-06, "loss": 2.2263, "step": 4932 }, { "epoch": 1.5463949843260187, "grad_norm": 16.245092391967773, "learning_rate": 2.4665000000000004e-06, "loss": 2.2703, "step": 4933 }, { "epoch": 1.5467084639498432, "grad_norm": 47.172786712646484, "learning_rate": 2.4670000000000003e-06, "loss": 2.4811, "step": 4934 }, { "epoch": 1.5470219435736676, "grad_norm": 36.98639678955078, "learning_rate": 2.4675e-06, "loss": 2.4616, "step": 4935 }, { "epoch": 1.547335423197492, "grad_norm": 88.0708236694336, "learning_rate": 2.468e-06, "loss": 3.2185, "step": 4936 }, { "epoch": 1.5476489028213165, "grad_norm": 20.56644058227539, "learning_rate": 2.4685e-06, "loss": 2.3528, "step": 4937 }, { "epoch": 1.547962382445141, "grad_norm": 22.679750442504883, "learning_rate": 2.4690000000000004e-06, "loss": 2.2352, "step": 4938 }, { "epoch": 1.5482758620689654, "grad_norm": 19.43711280822754, "learning_rate": 2.4695000000000002e-06, "loss": 2.2477, "step": 4939 }, { "epoch": 1.5485893416927898, "grad_norm": 55.72245788574219, "learning_rate": 2.47e-06, "loss": 2.6602, "step": 4940 }, { "epoch": 1.5489028213166143, "grad_norm": 44.908382415771484, "learning_rate": 2.4705000000000003e-06, "loss": 2.7444, "step": 4941 }, { "epoch": 1.5492163009404387, "grad_norm": 13.042838096618652, "learning_rate": 2.471e-06, "loss": 2.6583, "step": 4942 }, { "epoch": 1.5495297805642632, "grad_norm": 20.27886962890625, "learning_rate": 2.4715000000000004e-06, "loss": 2.3545, "step": 4943 }, { "epoch": 1.5498432601880878, "grad_norm": 17.01507568359375, "learning_rate": 2.4720000000000002e-06, "loss": 2.388, "step": 4944 }, { "epoch": 1.5501567398119123, "grad_norm": 12.890861511230469, "learning_rate": 2.4725e-06, "loss": 2.4163, "step": 4945 }, { "epoch": 1.5504702194357367, "grad_norm": 70.1278305053711, "learning_rate": 2.4730000000000003e-06, "loss": 2.9749, "step": 4946 }, { "epoch": 1.5507836990595611, "grad_norm": 21.87778091430664, "learning_rate": 2.4735e-06, "loss": 2.2962, "step": 4947 }, { "epoch": 1.5510971786833856, "grad_norm": 29.6918888092041, "learning_rate": 2.4740000000000004e-06, "loss": 2.5881, "step": 4948 }, { "epoch": 1.55141065830721, "grad_norm": 30.08953285217285, "learning_rate": 2.4745e-06, "loss": 2.466, "step": 4949 }, { "epoch": 1.5517241379310345, "grad_norm": 122.53662872314453, "learning_rate": 2.475e-06, "loss": 2.8568, "step": 4950 }, { "epoch": 1.552037617554859, "grad_norm": 11.565924644470215, "learning_rate": 2.4755000000000003e-06, "loss": 2.0998, "step": 4951 }, { "epoch": 1.5523510971786834, "grad_norm": 14.999190330505371, "learning_rate": 2.476e-06, "loss": 2.2514, "step": 4952 }, { "epoch": 1.5526645768025078, "grad_norm": 49.61758041381836, "learning_rate": 2.4765000000000003e-06, "loss": 2.2886, "step": 4953 }, { "epoch": 1.5529780564263322, "grad_norm": 38.12818908691406, "learning_rate": 2.477e-06, "loss": 2.3246, "step": 4954 }, { "epoch": 1.5532915360501567, "grad_norm": 19.00833511352539, "learning_rate": 2.4775e-06, "loss": 2.3158, "step": 4955 }, { "epoch": 1.5536050156739813, "grad_norm": 10.164694786071777, "learning_rate": 2.4780000000000002e-06, "loss": 2.3208, "step": 4956 }, { "epoch": 1.5539184952978058, "grad_norm": 19.940147399902344, "learning_rate": 2.4785e-06, "loss": 2.2967, "step": 4957 }, { "epoch": 1.5542319749216302, "grad_norm": 46.21580123901367, "learning_rate": 2.4790000000000003e-06, "loss": 2.3477, "step": 4958 }, { "epoch": 1.5545454545454547, "grad_norm": 13.118972778320312, "learning_rate": 2.4795e-06, "loss": 1.9755, "step": 4959 }, { "epoch": 1.5548589341692791, "grad_norm": 42.64442825317383, "learning_rate": 2.4800000000000004e-06, "loss": 2.5682, "step": 4960 }, { "epoch": 1.5551724137931036, "grad_norm": 13.624639511108398, "learning_rate": 2.4805000000000002e-06, "loss": 2.4254, "step": 4961 }, { "epoch": 1.555485893416928, "grad_norm": 21.184711456298828, "learning_rate": 2.481e-06, "loss": 2.2923, "step": 4962 }, { "epoch": 1.5557993730407524, "grad_norm": 15.5918607711792, "learning_rate": 2.4815000000000003e-06, "loss": 2.0308, "step": 4963 }, { "epoch": 1.5561128526645769, "grad_norm": 19.084026336669922, "learning_rate": 2.482e-06, "loss": 2.5449, "step": 4964 }, { "epoch": 1.5564263322884013, "grad_norm": 14.544201850891113, "learning_rate": 2.4825000000000004e-06, "loss": 2.2216, "step": 4965 }, { "epoch": 1.5567398119122258, "grad_norm": 51.081851959228516, "learning_rate": 2.483e-06, "loss": 2.478, "step": 4966 }, { "epoch": 1.5570532915360502, "grad_norm": 45.603790283203125, "learning_rate": 2.4835e-06, "loss": 2.3317, "step": 4967 }, { "epoch": 1.5573667711598747, "grad_norm": 23.09310531616211, "learning_rate": 2.4840000000000003e-06, "loss": 2.3886, "step": 4968 }, { "epoch": 1.557680250783699, "grad_norm": 36.93960952758789, "learning_rate": 2.4845e-06, "loss": 2.4891, "step": 4969 }, { "epoch": 1.5579937304075235, "grad_norm": 34.09645080566406, "learning_rate": 2.4850000000000003e-06, "loss": 3.6792, "step": 4970 }, { "epoch": 1.558307210031348, "grad_norm": 21.661108016967773, "learning_rate": 2.4855e-06, "loss": 2.4421, "step": 4971 }, { "epoch": 1.5586206896551724, "grad_norm": 31.483070373535156, "learning_rate": 2.486e-06, "loss": 3.076, "step": 4972 }, { "epoch": 1.5589341692789969, "grad_norm": 12.45557689666748, "learning_rate": 2.4865000000000002e-06, "loss": 2.2351, "step": 4973 }, { "epoch": 1.5592476489028213, "grad_norm": 19.75053596496582, "learning_rate": 2.487e-06, "loss": 2.4616, "step": 4974 }, { "epoch": 1.5595611285266457, "grad_norm": 26.347949981689453, "learning_rate": 2.4875000000000003e-06, "loss": 2.1774, "step": 4975 }, { "epoch": 1.5598746081504702, "grad_norm": 55.684993743896484, "learning_rate": 2.488e-06, "loss": 2.4876, "step": 4976 }, { "epoch": 1.5601880877742946, "grad_norm": 46.71628952026367, "learning_rate": 2.4885e-06, "loss": 2.73, "step": 4977 }, { "epoch": 1.560501567398119, "grad_norm": 18.326326370239258, "learning_rate": 2.4890000000000002e-06, "loss": 2.3658, "step": 4978 }, { "epoch": 1.5608150470219435, "grad_norm": 14.326557159423828, "learning_rate": 2.4895e-06, "loss": 2.0581, "step": 4979 }, { "epoch": 1.561128526645768, "grad_norm": 16.678058624267578, "learning_rate": 2.4900000000000003e-06, "loss": 2.356, "step": 4980 }, { "epoch": 1.5614420062695924, "grad_norm": 60.66374588012695, "learning_rate": 2.4905e-06, "loss": 2.6371, "step": 4981 }, { "epoch": 1.5617554858934168, "grad_norm": 19.981624603271484, "learning_rate": 2.491e-06, "loss": 2.2885, "step": 4982 }, { "epoch": 1.5620689655172413, "grad_norm": 161.30465698242188, "learning_rate": 2.4915e-06, "loss": 2.5611, "step": 4983 }, { "epoch": 1.5623824451410657, "grad_norm": 21.220386505126953, "learning_rate": 2.4920000000000005e-06, "loss": 2.3193, "step": 4984 }, { "epoch": 1.5626959247648902, "grad_norm": 114.13520050048828, "learning_rate": 2.4925000000000003e-06, "loss": 2.5983, "step": 4985 }, { "epoch": 1.5630094043887146, "grad_norm": 23.770221710205078, "learning_rate": 2.493e-06, "loss": 2.325, "step": 4986 }, { "epoch": 1.563322884012539, "grad_norm": 18.08496856689453, "learning_rate": 2.4935e-06, "loss": 2.4323, "step": 4987 }, { "epoch": 1.5636363636363635, "grad_norm": 35.3812370300293, "learning_rate": 2.494e-06, "loss": 2.4329, "step": 4988 }, { "epoch": 1.563949843260188, "grad_norm": 21.316783905029297, "learning_rate": 2.4945000000000004e-06, "loss": 2.3539, "step": 4989 }, { "epoch": 1.5642633228840124, "grad_norm": 17.182456970214844, "learning_rate": 2.4950000000000003e-06, "loss": 2.3882, "step": 4990 }, { "epoch": 1.564576802507837, "grad_norm": 18.844646453857422, "learning_rate": 2.4955e-06, "loss": 2.2841, "step": 4991 }, { "epoch": 1.5648902821316615, "grad_norm": 38.493377685546875, "learning_rate": 2.496e-06, "loss": 2.9194, "step": 4992 }, { "epoch": 1.565203761755486, "grad_norm": 14.904982566833496, "learning_rate": 2.4965e-06, "loss": 2.2876, "step": 4993 }, { "epoch": 1.5655172413793104, "grad_norm": 16.828540802001953, "learning_rate": 2.4970000000000004e-06, "loss": 2.4224, "step": 4994 }, { "epoch": 1.5658307210031348, "grad_norm": 25.141571044921875, "learning_rate": 2.4975000000000002e-06, "loss": 2.3427, "step": 4995 }, { "epoch": 1.5661442006269592, "grad_norm": 42.96630859375, "learning_rate": 2.498e-06, "loss": 2.5007, "step": 4996 }, { "epoch": 1.5664576802507837, "grad_norm": 39.685176849365234, "learning_rate": 2.4985000000000003e-06, "loss": 2.5697, "step": 4997 }, { "epoch": 1.5667711598746081, "grad_norm": 34.34861755371094, "learning_rate": 2.499e-06, "loss": 2.1571, "step": 4998 }, { "epoch": 1.5670846394984326, "grad_norm": 33.744346618652344, "learning_rate": 2.4995000000000004e-06, "loss": 2.2946, "step": 4999 }, { "epoch": 1.567398119122257, "grad_norm": 22.578693389892578, "learning_rate": 2.5e-06, "loss": 2.2518, "step": 5000 }, { "epoch": 1.5677115987460815, "grad_norm": 26.208200454711914, "learning_rate": 2.5005e-06, "loss": 2.383, "step": 5001 }, { "epoch": 1.5680250783699061, "grad_norm": 27.605117797851562, "learning_rate": 2.5010000000000003e-06, "loss": 2.265, "step": 5002 }, { "epoch": 1.5683385579937306, "grad_norm": 56.79507064819336, "learning_rate": 2.5015e-06, "loss": 2.8529, "step": 5003 }, { "epoch": 1.568652037617555, "grad_norm": 16.00729751586914, "learning_rate": 2.502e-06, "loss": 2.2934, "step": 5004 }, { "epoch": 1.5689655172413794, "grad_norm": 15.617908477783203, "learning_rate": 2.5024999999999998e-06, "loss": 2.713, "step": 5005 }, { "epoch": 1.569278996865204, "grad_norm": 25.584407806396484, "learning_rate": 2.5030000000000004e-06, "loss": 2.6503, "step": 5006 }, { "epoch": 1.5695924764890283, "grad_norm": 34.239315032958984, "learning_rate": 2.5035000000000003e-06, "loss": 2.1223, "step": 5007 }, { "epoch": 1.5699059561128528, "grad_norm": 42.1166877746582, "learning_rate": 2.5040000000000005e-06, "loss": 2.6302, "step": 5008 }, { "epoch": 1.5702194357366772, "grad_norm": 24.463972091674805, "learning_rate": 2.5045000000000003e-06, "loss": 2.5372, "step": 5009 }, { "epoch": 1.5705329153605017, "grad_norm": 50.24067687988281, "learning_rate": 2.505e-06, "loss": 2.6691, "step": 5010 }, { "epoch": 1.570846394984326, "grad_norm": 33.488521575927734, "learning_rate": 2.5055e-06, "loss": 2.6461, "step": 5011 }, { "epoch": 1.5711598746081505, "grad_norm": 13.341475486755371, "learning_rate": 2.5060000000000002e-06, "loss": 2.2295, "step": 5012 }, { "epoch": 1.571473354231975, "grad_norm": 20.728574752807617, "learning_rate": 2.5065e-06, "loss": 2.5787, "step": 5013 }, { "epoch": 1.5717868338557994, "grad_norm": 21.363683700561523, "learning_rate": 2.507e-06, "loss": 2.4587, "step": 5014 }, { "epoch": 1.5721003134796239, "grad_norm": 23.327157974243164, "learning_rate": 2.5075e-06, "loss": 2.6306, "step": 5015 }, { "epoch": 1.5724137931034483, "grad_norm": 12.774349212646484, "learning_rate": 2.5080000000000004e-06, "loss": 2.285, "step": 5016 }, { "epoch": 1.5727272727272728, "grad_norm": 15.191110610961914, "learning_rate": 2.5085e-06, "loss": 2.406, "step": 5017 }, { "epoch": 1.5730407523510972, "grad_norm": 36.179500579833984, "learning_rate": 2.5090000000000005e-06, "loss": 2.7932, "step": 5018 }, { "epoch": 1.5733542319749216, "grad_norm": 16.943391799926758, "learning_rate": 2.5095000000000003e-06, "loss": 2.5186, "step": 5019 }, { "epoch": 1.573667711598746, "grad_norm": 51.045005798339844, "learning_rate": 2.51e-06, "loss": 2.5958, "step": 5020 }, { "epoch": 1.5739811912225705, "grad_norm": 38.750431060791016, "learning_rate": 2.5105000000000004e-06, "loss": 3.3926, "step": 5021 }, { "epoch": 1.574294670846395, "grad_norm": 12.69948959350586, "learning_rate": 2.511e-06, "loss": 2.1353, "step": 5022 }, { "epoch": 1.5746081504702194, "grad_norm": 24.718772888183594, "learning_rate": 2.5115e-06, "loss": 2.467, "step": 5023 }, { "epoch": 1.5749216300940438, "grad_norm": 17.55669593811035, "learning_rate": 2.512e-06, "loss": 2.4588, "step": 5024 }, { "epoch": 1.5752351097178683, "grad_norm": 7.0970587730407715, "learning_rate": 2.5125e-06, "loss": 2.311, "step": 5025 }, { "epoch": 1.5755485893416927, "grad_norm": 25.72663688659668, "learning_rate": 2.5130000000000003e-06, "loss": 2.9244, "step": 5026 }, { "epoch": 1.5758620689655172, "grad_norm": 29.8355655670166, "learning_rate": 2.5135000000000006e-06, "loss": 2.3327, "step": 5027 }, { "epoch": 1.5761755485893416, "grad_norm": 14.82983684539795, "learning_rate": 2.5140000000000004e-06, "loss": 2.4196, "step": 5028 }, { "epoch": 1.576489028213166, "grad_norm": 23.94029998779297, "learning_rate": 2.5145000000000002e-06, "loss": 2.4763, "step": 5029 }, { "epoch": 1.5768025078369905, "grad_norm": 18.115259170532227, "learning_rate": 2.515e-06, "loss": 2.4923, "step": 5030 }, { "epoch": 1.577115987460815, "grad_norm": 12.026752471923828, "learning_rate": 2.5155000000000003e-06, "loss": 2.3874, "step": 5031 }, { "epoch": 1.5774294670846394, "grad_norm": 29.68210792541504, "learning_rate": 2.516e-06, "loss": 2.8552, "step": 5032 }, { "epoch": 1.5777429467084638, "grad_norm": 15.086202621459961, "learning_rate": 2.5165e-06, "loss": 2.1304, "step": 5033 }, { "epoch": 1.5780564263322883, "grad_norm": 24.435060501098633, "learning_rate": 2.5169999999999998e-06, "loss": 2.339, "step": 5034 }, { "epoch": 1.5783699059561127, "grad_norm": 18.48145866394043, "learning_rate": 2.5175e-06, "loss": 2.3635, "step": 5035 }, { "epoch": 1.5786833855799371, "grad_norm": 21.17127227783203, "learning_rate": 2.5180000000000003e-06, "loss": 2.18, "step": 5036 }, { "epoch": 1.5789968652037616, "grad_norm": 26.255979537963867, "learning_rate": 2.5185000000000005e-06, "loss": 2.5204, "step": 5037 }, { "epoch": 1.5793103448275863, "grad_norm": 260.4898986816406, "learning_rate": 2.5190000000000004e-06, "loss": 2.4303, "step": 5038 }, { "epoch": 1.5796238244514107, "grad_norm": 53.10416793823242, "learning_rate": 2.5195e-06, "loss": 2.4221, "step": 5039 }, { "epoch": 1.5799373040752351, "grad_norm": 18.819931030273438, "learning_rate": 2.52e-06, "loss": 2.5317, "step": 5040 }, { "epoch": 1.5802507836990596, "grad_norm": 26.026790618896484, "learning_rate": 2.5205000000000003e-06, "loss": 1.9606, "step": 5041 }, { "epoch": 1.580564263322884, "grad_norm": 23.0932559967041, "learning_rate": 2.521e-06, "loss": 2.3511, "step": 5042 }, { "epoch": 1.5808777429467085, "grad_norm": 20.482589721679688, "learning_rate": 2.5215e-06, "loss": 2.6349, "step": 5043 }, { "epoch": 1.581191222570533, "grad_norm": 22.782913208007812, "learning_rate": 2.522e-06, "loss": 2.5406, "step": 5044 }, { "epoch": 1.5815047021943573, "grad_norm": 17.503292083740234, "learning_rate": 2.5225e-06, "loss": 2.0932, "step": 5045 }, { "epoch": 1.5818181818181818, "grad_norm": 11.492012023925781, "learning_rate": 2.5230000000000002e-06, "loss": 2.4, "step": 5046 }, { "epoch": 1.5821316614420062, "grad_norm": 13.962905883789062, "learning_rate": 2.5235000000000005e-06, "loss": 2.2931, "step": 5047 }, { "epoch": 1.5824451410658307, "grad_norm": 13.378860473632812, "learning_rate": 2.5240000000000003e-06, "loss": 2.0573, "step": 5048 }, { "epoch": 1.5827586206896553, "grad_norm": 35.7122802734375, "learning_rate": 2.5245e-06, "loss": 2.5505, "step": 5049 }, { "epoch": 1.5830721003134798, "grad_norm": 19.624452590942383, "learning_rate": 2.5250000000000004e-06, "loss": 2.2378, "step": 5050 }, { "epoch": 1.5833855799373042, "grad_norm": 41.233428955078125, "learning_rate": 2.5255000000000002e-06, "loss": 2.5732, "step": 5051 }, { "epoch": 1.5836990595611287, "grad_norm": 15.03004264831543, "learning_rate": 2.526e-06, "loss": 2.6814, "step": 5052 }, { "epoch": 1.584012539184953, "grad_norm": 358.0602722167969, "learning_rate": 2.5265e-06, "loss": 2.7063, "step": 5053 }, { "epoch": 1.5843260188087775, "grad_norm": 33.94559860229492, "learning_rate": 2.527e-06, "loss": 2.775, "step": 5054 }, { "epoch": 1.584639498432602, "grad_norm": 14.883088111877441, "learning_rate": 2.5275e-06, "loss": 2.8067, "step": 5055 }, { "epoch": 1.5849529780564264, "grad_norm": 51.98908615112305, "learning_rate": 2.5280000000000006e-06, "loss": 2.34, "step": 5056 }, { "epoch": 1.5852664576802509, "grad_norm": 21.30936050415039, "learning_rate": 2.5285000000000004e-06, "loss": 2.4628, "step": 5057 }, { "epoch": 1.5855799373040753, "grad_norm": 18.566699981689453, "learning_rate": 2.5290000000000003e-06, "loss": 2.4734, "step": 5058 }, { "epoch": 1.5858934169278998, "grad_norm": 34.335670471191406, "learning_rate": 2.5295e-06, "loss": 2.5809, "step": 5059 }, { "epoch": 1.5862068965517242, "grad_norm": 27.54088592529297, "learning_rate": 2.5300000000000003e-06, "loss": 2.2176, "step": 5060 }, { "epoch": 1.5865203761755486, "grad_norm": 103.77539825439453, "learning_rate": 2.5305e-06, "loss": 2.7314, "step": 5061 }, { "epoch": 1.586833855799373, "grad_norm": 25.50882339477539, "learning_rate": 2.531e-06, "loss": 2.2157, "step": 5062 }, { "epoch": 1.5871473354231975, "grad_norm": 90.0643539428711, "learning_rate": 2.5315000000000002e-06, "loss": 2.1978, "step": 5063 }, { "epoch": 1.587460815047022, "grad_norm": 69.36904907226562, "learning_rate": 2.532e-06, "loss": 2.9566, "step": 5064 }, { "epoch": 1.5877742946708464, "grad_norm": 38.16106414794922, "learning_rate": 2.5325e-06, "loss": 2.3429, "step": 5065 }, { "epoch": 1.5880877742946709, "grad_norm": 29.278573989868164, "learning_rate": 2.5330000000000006e-06, "loss": 2.2119, "step": 5066 }, { "epoch": 1.5884012539184953, "grad_norm": 89.26683044433594, "learning_rate": 2.5335000000000004e-06, "loss": 2.4715, "step": 5067 }, { "epoch": 1.5887147335423197, "grad_norm": 21.57378578186035, "learning_rate": 2.5340000000000002e-06, "loss": 2.3059, "step": 5068 }, { "epoch": 1.5890282131661442, "grad_norm": 35.4509162902832, "learning_rate": 2.5345000000000005e-06, "loss": 2.3637, "step": 5069 }, { "epoch": 1.5893416927899686, "grad_norm": 21.950578689575195, "learning_rate": 2.5350000000000003e-06, "loss": 2.2167, "step": 5070 }, { "epoch": 1.589655172413793, "grad_norm": 52.572540283203125, "learning_rate": 2.5355e-06, "loss": 2.7111, "step": 5071 }, { "epoch": 1.5899686520376175, "grad_norm": 77.87381744384766, "learning_rate": 2.536e-06, "loss": 2.7671, "step": 5072 }, { "epoch": 1.590282131661442, "grad_norm": 15.252903938293457, "learning_rate": 2.5365e-06, "loss": 2.1193, "step": 5073 }, { "epoch": 1.5905956112852664, "grad_norm": 25.91535758972168, "learning_rate": 2.537e-06, "loss": 2.7424, "step": 5074 }, { "epoch": 1.5909090909090908, "grad_norm": 25.11017608642578, "learning_rate": 2.5375e-06, "loss": 2.8389, "step": 5075 }, { "epoch": 1.5912225705329153, "grad_norm": 15.80587100982666, "learning_rate": 2.5380000000000005e-06, "loss": 2.4494, "step": 5076 }, { "epoch": 1.5915360501567397, "grad_norm": 97.02642822265625, "learning_rate": 2.5385000000000003e-06, "loss": 2.6632, "step": 5077 }, { "epoch": 1.5918495297805642, "grad_norm": 16.840496063232422, "learning_rate": 2.539e-06, "loss": 2.4333, "step": 5078 }, { "epoch": 1.5921630094043886, "grad_norm": 27.470895767211914, "learning_rate": 2.5395000000000004e-06, "loss": 2.7408, "step": 5079 }, { "epoch": 1.592476489028213, "grad_norm": 11.823321342468262, "learning_rate": 2.5400000000000002e-06, "loss": 2.129, "step": 5080 }, { "epoch": 1.5927899686520375, "grad_norm": 43.7274169921875, "learning_rate": 2.5405e-06, "loss": 2.368, "step": 5081 }, { "epoch": 1.593103448275862, "grad_norm": 12.855867385864258, "learning_rate": 2.5410000000000003e-06, "loss": 2.1715, "step": 5082 }, { "epoch": 1.5934169278996864, "grad_norm": 12.842061996459961, "learning_rate": 2.5415e-06, "loss": 2.0901, "step": 5083 }, { "epoch": 1.5937304075235108, "grad_norm": 33.203914642333984, "learning_rate": 2.542e-06, "loss": 2.51, "step": 5084 }, { "epoch": 1.5940438871473355, "grad_norm": 15.070207595825195, "learning_rate": 2.5425e-06, "loss": 2.3985, "step": 5085 }, { "epoch": 1.59435736677116, "grad_norm": 24.810152053833008, "learning_rate": 2.5430000000000005e-06, "loss": 2.2131, "step": 5086 }, { "epoch": 1.5946708463949844, "grad_norm": 22.131668090820312, "learning_rate": 2.5435000000000003e-06, "loss": 2.276, "step": 5087 }, { "epoch": 1.5949843260188088, "grad_norm": 19.9857234954834, "learning_rate": 2.5440000000000005e-06, "loss": 2.5681, "step": 5088 }, { "epoch": 1.5952978056426332, "grad_norm": 33.42539978027344, "learning_rate": 2.5445000000000004e-06, "loss": 3.0566, "step": 5089 }, { "epoch": 1.5956112852664577, "grad_norm": 41.5034065246582, "learning_rate": 2.545e-06, "loss": 2.5322, "step": 5090 }, { "epoch": 1.5959247648902821, "grad_norm": 9.396846771240234, "learning_rate": 2.5455e-06, "loss": 2.5493, "step": 5091 }, { "epoch": 1.5962382445141066, "grad_norm": 17.213747024536133, "learning_rate": 2.5460000000000003e-06, "loss": 2.3311, "step": 5092 }, { "epoch": 1.596551724137931, "grad_norm": 25.562578201293945, "learning_rate": 2.5465e-06, "loss": 2.5074, "step": 5093 }, { "epoch": 1.5968652037617554, "grad_norm": 36.171207427978516, "learning_rate": 2.547e-06, "loss": 2.9993, "step": 5094 }, { "epoch": 1.59717868338558, "grad_norm": 19.676252365112305, "learning_rate": 2.5475e-06, "loss": 2.343, "step": 5095 }, { "epoch": 1.5974921630094046, "grad_norm": 14.119227409362793, "learning_rate": 2.5480000000000004e-06, "loss": 2.7787, "step": 5096 }, { "epoch": 1.597805642633229, "grad_norm": 23.950241088867188, "learning_rate": 2.5485000000000002e-06, "loss": 2.3251, "step": 5097 }, { "epoch": 1.5981191222570534, "grad_norm": 23.108810424804688, "learning_rate": 2.5490000000000005e-06, "loss": 2.6356, "step": 5098 }, { "epoch": 1.5984326018808779, "grad_norm": 12.062274932861328, "learning_rate": 2.5495000000000003e-06, "loss": 2.2522, "step": 5099 }, { "epoch": 1.5987460815047023, "grad_norm": 17.99264907836914, "learning_rate": 2.55e-06, "loss": 2.1981, "step": 5100 }, { "epoch": 1.5990595611285268, "grad_norm": 75.7223129272461, "learning_rate": 2.5505000000000004e-06, "loss": 2.5135, "step": 5101 }, { "epoch": 1.5993730407523512, "grad_norm": 16.194011688232422, "learning_rate": 2.5510000000000002e-06, "loss": 2.4676, "step": 5102 }, { "epoch": 1.5996865203761756, "grad_norm": 12.641945838928223, "learning_rate": 2.5515e-06, "loss": 2.1026, "step": 5103 }, { "epoch": 1.6, "grad_norm": 13.797097206115723, "learning_rate": 2.552e-06, "loss": 2.1526, "step": 5104 }, { "epoch": 1.6003134796238245, "grad_norm": 26.577638626098633, "learning_rate": 2.5525e-06, "loss": 2.4505, "step": 5105 }, { "epoch": 1.600626959247649, "grad_norm": 46.32077407836914, "learning_rate": 2.5530000000000004e-06, "loss": 2.4153, "step": 5106 }, { "epoch": 1.6009404388714734, "grad_norm": 33.867061614990234, "learning_rate": 2.5535000000000006e-06, "loss": 2.4845, "step": 5107 }, { "epoch": 1.6012539184952979, "grad_norm": 29.492074966430664, "learning_rate": 2.5540000000000004e-06, "loss": 2.2374, "step": 5108 }, { "epoch": 1.6015673981191223, "grad_norm": 15.982016563415527, "learning_rate": 2.5545000000000003e-06, "loss": 2.1256, "step": 5109 }, { "epoch": 1.6018808777429467, "grad_norm": 23.65928840637207, "learning_rate": 2.555e-06, "loss": 2.5441, "step": 5110 }, { "epoch": 1.6021943573667712, "grad_norm": 31.558923721313477, "learning_rate": 2.5555000000000004e-06, "loss": 2.4304, "step": 5111 }, { "epoch": 1.6025078369905956, "grad_norm": 22.344242095947266, "learning_rate": 2.556e-06, "loss": 2.3139, "step": 5112 }, { "epoch": 1.60282131661442, "grad_norm": 70.99754333496094, "learning_rate": 2.5565e-06, "loss": 3.1882, "step": 5113 }, { "epoch": 1.6031347962382445, "grad_norm": 11.966924667358398, "learning_rate": 2.557e-06, "loss": 2.2407, "step": 5114 }, { "epoch": 1.603448275862069, "grad_norm": 14.561931610107422, "learning_rate": 2.5575e-06, "loss": 2.5646, "step": 5115 }, { "epoch": 1.6037617554858934, "grad_norm": 18.54060173034668, "learning_rate": 2.5580000000000003e-06, "loss": 2.6713, "step": 5116 }, { "epoch": 1.6040752351097178, "grad_norm": 18.346193313598633, "learning_rate": 2.5585000000000006e-06, "loss": 2.2139, "step": 5117 }, { "epoch": 1.6043887147335423, "grad_norm": 16.431509017944336, "learning_rate": 2.5590000000000004e-06, "loss": 2.22, "step": 5118 }, { "epoch": 1.6047021943573667, "grad_norm": 17.47113037109375, "learning_rate": 2.5595000000000002e-06, "loss": 2.3653, "step": 5119 }, { "epoch": 1.6050156739811912, "grad_norm": 28.77997398376465, "learning_rate": 2.56e-06, "loss": 2.9677, "step": 5120 }, { "epoch": 1.6053291536050156, "grad_norm": 14.653223991394043, "learning_rate": 2.5605000000000003e-06, "loss": 2.2911, "step": 5121 }, { "epoch": 1.60564263322884, "grad_norm": 20.274412155151367, "learning_rate": 2.561e-06, "loss": 2.3815, "step": 5122 }, { "epoch": 1.6059561128526645, "grad_norm": 23.261613845825195, "learning_rate": 2.5615e-06, "loss": 2.4761, "step": 5123 }, { "epoch": 1.606269592476489, "grad_norm": 26.230159759521484, "learning_rate": 2.562e-06, "loss": 2.4104, "step": 5124 }, { "epoch": 1.6065830721003134, "grad_norm": 12.218854904174805, "learning_rate": 2.5625e-06, "loss": 2.2321, "step": 5125 }, { "epoch": 1.6068965517241378, "grad_norm": 30.996965408325195, "learning_rate": 2.563e-06, "loss": 2.6327, "step": 5126 }, { "epoch": 1.6072100313479623, "grad_norm": 21.883953094482422, "learning_rate": 2.5635000000000005e-06, "loss": 2.3285, "step": 5127 }, { "epoch": 1.6075235109717867, "grad_norm": 22.208730697631836, "learning_rate": 2.5640000000000004e-06, "loss": 2.2301, "step": 5128 }, { "epoch": 1.6078369905956111, "grad_norm": 41.55567169189453, "learning_rate": 2.5645e-06, "loss": 2.5755, "step": 5129 }, { "epoch": 1.6081504702194356, "grad_norm": 36.985572814941406, "learning_rate": 2.5650000000000004e-06, "loss": 2.9383, "step": 5130 }, { "epoch": 1.6084639498432602, "grad_norm": 10.896848678588867, "learning_rate": 2.5655000000000003e-06, "loss": 2.278, "step": 5131 }, { "epoch": 1.6087774294670847, "grad_norm": 26.881319046020508, "learning_rate": 2.566e-06, "loss": 2.5914, "step": 5132 }, { "epoch": 1.6090909090909091, "grad_norm": 43.129730224609375, "learning_rate": 2.5665e-06, "loss": 2.4239, "step": 5133 }, { "epoch": 1.6094043887147336, "grad_norm": 35.279144287109375, "learning_rate": 2.567e-06, "loss": 2.339, "step": 5134 }, { "epoch": 1.609717868338558, "grad_norm": 20.479421615600586, "learning_rate": 2.5675e-06, "loss": 2.5922, "step": 5135 }, { "epoch": 1.6100313479623825, "grad_norm": 47.79603576660156, "learning_rate": 2.568e-06, "loss": 2.3022, "step": 5136 }, { "epoch": 1.610344827586207, "grad_norm": 21.276691436767578, "learning_rate": 2.5685000000000005e-06, "loss": 2.4997, "step": 5137 }, { "epoch": 1.6106583072100313, "grad_norm": 32.518619537353516, "learning_rate": 2.5690000000000003e-06, "loss": 2.3542, "step": 5138 }, { "epoch": 1.6109717868338558, "grad_norm": 11.362214088439941, "learning_rate": 2.5695e-06, "loss": 2.4764, "step": 5139 }, { "epoch": 1.6112852664576802, "grad_norm": 18.297727584838867, "learning_rate": 2.5700000000000004e-06, "loss": 2.359, "step": 5140 }, { "epoch": 1.6115987460815047, "grad_norm": 35.51220703125, "learning_rate": 2.5705e-06, "loss": 2.5266, "step": 5141 }, { "epoch": 1.611912225705329, "grad_norm": 37.287601470947266, "learning_rate": 2.571e-06, "loss": 2.4847, "step": 5142 }, { "epoch": 1.6122257053291538, "grad_norm": 28.55080223083496, "learning_rate": 2.5715000000000003e-06, "loss": 2.6724, "step": 5143 }, { "epoch": 1.6125391849529782, "grad_norm": 24.460296630859375, "learning_rate": 2.572e-06, "loss": 2.4058, "step": 5144 }, { "epoch": 1.6128526645768027, "grad_norm": 13.860005378723145, "learning_rate": 2.5725e-06, "loss": 2.5218, "step": 5145 }, { "epoch": 1.613166144200627, "grad_norm": 20.911226272583008, "learning_rate": 2.5729999999999998e-06, "loss": 2.738, "step": 5146 }, { "epoch": 1.6134796238244515, "grad_norm": 20.03626251220703, "learning_rate": 2.5735000000000004e-06, "loss": 2.4337, "step": 5147 }, { "epoch": 1.613793103448276, "grad_norm": 15.790233612060547, "learning_rate": 2.5740000000000003e-06, "loss": 2.2734, "step": 5148 }, { "epoch": 1.6141065830721004, "grad_norm": 15.679661750793457, "learning_rate": 2.5745000000000005e-06, "loss": 2.7092, "step": 5149 }, { "epoch": 1.6144200626959249, "grad_norm": 16.354408264160156, "learning_rate": 2.5750000000000003e-06, "loss": 2.1285, "step": 5150 }, { "epoch": 1.6147335423197493, "grad_norm": 26.114641189575195, "learning_rate": 2.5755e-06, "loss": 2.459, "step": 5151 }, { "epoch": 1.6150470219435737, "grad_norm": 36.885169982910156, "learning_rate": 2.576e-06, "loss": 2.3374, "step": 5152 }, { "epoch": 1.6153605015673982, "grad_norm": 18.303848266601562, "learning_rate": 2.5765000000000002e-06, "loss": 2.8493, "step": 5153 }, { "epoch": 1.6156739811912226, "grad_norm": 29.80558204650879, "learning_rate": 2.577e-06, "loss": 2.535, "step": 5154 }, { "epoch": 1.615987460815047, "grad_norm": 12.516839981079102, "learning_rate": 2.5775e-06, "loss": 2.1776, "step": 5155 }, { "epoch": 1.6163009404388715, "grad_norm": 15.919252395629883, "learning_rate": 2.578e-06, "loss": 2.1872, "step": 5156 }, { "epoch": 1.616614420062696, "grad_norm": 11.775092124938965, "learning_rate": 2.5785000000000004e-06, "loss": 2.1423, "step": 5157 }, { "epoch": 1.6169278996865204, "grad_norm": 31.506715774536133, "learning_rate": 2.579e-06, "loss": 2.2092, "step": 5158 }, { "epoch": 1.6172413793103448, "grad_norm": 43.23335266113281, "learning_rate": 2.5795000000000005e-06, "loss": 2.4301, "step": 5159 }, { "epoch": 1.6175548589341693, "grad_norm": 17.751665115356445, "learning_rate": 2.5800000000000003e-06, "loss": 2.2749, "step": 5160 }, { "epoch": 1.6178683385579937, "grad_norm": 34.12142562866211, "learning_rate": 2.5805e-06, "loss": 2.522, "step": 5161 }, { "epoch": 1.6181818181818182, "grad_norm": 25.955385208129883, "learning_rate": 2.5810000000000004e-06, "loss": 2.3837, "step": 5162 }, { "epoch": 1.6184952978056426, "grad_norm": 18.65509605407715, "learning_rate": 2.5815e-06, "loss": 2.3719, "step": 5163 }, { "epoch": 1.618808777429467, "grad_norm": 13.978825569152832, "learning_rate": 2.582e-06, "loss": 2.1543, "step": 5164 }, { "epoch": 1.6191222570532915, "grad_norm": 32.66872024536133, "learning_rate": 2.5825e-06, "loss": 2.4058, "step": 5165 }, { "epoch": 1.619435736677116, "grad_norm": 13.46345329284668, "learning_rate": 2.583e-06, "loss": 2.4266, "step": 5166 }, { "epoch": 1.6197492163009404, "grad_norm": 21.65625, "learning_rate": 2.5835000000000003e-06, "loss": 2.2746, "step": 5167 }, { "epoch": 1.6200626959247648, "grad_norm": 47.85426712036133, "learning_rate": 2.5840000000000006e-06, "loss": 2.7069, "step": 5168 }, { "epoch": 1.6203761755485893, "grad_norm": 12.973807334899902, "learning_rate": 2.5845000000000004e-06, "loss": 2.5359, "step": 5169 }, { "epoch": 1.6206896551724137, "grad_norm": 59.425838470458984, "learning_rate": 2.5850000000000002e-06, "loss": 2.5992, "step": 5170 }, { "epoch": 1.6210031347962381, "grad_norm": 42.35601043701172, "learning_rate": 2.5855e-06, "loss": 3.1424, "step": 5171 }, { "epoch": 1.6213166144200626, "grad_norm": 18.894989013671875, "learning_rate": 2.5860000000000003e-06, "loss": 2.4605, "step": 5172 }, { "epoch": 1.621630094043887, "grad_norm": 24.31694221496582, "learning_rate": 2.5865e-06, "loss": 2.3184, "step": 5173 }, { "epoch": 1.6219435736677115, "grad_norm": 44.74848175048828, "learning_rate": 2.587e-06, "loss": 2.7195, "step": 5174 }, { "epoch": 1.622257053291536, "grad_norm": 49.74072265625, "learning_rate": 2.5875000000000002e-06, "loss": 2.1722, "step": 5175 }, { "epoch": 1.6225705329153604, "grad_norm": 48.17268753051758, "learning_rate": 2.588e-06, "loss": 2.392, "step": 5176 }, { "epoch": 1.6228840125391848, "grad_norm": 10.797576904296875, "learning_rate": 2.5885000000000003e-06, "loss": 2.3357, "step": 5177 }, { "epoch": 1.6231974921630095, "grad_norm": 18.476648330688477, "learning_rate": 2.5890000000000005e-06, "loss": 2.2672, "step": 5178 }, { "epoch": 1.623510971786834, "grad_norm": 18.288114547729492, "learning_rate": 2.5895000000000004e-06, "loss": 2.0863, "step": 5179 }, { "epoch": 1.6238244514106583, "grad_norm": 19.165102005004883, "learning_rate": 2.59e-06, "loss": 2.6391, "step": 5180 }, { "epoch": 1.6241379310344828, "grad_norm": 14.676997184753418, "learning_rate": 2.5905000000000004e-06, "loss": 2.2653, "step": 5181 }, { "epoch": 1.6244514106583072, "grad_norm": 20.614850997924805, "learning_rate": 2.5910000000000003e-06, "loss": 2.2439, "step": 5182 }, { "epoch": 1.6247648902821317, "grad_norm": 97.47027587890625, "learning_rate": 2.5915e-06, "loss": 2.2794, "step": 5183 }, { "epoch": 1.625078369905956, "grad_norm": 24.965999603271484, "learning_rate": 2.592e-06, "loss": 2.1568, "step": 5184 }, { "epoch": 1.6253918495297806, "grad_norm": 29.646013259887695, "learning_rate": 2.5925e-06, "loss": 2.3762, "step": 5185 }, { "epoch": 1.625705329153605, "grad_norm": 12.286757469177246, "learning_rate": 2.593e-06, "loss": 2.0198, "step": 5186 }, { "epoch": 1.6260188087774294, "grad_norm": 13.670647621154785, "learning_rate": 2.5935000000000007e-06, "loss": 2.2541, "step": 5187 }, { "epoch": 1.6263322884012539, "grad_norm": 14.53826904296875, "learning_rate": 2.5940000000000005e-06, "loss": 2.3772, "step": 5188 }, { "epoch": 1.6266457680250783, "grad_norm": 14.540467262268066, "learning_rate": 2.5945000000000003e-06, "loss": 2.2134, "step": 5189 }, { "epoch": 1.626959247648903, "grad_norm": 11.156545639038086, "learning_rate": 2.595e-06, "loss": 2.4183, "step": 5190 }, { "epoch": 1.6272727272727274, "grad_norm": 39.29147720336914, "learning_rate": 2.5955000000000004e-06, "loss": 2.5317, "step": 5191 }, { "epoch": 1.6275862068965519, "grad_norm": 23.9459171295166, "learning_rate": 2.5960000000000002e-06, "loss": 2.36, "step": 5192 }, { "epoch": 1.6278996865203763, "grad_norm": 17.24390411376953, "learning_rate": 2.5965e-06, "loss": 2.7135, "step": 5193 }, { "epoch": 1.6282131661442008, "grad_norm": 18.405895233154297, "learning_rate": 2.597e-06, "loss": 2.3673, "step": 5194 }, { "epoch": 1.6285266457680252, "grad_norm": 13.795333862304688, "learning_rate": 2.5975e-06, "loss": 2.3724, "step": 5195 }, { "epoch": 1.6288401253918496, "grad_norm": 20.32740592956543, "learning_rate": 2.598e-06, "loss": 2.2114, "step": 5196 }, { "epoch": 1.629153605015674, "grad_norm": 11.09119987487793, "learning_rate": 2.5985000000000006e-06, "loss": 2.2799, "step": 5197 }, { "epoch": 1.6294670846394985, "grad_norm": 37.84120559692383, "learning_rate": 2.5990000000000004e-06, "loss": 2.7828, "step": 5198 }, { "epoch": 1.629780564263323, "grad_norm": 16.424787521362305, "learning_rate": 2.5995000000000003e-06, "loss": 2.335, "step": 5199 }, { "epoch": 1.6300940438871474, "grad_norm": 19.634714126586914, "learning_rate": 2.6e-06, "loss": 2.3331, "step": 5200 }, { "epoch": 1.6304075235109718, "grad_norm": 16.779815673828125, "learning_rate": 2.6005000000000003e-06, "loss": 2.168, "step": 5201 }, { "epoch": 1.6307210031347963, "grad_norm": 27.600574493408203, "learning_rate": 2.601e-06, "loss": 2.5557, "step": 5202 }, { "epoch": 1.6310344827586207, "grad_norm": 38.47439193725586, "learning_rate": 2.6015e-06, "loss": 2.7327, "step": 5203 }, { "epoch": 1.6313479623824452, "grad_norm": 19.682342529296875, "learning_rate": 2.6020000000000002e-06, "loss": 2.5308, "step": 5204 }, { "epoch": 1.6316614420062696, "grad_norm": 53.40803909301758, "learning_rate": 2.6025e-06, "loss": 2.278, "step": 5205 }, { "epoch": 1.631974921630094, "grad_norm": 16.326505661010742, "learning_rate": 2.603e-06, "loss": 2.6542, "step": 5206 }, { "epoch": 1.6322884012539185, "grad_norm": 13.8661527633667, "learning_rate": 2.6035000000000006e-06, "loss": 2.1928, "step": 5207 }, { "epoch": 1.632601880877743, "grad_norm": 16.33017349243164, "learning_rate": 2.6040000000000004e-06, "loss": 2.3582, "step": 5208 }, { "epoch": 1.6329153605015674, "grad_norm": 51.326820373535156, "learning_rate": 2.6045000000000002e-06, "loss": 2.0451, "step": 5209 }, { "epoch": 1.6332288401253918, "grad_norm": 17.90423011779785, "learning_rate": 2.6050000000000005e-06, "loss": 2.5162, "step": 5210 }, { "epoch": 1.6335423197492163, "grad_norm": 40.12839126586914, "learning_rate": 2.6055000000000003e-06, "loss": 2.2756, "step": 5211 }, { "epoch": 1.6338557993730407, "grad_norm": 17.745290756225586, "learning_rate": 2.606e-06, "loss": 2.1294, "step": 5212 }, { "epoch": 1.6341692789968651, "grad_norm": 13.970556259155273, "learning_rate": 2.6065e-06, "loss": 2.1501, "step": 5213 }, { "epoch": 1.6344827586206896, "grad_norm": 22.84992027282715, "learning_rate": 2.607e-06, "loss": 2.8207, "step": 5214 }, { "epoch": 1.634796238244514, "grad_norm": 43.31232452392578, "learning_rate": 2.6075e-06, "loss": 3.122, "step": 5215 }, { "epoch": 1.6351097178683385, "grad_norm": 14.463607788085938, "learning_rate": 2.608e-06, "loss": 2.2365, "step": 5216 }, { "epoch": 1.635423197492163, "grad_norm": 23.331586837768555, "learning_rate": 2.6085000000000005e-06, "loss": 2.984, "step": 5217 }, { "epoch": 1.6357366771159874, "grad_norm": 10.62083625793457, "learning_rate": 2.6090000000000003e-06, "loss": 2.1679, "step": 5218 }, { "epoch": 1.6360501567398118, "grad_norm": 40.45970153808594, "learning_rate": 2.6095e-06, "loss": 2.3978, "step": 5219 }, { "epoch": 1.6363636363636362, "grad_norm": 12.194794654846191, "learning_rate": 2.6100000000000004e-06, "loss": 2.4125, "step": 5220 }, { "epoch": 1.6366771159874607, "grad_norm": 12.354119300842285, "learning_rate": 2.6105000000000002e-06, "loss": 2.2636, "step": 5221 }, { "epoch": 1.6369905956112851, "grad_norm": 15.73239517211914, "learning_rate": 2.611e-06, "loss": 2.2861, "step": 5222 }, { "epoch": 1.6373040752351096, "grad_norm": 13.624801635742188, "learning_rate": 2.6115000000000003e-06, "loss": 2.3806, "step": 5223 }, { "epoch": 1.637617554858934, "grad_norm": 38.89716720581055, "learning_rate": 2.612e-06, "loss": 2.7064, "step": 5224 }, { "epoch": 1.6379310344827587, "grad_norm": 104.40101623535156, "learning_rate": 2.6125e-06, "loss": 3.6707, "step": 5225 }, { "epoch": 1.6382445141065831, "grad_norm": 15.195369720458984, "learning_rate": 2.613e-06, "loss": 2.2893, "step": 5226 }, { "epoch": 1.6385579937304076, "grad_norm": 63.90957260131836, "learning_rate": 2.6135000000000005e-06, "loss": 2.4651, "step": 5227 }, { "epoch": 1.638871473354232, "grad_norm": 13.292878150939941, "learning_rate": 2.6140000000000003e-06, "loss": 2.3637, "step": 5228 }, { "epoch": 1.6391849529780564, "grad_norm": 25.40592384338379, "learning_rate": 2.6145000000000005e-06, "loss": 2.0852, "step": 5229 }, { "epoch": 1.6394984326018809, "grad_norm": 13.220362663269043, "learning_rate": 2.6150000000000004e-06, "loss": 2.1517, "step": 5230 }, { "epoch": 1.6398119122257053, "grad_norm": 11.700094223022461, "learning_rate": 2.6155e-06, "loss": 2.3451, "step": 5231 }, { "epoch": 1.6401253918495298, "grad_norm": 15.420618057250977, "learning_rate": 2.616e-06, "loss": 2.141, "step": 5232 }, { "epoch": 1.6404388714733542, "grad_norm": 17.454343795776367, "learning_rate": 2.6165000000000003e-06, "loss": 2.2411, "step": 5233 }, { "epoch": 1.6407523510971787, "grad_norm": 13.41075325012207, "learning_rate": 2.617e-06, "loss": 2.3167, "step": 5234 }, { "epoch": 1.641065830721003, "grad_norm": 24.065536499023438, "learning_rate": 2.6175e-06, "loss": 2.5559, "step": 5235 }, { "epoch": 1.6413793103448275, "grad_norm": 33.53857421875, "learning_rate": 2.618e-06, "loss": 2.0863, "step": 5236 }, { "epoch": 1.6416927899686522, "grad_norm": 69.07901763916016, "learning_rate": 2.6185000000000004e-06, "loss": 2.4614, "step": 5237 }, { "epoch": 1.6420062695924766, "grad_norm": 29.503938674926758, "learning_rate": 2.6190000000000003e-06, "loss": 2.262, "step": 5238 }, { "epoch": 1.642319749216301, "grad_norm": 13.055183410644531, "learning_rate": 2.6195000000000005e-06, "loss": 2.0189, "step": 5239 }, { "epoch": 1.6426332288401255, "grad_norm": 19.09827995300293, "learning_rate": 2.6200000000000003e-06, "loss": 2.3149, "step": 5240 }, { "epoch": 1.64294670846395, "grad_norm": 16.848682403564453, "learning_rate": 2.6205e-06, "loss": 2.3306, "step": 5241 }, { "epoch": 1.6432601880877744, "grad_norm": 37.780418395996094, "learning_rate": 2.6210000000000004e-06, "loss": 2.4372, "step": 5242 }, { "epoch": 1.6435736677115989, "grad_norm": 14.445202827453613, "learning_rate": 2.6215000000000002e-06, "loss": 2.3163, "step": 5243 }, { "epoch": 1.6438871473354233, "grad_norm": 32.84733200073242, "learning_rate": 2.622e-06, "loss": 2.4044, "step": 5244 }, { "epoch": 1.6442006269592477, "grad_norm": 11.114374160766602, "learning_rate": 2.6225e-06, "loss": 2.394, "step": 5245 }, { "epoch": 1.6445141065830722, "grad_norm": 11.89560317993164, "learning_rate": 2.623e-06, "loss": 2.3219, "step": 5246 }, { "epoch": 1.6448275862068966, "grad_norm": 39.383079528808594, "learning_rate": 2.6235000000000004e-06, "loss": 2.4139, "step": 5247 }, { "epoch": 1.645141065830721, "grad_norm": 59.19121551513672, "learning_rate": 2.6240000000000006e-06, "loss": 2.2169, "step": 5248 }, { "epoch": 1.6454545454545455, "grad_norm": 34.60609436035156, "learning_rate": 2.6245000000000005e-06, "loss": 2.5338, "step": 5249 }, { "epoch": 1.64576802507837, "grad_norm": 21.994110107421875, "learning_rate": 2.6250000000000003e-06, "loss": 2.6902, "step": 5250 }, { "epoch": 1.6460815047021944, "grad_norm": 17.888416290283203, "learning_rate": 2.6255e-06, "loss": 2.2012, "step": 5251 }, { "epoch": 1.6463949843260188, "grad_norm": 126.86553192138672, "learning_rate": 2.6260000000000004e-06, "loss": 2.5988, "step": 5252 }, { "epoch": 1.6467084639498433, "grad_norm": 20.460132598876953, "learning_rate": 2.6265e-06, "loss": 2.1957, "step": 5253 }, { "epoch": 1.6470219435736677, "grad_norm": 16.447261810302734, "learning_rate": 2.627e-06, "loss": 2.5429, "step": 5254 }, { "epoch": 1.6473354231974922, "grad_norm": 23.05217170715332, "learning_rate": 2.6275000000000003e-06, "loss": 2.5669, "step": 5255 }, { "epoch": 1.6476489028213166, "grad_norm": 27.731290817260742, "learning_rate": 2.628e-06, "loss": 2.3274, "step": 5256 }, { "epoch": 1.647962382445141, "grad_norm": 18.66347312927246, "learning_rate": 2.6285e-06, "loss": 2.2186, "step": 5257 }, { "epoch": 1.6482758620689655, "grad_norm": 12.416943550109863, "learning_rate": 2.6290000000000006e-06, "loss": 2.0702, "step": 5258 }, { "epoch": 1.64858934169279, "grad_norm": 58.2420539855957, "learning_rate": 2.6295000000000004e-06, "loss": 2.832, "step": 5259 }, { "epoch": 1.6489028213166144, "grad_norm": 18.872852325439453, "learning_rate": 2.6300000000000002e-06, "loss": 2.4728, "step": 5260 }, { "epoch": 1.6492163009404388, "grad_norm": 27.66944122314453, "learning_rate": 2.6305000000000005e-06, "loss": 2.3424, "step": 5261 }, { "epoch": 1.6495297805642632, "grad_norm": 21.679662704467773, "learning_rate": 2.6310000000000003e-06, "loss": 1.9519, "step": 5262 }, { "epoch": 1.6498432601880877, "grad_norm": 31.611692428588867, "learning_rate": 2.6315e-06, "loss": 2.9403, "step": 5263 }, { "epoch": 1.6501567398119121, "grad_norm": 31.275747299194336, "learning_rate": 2.632e-06, "loss": 2.1951, "step": 5264 }, { "epoch": 1.6504702194357366, "grad_norm": 15.326471328735352, "learning_rate": 2.6325e-06, "loss": 2.2404, "step": 5265 }, { "epoch": 1.650783699059561, "grad_norm": 16.914216995239258, "learning_rate": 2.633e-06, "loss": 2.2977, "step": 5266 }, { "epoch": 1.6510971786833855, "grad_norm": 14.189790725708008, "learning_rate": 2.6335e-06, "loss": 2.2385, "step": 5267 }, { "epoch": 1.65141065830721, "grad_norm": 74.67144775390625, "learning_rate": 2.6340000000000005e-06, "loss": 2.6448, "step": 5268 }, { "epoch": 1.6517241379310343, "grad_norm": 7.973588466644287, "learning_rate": 2.6345000000000004e-06, "loss": 2.3253, "step": 5269 }, { "epoch": 1.6520376175548588, "grad_norm": 17.322906494140625, "learning_rate": 2.635e-06, "loss": 2.3823, "step": 5270 }, { "epoch": 1.6523510971786832, "grad_norm": 20.03339385986328, "learning_rate": 2.6355000000000004e-06, "loss": 2.4951, "step": 5271 }, { "epoch": 1.652664576802508, "grad_norm": 35.52084732055664, "learning_rate": 2.6360000000000003e-06, "loss": 2.3325, "step": 5272 }, { "epoch": 1.6529780564263323, "grad_norm": 16.430076599121094, "learning_rate": 2.6365e-06, "loss": 2.3975, "step": 5273 }, { "epoch": 1.6532915360501568, "grad_norm": 14.450660705566406, "learning_rate": 2.637e-06, "loss": 2.2319, "step": 5274 }, { "epoch": 1.6536050156739812, "grad_norm": 16.825803756713867, "learning_rate": 2.6375e-06, "loss": 2.1143, "step": 5275 }, { "epoch": 1.6539184952978057, "grad_norm": 38.360015869140625, "learning_rate": 2.638e-06, "loss": 2.2986, "step": 5276 }, { "epoch": 1.65423197492163, "grad_norm": 19.999114990234375, "learning_rate": 2.6385e-06, "loss": 2.7058, "step": 5277 }, { "epoch": 1.6545454545454545, "grad_norm": 22.12860107421875, "learning_rate": 2.6390000000000005e-06, "loss": 2.3911, "step": 5278 }, { "epoch": 1.654858934169279, "grad_norm": 11.342117309570312, "learning_rate": 2.6395000000000003e-06, "loss": 2.0603, "step": 5279 }, { "epoch": 1.6551724137931034, "grad_norm": 25.63351821899414, "learning_rate": 2.64e-06, "loss": 2.3815, "step": 5280 }, { "epoch": 1.6554858934169279, "grad_norm": 10.051604270935059, "learning_rate": 2.6405000000000004e-06, "loss": 2.2003, "step": 5281 }, { "epoch": 1.6557993730407523, "grad_norm": 14.631726264953613, "learning_rate": 2.641e-06, "loss": 2.3557, "step": 5282 }, { "epoch": 1.6561128526645768, "grad_norm": 21.072256088256836, "learning_rate": 2.6415e-06, "loss": 2.607, "step": 5283 }, { "epoch": 1.6564263322884014, "grad_norm": 21.111793518066406, "learning_rate": 2.6420000000000003e-06, "loss": 2.3936, "step": 5284 }, { "epoch": 1.6567398119122259, "grad_norm": 31.564790725708008, "learning_rate": 2.6425e-06, "loss": 2.3416, "step": 5285 }, { "epoch": 1.6570532915360503, "grad_norm": 15.514556884765625, "learning_rate": 2.643e-06, "loss": 2.462, "step": 5286 }, { "epoch": 1.6573667711598747, "grad_norm": 49.09897994995117, "learning_rate": 2.6434999999999998e-06, "loss": 2.4764, "step": 5287 }, { "epoch": 1.6576802507836992, "grad_norm": 22.08146095275879, "learning_rate": 2.6440000000000004e-06, "loss": 2.3708, "step": 5288 }, { "epoch": 1.6579937304075236, "grad_norm": 37.27674102783203, "learning_rate": 2.6445000000000003e-06, "loss": 2.382, "step": 5289 }, { "epoch": 1.658307210031348, "grad_norm": 102.43901062011719, "learning_rate": 2.6450000000000005e-06, "loss": 2.4155, "step": 5290 }, { "epoch": 1.6586206896551725, "grad_norm": 22.695934295654297, "learning_rate": 2.6455000000000003e-06, "loss": 2.4774, "step": 5291 }, { "epoch": 1.658934169278997, "grad_norm": 57.112098693847656, "learning_rate": 2.646e-06, "loss": 4.0764, "step": 5292 }, { "epoch": 1.6592476489028214, "grad_norm": 18.36469268798828, "learning_rate": 2.6465e-06, "loss": 2.3275, "step": 5293 }, { "epoch": 1.6595611285266458, "grad_norm": 32.93498992919922, "learning_rate": 2.6470000000000002e-06, "loss": 2.4024, "step": 5294 }, { "epoch": 1.6598746081504703, "grad_norm": 15.336421966552734, "learning_rate": 2.6475e-06, "loss": 2.4848, "step": 5295 }, { "epoch": 1.6601880877742947, "grad_norm": 14.629399299621582, "learning_rate": 2.648e-06, "loss": 2.3644, "step": 5296 }, { "epoch": 1.6605015673981192, "grad_norm": 12.701495170593262, "learning_rate": 2.6485e-06, "loss": 2.1344, "step": 5297 }, { "epoch": 1.6608150470219436, "grad_norm": 13.902689933776855, "learning_rate": 2.6490000000000004e-06, "loss": 2.3507, "step": 5298 }, { "epoch": 1.661128526645768, "grad_norm": 21.512542724609375, "learning_rate": 2.6495000000000002e-06, "loss": 2.6414, "step": 5299 }, { "epoch": 1.6614420062695925, "grad_norm": 25.147584915161133, "learning_rate": 2.6500000000000005e-06, "loss": 2.5541, "step": 5300 }, { "epoch": 1.661755485893417, "grad_norm": 13.754753112792969, "learning_rate": 2.6505000000000003e-06, "loss": 2.1547, "step": 5301 }, { "epoch": 1.6620689655172414, "grad_norm": 12.000065803527832, "learning_rate": 2.651e-06, "loss": 2.3954, "step": 5302 }, { "epoch": 1.6623824451410658, "grad_norm": 26.096010208129883, "learning_rate": 2.6515000000000004e-06, "loss": 2.3964, "step": 5303 }, { "epoch": 1.6626959247648903, "grad_norm": 12.003847122192383, "learning_rate": 2.652e-06, "loss": 2.0897, "step": 5304 }, { "epoch": 1.6630094043887147, "grad_norm": 9.624919891357422, "learning_rate": 2.6525e-06, "loss": 2.3602, "step": 5305 }, { "epoch": 1.6633228840125391, "grad_norm": 17.856704711914062, "learning_rate": 2.653e-06, "loss": 2.5554, "step": 5306 }, { "epoch": 1.6636363636363636, "grad_norm": 13.328954696655273, "learning_rate": 2.6535e-06, "loss": 2.316, "step": 5307 }, { "epoch": 1.663949843260188, "grad_norm": 18.4957218170166, "learning_rate": 2.6540000000000003e-06, "loss": 2.3456, "step": 5308 }, { "epoch": 1.6642633228840125, "grad_norm": 27.82377815246582, "learning_rate": 2.6545000000000006e-06, "loss": 2.9891, "step": 5309 }, { "epoch": 1.664576802507837, "grad_norm": 23.645793914794922, "learning_rate": 2.6550000000000004e-06, "loss": 2.3558, "step": 5310 }, { "epoch": 1.6648902821316613, "grad_norm": 12.790231704711914, "learning_rate": 2.6555000000000002e-06, "loss": 2.1346, "step": 5311 }, { "epoch": 1.6652037617554858, "grad_norm": 33.11914825439453, "learning_rate": 2.656e-06, "loss": 2.3526, "step": 5312 }, { "epoch": 1.6655172413793102, "grad_norm": 17.447622299194336, "learning_rate": 2.6565000000000003e-06, "loss": 2.4022, "step": 5313 }, { "epoch": 1.6658307210031347, "grad_norm": 25.620925903320312, "learning_rate": 2.657e-06, "loss": 2.3914, "step": 5314 }, { "epoch": 1.6661442006269591, "grad_norm": 17.788557052612305, "learning_rate": 2.6575e-06, "loss": 2.2504, "step": 5315 }, { "epoch": 1.6664576802507836, "grad_norm": 34.506771087646484, "learning_rate": 2.6580000000000002e-06, "loss": 2.6078, "step": 5316 }, { "epoch": 1.666771159874608, "grad_norm": 15.220390319824219, "learning_rate": 2.6585e-06, "loss": 2.3253, "step": 5317 }, { "epoch": 1.6670846394984324, "grad_norm": 60.35187911987305, "learning_rate": 2.6590000000000003e-06, "loss": 2.5263, "step": 5318 }, { "epoch": 1.667398119122257, "grad_norm": 11.838387489318848, "learning_rate": 2.6595000000000005e-06, "loss": 2.2337, "step": 5319 }, { "epoch": 1.6677115987460815, "grad_norm": 13.644763946533203, "learning_rate": 2.6600000000000004e-06, "loss": 2.4781, "step": 5320 }, { "epoch": 1.668025078369906, "grad_norm": 28.51342010498047, "learning_rate": 2.6605e-06, "loss": 2.7712, "step": 5321 }, { "epoch": 1.6683385579937304, "grad_norm": 14.968554496765137, "learning_rate": 2.6610000000000004e-06, "loss": 2.4118, "step": 5322 }, { "epoch": 1.6686520376175549, "grad_norm": 16.145463943481445, "learning_rate": 2.6615000000000003e-06, "loss": 2.3086, "step": 5323 }, { "epoch": 1.6689655172413793, "grad_norm": 24.526046752929688, "learning_rate": 2.662e-06, "loss": 2.1476, "step": 5324 }, { "epoch": 1.6692789968652038, "grad_norm": 25.00034523010254, "learning_rate": 2.6625e-06, "loss": 2.2905, "step": 5325 }, { "epoch": 1.6695924764890282, "grad_norm": 39.29100036621094, "learning_rate": 2.663e-06, "loss": 2.4597, "step": 5326 }, { "epoch": 1.6699059561128526, "grad_norm": 31.492595672607422, "learning_rate": 2.6635e-06, "loss": 2.418, "step": 5327 }, { "epoch": 1.670219435736677, "grad_norm": 56.26333999633789, "learning_rate": 2.6640000000000007e-06, "loss": 2.2798, "step": 5328 }, { "epoch": 1.6705329153605015, "grad_norm": 19.510744094848633, "learning_rate": 2.6645000000000005e-06, "loss": 2.5782, "step": 5329 }, { "epoch": 1.670846394984326, "grad_norm": 21.897645950317383, "learning_rate": 2.6650000000000003e-06, "loss": 2.3679, "step": 5330 }, { "epoch": 1.6711598746081506, "grad_norm": 16.739225387573242, "learning_rate": 2.6655e-06, "loss": 2.4401, "step": 5331 }, { "epoch": 1.671473354231975, "grad_norm": 17.856504440307617, "learning_rate": 2.6660000000000004e-06, "loss": 2.3782, "step": 5332 }, { "epoch": 1.6717868338557995, "grad_norm": 16.302274703979492, "learning_rate": 2.6665000000000002e-06, "loss": 2.2191, "step": 5333 }, { "epoch": 1.672100313479624, "grad_norm": 12.230741500854492, "learning_rate": 2.667e-06, "loss": 2.4313, "step": 5334 }, { "epoch": 1.6724137931034484, "grad_norm": 37.70825958251953, "learning_rate": 2.6675000000000003e-06, "loss": 2.8394, "step": 5335 }, { "epoch": 1.6727272727272728, "grad_norm": 19.169376373291016, "learning_rate": 2.668e-06, "loss": 2.4162, "step": 5336 }, { "epoch": 1.6730407523510973, "grad_norm": 143.8944549560547, "learning_rate": 2.6685e-06, "loss": 2.3096, "step": 5337 }, { "epoch": 1.6733542319749217, "grad_norm": 25.016977310180664, "learning_rate": 2.6690000000000006e-06, "loss": 2.3672, "step": 5338 }, { "epoch": 1.6736677115987462, "grad_norm": 11.875692367553711, "learning_rate": 2.6695000000000004e-06, "loss": 2.2422, "step": 5339 }, { "epoch": 1.6739811912225706, "grad_norm": 14.99013614654541, "learning_rate": 2.6700000000000003e-06, "loss": 2.1051, "step": 5340 }, { "epoch": 1.674294670846395, "grad_norm": 7.908255100250244, "learning_rate": 2.6705000000000005e-06, "loss": 2.2055, "step": 5341 }, { "epoch": 1.6746081504702195, "grad_norm": 29.556995391845703, "learning_rate": 2.6710000000000003e-06, "loss": 2.1735, "step": 5342 }, { "epoch": 1.674921630094044, "grad_norm": 14.469071388244629, "learning_rate": 2.6715e-06, "loss": 2.2748, "step": 5343 }, { "epoch": 1.6752351097178684, "grad_norm": 12.449281692504883, "learning_rate": 2.672e-06, "loss": 2.3401, "step": 5344 }, { "epoch": 1.6755485893416928, "grad_norm": 12.840116500854492, "learning_rate": 2.6725000000000002e-06, "loss": 2.5153, "step": 5345 }, { "epoch": 1.6758620689655173, "grad_norm": 13.369140625, "learning_rate": 2.673e-06, "loss": 2.445, "step": 5346 }, { "epoch": 1.6761755485893417, "grad_norm": 26.555614471435547, "learning_rate": 2.6735e-06, "loss": 2.3638, "step": 5347 }, { "epoch": 1.6764890282131661, "grad_norm": 13.960328102111816, "learning_rate": 2.6740000000000006e-06, "loss": 2.1247, "step": 5348 }, { "epoch": 1.6768025078369906, "grad_norm": 17.971302032470703, "learning_rate": 2.6745000000000004e-06, "loss": 2.352, "step": 5349 }, { "epoch": 1.677115987460815, "grad_norm": 15.672755241394043, "learning_rate": 2.6750000000000002e-06, "loss": 2.0947, "step": 5350 }, { "epoch": 1.6774294670846395, "grad_norm": 27.532419204711914, "learning_rate": 2.6755000000000005e-06, "loss": 3.1038, "step": 5351 }, { "epoch": 1.677742946708464, "grad_norm": 32.5295295715332, "learning_rate": 2.6760000000000003e-06, "loss": 2.0789, "step": 5352 }, { "epoch": 1.6780564263322884, "grad_norm": 10.138176918029785, "learning_rate": 2.6765e-06, "loss": 2.6218, "step": 5353 }, { "epoch": 1.6783699059561128, "grad_norm": 22.55307960510254, "learning_rate": 2.677e-06, "loss": 1.9356, "step": 5354 }, { "epoch": 1.6786833855799372, "grad_norm": 47.21834182739258, "learning_rate": 2.6775e-06, "loss": 2.6484, "step": 5355 }, { "epoch": 1.6789968652037617, "grad_norm": 52.77558517456055, "learning_rate": 2.678e-06, "loss": 2.319, "step": 5356 }, { "epoch": 1.6793103448275861, "grad_norm": 13.631694793701172, "learning_rate": 2.6785e-06, "loss": 2.0864, "step": 5357 }, { "epoch": 1.6796238244514106, "grad_norm": 17.103946685791016, "learning_rate": 2.6790000000000005e-06, "loss": 2.2015, "step": 5358 }, { "epoch": 1.679937304075235, "grad_norm": 11.737060546875, "learning_rate": 2.6795000000000003e-06, "loss": 2.291, "step": 5359 }, { "epoch": 1.6802507836990594, "grad_norm": 34.89884948730469, "learning_rate": 2.68e-06, "loss": 2.3886, "step": 5360 }, { "epoch": 1.6805642633228839, "grad_norm": 25.996509552001953, "learning_rate": 2.6805000000000004e-06, "loss": 2.4901, "step": 5361 }, { "epoch": 1.6808777429467083, "grad_norm": 11.781593322753906, "learning_rate": 2.6810000000000003e-06, "loss": 2.2829, "step": 5362 }, { "epoch": 1.6811912225705328, "grad_norm": 21.603439331054688, "learning_rate": 2.6815e-06, "loss": 2.2104, "step": 5363 }, { "epoch": 1.6815047021943572, "grad_norm": 24.865238189697266, "learning_rate": 2.6820000000000003e-06, "loss": 2.3377, "step": 5364 }, { "epoch": 1.6818181818181817, "grad_norm": 16.560691833496094, "learning_rate": 2.6825e-06, "loss": 2.4748, "step": 5365 }, { "epoch": 1.6821316614420063, "grad_norm": 14.380731582641602, "learning_rate": 2.683e-06, "loss": 2.217, "step": 5366 }, { "epoch": 1.6824451410658308, "grad_norm": 13.259237289428711, "learning_rate": 2.6835e-06, "loss": 2.2073, "step": 5367 }, { "epoch": 1.6827586206896552, "grad_norm": 24.348876953125, "learning_rate": 2.6840000000000005e-06, "loss": 2.5637, "step": 5368 }, { "epoch": 1.6830721003134796, "grad_norm": 14.206762313842773, "learning_rate": 2.6845000000000003e-06, "loss": 2.3421, "step": 5369 }, { "epoch": 1.683385579937304, "grad_norm": 20.449251174926758, "learning_rate": 2.6850000000000006e-06, "loss": 2.2867, "step": 5370 }, { "epoch": 1.6836990595611285, "grad_norm": 29.787328720092773, "learning_rate": 2.6855000000000004e-06, "loss": 2.7544, "step": 5371 }, { "epoch": 1.684012539184953, "grad_norm": 107.39007568359375, "learning_rate": 2.686e-06, "loss": 2.6798, "step": 5372 }, { "epoch": 1.6843260188087774, "grad_norm": 10.828439712524414, "learning_rate": 2.6865e-06, "loss": 2.1149, "step": 5373 }, { "epoch": 1.6846394984326019, "grad_norm": 17.259565353393555, "learning_rate": 2.6870000000000003e-06, "loss": 2.4487, "step": 5374 }, { "epoch": 1.6849529780564263, "grad_norm": 9.159259796142578, "learning_rate": 2.6875e-06, "loss": 2.1542, "step": 5375 }, { "epoch": 1.6852664576802507, "grad_norm": 9.4286470413208, "learning_rate": 2.688e-06, "loss": 2.1916, "step": 5376 }, { "epoch": 1.6855799373040752, "grad_norm": 10.63680362701416, "learning_rate": 2.6885e-06, "loss": 2.1223, "step": 5377 }, { "epoch": 1.6858934169278998, "grad_norm": 15.974871635437012, "learning_rate": 2.689e-06, "loss": 2.473, "step": 5378 }, { "epoch": 1.6862068965517243, "grad_norm": 16.37948989868164, "learning_rate": 2.6895000000000003e-06, "loss": 2.6459, "step": 5379 }, { "epoch": 1.6865203761755487, "grad_norm": 17.967891693115234, "learning_rate": 2.6900000000000005e-06, "loss": 2.246, "step": 5380 }, { "epoch": 1.6868338557993732, "grad_norm": 15.976727485656738, "learning_rate": 2.6905000000000003e-06, "loss": 2.1751, "step": 5381 }, { "epoch": 1.6871473354231976, "grad_norm": 27.1754207611084, "learning_rate": 2.691e-06, "loss": 2.7939, "step": 5382 }, { "epoch": 1.687460815047022, "grad_norm": 14.75679874420166, "learning_rate": 2.6915000000000004e-06, "loss": 2.4285, "step": 5383 }, { "epoch": 1.6877742946708465, "grad_norm": 38.64543914794922, "learning_rate": 2.6920000000000002e-06, "loss": 2.2422, "step": 5384 }, { "epoch": 1.688087774294671, "grad_norm": 64.94852447509766, "learning_rate": 2.6925e-06, "loss": 2.3804, "step": 5385 }, { "epoch": 1.6884012539184954, "grad_norm": 17.144126892089844, "learning_rate": 2.693e-06, "loss": 2.2281, "step": 5386 }, { "epoch": 1.6887147335423198, "grad_norm": 29.98297882080078, "learning_rate": 2.6935e-06, "loss": 2.3383, "step": 5387 }, { "epoch": 1.6890282131661443, "grad_norm": 20.504491806030273, "learning_rate": 2.694e-06, "loss": 2.2015, "step": 5388 }, { "epoch": 1.6893416927899687, "grad_norm": 23.511560440063477, "learning_rate": 2.6945000000000006e-06, "loss": 2.2689, "step": 5389 }, { "epoch": 1.6896551724137931, "grad_norm": 13.887767791748047, "learning_rate": 2.6950000000000005e-06, "loss": 2.3898, "step": 5390 }, { "epoch": 1.6899686520376176, "grad_norm": 14.888626098632812, "learning_rate": 2.6955000000000003e-06, "loss": 2.3539, "step": 5391 }, { "epoch": 1.690282131661442, "grad_norm": 62.88370895385742, "learning_rate": 2.696e-06, "loss": 2.6187, "step": 5392 }, { "epoch": 1.6905956112852665, "grad_norm": 13.9352445602417, "learning_rate": 2.6965000000000004e-06, "loss": 2.3166, "step": 5393 }, { "epoch": 1.690909090909091, "grad_norm": 21.216232299804688, "learning_rate": 2.697e-06, "loss": 2.1817, "step": 5394 }, { "epoch": 1.6912225705329154, "grad_norm": 13.734606742858887, "learning_rate": 2.6975e-06, "loss": 2.5173, "step": 5395 }, { "epoch": 1.6915360501567398, "grad_norm": 68.65769958496094, "learning_rate": 2.6980000000000003e-06, "loss": 2.5343, "step": 5396 }, { "epoch": 1.6918495297805642, "grad_norm": 20.774269104003906, "learning_rate": 2.6985e-06, "loss": 2.9843, "step": 5397 }, { "epoch": 1.6921630094043887, "grad_norm": 37.50685119628906, "learning_rate": 2.699e-06, "loss": 2.4088, "step": 5398 }, { "epoch": 1.6924764890282131, "grad_norm": 137.178955078125, "learning_rate": 2.6995000000000006e-06, "loss": 3.5242, "step": 5399 }, { "epoch": 1.6927899686520376, "grad_norm": 47.885902404785156, "learning_rate": 2.7000000000000004e-06, "loss": 2.2528, "step": 5400 }, { "epoch": 1.693103448275862, "grad_norm": 18.669857025146484, "learning_rate": 2.7005000000000002e-06, "loss": 2.3694, "step": 5401 }, { "epoch": 1.6934169278996865, "grad_norm": 91.53276062011719, "learning_rate": 2.7010000000000005e-06, "loss": 2.7944, "step": 5402 }, { "epoch": 1.693730407523511, "grad_norm": 25.30732536315918, "learning_rate": 2.7015000000000003e-06, "loss": 2.3734, "step": 5403 }, { "epoch": 1.6940438871473353, "grad_norm": 22.524877548217773, "learning_rate": 2.702e-06, "loss": 2.5223, "step": 5404 }, { "epoch": 1.6943573667711598, "grad_norm": 14.78250503540039, "learning_rate": 2.7025e-06, "loss": 2.1247, "step": 5405 }, { "epoch": 1.6946708463949842, "grad_norm": 14.238978385925293, "learning_rate": 2.703e-06, "loss": 2.525, "step": 5406 }, { "epoch": 1.6949843260188087, "grad_norm": 27.516679763793945, "learning_rate": 2.7035e-06, "loss": 2.345, "step": 5407 }, { "epoch": 1.695297805642633, "grad_norm": 219.90899658203125, "learning_rate": 2.704e-06, "loss": 2.7112, "step": 5408 }, { "epoch": 1.6956112852664575, "grad_norm": 12.407668113708496, "learning_rate": 2.7045000000000005e-06, "loss": 2.3901, "step": 5409 }, { "epoch": 1.695924764890282, "grad_norm": 39.14141845703125, "learning_rate": 2.7050000000000004e-06, "loss": 2.2635, "step": 5410 }, { "epoch": 1.6962382445141064, "grad_norm": 16.694364547729492, "learning_rate": 2.7055e-06, "loss": 2.7107, "step": 5411 }, { "epoch": 1.6965517241379309, "grad_norm": 14.390851020812988, "learning_rate": 2.7060000000000004e-06, "loss": 2.1942, "step": 5412 }, { "epoch": 1.6968652037617555, "grad_norm": 18.806320190429688, "learning_rate": 2.7065000000000003e-06, "loss": 2.4038, "step": 5413 }, { "epoch": 1.69717868338558, "grad_norm": 13.088409423828125, "learning_rate": 2.707e-06, "loss": 2.1712, "step": 5414 }, { "epoch": 1.6974921630094044, "grad_norm": 19.5601749420166, "learning_rate": 2.7075000000000003e-06, "loss": 2.8417, "step": 5415 }, { "epoch": 1.6978056426332289, "grad_norm": 40.176918029785156, "learning_rate": 2.708e-06, "loss": 2.4874, "step": 5416 }, { "epoch": 1.6981191222570533, "grad_norm": 20.487564086914062, "learning_rate": 2.7085e-06, "loss": 2.2593, "step": 5417 }, { "epoch": 1.6984326018808777, "grad_norm": 13.035333633422852, "learning_rate": 2.709e-06, "loss": 2.2706, "step": 5418 }, { "epoch": 1.6987460815047022, "grad_norm": 10.95355224609375, "learning_rate": 2.7095000000000005e-06, "loss": 2.2284, "step": 5419 }, { "epoch": 1.6990595611285266, "grad_norm": 13.154052734375, "learning_rate": 2.7100000000000003e-06, "loss": 2.3685, "step": 5420 }, { "epoch": 1.699373040752351, "grad_norm": 26.198123931884766, "learning_rate": 2.7105000000000006e-06, "loss": 2.5497, "step": 5421 }, { "epoch": 1.6996865203761755, "grad_norm": 36.201995849609375, "learning_rate": 2.7110000000000004e-06, "loss": 3.0085, "step": 5422 }, { "epoch": 1.7, "grad_norm": 73.30572509765625, "learning_rate": 2.7115000000000002e-06, "loss": 1.8637, "step": 5423 }, { "epoch": 1.7003134796238244, "grad_norm": 59.0308723449707, "learning_rate": 2.712e-06, "loss": 2.7115, "step": 5424 }, { "epoch": 1.700626959247649, "grad_norm": 47.59678268432617, "learning_rate": 2.7125000000000003e-06, "loss": 2.8762, "step": 5425 }, { "epoch": 1.7009404388714735, "grad_norm": 11.48272705078125, "learning_rate": 2.713e-06, "loss": 2.2331, "step": 5426 }, { "epoch": 1.701253918495298, "grad_norm": 20.130481719970703, "learning_rate": 2.7135e-06, "loss": 2.4474, "step": 5427 }, { "epoch": 1.7015673981191224, "grad_norm": 15.507676124572754, "learning_rate": 2.7139999999999998e-06, "loss": 2.312, "step": 5428 }, { "epoch": 1.7018808777429468, "grad_norm": 15.507573127746582, "learning_rate": 2.7145000000000004e-06, "loss": 2.1403, "step": 5429 }, { "epoch": 1.7021943573667713, "grad_norm": 23.145816802978516, "learning_rate": 2.7150000000000003e-06, "loss": 2.5529, "step": 5430 }, { "epoch": 1.7025078369905957, "grad_norm": 38.59257888793945, "learning_rate": 2.7155000000000005e-06, "loss": 2.3534, "step": 5431 }, { "epoch": 1.7028213166144202, "grad_norm": 24.967159271240234, "learning_rate": 2.7160000000000003e-06, "loss": 2.9059, "step": 5432 }, { "epoch": 1.7031347962382446, "grad_norm": 29.68940544128418, "learning_rate": 2.7165e-06, "loss": 2.2956, "step": 5433 }, { "epoch": 1.703448275862069, "grad_norm": 23.613834381103516, "learning_rate": 2.717e-06, "loss": 2.2259, "step": 5434 }, { "epoch": 1.7037617554858935, "grad_norm": 18.64572525024414, "learning_rate": 2.7175000000000002e-06, "loss": 2.419, "step": 5435 }, { "epoch": 1.704075235109718, "grad_norm": 16.19615936279297, "learning_rate": 2.718e-06, "loss": 2.9591, "step": 5436 }, { "epoch": 1.7043887147335424, "grad_norm": 18.514894485473633, "learning_rate": 2.7185e-06, "loss": 2.215, "step": 5437 }, { "epoch": 1.7047021943573668, "grad_norm": 18.013778686523438, "learning_rate": 2.719e-06, "loss": 2.3559, "step": 5438 }, { "epoch": 1.7050156739811912, "grad_norm": 20.540124893188477, "learning_rate": 2.7195000000000004e-06, "loss": 2.5045, "step": 5439 }, { "epoch": 1.7053291536050157, "grad_norm": 13.501602172851562, "learning_rate": 2.7200000000000002e-06, "loss": 1.9096, "step": 5440 }, { "epoch": 1.7056426332288401, "grad_norm": 11.370809555053711, "learning_rate": 2.7205000000000005e-06, "loss": 2.1822, "step": 5441 }, { "epoch": 1.7059561128526646, "grad_norm": 21.86370086669922, "learning_rate": 2.7210000000000003e-06, "loss": 2.1602, "step": 5442 }, { "epoch": 1.706269592476489, "grad_norm": 30.073101043701172, "learning_rate": 2.7215e-06, "loss": 2.6491, "step": 5443 }, { "epoch": 1.7065830721003135, "grad_norm": 17.897363662719727, "learning_rate": 2.7220000000000004e-06, "loss": 2.2375, "step": 5444 }, { "epoch": 1.706896551724138, "grad_norm": 11.158616065979004, "learning_rate": 2.7225e-06, "loss": 2.0454, "step": 5445 }, { "epoch": 1.7072100313479623, "grad_norm": 8.605175018310547, "learning_rate": 2.723e-06, "loss": 2.2267, "step": 5446 }, { "epoch": 1.7075235109717868, "grad_norm": 27.836732864379883, "learning_rate": 2.7235e-06, "loss": 2.3088, "step": 5447 }, { "epoch": 1.7078369905956112, "grad_norm": 14.671130180358887, "learning_rate": 2.724e-06, "loss": 2.7475, "step": 5448 }, { "epoch": 1.7081504702194357, "grad_norm": 20.618295669555664, "learning_rate": 2.7245000000000003e-06, "loss": 2.8887, "step": 5449 }, { "epoch": 1.70846394984326, "grad_norm": 38.52854537963867, "learning_rate": 2.7250000000000006e-06, "loss": 2.4788, "step": 5450 }, { "epoch": 1.7087774294670846, "grad_norm": 11.359956741333008, "learning_rate": 2.7255000000000004e-06, "loss": 2.3325, "step": 5451 }, { "epoch": 1.709090909090909, "grad_norm": 36.49653625488281, "learning_rate": 2.7260000000000002e-06, "loss": 2.5183, "step": 5452 }, { "epoch": 1.7094043887147334, "grad_norm": 25.822084426879883, "learning_rate": 2.7265e-06, "loss": 2.372, "step": 5453 }, { "epoch": 1.7097178683385579, "grad_norm": 85.20195770263672, "learning_rate": 2.7270000000000003e-06, "loss": 2.3598, "step": 5454 }, { "epoch": 1.7100313479623823, "grad_norm": 12.452218055725098, "learning_rate": 2.7275e-06, "loss": 2.493, "step": 5455 }, { "epoch": 1.7103448275862068, "grad_norm": 10.54835319519043, "learning_rate": 2.728e-06, "loss": 2.2473, "step": 5456 }, { "epoch": 1.7106583072100312, "grad_norm": 12.409470558166504, "learning_rate": 2.7285000000000002e-06, "loss": 2.4374, "step": 5457 }, { "epoch": 1.7109717868338556, "grad_norm": 12.735477447509766, "learning_rate": 2.729e-06, "loss": 2.047, "step": 5458 }, { "epoch": 1.71128526645768, "grad_norm": 12.568780899047852, "learning_rate": 2.7295000000000003e-06, "loss": 2.2044, "step": 5459 }, { "epoch": 1.7115987460815048, "grad_norm": 11.628466606140137, "learning_rate": 2.7300000000000005e-06, "loss": 2.1427, "step": 5460 }, { "epoch": 1.7119122257053292, "grad_norm": 34.290645599365234, "learning_rate": 2.7305000000000004e-06, "loss": 2.2639, "step": 5461 }, { "epoch": 1.7122257053291536, "grad_norm": 16.477323532104492, "learning_rate": 2.731e-06, "loss": 2.4689, "step": 5462 }, { "epoch": 1.712539184952978, "grad_norm": 15.450041770935059, "learning_rate": 2.7315000000000004e-06, "loss": 2.4502, "step": 5463 }, { "epoch": 1.7128526645768025, "grad_norm": 45.1892204284668, "learning_rate": 2.7320000000000003e-06, "loss": 2.48, "step": 5464 }, { "epoch": 1.713166144200627, "grad_norm": 16.953018188476562, "learning_rate": 2.7325e-06, "loss": 1.9744, "step": 5465 }, { "epoch": 1.7134796238244514, "grad_norm": 14.956791877746582, "learning_rate": 2.733e-06, "loss": 2.1789, "step": 5466 }, { "epoch": 1.7137931034482758, "grad_norm": 8.133874893188477, "learning_rate": 2.7335e-06, "loss": 2.0267, "step": 5467 }, { "epoch": 1.7141065830721003, "grad_norm": 17.319612503051758, "learning_rate": 2.734e-06, "loss": 2.1611, "step": 5468 }, { "epoch": 1.7144200626959247, "grad_norm": 13.33533763885498, "learning_rate": 2.7345000000000007e-06, "loss": 2.2603, "step": 5469 }, { "epoch": 1.7147335423197492, "grad_norm": 18.363788604736328, "learning_rate": 2.7350000000000005e-06, "loss": 2.5405, "step": 5470 }, { "epoch": 1.7150470219435736, "grad_norm": 17.000051498413086, "learning_rate": 2.7355000000000003e-06, "loss": 2.9102, "step": 5471 }, { "epoch": 1.7153605015673983, "grad_norm": 10.053338050842285, "learning_rate": 2.736e-06, "loss": 2.263, "step": 5472 }, { "epoch": 1.7156739811912227, "grad_norm": 42.593536376953125, "learning_rate": 2.7365000000000004e-06, "loss": 2.4529, "step": 5473 }, { "epoch": 1.7159874608150472, "grad_norm": 12.775186538696289, "learning_rate": 2.7370000000000002e-06, "loss": 2.4782, "step": 5474 }, { "epoch": 1.7163009404388716, "grad_norm": 23.6019287109375, "learning_rate": 2.7375e-06, "loss": 2.2155, "step": 5475 }, { "epoch": 1.716614420062696, "grad_norm": 12.975761413574219, "learning_rate": 2.7380000000000003e-06, "loss": 2.4409, "step": 5476 }, { "epoch": 1.7169278996865205, "grad_norm": 14.86938762664795, "learning_rate": 2.7385e-06, "loss": 2.3281, "step": 5477 }, { "epoch": 1.717241379310345, "grad_norm": 13.768951416015625, "learning_rate": 2.739e-06, "loss": 2.267, "step": 5478 }, { "epoch": 1.7175548589341694, "grad_norm": 33.40892791748047, "learning_rate": 2.7395000000000006e-06, "loss": 2.019, "step": 5479 }, { "epoch": 1.7178683385579938, "grad_norm": 18.362037658691406, "learning_rate": 2.7400000000000004e-06, "loss": 2.4633, "step": 5480 }, { "epoch": 1.7181818181818183, "grad_norm": 20.025041580200195, "learning_rate": 2.7405000000000003e-06, "loss": 2.3384, "step": 5481 }, { "epoch": 1.7184952978056427, "grad_norm": 18.812082290649414, "learning_rate": 2.7410000000000005e-06, "loss": 2.2198, "step": 5482 }, { "epoch": 1.7188087774294671, "grad_norm": 20.752206802368164, "learning_rate": 2.7415000000000003e-06, "loss": 2.2079, "step": 5483 }, { "epoch": 1.7191222570532916, "grad_norm": 11.19144344329834, "learning_rate": 2.742e-06, "loss": 2.5804, "step": 5484 }, { "epoch": 1.719435736677116, "grad_norm": 20.350521087646484, "learning_rate": 2.7425e-06, "loss": 2.561, "step": 5485 }, { "epoch": 1.7197492163009405, "grad_norm": 40.37380599975586, "learning_rate": 2.7430000000000002e-06, "loss": 2.5229, "step": 5486 }, { "epoch": 1.720062695924765, "grad_norm": 51.770198822021484, "learning_rate": 2.7435e-06, "loss": 2.4845, "step": 5487 }, { "epoch": 1.7203761755485893, "grad_norm": 21.724157333374023, "learning_rate": 2.744e-06, "loss": 2.0513, "step": 5488 }, { "epoch": 1.7206896551724138, "grad_norm": 9.700921058654785, "learning_rate": 2.7445000000000006e-06, "loss": 2.2333, "step": 5489 }, { "epoch": 1.7210031347962382, "grad_norm": 33.1986198425293, "learning_rate": 2.7450000000000004e-06, "loss": 2.3553, "step": 5490 }, { "epoch": 1.7213166144200627, "grad_norm": 26.239349365234375, "learning_rate": 2.7455000000000002e-06, "loss": 2.4481, "step": 5491 }, { "epoch": 1.7216300940438871, "grad_norm": 25.467309951782227, "learning_rate": 2.7460000000000005e-06, "loss": 2.5357, "step": 5492 }, { "epoch": 1.7219435736677116, "grad_norm": 9.850618362426758, "learning_rate": 2.7465000000000003e-06, "loss": 2.0866, "step": 5493 }, { "epoch": 1.722257053291536, "grad_norm": 18.098697662353516, "learning_rate": 2.747e-06, "loss": 2.303, "step": 5494 }, { "epoch": 1.7225705329153604, "grad_norm": 27.904993057250977, "learning_rate": 2.7475000000000004e-06, "loss": 2.2298, "step": 5495 }, { "epoch": 1.7228840125391849, "grad_norm": 11.535282135009766, "learning_rate": 2.748e-06, "loss": 1.9315, "step": 5496 }, { "epoch": 1.7231974921630093, "grad_norm": 20.059200286865234, "learning_rate": 2.7485e-06, "loss": 2.3591, "step": 5497 }, { "epoch": 1.7235109717868338, "grad_norm": 12.67073917388916, "learning_rate": 2.749e-06, "loss": 2.4145, "step": 5498 }, { "epoch": 1.7238244514106582, "grad_norm": 21.71510124206543, "learning_rate": 2.7495000000000005e-06, "loss": 2.323, "step": 5499 }, { "epoch": 1.7241379310344827, "grad_norm": 65.52584075927734, "learning_rate": 2.7500000000000004e-06, "loss": 2.7043, "step": 5500 }, { "epoch": 1.724451410658307, "grad_norm": 11.966585159301758, "learning_rate": 2.7505000000000006e-06, "loss": 2.2303, "step": 5501 }, { "epoch": 1.7247648902821315, "grad_norm": 34.113651275634766, "learning_rate": 2.7510000000000004e-06, "loss": 2.1599, "step": 5502 }, { "epoch": 1.725078369905956, "grad_norm": 33.491355895996094, "learning_rate": 2.7515000000000003e-06, "loss": 2.5904, "step": 5503 }, { "epoch": 1.7253918495297804, "grad_norm": 16.249862670898438, "learning_rate": 2.752e-06, "loss": 2.6213, "step": 5504 }, { "epoch": 1.7257053291536049, "grad_norm": 20.483736038208008, "learning_rate": 2.7525000000000003e-06, "loss": 2.2492, "step": 5505 }, { "epoch": 1.7260188087774293, "grad_norm": 15.815305709838867, "learning_rate": 2.753e-06, "loss": 2.0444, "step": 5506 }, { "epoch": 1.726332288401254, "grad_norm": 8.834308624267578, "learning_rate": 2.7535e-06, "loss": 2.2711, "step": 5507 }, { "epoch": 1.7266457680250784, "grad_norm": 29.914960861206055, "learning_rate": 2.754e-06, "loss": 2.2056, "step": 5508 }, { "epoch": 1.7269592476489029, "grad_norm": 22.222673416137695, "learning_rate": 2.7545e-06, "loss": 2.6003, "step": 5509 }, { "epoch": 1.7272727272727273, "grad_norm": 59.530555725097656, "learning_rate": 2.7550000000000003e-06, "loss": 3.0075, "step": 5510 }, { "epoch": 1.7275862068965517, "grad_norm": 10.841995239257812, "learning_rate": 2.7555000000000006e-06, "loss": 2.108, "step": 5511 }, { "epoch": 1.7278996865203762, "grad_norm": 39.829383850097656, "learning_rate": 2.7560000000000004e-06, "loss": 2.0235, "step": 5512 }, { "epoch": 1.7282131661442006, "grad_norm": 9.398069381713867, "learning_rate": 2.7565e-06, "loss": 2.2911, "step": 5513 }, { "epoch": 1.728526645768025, "grad_norm": 17.04472541809082, "learning_rate": 2.757e-06, "loss": 2.0957, "step": 5514 }, { "epoch": 1.7288401253918495, "grad_norm": 21.816059112548828, "learning_rate": 2.7575000000000003e-06, "loss": 1.9723, "step": 5515 }, { "epoch": 1.729153605015674, "grad_norm": 16.998680114746094, "learning_rate": 2.758e-06, "loss": 2.2045, "step": 5516 }, { "epoch": 1.7294670846394984, "grad_norm": 16.111724853515625, "learning_rate": 2.7585e-06, "loss": 2.7597, "step": 5517 }, { "epoch": 1.729780564263323, "grad_norm": 20.748037338256836, "learning_rate": 2.759e-06, "loss": 2.6406, "step": 5518 }, { "epoch": 1.7300940438871475, "grad_norm": 47.98612976074219, "learning_rate": 2.7595e-06, "loss": 2.7046, "step": 5519 }, { "epoch": 1.730407523510972, "grad_norm": 15.667613983154297, "learning_rate": 2.7600000000000003e-06, "loss": 2.359, "step": 5520 }, { "epoch": 1.7307210031347964, "grad_norm": 10.290972709655762, "learning_rate": 2.7605000000000005e-06, "loss": 2.3374, "step": 5521 }, { "epoch": 1.7310344827586208, "grad_norm": 39.547481536865234, "learning_rate": 2.7610000000000003e-06, "loss": 2.2134, "step": 5522 }, { "epoch": 1.7313479623824453, "grad_norm": 9.93276309967041, "learning_rate": 2.7615e-06, "loss": 2.5338, "step": 5523 }, { "epoch": 1.7316614420062697, "grad_norm": 36.52473831176758, "learning_rate": 2.7620000000000004e-06, "loss": 2.8736, "step": 5524 }, { "epoch": 1.7319749216300941, "grad_norm": 16.293170928955078, "learning_rate": 2.7625000000000002e-06, "loss": 2.0691, "step": 5525 }, { "epoch": 1.7322884012539186, "grad_norm": 15.633368492126465, "learning_rate": 2.763e-06, "loss": 2.2657, "step": 5526 }, { "epoch": 1.732601880877743, "grad_norm": 11.91658878326416, "learning_rate": 2.7635e-06, "loss": 2.2018, "step": 5527 }, { "epoch": 1.7329153605015675, "grad_norm": 70.28834533691406, "learning_rate": 2.764e-06, "loss": 2.817, "step": 5528 }, { "epoch": 1.733228840125392, "grad_norm": 29.530969619750977, "learning_rate": 2.7645e-06, "loss": 2.3077, "step": 5529 }, { "epoch": 1.7335423197492164, "grad_norm": 13.929141998291016, "learning_rate": 2.7650000000000006e-06, "loss": 2.6901, "step": 5530 }, { "epoch": 1.7338557993730408, "grad_norm": 15.058401107788086, "learning_rate": 2.7655000000000005e-06, "loss": 2.2645, "step": 5531 }, { "epoch": 1.7341692789968652, "grad_norm": 17.164827346801758, "learning_rate": 2.7660000000000003e-06, "loss": 2.3818, "step": 5532 }, { "epoch": 1.7344827586206897, "grad_norm": 17.328838348388672, "learning_rate": 2.7665e-06, "loss": 2.4011, "step": 5533 }, { "epoch": 1.7347962382445141, "grad_norm": 11.72863483428955, "learning_rate": 2.7670000000000004e-06, "loss": 2.5156, "step": 5534 }, { "epoch": 1.7351097178683386, "grad_norm": 15.69046688079834, "learning_rate": 2.7675e-06, "loss": 2.2565, "step": 5535 }, { "epoch": 1.735423197492163, "grad_norm": 10.286733627319336, "learning_rate": 2.768e-06, "loss": 2.2191, "step": 5536 }, { "epoch": 1.7357366771159874, "grad_norm": 12.609308242797852, "learning_rate": 2.7685000000000003e-06, "loss": 2.3434, "step": 5537 }, { "epoch": 1.736050156739812, "grad_norm": 10.516945838928223, "learning_rate": 2.769e-06, "loss": 2.2088, "step": 5538 }, { "epoch": 1.7363636363636363, "grad_norm": 20.95050048828125, "learning_rate": 2.7695e-06, "loss": 2.4958, "step": 5539 }, { "epoch": 1.7366771159874608, "grad_norm": 10.20532512664795, "learning_rate": 2.7700000000000006e-06, "loss": 2.5765, "step": 5540 }, { "epoch": 1.7369905956112852, "grad_norm": 11.721442222595215, "learning_rate": 2.7705000000000004e-06, "loss": 2.336, "step": 5541 }, { "epoch": 1.7373040752351097, "grad_norm": 21.627376556396484, "learning_rate": 2.7710000000000002e-06, "loss": 2.3839, "step": 5542 }, { "epoch": 1.737617554858934, "grad_norm": 17.041444778442383, "learning_rate": 2.7715000000000005e-06, "loss": 2.4114, "step": 5543 }, { "epoch": 1.7379310344827585, "grad_norm": 37.384986877441406, "learning_rate": 2.7720000000000003e-06, "loss": 2.3266, "step": 5544 }, { "epoch": 1.738244514106583, "grad_norm": 34.521976470947266, "learning_rate": 2.7725e-06, "loss": 2.3097, "step": 5545 }, { "epoch": 1.7385579937304074, "grad_norm": 37.94413375854492, "learning_rate": 2.773e-06, "loss": 2.3119, "step": 5546 }, { "epoch": 1.7388714733542319, "grad_norm": 13.897578239440918, "learning_rate": 2.7735e-06, "loss": 2.3442, "step": 5547 }, { "epoch": 1.7391849529780563, "grad_norm": 26.50729751586914, "learning_rate": 2.774e-06, "loss": 2.1502, "step": 5548 }, { "epoch": 1.7394984326018808, "grad_norm": 12.668210983276367, "learning_rate": 2.7745e-06, "loss": 2.1628, "step": 5549 }, { "epoch": 1.7398119122257052, "grad_norm": 9.052404403686523, "learning_rate": 2.7750000000000005e-06, "loss": 2.1179, "step": 5550 }, { "epoch": 1.7401253918495296, "grad_norm": 16.18061637878418, "learning_rate": 2.7755000000000004e-06, "loss": 2.0622, "step": 5551 }, { "epoch": 1.740438871473354, "grad_norm": 45.02144241333008, "learning_rate": 2.776e-06, "loss": 2.3295, "step": 5552 }, { "epoch": 1.7407523510971785, "grad_norm": 13.236767768859863, "learning_rate": 2.7765000000000004e-06, "loss": 2.2401, "step": 5553 }, { "epoch": 1.7410658307210032, "grad_norm": 20.257795333862305, "learning_rate": 2.7770000000000003e-06, "loss": 2.545, "step": 5554 }, { "epoch": 1.7413793103448276, "grad_norm": 11.673392295837402, "learning_rate": 2.7775e-06, "loss": 2.0406, "step": 5555 }, { "epoch": 1.741692789968652, "grad_norm": 14.821595191955566, "learning_rate": 2.7780000000000003e-06, "loss": 2.215, "step": 5556 }, { "epoch": 1.7420062695924765, "grad_norm": 58.19940948486328, "learning_rate": 2.7785e-06, "loss": 2.8323, "step": 5557 }, { "epoch": 1.742319749216301, "grad_norm": 30.191211700439453, "learning_rate": 2.779e-06, "loss": 2.2626, "step": 5558 }, { "epoch": 1.7426332288401254, "grad_norm": 8.197135925292969, "learning_rate": 2.7795e-06, "loss": 2.214, "step": 5559 }, { "epoch": 1.7429467084639498, "grad_norm": 15.255378723144531, "learning_rate": 2.7800000000000005e-06, "loss": 2.1536, "step": 5560 }, { "epoch": 1.7432601880877743, "grad_norm": 86.08578491210938, "learning_rate": 2.7805000000000003e-06, "loss": 2.4542, "step": 5561 }, { "epoch": 1.7435736677115987, "grad_norm": 17.217504501342773, "learning_rate": 2.7810000000000006e-06, "loss": 2.9291, "step": 5562 }, { "epoch": 1.7438871473354232, "grad_norm": 65.4100112915039, "learning_rate": 2.7815000000000004e-06, "loss": 2.2653, "step": 5563 }, { "epoch": 1.7442006269592476, "grad_norm": 17.314762115478516, "learning_rate": 2.7820000000000002e-06, "loss": 2.2533, "step": 5564 }, { "epoch": 1.7445141065830723, "grad_norm": 26.037796020507812, "learning_rate": 2.7825e-06, "loss": 2.3206, "step": 5565 }, { "epoch": 1.7448275862068967, "grad_norm": 9.311919212341309, "learning_rate": 2.7830000000000003e-06, "loss": 2.3518, "step": 5566 }, { "epoch": 1.7451410658307211, "grad_norm": 20.615209579467773, "learning_rate": 2.7835e-06, "loss": 2.1305, "step": 5567 }, { "epoch": 1.7454545454545456, "grad_norm": 15.762939453125, "learning_rate": 2.784e-06, "loss": 2.4987, "step": 5568 }, { "epoch": 1.74576802507837, "grad_norm": 18.63046646118164, "learning_rate": 2.7845e-06, "loss": 2.0101, "step": 5569 }, { "epoch": 1.7460815047021945, "grad_norm": 11.48450756072998, "learning_rate": 2.7850000000000004e-06, "loss": 2.2357, "step": 5570 }, { "epoch": 1.746394984326019, "grad_norm": 96.20525360107422, "learning_rate": 2.7855000000000003e-06, "loss": 2.1663, "step": 5571 }, { "epoch": 1.7467084639498434, "grad_norm": 17.738521575927734, "learning_rate": 2.7860000000000005e-06, "loss": 2.1544, "step": 5572 }, { "epoch": 1.7470219435736678, "grad_norm": 15.63515853881836, "learning_rate": 2.7865000000000003e-06, "loss": 2.4514, "step": 5573 }, { "epoch": 1.7473354231974922, "grad_norm": 22.114669799804688, "learning_rate": 2.787e-06, "loss": 2.0831, "step": 5574 }, { "epoch": 1.7476489028213167, "grad_norm": 31.60894012451172, "learning_rate": 2.7875000000000004e-06, "loss": 2.2297, "step": 5575 }, { "epoch": 1.7479623824451411, "grad_norm": 14.62855339050293, "learning_rate": 2.7880000000000002e-06, "loss": 2.2535, "step": 5576 }, { "epoch": 1.7482758620689656, "grad_norm": 14.154539108276367, "learning_rate": 2.7885e-06, "loss": 2.3844, "step": 5577 }, { "epoch": 1.74858934169279, "grad_norm": 91.77095794677734, "learning_rate": 2.789e-06, "loss": 2.9025, "step": 5578 }, { "epoch": 1.7489028213166145, "grad_norm": 30.688533782958984, "learning_rate": 2.7895e-06, "loss": 2.7722, "step": 5579 }, { "epoch": 1.749216300940439, "grad_norm": 11.394512176513672, "learning_rate": 2.7900000000000004e-06, "loss": 2.3959, "step": 5580 }, { "epoch": 1.7495297805642633, "grad_norm": 43.727603912353516, "learning_rate": 2.7905000000000006e-06, "loss": 2.2847, "step": 5581 }, { "epoch": 1.7498432601880878, "grad_norm": 18.254802703857422, "learning_rate": 2.7910000000000005e-06, "loss": 2.4319, "step": 5582 }, { "epoch": 1.7501567398119122, "grad_norm": 12.743124961853027, "learning_rate": 2.7915000000000003e-06, "loss": 2.4469, "step": 5583 }, { "epoch": 1.7504702194357367, "grad_norm": 29.064430236816406, "learning_rate": 2.792e-06, "loss": 2.3628, "step": 5584 }, { "epoch": 1.750783699059561, "grad_norm": 16.731903076171875, "learning_rate": 2.7925000000000004e-06, "loss": 2.3773, "step": 5585 }, { "epoch": 1.7510971786833855, "grad_norm": 24.20734977722168, "learning_rate": 2.793e-06, "loss": 2.5311, "step": 5586 }, { "epoch": 1.7510971786833855, "eval_loss": 2.7825205326080322, "eval_runtime": 20.9022, "eval_samples_per_second": 128.551, "eval_steps_per_second": 8.037, "step": 5586 }, { "epoch": 1.75141065830721, "grad_norm": 36.4649658203125, "learning_rate": 2.7935e-06, "loss": 2.6414, "step": 5587 }, { "epoch": 1.7517241379310344, "grad_norm": 15.779841423034668, "learning_rate": 2.794e-06, "loss": 2.2283, "step": 5588 }, { "epoch": 1.7520376175548589, "grad_norm": 16.722549438476562, "learning_rate": 2.7945e-06, "loss": 1.9054, "step": 5589 }, { "epoch": 1.7523510971786833, "grad_norm": 21.327329635620117, "learning_rate": 2.7950000000000003e-06, "loss": 2.4145, "step": 5590 }, { "epoch": 1.7526645768025078, "grad_norm": 15.905557632446289, "learning_rate": 2.7955000000000006e-06, "loss": 2.3218, "step": 5591 }, { "epoch": 1.7529780564263322, "grad_norm": 38.292198181152344, "learning_rate": 2.7960000000000004e-06, "loss": 2.2133, "step": 5592 }, { "epoch": 1.7532915360501566, "grad_norm": 16.73972511291504, "learning_rate": 2.7965000000000002e-06, "loss": 2.2877, "step": 5593 }, { "epoch": 1.753605015673981, "grad_norm": 33.32880783081055, "learning_rate": 2.797e-06, "loss": 2.4803, "step": 5594 }, { "epoch": 1.7539184952978055, "grad_norm": 12.935139656066895, "learning_rate": 2.7975000000000003e-06, "loss": 2.2784, "step": 5595 }, { "epoch": 1.75423197492163, "grad_norm": 65.7669448852539, "learning_rate": 2.798e-06, "loss": 2.5523, "step": 5596 }, { "epoch": 1.7545454545454544, "grad_norm": 23.66747283935547, "learning_rate": 2.7985e-06, "loss": 2.8021, "step": 5597 }, { "epoch": 1.7548589341692789, "grad_norm": 19.805585861206055, "learning_rate": 2.7990000000000002e-06, "loss": 2.7297, "step": 5598 }, { "epoch": 1.7551724137931033, "grad_norm": 19.813541412353516, "learning_rate": 2.7995e-06, "loss": 2.6171, "step": 5599 }, { "epoch": 1.7554858934169277, "grad_norm": 10.666125297546387, "learning_rate": 2.8000000000000003e-06, "loss": 2.1845, "step": 5600 }, { "epoch": 1.7557993730407524, "grad_norm": 14.279791831970215, "learning_rate": 2.8005000000000005e-06, "loss": 2.2536, "step": 5601 }, { "epoch": 1.7561128526645768, "grad_norm": 11.552290916442871, "learning_rate": 2.8010000000000004e-06, "loss": 1.9127, "step": 5602 }, { "epoch": 1.7564263322884013, "grad_norm": 10.985512733459473, "learning_rate": 2.8015e-06, "loss": 2.3382, "step": 5603 }, { "epoch": 1.7567398119122257, "grad_norm": 35.681644439697266, "learning_rate": 2.8020000000000004e-06, "loss": 2.9624, "step": 5604 }, { "epoch": 1.7570532915360502, "grad_norm": 24.726045608520508, "learning_rate": 2.8025000000000003e-06, "loss": 2.3681, "step": 5605 }, { "epoch": 1.7573667711598746, "grad_norm": 38.52384567260742, "learning_rate": 2.803e-06, "loss": 2.2576, "step": 5606 }, { "epoch": 1.757680250783699, "grad_norm": 21.216524124145508, "learning_rate": 2.8035e-06, "loss": 3.0623, "step": 5607 }, { "epoch": 1.7579937304075235, "grad_norm": 13.406828880310059, "learning_rate": 2.804e-06, "loss": 2.317, "step": 5608 }, { "epoch": 1.758307210031348, "grad_norm": 40.402992248535156, "learning_rate": 2.8045e-06, "loss": 2.461, "step": 5609 }, { "epoch": 1.7586206896551724, "grad_norm": 14.416365623474121, "learning_rate": 2.8050000000000007e-06, "loss": 2.4303, "step": 5610 }, { "epoch": 1.7589341692789968, "grad_norm": 38.3383903503418, "learning_rate": 2.8055000000000005e-06, "loss": 2.8362, "step": 5611 }, { "epoch": 1.7592476489028215, "grad_norm": 14.47104549407959, "learning_rate": 2.8060000000000003e-06, "loss": 2.3346, "step": 5612 }, { "epoch": 1.759561128526646, "grad_norm": 30.86193084716797, "learning_rate": 2.8065e-06, "loss": 2.4582, "step": 5613 }, { "epoch": 1.7598746081504704, "grad_norm": 22.559818267822266, "learning_rate": 2.8070000000000004e-06, "loss": 2.6912, "step": 5614 }, { "epoch": 1.7601880877742948, "grad_norm": 16.12966537475586, "learning_rate": 2.8075000000000002e-06, "loss": 2.3455, "step": 5615 }, { "epoch": 1.7605015673981192, "grad_norm": 20.706560134887695, "learning_rate": 2.808e-06, "loss": 2.2846, "step": 5616 }, { "epoch": 1.7608150470219437, "grad_norm": 25.860445022583008, "learning_rate": 2.8085000000000003e-06, "loss": 2.3389, "step": 5617 }, { "epoch": 1.7611285266457681, "grad_norm": 20.793603897094727, "learning_rate": 2.809e-06, "loss": 2.3378, "step": 5618 }, { "epoch": 1.7614420062695926, "grad_norm": 14.1869535446167, "learning_rate": 2.8095e-06, "loss": 2.5403, "step": 5619 }, { "epoch": 1.761755485893417, "grad_norm": 17.897798538208008, "learning_rate": 2.8100000000000006e-06, "loss": 2.4763, "step": 5620 }, { "epoch": 1.7620689655172415, "grad_norm": 14.624987602233887, "learning_rate": 2.8105000000000005e-06, "loss": 2.2365, "step": 5621 }, { "epoch": 1.762382445141066, "grad_norm": 51.904640197753906, "learning_rate": 2.8110000000000003e-06, "loss": 2.3868, "step": 5622 }, { "epoch": 1.7626959247648903, "grad_norm": 14.12952995300293, "learning_rate": 2.8115000000000005e-06, "loss": 2.2413, "step": 5623 }, { "epoch": 1.7630094043887148, "grad_norm": 34.63715744018555, "learning_rate": 2.8120000000000004e-06, "loss": 2.5888, "step": 5624 }, { "epoch": 1.7633228840125392, "grad_norm": 53.19129943847656, "learning_rate": 2.8125e-06, "loss": 2.4327, "step": 5625 }, { "epoch": 1.7636363636363637, "grad_norm": 26.255016326904297, "learning_rate": 2.813e-06, "loss": 2.3355, "step": 5626 }, { "epoch": 1.763949843260188, "grad_norm": 21.07077407836914, "learning_rate": 2.8135000000000003e-06, "loss": 2.2259, "step": 5627 }, { "epoch": 1.7642633228840126, "grad_norm": 27.582504272460938, "learning_rate": 2.814e-06, "loss": 2.3778, "step": 5628 }, { "epoch": 1.764576802507837, "grad_norm": 15.880824089050293, "learning_rate": 2.8145e-06, "loss": 2.355, "step": 5629 }, { "epoch": 1.7648902821316614, "grad_norm": 10.928897857666016, "learning_rate": 2.815e-06, "loss": 2.1491, "step": 5630 }, { "epoch": 1.7652037617554859, "grad_norm": 26.47492027282715, "learning_rate": 2.8155000000000004e-06, "loss": 2.2021, "step": 5631 }, { "epoch": 1.7655172413793103, "grad_norm": 193.11053466796875, "learning_rate": 2.8160000000000002e-06, "loss": 2.6495, "step": 5632 }, { "epoch": 1.7658307210031348, "grad_norm": 26.360347747802734, "learning_rate": 2.8165000000000005e-06, "loss": 2.2046, "step": 5633 }, { "epoch": 1.7661442006269592, "grad_norm": 22.70421028137207, "learning_rate": 2.8170000000000003e-06, "loss": 2.2962, "step": 5634 }, { "epoch": 1.7664576802507836, "grad_norm": 14.537629127502441, "learning_rate": 2.8175e-06, "loss": 2.3563, "step": 5635 }, { "epoch": 1.766771159874608, "grad_norm": 31.513465881347656, "learning_rate": 2.8180000000000004e-06, "loss": 2.3216, "step": 5636 }, { "epoch": 1.7670846394984325, "grad_norm": 15.880422592163086, "learning_rate": 2.8185e-06, "loss": 2.4413, "step": 5637 }, { "epoch": 1.767398119122257, "grad_norm": 18.613235473632812, "learning_rate": 2.819e-06, "loss": 2.5158, "step": 5638 }, { "epoch": 1.7677115987460814, "grad_norm": 14.459510803222656, "learning_rate": 2.8195e-06, "loss": 2.3783, "step": 5639 }, { "epoch": 1.7680250783699059, "grad_norm": 11.37955093383789, "learning_rate": 2.82e-06, "loss": 2.141, "step": 5640 }, { "epoch": 1.7683385579937303, "grad_norm": 20.611160278320312, "learning_rate": 2.8205000000000004e-06, "loss": 2.4464, "step": 5641 }, { "epoch": 1.7686520376175547, "grad_norm": 82.4933090209961, "learning_rate": 2.8210000000000006e-06, "loss": 2.3389, "step": 5642 }, { "epoch": 1.7689655172413792, "grad_norm": 15.467883110046387, "learning_rate": 2.8215000000000004e-06, "loss": 2.4918, "step": 5643 }, { "epoch": 1.7692789968652036, "grad_norm": 9.933538436889648, "learning_rate": 2.8220000000000003e-06, "loss": 2.1153, "step": 5644 }, { "epoch": 1.769592476489028, "grad_norm": 40.362850189208984, "learning_rate": 2.8225e-06, "loss": 2.2969, "step": 5645 }, { "epoch": 1.7699059561128525, "grad_norm": 35.1439094543457, "learning_rate": 2.8230000000000003e-06, "loss": 2.2496, "step": 5646 }, { "epoch": 1.770219435736677, "grad_norm": 16.86162757873535, "learning_rate": 2.8235e-06, "loss": 2.4232, "step": 5647 }, { "epoch": 1.7705329153605016, "grad_norm": 22.56192970275879, "learning_rate": 2.824e-06, "loss": 2.3042, "step": 5648 }, { "epoch": 1.770846394984326, "grad_norm": 30.514720916748047, "learning_rate": 2.8245e-06, "loss": 2.6457, "step": 5649 }, { "epoch": 1.7711598746081505, "grad_norm": 10.613741874694824, "learning_rate": 2.825e-06, "loss": 2.3016, "step": 5650 }, { "epoch": 1.771473354231975, "grad_norm": 15.156309127807617, "learning_rate": 2.8255000000000003e-06, "loss": 2.231, "step": 5651 }, { "epoch": 1.7717868338557994, "grad_norm": 81.3517074584961, "learning_rate": 2.8260000000000006e-06, "loss": 2.0737, "step": 5652 }, { "epoch": 1.7721003134796238, "grad_norm": 14.351300239562988, "learning_rate": 2.8265000000000004e-06, "loss": 2.106, "step": 5653 }, { "epoch": 1.7724137931034483, "grad_norm": 14.630460739135742, "learning_rate": 2.827e-06, "loss": 2.2051, "step": 5654 }, { "epoch": 1.7727272727272727, "grad_norm": 14.13293170928955, "learning_rate": 2.8275e-06, "loss": 2.3056, "step": 5655 }, { "epoch": 1.7730407523510971, "grad_norm": 7.3408026695251465, "learning_rate": 2.8280000000000003e-06, "loss": 2.1562, "step": 5656 }, { "epoch": 1.7733542319749216, "grad_norm": 86.32215881347656, "learning_rate": 2.8285e-06, "loss": 2.5063, "step": 5657 }, { "epoch": 1.773667711598746, "grad_norm": 15.42455768585205, "learning_rate": 2.829e-06, "loss": 2.2239, "step": 5658 }, { "epoch": 1.7739811912225707, "grad_norm": 18.74974822998047, "learning_rate": 2.8295e-06, "loss": 2.3227, "step": 5659 }, { "epoch": 1.7742946708463951, "grad_norm": 16.02386474609375, "learning_rate": 2.83e-06, "loss": 2.3212, "step": 5660 }, { "epoch": 1.7746081504702196, "grad_norm": 12.063882827758789, "learning_rate": 2.8305000000000003e-06, "loss": 2.3197, "step": 5661 }, { "epoch": 1.774921630094044, "grad_norm": 11.073810577392578, "learning_rate": 2.8310000000000005e-06, "loss": 2.2106, "step": 5662 }, { "epoch": 1.7752351097178685, "grad_norm": 35.60942840576172, "learning_rate": 2.8315000000000003e-06, "loss": 2.3919, "step": 5663 }, { "epoch": 1.775548589341693, "grad_norm": 11.828412055969238, "learning_rate": 2.832e-06, "loss": 2.1175, "step": 5664 }, { "epoch": 1.7758620689655173, "grad_norm": 42.694332122802734, "learning_rate": 2.8325000000000004e-06, "loss": 2.7357, "step": 5665 }, { "epoch": 1.7761755485893418, "grad_norm": 12.503061294555664, "learning_rate": 2.8330000000000002e-06, "loss": 2.5552, "step": 5666 }, { "epoch": 1.7764890282131662, "grad_norm": 11.709190368652344, "learning_rate": 2.8335e-06, "loss": 2.1738, "step": 5667 }, { "epoch": 1.7768025078369907, "grad_norm": 25.45659637451172, "learning_rate": 2.834e-06, "loss": 2.4358, "step": 5668 }, { "epoch": 1.7771159874608151, "grad_norm": 38.15581512451172, "learning_rate": 2.8345e-06, "loss": 2.1994, "step": 5669 }, { "epoch": 1.7774294670846396, "grad_norm": 65.19048309326172, "learning_rate": 2.835e-06, "loss": 4.0908, "step": 5670 }, { "epoch": 1.777742946708464, "grad_norm": 11.405741691589355, "learning_rate": 2.8355000000000006e-06, "loss": 2.2353, "step": 5671 }, { "epoch": 1.7780564263322884, "grad_norm": 16.600887298583984, "learning_rate": 2.8360000000000005e-06, "loss": 1.9491, "step": 5672 }, { "epoch": 1.7783699059561129, "grad_norm": 19.279136657714844, "learning_rate": 2.8365000000000003e-06, "loss": 2.1248, "step": 5673 }, { "epoch": 1.7786833855799373, "grad_norm": 20.416065216064453, "learning_rate": 2.837e-06, "loss": 2.6189, "step": 5674 }, { "epoch": 1.7789968652037618, "grad_norm": 12.208967208862305, "learning_rate": 2.8375000000000004e-06, "loss": 2.1409, "step": 5675 }, { "epoch": 1.7793103448275862, "grad_norm": 92.15316009521484, "learning_rate": 2.838e-06, "loss": 3.3053, "step": 5676 }, { "epoch": 1.7796238244514107, "grad_norm": 62.69886016845703, "learning_rate": 2.8385e-06, "loss": 3.2674, "step": 5677 }, { "epoch": 1.779937304075235, "grad_norm": 36.62449264526367, "learning_rate": 2.8390000000000003e-06, "loss": 2.3739, "step": 5678 }, { "epoch": 1.7802507836990595, "grad_norm": 30.333797454833984, "learning_rate": 2.8395e-06, "loss": 2.167, "step": 5679 }, { "epoch": 1.780564263322884, "grad_norm": 18.453176498413086, "learning_rate": 2.84e-06, "loss": 2.1609, "step": 5680 }, { "epoch": 1.7808777429467084, "grad_norm": 13.545697212219238, "learning_rate": 2.8405000000000006e-06, "loss": 2.1917, "step": 5681 }, { "epoch": 1.7811912225705329, "grad_norm": 37.35604476928711, "learning_rate": 2.8410000000000004e-06, "loss": 2.1792, "step": 5682 }, { "epoch": 1.7815047021943573, "grad_norm": 39.16984558105469, "learning_rate": 2.8415000000000002e-06, "loss": 2.2692, "step": 5683 }, { "epoch": 1.7818181818181817, "grad_norm": 45.23867416381836, "learning_rate": 2.8420000000000005e-06, "loss": 2.1975, "step": 5684 }, { "epoch": 1.7821316614420062, "grad_norm": 16.841379165649414, "learning_rate": 2.8425000000000003e-06, "loss": 2.2613, "step": 5685 }, { "epoch": 1.7824451410658306, "grad_norm": 14.964749336242676, "learning_rate": 2.843e-06, "loss": 2.6653, "step": 5686 }, { "epoch": 1.782758620689655, "grad_norm": 45.70800018310547, "learning_rate": 2.8435e-06, "loss": 2.3517, "step": 5687 }, { "epoch": 1.7830721003134795, "grad_norm": 26.868568420410156, "learning_rate": 2.8440000000000002e-06, "loss": 2.0901, "step": 5688 }, { "epoch": 1.783385579937304, "grad_norm": 67.97185516357422, "learning_rate": 2.8445e-06, "loss": 2.3446, "step": 5689 }, { "epoch": 1.7836990595611284, "grad_norm": 16.462514877319336, "learning_rate": 2.845e-06, "loss": 2.2372, "step": 5690 }, { "epoch": 1.7840125391849528, "grad_norm": 20.244930267333984, "learning_rate": 2.8455000000000005e-06, "loss": 2.4724, "step": 5691 }, { "epoch": 1.7843260188087773, "grad_norm": 31.71061134338379, "learning_rate": 2.8460000000000004e-06, "loss": 2.4825, "step": 5692 }, { "epoch": 1.7846394984326017, "grad_norm": 15.576581954956055, "learning_rate": 2.8465e-06, "loss": 1.9315, "step": 5693 }, { "epoch": 1.7849529780564264, "grad_norm": 21.8217716217041, "learning_rate": 2.8470000000000004e-06, "loss": 2.428, "step": 5694 }, { "epoch": 1.7852664576802508, "grad_norm": 12.862072944641113, "learning_rate": 2.8475000000000003e-06, "loss": 2.0832, "step": 5695 }, { "epoch": 1.7855799373040753, "grad_norm": 26.172510147094727, "learning_rate": 2.848e-06, "loss": 2.2061, "step": 5696 }, { "epoch": 1.7858934169278997, "grad_norm": 9.237683296203613, "learning_rate": 2.8485000000000003e-06, "loss": 2.2316, "step": 5697 }, { "epoch": 1.7862068965517242, "grad_norm": 34.316192626953125, "learning_rate": 2.849e-06, "loss": 2.188, "step": 5698 }, { "epoch": 1.7865203761755486, "grad_norm": 41.66139221191406, "learning_rate": 2.8495e-06, "loss": 2.1467, "step": 5699 }, { "epoch": 1.786833855799373, "grad_norm": 24.990121841430664, "learning_rate": 2.85e-06, "loss": 2.2904, "step": 5700 }, { "epoch": 1.7871473354231975, "grad_norm": 15.360755920410156, "learning_rate": 2.8505000000000005e-06, "loss": 2.3998, "step": 5701 }, { "epoch": 1.787460815047022, "grad_norm": 13.665358543395996, "learning_rate": 2.8510000000000003e-06, "loss": 1.9507, "step": 5702 }, { "epoch": 1.7877742946708464, "grad_norm": 16.122602462768555, "learning_rate": 2.8515000000000006e-06, "loss": 2.5211, "step": 5703 }, { "epoch": 1.7880877742946708, "grad_norm": 18.225542068481445, "learning_rate": 2.8520000000000004e-06, "loss": 2.325, "step": 5704 }, { "epoch": 1.7884012539184952, "grad_norm": 10.050536155700684, "learning_rate": 2.8525000000000002e-06, "loss": 2.3043, "step": 5705 }, { "epoch": 1.78871473354232, "grad_norm": 15.066789627075195, "learning_rate": 2.853e-06, "loss": 2.0683, "step": 5706 }, { "epoch": 1.7890282131661444, "grad_norm": 22.185789108276367, "learning_rate": 2.8535000000000003e-06, "loss": 2.4388, "step": 5707 }, { "epoch": 1.7893416927899688, "grad_norm": 29.357311248779297, "learning_rate": 2.854e-06, "loss": 2.3013, "step": 5708 }, { "epoch": 1.7896551724137932, "grad_norm": 40.66057205200195, "learning_rate": 2.8545e-06, "loss": 2.5906, "step": 5709 }, { "epoch": 1.7899686520376177, "grad_norm": 26.080907821655273, "learning_rate": 2.855e-06, "loss": 2.2303, "step": 5710 }, { "epoch": 1.7902821316614421, "grad_norm": 21.70749282836914, "learning_rate": 2.8555000000000004e-06, "loss": 2.2158, "step": 5711 }, { "epoch": 1.7905956112852666, "grad_norm": 12.691455841064453, "learning_rate": 2.8560000000000003e-06, "loss": 2.4867, "step": 5712 }, { "epoch": 1.790909090909091, "grad_norm": 13.39187240600586, "learning_rate": 2.8565000000000005e-06, "loss": 2.1281, "step": 5713 }, { "epoch": 1.7912225705329154, "grad_norm": 75.04694366455078, "learning_rate": 2.8570000000000003e-06, "loss": 2.0658, "step": 5714 }, { "epoch": 1.79153605015674, "grad_norm": 27.390806198120117, "learning_rate": 2.8575e-06, "loss": 2.898, "step": 5715 }, { "epoch": 1.7918495297805643, "grad_norm": 22.34684181213379, "learning_rate": 2.8580000000000004e-06, "loss": 2.4863, "step": 5716 }, { "epoch": 1.7921630094043888, "grad_norm": 19.219614028930664, "learning_rate": 2.8585000000000002e-06, "loss": 2.5412, "step": 5717 }, { "epoch": 1.7924764890282132, "grad_norm": 16.021224975585938, "learning_rate": 2.859e-06, "loss": 2.3316, "step": 5718 }, { "epoch": 1.7927899686520377, "grad_norm": 8.465279579162598, "learning_rate": 2.8595e-06, "loss": 2.2175, "step": 5719 }, { "epoch": 1.793103448275862, "grad_norm": 15.827347755432129, "learning_rate": 2.86e-06, "loss": 2.2974, "step": 5720 }, { "epoch": 1.7934169278996865, "grad_norm": 15.751001358032227, "learning_rate": 2.8605000000000004e-06, "loss": 2.335, "step": 5721 }, { "epoch": 1.793730407523511, "grad_norm": 15.49514389038086, "learning_rate": 2.8610000000000006e-06, "loss": 2.2435, "step": 5722 }, { "epoch": 1.7940438871473354, "grad_norm": 53.53825759887695, "learning_rate": 2.8615000000000005e-06, "loss": 2.6453, "step": 5723 }, { "epoch": 1.7943573667711599, "grad_norm": 19.740135192871094, "learning_rate": 2.8620000000000003e-06, "loss": 2.3032, "step": 5724 }, { "epoch": 1.7946708463949843, "grad_norm": 40.47820281982422, "learning_rate": 2.8625e-06, "loss": 2.433, "step": 5725 }, { "epoch": 1.7949843260188088, "grad_norm": 13.667232513427734, "learning_rate": 2.8630000000000004e-06, "loss": 2.0633, "step": 5726 }, { "epoch": 1.7952978056426332, "grad_norm": 133.35665893554688, "learning_rate": 2.8635e-06, "loss": 2.2366, "step": 5727 }, { "epoch": 1.7956112852664576, "grad_norm": 19.358415603637695, "learning_rate": 2.864e-06, "loss": 2.5498, "step": 5728 }, { "epoch": 1.795924764890282, "grad_norm": 69.85310363769531, "learning_rate": 2.8645e-06, "loss": 2.8183, "step": 5729 }, { "epoch": 1.7962382445141065, "grad_norm": 21.108644485473633, "learning_rate": 2.865e-06, "loss": 2.4494, "step": 5730 }, { "epoch": 1.796551724137931, "grad_norm": 12.527911186218262, "learning_rate": 2.8655000000000003e-06, "loss": 2.2376, "step": 5731 }, { "epoch": 1.7968652037617554, "grad_norm": 18.049640655517578, "learning_rate": 2.8660000000000006e-06, "loss": 2.2159, "step": 5732 }, { "epoch": 1.7971786833855798, "grad_norm": 14.708891868591309, "learning_rate": 2.8665000000000004e-06, "loss": 2.0556, "step": 5733 }, { "epoch": 1.7974921630094043, "grad_norm": 20.81380271911621, "learning_rate": 2.8670000000000002e-06, "loss": 2.6582, "step": 5734 }, { "epoch": 1.7978056426332287, "grad_norm": 16.69767951965332, "learning_rate": 2.8675e-06, "loss": 2.5488, "step": 5735 }, { "epoch": 1.7981191222570532, "grad_norm": 25.40092658996582, "learning_rate": 2.8680000000000003e-06, "loss": 2.279, "step": 5736 }, { "epoch": 1.7984326018808776, "grad_norm": 8.9202241897583, "learning_rate": 2.8685e-06, "loss": 2.3573, "step": 5737 }, { "epoch": 1.798746081504702, "grad_norm": 26.168312072753906, "learning_rate": 2.869e-06, "loss": 2.3775, "step": 5738 }, { "epoch": 1.7990595611285265, "grad_norm": 14.53439712524414, "learning_rate": 2.8695000000000002e-06, "loss": 2.5065, "step": 5739 }, { "epoch": 1.799373040752351, "grad_norm": 16.31795883178711, "learning_rate": 2.87e-06, "loss": 2.4633, "step": 5740 }, { "epoch": 1.7996865203761756, "grad_norm": 28.0877742767334, "learning_rate": 2.8705000000000003e-06, "loss": 2.4924, "step": 5741 }, { "epoch": 1.8, "grad_norm": 18.114452362060547, "learning_rate": 2.8710000000000005e-06, "loss": 2.3432, "step": 5742 }, { "epoch": 1.8003134796238245, "grad_norm": 16.870342254638672, "learning_rate": 2.8715000000000004e-06, "loss": 2.624, "step": 5743 }, { "epoch": 1.800626959247649, "grad_norm": 11.62446403503418, "learning_rate": 2.872e-06, "loss": 2.2845, "step": 5744 }, { "epoch": 1.8009404388714734, "grad_norm": 13.54909896850586, "learning_rate": 2.8725000000000004e-06, "loss": 2.8479, "step": 5745 }, { "epoch": 1.8012539184952978, "grad_norm": 11.23021125793457, "learning_rate": 2.8730000000000003e-06, "loss": 2.3157, "step": 5746 }, { "epoch": 1.8015673981191223, "grad_norm": 24.773460388183594, "learning_rate": 2.8735e-06, "loss": 2.1324, "step": 5747 }, { "epoch": 1.8018808777429467, "grad_norm": 28.940399169921875, "learning_rate": 2.874e-06, "loss": 2.3426, "step": 5748 }, { "epoch": 1.8021943573667711, "grad_norm": 16.026710510253906, "learning_rate": 2.8745e-06, "loss": 2.1109, "step": 5749 }, { "epoch": 1.8025078369905956, "grad_norm": 24.745437622070312, "learning_rate": 2.875e-06, "loss": 2.5266, "step": 5750 }, { "epoch": 1.80282131661442, "grad_norm": 12.666834831237793, "learning_rate": 2.8755e-06, "loss": 2.153, "step": 5751 }, { "epoch": 1.8031347962382445, "grad_norm": 21.118148803710938, "learning_rate": 2.8760000000000005e-06, "loss": 2.0629, "step": 5752 }, { "epoch": 1.8034482758620691, "grad_norm": 40.68550491333008, "learning_rate": 2.8765000000000003e-06, "loss": 2.7818, "step": 5753 }, { "epoch": 1.8037617554858936, "grad_norm": 8.309558868408203, "learning_rate": 2.877e-06, "loss": 2.3525, "step": 5754 }, { "epoch": 1.804075235109718, "grad_norm": 18.94241714477539, "learning_rate": 2.8775000000000004e-06, "loss": 2.2492, "step": 5755 }, { "epoch": 1.8043887147335425, "grad_norm": 77.01664733886719, "learning_rate": 2.8780000000000002e-06, "loss": 2.0521, "step": 5756 }, { "epoch": 1.804702194357367, "grad_norm": 38.76837158203125, "learning_rate": 2.8785e-06, "loss": 3.3167, "step": 5757 }, { "epoch": 1.8050156739811913, "grad_norm": 24.91385841369629, "learning_rate": 2.8790000000000003e-06, "loss": 2.5723, "step": 5758 }, { "epoch": 1.8053291536050158, "grad_norm": 35.777259826660156, "learning_rate": 2.8795e-06, "loss": 2.3896, "step": 5759 }, { "epoch": 1.8056426332288402, "grad_norm": 39.020076751708984, "learning_rate": 2.88e-06, "loss": 2.5884, "step": 5760 }, { "epoch": 1.8059561128526647, "grad_norm": 31.04021453857422, "learning_rate": 2.8804999999999998e-06, "loss": 2.3261, "step": 5761 }, { "epoch": 1.806269592476489, "grad_norm": 17.670146942138672, "learning_rate": 2.8810000000000005e-06, "loss": 2.0315, "step": 5762 }, { "epoch": 1.8065830721003135, "grad_norm": 21.669349670410156, "learning_rate": 2.8815000000000003e-06, "loss": 2.3209, "step": 5763 }, { "epoch": 1.806896551724138, "grad_norm": 21.78011703491211, "learning_rate": 2.8820000000000005e-06, "loss": 2.4031, "step": 5764 }, { "epoch": 1.8072100313479624, "grad_norm": 16.515914916992188, "learning_rate": 2.8825000000000004e-06, "loss": 2.2467, "step": 5765 }, { "epoch": 1.8075235109717869, "grad_norm": 14.0270414352417, "learning_rate": 2.883e-06, "loss": 2.7854, "step": 5766 }, { "epoch": 1.8078369905956113, "grad_norm": 20.48775291442871, "learning_rate": 2.8835e-06, "loss": 2.5621, "step": 5767 }, { "epoch": 1.8081504702194358, "grad_norm": 11.038260459899902, "learning_rate": 2.8840000000000003e-06, "loss": 2.1779, "step": 5768 }, { "epoch": 1.8084639498432602, "grad_norm": 19.993820190429688, "learning_rate": 2.8845e-06, "loss": 2.2544, "step": 5769 }, { "epoch": 1.8087774294670846, "grad_norm": 18.76875114440918, "learning_rate": 2.885e-06, "loss": 2.74, "step": 5770 }, { "epoch": 1.809090909090909, "grad_norm": 29.65410804748535, "learning_rate": 2.8855e-06, "loss": 2.3125, "step": 5771 }, { "epoch": 1.8094043887147335, "grad_norm": 8.846988677978516, "learning_rate": 2.8860000000000004e-06, "loss": 2.4192, "step": 5772 }, { "epoch": 1.809717868338558, "grad_norm": 15.321855545043945, "learning_rate": 2.8865000000000002e-06, "loss": 2.3592, "step": 5773 }, { "epoch": 1.8100313479623824, "grad_norm": 68.62010955810547, "learning_rate": 2.8870000000000005e-06, "loss": 2.7464, "step": 5774 }, { "epoch": 1.8103448275862069, "grad_norm": 11.731450080871582, "learning_rate": 2.8875000000000003e-06, "loss": 2.1994, "step": 5775 }, { "epoch": 1.8106583072100313, "grad_norm": 22.625207901000977, "learning_rate": 2.888e-06, "loss": 2.3414, "step": 5776 }, { "epoch": 1.8109717868338557, "grad_norm": 12.00716781616211, "learning_rate": 2.8885000000000004e-06, "loss": 2.2781, "step": 5777 }, { "epoch": 1.8112852664576802, "grad_norm": 15.153823852539062, "learning_rate": 2.889e-06, "loss": 2.3895, "step": 5778 }, { "epoch": 1.8115987460815046, "grad_norm": 33.36376190185547, "learning_rate": 2.8895e-06, "loss": 2.9956, "step": 5779 }, { "epoch": 1.811912225705329, "grad_norm": 16.726226806640625, "learning_rate": 2.89e-06, "loss": 2.2121, "step": 5780 }, { "epoch": 1.8122257053291535, "grad_norm": 102.48206329345703, "learning_rate": 2.8905e-06, "loss": 2.3842, "step": 5781 }, { "epoch": 1.812539184952978, "grad_norm": 11.28581714630127, "learning_rate": 2.8910000000000004e-06, "loss": 2.3171, "step": 5782 }, { "epoch": 1.8128526645768024, "grad_norm": 13.72201156616211, "learning_rate": 2.8915000000000006e-06, "loss": 2.279, "step": 5783 }, { "epoch": 1.8131661442006268, "grad_norm": 23.668373107910156, "learning_rate": 2.8920000000000004e-06, "loss": 2.2777, "step": 5784 }, { "epoch": 1.8134796238244513, "grad_norm": 18.73801040649414, "learning_rate": 2.8925000000000003e-06, "loss": 2.475, "step": 5785 }, { "epoch": 1.8137931034482757, "grad_norm": 22.289701461791992, "learning_rate": 2.893e-06, "loss": 2.302, "step": 5786 }, { "epoch": 1.8141065830721002, "grad_norm": 12.893795013427734, "learning_rate": 2.8935000000000003e-06, "loss": 2.3009, "step": 5787 }, { "epoch": 1.8144200626959248, "grad_norm": 21.486906051635742, "learning_rate": 2.894e-06, "loss": 2.6981, "step": 5788 }, { "epoch": 1.8147335423197493, "grad_norm": 19.843420028686523, "learning_rate": 2.8945e-06, "loss": 1.9245, "step": 5789 }, { "epoch": 1.8150470219435737, "grad_norm": 16.991474151611328, "learning_rate": 2.8950000000000002e-06, "loss": 2.3261, "step": 5790 }, { "epoch": 1.8153605015673981, "grad_norm": 15.307354927062988, "learning_rate": 2.8955e-06, "loss": 2.215, "step": 5791 }, { "epoch": 1.8156739811912226, "grad_norm": 18.179475784301758, "learning_rate": 2.8960000000000003e-06, "loss": 2.4475, "step": 5792 }, { "epoch": 1.815987460815047, "grad_norm": 16.987171173095703, "learning_rate": 2.8965000000000006e-06, "loss": 2.2285, "step": 5793 }, { "epoch": 1.8163009404388715, "grad_norm": 11.264395713806152, "learning_rate": 2.8970000000000004e-06, "loss": 2.2343, "step": 5794 }, { "epoch": 1.816614420062696, "grad_norm": 28.56736946105957, "learning_rate": 2.8975e-06, "loss": 2.4624, "step": 5795 }, { "epoch": 1.8169278996865204, "grad_norm": 11.755648612976074, "learning_rate": 2.8980000000000005e-06, "loss": 2.2397, "step": 5796 }, { "epoch": 1.8172413793103448, "grad_norm": 9.479302406311035, "learning_rate": 2.8985000000000003e-06, "loss": 2.146, "step": 5797 }, { "epoch": 1.8175548589341692, "grad_norm": 82.61997985839844, "learning_rate": 2.899e-06, "loss": 2.3617, "step": 5798 }, { "epoch": 1.8178683385579937, "grad_norm": 10.157063484191895, "learning_rate": 2.8995e-06, "loss": 2.0932, "step": 5799 }, { "epoch": 1.8181818181818183, "grad_norm": 11.576388359069824, "learning_rate": 2.9e-06, "loss": 2.2464, "step": 5800 }, { "epoch": 1.8184952978056428, "grad_norm": 14.041361808776855, "learning_rate": 2.9005e-06, "loss": 2.2786, "step": 5801 }, { "epoch": 1.8188087774294672, "grad_norm": 44.641910552978516, "learning_rate": 2.9010000000000007e-06, "loss": 2.2961, "step": 5802 }, { "epoch": 1.8191222570532917, "grad_norm": 15.94238567352295, "learning_rate": 2.9015000000000005e-06, "loss": 1.8881, "step": 5803 }, { "epoch": 1.8194357366771161, "grad_norm": 46.97993087768555, "learning_rate": 2.9020000000000003e-06, "loss": 2.2066, "step": 5804 }, { "epoch": 1.8197492163009406, "grad_norm": 11.219070434570312, "learning_rate": 2.9025e-06, "loss": 2.0557, "step": 5805 }, { "epoch": 1.820062695924765, "grad_norm": 44.40027618408203, "learning_rate": 2.9030000000000004e-06, "loss": 2.4332, "step": 5806 }, { "epoch": 1.8203761755485894, "grad_norm": 23.074581146240234, "learning_rate": 2.9035000000000002e-06, "loss": 2.328, "step": 5807 }, { "epoch": 1.8206896551724139, "grad_norm": 23.892196655273438, "learning_rate": 2.904e-06, "loss": 2.2622, "step": 5808 }, { "epoch": 1.8210031347962383, "grad_norm": 68.89410400390625, "learning_rate": 2.9045e-06, "loss": 3.0924, "step": 5809 }, { "epoch": 1.8213166144200628, "grad_norm": 11.803862571716309, "learning_rate": 2.905e-06, "loss": 2.1911, "step": 5810 }, { "epoch": 1.8216300940438872, "grad_norm": 25.095535278320312, "learning_rate": 2.9055e-06, "loss": 2.3491, "step": 5811 }, { "epoch": 1.8219435736677116, "grad_norm": 13.671652793884277, "learning_rate": 2.9060000000000006e-06, "loss": 2.2188, "step": 5812 }, { "epoch": 1.822257053291536, "grad_norm": 25.322555541992188, "learning_rate": 2.9065000000000005e-06, "loss": 2.3493, "step": 5813 }, { "epoch": 1.8225705329153605, "grad_norm": 27.850494384765625, "learning_rate": 2.9070000000000003e-06, "loss": 2.192, "step": 5814 }, { "epoch": 1.822884012539185, "grad_norm": 23.03537368774414, "learning_rate": 2.9075e-06, "loss": 2.2677, "step": 5815 }, { "epoch": 1.8231974921630094, "grad_norm": 63.14573287963867, "learning_rate": 2.9080000000000004e-06, "loss": 2.7937, "step": 5816 }, { "epoch": 1.8235109717868339, "grad_norm": 93.47972869873047, "learning_rate": 2.9085e-06, "loss": 2.2098, "step": 5817 }, { "epoch": 1.8238244514106583, "grad_norm": 24.212627410888672, "learning_rate": 2.909e-06, "loss": 2.2872, "step": 5818 }, { "epoch": 1.8241379310344827, "grad_norm": 11.589085578918457, "learning_rate": 2.9095000000000003e-06, "loss": 2.227, "step": 5819 }, { "epoch": 1.8244514106583072, "grad_norm": 29.045310974121094, "learning_rate": 2.91e-06, "loss": 2.272, "step": 5820 }, { "epoch": 1.8247648902821316, "grad_norm": 112.3254623413086, "learning_rate": 2.9105e-06, "loss": 2.721, "step": 5821 }, { "epoch": 1.825078369905956, "grad_norm": 16.254894256591797, "learning_rate": 2.9110000000000006e-06, "loss": 2.5039, "step": 5822 }, { "epoch": 1.8253918495297805, "grad_norm": 18.320463180541992, "learning_rate": 2.9115000000000004e-06, "loss": 2.1283, "step": 5823 }, { "epoch": 1.825705329153605, "grad_norm": 16.942607879638672, "learning_rate": 2.9120000000000002e-06, "loss": 2.5775, "step": 5824 }, { "epoch": 1.8260188087774294, "grad_norm": 44.85321807861328, "learning_rate": 2.9125000000000005e-06, "loss": 2.0103, "step": 5825 }, { "epoch": 1.8263322884012538, "grad_norm": 16.66575050354004, "learning_rate": 2.9130000000000003e-06, "loss": 1.7175, "step": 5826 }, { "epoch": 1.8266457680250783, "grad_norm": 9.837501525878906, "learning_rate": 2.9135e-06, "loss": 2.1418, "step": 5827 }, { "epoch": 1.8269592476489027, "grad_norm": 75.40845489501953, "learning_rate": 2.914e-06, "loss": 2.7283, "step": 5828 }, { "epoch": 1.8272727272727272, "grad_norm": 23.27977752685547, "learning_rate": 2.9145000000000002e-06, "loss": 2.1375, "step": 5829 }, { "epoch": 1.8275862068965516, "grad_norm": 27.287527084350586, "learning_rate": 2.915e-06, "loss": 2.1722, "step": 5830 }, { "epoch": 1.827899686520376, "grad_norm": 23.660470962524414, "learning_rate": 2.9155e-06, "loss": 2.3881, "step": 5831 }, { "epoch": 1.8282131661442005, "grad_norm": 13.798928260803223, "learning_rate": 2.9160000000000005e-06, "loss": 2.1995, "step": 5832 }, { "epoch": 1.828526645768025, "grad_norm": 72.20987701416016, "learning_rate": 2.9165000000000004e-06, "loss": 2.6333, "step": 5833 }, { "epoch": 1.8288401253918494, "grad_norm": 19.457874298095703, "learning_rate": 2.917e-06, "loss": 2.2748, "step": 5834 }, { "epoch": 1.829153605015674, "grad_norm": 16.82941436767578, "learning_rate": 2.9175000000000004e-06, "loss": 2.2334, "step": 5835 }, { "epoch": 1.8294670846394985, "grad_norm": 46.999229431152344, "learning_rate": 2.9180000000000003e-06, "loss": 2.0499, "step": 5836 }, { "epoch": 1.829780564263323, "grad_norm": 11.347408294677734, "learning_rate": 2.9185e-06, "loss": 2.1089, "step": 5837 }, { "epoch": 1.8300940438871474, "grad_norm": 7.4026665687561035, "learning_rate": 2.9190000000000003e-06, "loss": 2.3338, "step": 5838 }, { "epoch": 1.8304075235109718, "grad_norm": 17.908647537231445, "learning_rate": 2.9195e-06, "loss": 2.3173, "step": 5839 }, { "epoch": 1.8307210031347962, "grad_norm": 36.890201568603516, "learning_rate": 2.92e-06, "loss": 2.7883, "step": 5840 }, { "epoch": 1.8310344827586207, "grad_norm": 19.727296829223633, "learning_rate": 2.9205e-06, "loss": 2.3895, "step": 5841 }, { "epoch": 1.8313479623824451, "grad_norm": 18.831193923950195, "learning_rate": 2.9210000000000005e-06, "loss": 2.6153, "step": 5842 }, { "epoch": 1.8316614420062696, "grad_norm": 79.4319839477539, "learning_rate": 2.9215000000000003e-06, "loss": 2.4548, "step": 5843 }, { "epoch": 1.831974921630094, "grad_norm": 14.499472618103027, "learning_rate": 2.9220000000000006e-06, "loss": 2.1656, "step": 5844 }, { "epoch": 1.8322884012539185, "grad_norm": 16.53804588317871, "learning_rate": 2.9225000000000004e-06, "loss": 2.2322, "step": 5845 }, { "epoch": 1.832601880877743, "grad_norm": 22.1581974029541, "learning_rate": 2.9230000000000002e-06, "loss": 2.0958, "step": 5846 }, { "epoch": 1.8329153605015676, "grad_norm": 31.330753326416016, "learning_rate": 2.9235e-06, "loss": 2.1026, "step": 5847 }, { "epoch": 1.833228840125392, "grad_norm": 19.017539978027344, "learning_rate": 2.9240000000000003e-06, "loss": 2.2064, "step": 5848 }, { "epoch": 1.8335423197492164, "grad_norm": 23.104690551757812, "learning_rate": 2.9245e-06, "loss": 2.5904, "step": 5849 }, { "epoch": 1.8338557993730409, "grad_norm": 24.975191116333008, "learning_rate": 2.925e-06, "loss": 1.9996, "step": 5850 }, { "epoch": 1.8341692789968653, "grad_norm": 15.302990913391113, "learning_rate": 2.9255e-06, "loss": 2.1248, "step": 5851 }, { "epoch": 1.8344827586206898, "grad_norm": 21.22588348388672, "learning_rate": 2.9260000000000004e-06, "loss": 2.5368, "step": 5852 }, { "epoch": 1.8347962382445142, "grad_norm": 184.5251007080078, "learning_rate": 2.9265000000000003e-06, "loss": 2.5012, "step": 5853 }, { "epoch": 1.8351097178683387, "grad_norm": 10.427804946899414, "learning_rate": 2.9270000000000005e-06, "loss": 2.2236, "step": 5854 }, { "epoch": 1.835423197492163, "grad_norm": 22.081449508666992, "learning_rate": 2.9275000000000003e-06, "loss": 2.3903, "step": 5855 }, { "epoch": 1.8357366771159875, "grad_norm": 11.947712898254395, "learning_rate": 2.928e-06, "loss": 2.1356, "step": 5856 }, { "epoch": 1.836050156739812, "grad_norm": 15.462569236755371, "learning_rate": 2.9285000000000004e-06, "loss": 2.1701, "step": 5857 }, { "epoch": 1.8363636363636364, "grad_norm": 13.894822120666504, "learning_rate": 2.9290000000000002e-06, "loss": 1.9988, "step": 5858 }, { "epoch": 1.8366771159874609, "grad_norm": 15.583471298217773, "learning_rate": 2.9295e-06, "loss": 2.6659, "step": 5859 }, { "epoch": 1.8369905956112853, "grad_norm": 15.877820014953613, "learning_rate": 2.93e-06, "loss": 2.5541, "step": 5860 }, { "epoch": 1.8373040752351097, "grad_norm": 13.1459379196167, "learning_rate": 2.9305e-06, "loss": 2.1646, "step": 5861 }, { "epoch": 1.8376175548589342, "grad_norm": 15.416597366333008, "learning_rate": 2.9310000000000004e-06, "loss": 2.2812, "step": 5862 }, { "epoch": 1.8379310344827586, "grad_norm": 32.908973693847656, "learning_rate": 2.9315000000000006e-06, "loss": 2.3487, "step": 5863 }, { "epoch": 1.838244514106583, "grad_norm": 9.497340202331543, "learning_rate": 2.9320000000000005e-06, "loss": 2.2336, "step": 5864 }, { "epoch": 1.8385579937304075, "grad_norm": 23.483781814575195, "learning_rate": 2.9325000000000003e-06, "loss": 2.5881, "step": 5865 }, { "epoch": 1.838871473354232, "grad_norm": 13.98792552947998, "learning_rate": 2.933e-06, "loss": 2.2051, "step": 5866 }, { "epoch": 1.8391849529780564, "grad_norm": 13.942326545715332, "learning_rate": 2.9335000000000004e-06, "loss": 2.2039, "step": 5867 }, { "epoch": 1.8394984326018808, "grad_norm": 10.772849082946777, "learning_rate": 2.934e-06, "loss": 2.1795, "step": 5868 }, { "epoch": 1.8398119122257053, "grad_norm": 22.885339736938477, "learning_rate": 2.9345e-06, "loss": 2.2539, "step": 5869 }, { "epoch": 1.8401253918495297, "grad_norm": 13.527241706848145, "learning_rate": 2.9350000000000003e-06, "loss": 2.2251, "step": 5870 }, { "epoch": 1.8404388714733542, "grad_norm": 32.79869079589844, "learning_rate": 2.9355e-06, "loss": 2.3778, "step": 5871 }, { "epoch": 1.8407523510971786, "grad_norm": 87.69104766845703, "learning_rate": 2.9360000000000003e-06, "loss": 2.5287, "step": 5872 }, { "epoch": 1.841065830721003, "grad_norm": 8.843754768371582, "learning_rate": 2.9365000000000006e-06, "loss": 2.1424, "step": 5873 }, { "epoch": 1.8413793103448275, "grad_norm": 20.374862670898438, "learning_rate": 2.9370000000000004e-06, "loss": 2.5718, "step": 5874 }, { "epoch": 1.841692789968652, "grad_norm": 13.062189102172852, "learning_rate": 2.9375000000000003e-06, "loss": 2.1401, "step": 5875 }, { "epoch": 1.8420062695924764, "grad_norm": 15.953989028930664, "learning_rate": 2.9380000000000005e-06, "loss": 2.4322, "step": 5876 }, { "epoch": 1.8423197492163008, "grad_norm": 11.689451217651367, "learning_rate": 2.9385000000000003e-06, "loss": 2.2732, "step": 5877 }, { "epoch": 1.8426332288401253, "grad_norm": 14.597411155700684, "learning_rate": 2.939e-06, "loss": 2.0253, "step": 5878 }, { "epoch": 1.8429467084639497, "grad_norm": 15.932793617248535, "learning_rate": 2.9395e-06, "loss": 2.3347, "step": 5879 }, { "epoch": 1.8432601880877741, "grad_norm": 16.13714599609375, "learning_rate": 2.9400000000000002e-06, "loss": 2.2366, "step": 5880 }, { "epoch": 1.8435736677115986, "grad_norm": 87.59251403808594, "learning_rate": 2.9405e-06, "loss": 2.5404, "step": 5881 }, { "epoch": 1.8438871473354232, "grad_norm": 13.624130249023438, "learning_rate": 2.941e-06, "loss": 2.4282, "step": 5882 }, { "epoch": 1.8442006269592477, "grad_norm": 15.835795402526855, "learning_rate": 2.9415000000000006e-06, "loss": 2.9211, "step": 5883 }, { "epoch": 1.8445141065830721, "grad_norm": 23.732357025146484, "learning_rate": 2.9420000000000004e-06, "loss": 1.9677, "step": 5884 }, { "epoch": 1.8448275862068966, "grad_norm": 12.846848487854004, "learning_rate": 2.9425e-06, "loss": 2.6445, "step": 5885 }, { "epoch": 1.845141065830721, "grad_norm": 96.78413391113281, "learning_rate": 2.9430000000000005e-06, "loss": 2.6282, "step": 5886 }, { "epoch": 1.8454545454545455, "grad_norm": 57.51097106933594, "learning_rate": 2.9435000000000003e-06, "loss": 2.7437, "step": 5887 }, { "epoch": 1.84576802507837, "grad_norm": 12.915363311767578, "learning_rate": 2.944e-06, "loss": 2.2672, "step": 5888 }, { "epoch": 1.8460815047021943, "grad_norm": 40.32135009765625, "learning_rate": 2.9445e-06, "loss": 2.1966, "step": 5889 }, { "epoch": 1.8463949843260188, "grad_norm": 11.176884651184082, "learning_rate": 2.945e-06, "loss": 2.4467, "step": 5890 }, { "epoch": 1.8467084639498432, "grad_norm": 12.287383079528809, "learning_rate": 2.9455e-06, "loss": 2.1618, "step": 5891 }, { "epoch": 1.8470219435736677, "grad_norm": 9.98103141784668, "learning_rate": 2.946e-06, "loss": 2.647, "step": 5892 }, { "epoch": 1.847335423197492, "grad_norm": 24.924226760864258, "learning_rate": 2.9465000000000005e-06, "loss": 2.3428, "step": 5893 }, { "epoch": 1.8476489028213168, "grad_norm": 18.21531105041504, "learning_rate": 2.9470000000000003e-06, "loss": 2.1155, "step": 5894 }, { "epoch": 1.8479623824451412, "grad_norm": 8.226020812988281, "learning_rate": 2.9475e-06, "loss": 2.15, "step": 5895 }, { "epoch": 1.8482758620689657, "grad_norm": 15.366581916809082, "learning_rate": 2.9480000000000004e-06, "loss": 2.0918, "step": 5896 }, { "epoch": 1.84858934169279, "grad_norm": 27.560998916625977, "learning_rate": 2.9485000000000002e-06, "loss": 2.3529, "step": 5897 }, { "epoch": 1.8489028213166145, "grad_norm": 11.260773658752441, "learning_rate": 2.949e-06, "loss": 2.1609, "step": 5898 }, { "epoch": 1.849216300940439, "grad_norm": 16.45672607421875, "learning_rate": 2.9495000000000003e-06, "loss": 2.0179, "step": 5899 }, { "epoch": 1.8495297805642634, "grad_norm": 11.358928680419922, "learning_rate": 2.95e-06, "loss": 2.5883, "step": 5900 }, { "epoch": 1.8498432601880879, "grad_norm": 13.168031692504883, "learning_rate": 2.9505e-06, "loss": 2.1458, "step": 5901 }, { "epoch": 1.8501567398119123, "grad_norm": 22.90192413330078, "learning_rate": 2.9509999999999998e-06, "loss": 2.3417, "step": 5902 }, { "epoch": 1.8504702194357368, "grad_norm": 11.460537910461426, "learning_rate": 2.9515000000000005e-06, "loss": 2.3871, "step": 5903 }, { "epoch": 1.8507836990595612, "grad_norm": 116.56676483154297, "learning_rate": 2.9520000000000003e-06, "loss": 2.7412, "step": 5904 }, { "epoch": 1.8510971786833856, "grad_norm": 17.3189640045166, "learning_rate": 2.9525000000000005e-06, "loss": 2.1462, "step": 5905 }, { "epoch": 1.85141065830721, "grad_norm": 25.294736862182617, "learning_rate": 2.9530000000000004e-06, "loss": 2.2008, "step": 5906 }, { "epoch": 1.8517241379310345, "grad_norm": 13.655937194824219, "learning_rate": 2.9535e-06, "loss": 2.2359, "step": 5907 }, { "epoch": 1.852037617554859, "grad_norm": 51.57292938232422, "learning_rate": 2.954e-06, "loss": 2.4588, "step": 5908 }, { "epoch": 1.8523510971786834, "grad_norm": 17.087007522583008, "learning_rate": 2.9545000000000003e-06, "loss": 2.1137, "step": 5909 }, { "epoch": 1.8526645768025078, "grad_norm": 20.294261932373047, "learning_rate": 2.955e-06, "loss": 2.1156, "step": 5910 }, { "epoch": 1.8529780564263323, "grad_norm": 12.528114318847656, "learning_rate": 2.9555e-06, "loss": 1.9428, "step": 5911 }, { "epoch": 1.8532915360501567, "grad_norm": 25.253681182861328, "learning_rate": 2.956e-06, "loss": 2.5651, "step": 5912 }, { "epoch": 1.8536050156739812, "grad_norm": 19.341289520263672, "learning_rate": 2.9565000000000004e-06, "loss": 2.2829, "step": 5913 }, { "epoch": 1.8539184952978056, "grad_norm": 24.48165512084961, "learning_rate": 2.9570000000000002e-06, "loss": 2.2775, "step": 5914 }, { "epoch": 1.85423197492163, "grad_norm": 124.07843780517578, "learning_rate": 2.9575000000000005e-06, "loss": 4.0833, "step": 5915 }, { "epoch": 1.8545454545454545, "grad_norm": 44.97792053222656, "learning_rate": 2.9580000000000003e-06, "loss": 2.6771, "step": 5916 }, { "epoch": 1.854858934169279, "grad_norm": 37.24640655517578, "learning_rate": 2.9585e-06, "loss": 3.0157, "step": 5917 }, { "epoch": 1.8551724137931034, "grad_norm": 19.38148307800293, "learning_rate": 2.9590000000000004e-06, "loss": 2.6838, "step": 5918 }, { "epoch": 1.8554858934169278, "grad_norm": 20.10344886779785, "learning_rate": 2.9595e-06, "loss": 2.3734, "step": 5919 }, { "epoch": 1.8557993730407523, "grad_norm": 28.982986450195312, "learning_rate": 2.96e-06, "loss": 2.5651, "step": 5920 }, { "epoch": 1.8561128526645767, "grad_norm": 12.973660469055176, "learning_rate": 2.9605e-06, "loss": 2.204, "step": 5921 }, { "epoch": 1.8564263322884011, "grad_norm": 10.518831253051758, "learning_rate": 2.961e-06, "loss": 2.1735, "step": 5922 }, { "epoch": 1.8567398119122256, "grad_norm": 13.175191879272461, "learning_rate": 2.9615000000000004e-06, "loss": 2.236, "step": 5923 }, { "epoch": 1.85705329153605, "grad_norm": 23.340044021606445, "learning_rate": 2.9620000000000006e-06, "loss": 2.7431, "step": 5924 }, { "epoch": 1.8573667711598745, "grad_norm": 19.4545955657959, "learning_rate": 2.9625000000000004e-06, "loss": 2.6503, "step": 5925 }, { "epoch": 1.857680250783699, "grad_norm": 12.167764663696289, "learning_rate": 2.9630000000000003e-06, "loss": 2.2549, "step": 5926 }, { "epoch": 1.8579937304075234, "grad_norm": 20.977935791015625, "learning_rate": 2.9635e-06, "loss": 2.432, "step": 5927 }, { "epoch": 1.8583072100313478, "grad_norm": 22.710613250732422, "learning_rate": 2.9640000000000003e-06, "loss": 2.5537, "step": 5928 }, { "epoch": 1.8586206896551725, "grad_norm": 21.324052810668945, "learning_rate": 2.9645e-06, "loss": 3.0388, "step": 5929 }, { "epoch": 1.858934169278997, "grad_norm": 12.704240798950195, "learning_rate": 2.965e-06, "loss": 2.2973, "step": 5930 }, { "epoch": 1.8592476489028213, "grad_norm": 23.687358856201172, "learning_rate": 2.9655000000000002e-06, "loss": 2.4357, "step": 5931 }, { "epoch": 1.8595611285266458, "grad_norm": 12.968914031982422, "learning_rate": 2.966e-06, "loss": 2.3901, "step": 5932 }, { "epoch": 1.8598746081504702, "grad_norm": 35.27592849731445, "learning_rate": 2.9665000000000003e-06, "loss": 2.6777, "step": 5933 }, { "epoch": 1.8601880877742947, "grad_norm": 12.709157943725586, "learning_rate": 2.9670000000000006e-06, "loss": 2.2181, "step": 5934 }, { "epoch": 1.8605015673981191, "grad_norm": 18.975936889648438, "learning_rate": 2.9675000000000004e-06, "loss": 2.5685, "step": 5935 }, { "epoch": 1.8608150470219436, "grad_norm": 17.399898529052734, "learning_rate": 2.9680000000000002e-06, "loss": 2.0833, "step": 5936 }, { "epoch": 1.861128526645768, "grad_norm": 13.010747909545898, "learning_rate": 2.9685000000000005e-06, "loss": 2.0245, "step": 5937 }, { "epoch": 1.8614420062695924, "grad_norm": 11.078125953674316, "learning_rate": 2.9690000000000003e-06, "loss": 2.1143, "step": 5938 }, { "epoch": 1.8617554858934169, "grad_norm": 27.35584831237793, "learning_rate": 2.9695e-06, "loss": 2.5801, "step": 5939 }, { "epoch": 1.8620689655172413, "grad_norm": 15.091031074523926, "learning_rate": 2.97e-06, "loss": 2.3546, "step": 5940 }, { "epoch": 1.862382445141066, "grad_norm": 9.473328590393066, "learning_rate": 2.9705e-06, "loss": 2.1142, "step": 5941 }, { "epoch": 1.8626959247648904, "grad_norm": 19.720460891723633, "learning_rate": 2.971e-06, "loss": 3.0768, "step": 5942 }, { "epoch": 1.8630094043887149, "grad_norm": 16.488967895507812, "learning_rate": 2.9715000000000007e-06, "loss": 2.4286, "step": 5943 }, { "epoch": 1.8633228840125393, "grad_norm": 22.054641723632812, "learning_rate": 2.9720000000000005e-06, "loss": 2.2125, "step": 5944 }, { "epoch": 1.8636363636363638, "grad_norm": 11.244823455810547, "learning_rate": 2.9725000000000003e-06, "loss": 2.3011, "step": 5945 }, { "epoch": 1.8639498432601882, "grad_norm": 13.354893684387207, "learning_rate": 2.973e-06, "loss": 2.172, "step": 5946 }, { "epoch": 1.8642633228840126, "grad_norm": 10.610267639160156, "learning_rate": 2.9735000000000004e-06, "loss": 2.3918, "step": 5947 }, { "epoch": 1.864576802507837, "grad_norm": 18.93093490600586, "learning_rate": 2.9740000000000002e-06, "loss": 2.3002, "step": 5948 }, { "epoch": 1.8648902821316615, "grad_norm": 14.058737754821777, "learning_rate": 2.9745e-06, "loss": 2.2429, "step": 5949 }, { "epoch": 1.865203761755486, "grad_norm": 11.233428955078125, "learning_rate": 2.9750000000000003e-06, "loss": 2.0587, "step": 5950 }, { "epoch": 1.8655172413793104, "grad_norm": 15.62847900390625, "learning_rate": 2.9755e-06, "loss": 2.2816, "step": 5951 }, { "epoch": 1.8658307210031349, "grad_norm": 11.445661544799805, "learning_rate": 2.976e-06, "loss": 2.0684, "step": 5952 }, { "epoch": 1.8661442006269593, "grad_norm": 43.25938415527344, "learning_rate": 2.9765000000000006e-06, "loss": 2.6958, "step": 5953 }, { "epoch": 1.8664576802507837, "grad_norm": 16.91835594177246, "learning_rate": 2.9770000000000005e-06, "loss": 2.2447, "step": 5954 }, { "epoch": 1.8667711598746082, "grad_norm": 81.19352722167969, "learning_rate": 2.9775000000000003e-06, "loss": 2.3908, "step": 5955 }, { "epoch": 1.8670846394984326, "grad_norm": 14.449911117553711, "learning_rate": 2.9780000000000005e-06, "loss": 2.2353, "step": 5956 }, { "epoch": 1.867398119122257, "grad_norm": 15.223895072937012, "learning_rate": 2.9785000000000004e-06, "loss": 2.3473, "step": 5957 }, { "epoch": 1.8677115987460815, "grad_norm": 10.657572746276855, "learning_rate": 2.979e-06, "loss": 2.4062, "step": 5958 }, { "epoch": 1.868025078369906, "grad_norm": 20.128049850463867, "learning_rate": 2.9795e-06, "loss": 2.1415, "step": 5959 }, { "epoch": 1.8683385579937304, "grad_norm": 19.867464065551758, "learning_rate": 2.9800000000000003e-06, "loss": 2.3222, "step": 5960 }, { "epoch": 1.8686520376175548, "grad_norm": 37.976341247558594, "learning_rate": 2.9805e-06, "loss": 3.0599, "step": 5961 }, { "epoch": 1.8689655172413793, "grad_norm": 13.279169082641602, "learning_rate": 2.981e-06, "loss": 2.72, "step": 5962 }, { "epoch": 1.8692789968652037, "grad_norm": 29.617717742919922, "learning_rate": 2.9815000000000006e-06, "loss": 3.9811, "step": 5963 }, { "epoch": 1.8695924764890282, "grad_norm": 13.578438758850098, "learning_rate": 2.9820000000000004e-06, "loss": 2.1697, "step": 5964 }, { "epoch": 1.8699059561128526, "grad_norm": 22.47218132019043, "learning_rate": 2.9825000000000002e-06, "loss": 2.1429, "step": 5965 }, { "epoch": 1.870219435736677, "grad_norm": 68.62042999267578, "learning_rate": 2.9830000000000005e-06, "loss": 2.5068, "step": 5966 }, { "epoch": 1.8705329153605015, "grad_norm": 14.457962036132812, "learning_rate": 2.9835000000000003e-06, "loss": 2.4458, "step": 5967 }, { "epoch": 1.870846394984326, "grad_norm": 46.994361877441406, "learning_rate": 2.984e-06, "loss": 2.5197, "step": 5968 }, { "epoch": 1.8711598746081504, "grad_norm": 8.362174034118652, "learning_rate": 2.9845e-06, "loss": 2.101, "step": 5969 }, { "epoch": 1.8714733542319748, "grad_norm": 31.05512046813965, "learning_rate": 2.9850000000000002e-06, "loss": 2.975, "step": 5970 }, { "epoch": 1.8717868338557992, "grad_norm": 14.238176345825195, "learning_rate": 2.9855e-06, "loss": 1.9055, "step": 5971 }, { "epoch": 1.8721003134796237, "grad_norm": 10.674015998840332, "learning_rate": 2.986e-06, "loss": 2.0822, "step": 5972 }, { "epoch": 1.8724137931034481, "grad_norm": 21.955867767333984, "learning_rate": 2.9865000000000005e-06, "loss": 2.4844, "step": 5973 }, { "epoch": 1.8727272727272726, "grad_norm": 51.78453063964844, "learning_rate": 2.9870000000000004e-06, "loss": 2.5129, "step": 5974 }, { "epoch": 1.873040752351097, "grad_norm": 11.626786231994629, "learning_rate": 2.9875e-06, "loss": 2.1576, "step": 5975 }, { "epoch": 1.8733542319749217, "grad_norm": 11.270774841308594, "learning_rate": 2.9880000000000004e-06, "loss": 2.5178, "step": 5976 }, { "epoch": 1.8736677115987461, "grad_norm": 12.440287590026855, "learning_rate": 2.9885000000000003e-06, "loss": 2.2678, "step": 5977 }, { "epoch": 1.8739811912225706, "grad_norm": 22.860023498535156, "learning_rate": 2.989e-06, "loss": 2.2348, "step": 5978 }, { "epoch": 1.874294670846395, "grad_norm": 45.02328109741211, "learning_rate": 2.9895000000000003e-06, "loss": 2.6181, "step": 5979 }, { "epoch": 1.8746081504702194, "grad_norm": 10.591276168823242, "learning_rate": 2.99e-06, "loss": 2.3716, "step": 5980 }, { "epoch": 1.874921630094044, "grad_norm": 27.081436157226562, "learning_rate": 2.9905e-06, "loss": 2.1792, "step": 5981 }, { "epoch": 1.8752351097178683, "grad_norm": 17.257184982299805, "learning_rate": 2.991e-06, "loss": 2.4155, "step": 5982 }, { "epoch": 1.8755485893416928, "grad_norm": 8.944378852844238, "learning_rate": 2.9915000000000005e-06, "loss": 2.0324, "step": 5983 }, { "epoch": 1.8758620689655172, "grad_norm": 13.029302597045898, "learning_rate": 2.9920000000000003e-06, "loss": 2.0769, "step": 5984 }, { "epoch": 1.8761755485893417, "grad_norm": 93.2248306274414, "learning_rate": 2.9925000000000006e-06, "loss": 2.7404, "step": 5985 }, { "epoch": 1.876489028213166, "grad_norm": 11.743170738220215, "learning_rate": 2.9930000000000004e-06, "loss": 2.0512, "step": 5986 }, { "epoch": 1.8768025078369905, "grad_norm": 13.233771324157715, "learning_rate": 2.9935000000000002e-06, "loss": 1.9644, "step": 5987 }, { "epoch": 1.8771159874608152, "grad_norm": 12.37627124786377, "learning_rate": 2.994e-06, "loss": 2.49, "step": 5988 }, { "epoch": 1.8774294670846396, "grad_norm": 19.133132934570312, "learning_rate": 2.9945000000000003e-06, "loss": 2.2471, "step": 5989 }, { "epoch": 1.877742946708464, "grad_norm": 14.256895065307617, "learning_rate": 2.995e-06, "loss": 2.1018, "step": 5990 }, { "epoch": 1.8780564263322885, "grad_norm": 17.488203048706055, "learning_rate": 2.9955e-06, "loss": 2.121, "step": 5991 }, { "epoch": 1.878369905956113, "grad_norm": 9.792834281921387, "learning_rate": 2.996e-06, "loss": 2.1461, "step": 5992 }, { "epoch": 1.8786833855799374, "grad_norm": 15.044869422912598, "learning_rate": 2.9965000000000004e-06, "loss": 2.1132, "step": 5993 }, { "epoch": 1.8789968652037619, "grad_norm": 12.447378158569336, "learning_rate": 2.9970000000000003e-06, "loss": 2.1201, "step": 5994 }, { "epoch": 1.8793103448275863, "grad_norm": 11.574535369873047, "learning_rate": 2.9975000000000005e-06, "loss": 2.2942, "step": 5995 }, { "epoch": 1.8796238244514107, "grad_norm": 18.180683135986328, "learning_rate": 2.9980000000000003e-06, "loss": 2.2917, "step": 5996 }, { "epoch": 1.8799373040752352, "grad_norm": 15.773523330688477, "learning_rate": 2.9985e-06, "loss": 2.1649, "step": 5997 }, { "epoch": 1.8802507836990596, "grad_norm": 12.05911922454834, "learning_rate": 2.9990000000000004e-06, "loss": 2.2594, "step": 5998 }, { "epoch": 1.880564263322884, "grad_norm": 13.821588516235352, "learning_rate": 2.9995000000000003e-06, "loss": 2.283, "step": 5999 }, { "epoch": 1.8808777429467085, "grad_norm": 9.15234661102295, "learning_rate": 3e-06, "loss": 2.194, "step": 6000 }, { "epoch": 1.881191222570533, "grad_norm": 18.25379753112793, "learning_rate": 3.0005e-06, "loss": 2.3789, "step": 6001 }, { "epoch": 1.8815047021943574, "grad_norm": 14.774020195007324, "learning_rate": 3.001e-06, "loss": 2.2303, "step": 6002 }, { "epoch": 1.8818181818181818, "grad_norm": 7.372211456298828, "learning_rate": 3.0015e-06, "loss": 2.0202, "step": 6003 }, { "epoch": 1.8821316614420063, "grad_norm": 29.366544723510742, "learning_rate": 3.0020000000000006e-06, "loss": 2.7209, "step": 6004 }, { "epoch": 1.8824451410658307, "grad_norm": 20.27481460571289, "learning_rate": 3.0025000000000005e-06, "loss": 2.2696, "step": 6005 }, { "epoch": 1.8827586206896552, "grad_norm": 17.354726791381836, "learning_rate": 3.0030000000000003e-06, "loss": 2.2796, "step": 6006 }, { "epoch": 1.8830721003134796, "grad_norm": 11.225373268127441, "learning_rate": 3.0035e-06, "loss": 2.4752, "step": 6007 }, { "epoch": 1.883385579937304, "grad_norm": 13.302656173706055, "learning_rate": 3.0040000000000004e-06, "loss": 2.2338, "step": 6008 }, { "epoch": 1.8836990595611285, "grad_norm": 36.54410934448242, "learning_rate": 3.0045e-06, "loss": 2.5358, "step": 6009 }, { "epoch": 1.884012539184953, "grad_norm": 12.383371353149414, "learning_rate": 3.005e-06, "loss": 2.2359, "step": 6010 }, { "epoch": 1.8843260188087774, "grad_norm": 10.54360294342041, "learning_rate": 3.0055000000000003e-06, "loss": 2.1926, "step": 6011 }, { "epoch": 1.8846394984326018, "grad_norm": 13.832954406738281, "learning_rate": 3.006e-06, "loss": 2.343, "step": 6012 }, { "epoch": 1.8849529780564263, "grad_norm": 18.631515502929688, "learning_rate": 3.0065e-06, "loss": 2.1568, "step": 6013 }, { "epoch": 1.8852664576802507, "grad_norm": 14.329492568969727, "learning_rate": 3.0070000000000006e-06, "loss": 2.2601, "step": 6014 }, { "epoch": 1.8855799373040751, "grad_norm": 64.08553314208984, "learning_rate": 3.0075000000000004e-06, "loss": 2.8621, "step": 6015 }, { "epoch": 1.8858934169278996, "grad_norm": 15.698051452636719, "learning_rate": 3.0080000000000003e-06, "loss": 2.3793, "step": 6016 }, { "epoch": 1.886206896551724, "grad_norm": 64.55384826660156, "learning_rate": 3.0085000000000005e-06, "loss": 2.6757, "step": 6017 }, { "epoch": 1.8865203761755485, "grad_norm": 96.30728149414062, "learning_rate": 3.0090000000000003e-06, "loss": 2.4954, "step": 6018 }, { "epoch": 1.886833855799373, "grad_norm": 46.265907287597656, "learning_rate": 3.0095e-06, "loss": 2.3806, "step": 6019 }, { "epoch": 1.8871473354231973, "grad_norm": 19.963144302368164, "learning_rate": 3.01e-06, "loss": 1.9748, "step": 6020 }, { "epoch": 1.8874608150470218, "grad_norm": 36.911582946777344, "learning_rate": 3.0105000000000002e-06, "loss": 2.2604, "step": 6021 }, { "epoch": 1.8877742946708462, "grad_norm": 14.679679870605469, "learning_rate": 3.011e-06, "loss": 2.3641, "step": 6022 }, { "epoch": 1.888087774294671, "grad_norm": 17.748821258544922, "learning_rate": 3.0115e-06, "loss": 2.8093, "step": 6023 }, { "epoch": 1.8884012539184953, "grad_norm": 59.87860870361328, "learning_rate": 3.0120000000000006e-06, "loss": 3.2332, "step": 6024 }, { "epoch": 1.8887147335423198, "grad_norm": 14.679272651672363, "learning_rate": 3.0125000000000004e-06, "loss": 2.2552, "step": 6025 }, { "epoch": 1.8890282131661442, "grad_norm": 12.415968894958496, "learning_rate": 3.013e-06, "loss": 2.2754, "step": 6026 }, { "epoch": 1.8893416927899687, "grad_norm": 25.21853256225586, "learning_rate": 3.0135000000000005e-06, "loss": 2.2498, "step": 6027 }, { "epoch": 1.889655172413793, "grad_norm": 66.48529815673828, "learning_rate": 3.0140000000000003e-06, "loss": 2.3706, "step": 6028 }, { "epoch": 1.8899686520376175, "grad_norm": 39.09945297241211, "learning_rate": 3.0145e-06, "loss": 2.4398, "step": 6029 }, { "epoch": 1.890282131661442, "grad_norm": 18.697851181030273, "learning_rate": 3.0150000000000004e-06, "loss": 2.4102, "step": 6030 }, { "epoch": 1.8905956112852664, "grad_norm": 19.043466567993164, "learning_rate": 3.0155e-06, "loss": 2.2626, "step": 6031 }, { "epoch": 1.8909090909090909, "grad_norm": 31.95065689086914, "learning_rate": 3.016e-06, "loss": 2.3616, "step": 6032 }, { "epoch": 1.8912225705329153, "grad_norm": 15.939237594604492, "learning_rate": 3.0165e-06, "loss": 2.207, "step": 6033 }, { "epoch": 1.8915360501567398, "grad_norm": 40.93842315673828, "learning_rate": 3.0170000000000005e-06, "loss": 2.2072, "step": 6034 }, { "epoch": 1.8918495297805644, "grad_norm": 20.635238647460938, "learning_rate": 3.0175000000000003e-06, "loss": 2.5603, "step": 6035 }, { "epoch": 1.8921630094043889, "grad_norm": 14.814455032348633, "learning_rate": 3.0180000000000006e-06, "loss": 2.4815, "step": 6036 }, { "epoch": 1.8924764890282133, "grad_norm": 11.78087043762207, "learning_rate": 3.0185000000000004e-06, "loss": 2.332, "step": 6037 }, { "epoch": 1.8927899686520377, "grad_norm": 9.055764198303223, "learning_rate": 3.0190000000000002e-06, "loss": 2.1643, "step": 6038 }, { "epoch": 1.8931034482758622, "grad_norm": 62.65747833251953, "learning_rate": 3.0195e-06, "loss": 2.3768, "step": 6039 }, { "epoch": 1.8934169278996866, "grad_norm": 10.645723342895508, "learning_rate": 3.0200000000000003e-06, "loss": 2.4308, "step": 6040 }, { "epoch": 1.893730407523511, "grad_norm": 7.89349889755249, "learning_rate": 3.0205e-06, "loss": 2.5163, "step": 6041 }, { "epoch": 1.8940438871473355, "grad_norm": 56.9163703918457, "learning_rate": 3.021e-06, "loss": 3.444, "step": 6042 }, { "epoch": 1.89435736677116, "grad_norm": 9.060624122619629, "learning_rate": 3.0214999999999998e-06, "loss": 2.0517, "step": 6043 }, { "epoch": 1.8946708463949844, "grad_norm": 18.83366584777832, "learning_rate": 3.0220000000000005e-06, "loss": 2.2112, "step": 6044 }, { "epoch": 1.8949843260188088, "grad_norm": 15.034554481506348, "learning_rate": 3.0225000000000003e-06, "loss": 2.1677, "step": 6045 }, { "epoch": 1.8952978056426333, "grad_norm": 19.451780319213867, "learning_rate": 3.0230000000000005e-06, "loss": 2.4123, "step": 6046 }, { "epoch": 1.8956112852664577, "grad_norm": 12.623496055603027, "learning_rate": 3.0235000000000004e-06, "loss": 2.629, "step": 6047 }, { "epoch": 1.8959247648902822, "grad_norm": 17.31379508972168, "learning_rate": 3.024e-06, "loss": 2.3774, "step": 6048 }, { "epoch": 1.8962382445141066, "grad_norm": 77.49359893798828, "learning_rate": 3.0245e-06, "loss": 3.0316, "step": 6049 }, { "epoch": 1.896551724137931, "grad_norm": 13.705031394958496, "learning_rate": 3.0250000000000003e-06, "loss": 2.1171, "step": 6050 }, { "epoch": 1.8968652037617555, "grad_norm": 17.037595748901367, "learning_rate": 3.0255e-06, "loss": 2.2576, "step": 6051 }, { "epoch": 1.89717868338558, "grad_norm": 7.457324981689453, "learning_rate": 3.026e-06, "loss": 2.1425, "step": 6052 }, { "epoch": 1.8974921630094044, "grad_norm": 60.89271545410156, "learning_rate": 3.0265e-06, "loss": 2.7768, "step": 6053 }, { "epoch": 1.8978056426332288, "grad_norm": 17.38759422302246, "learning_rate": 3.0270000000000004e-06, "loss": 1.9765, "step": 6054 }, { "epoch": 1.8981191222570533, "grad_norm": 13.200677871704102, "learning_rate": 3.0275000000000002e-06, "loss": 2.2556, "step": 6055 }, { "epoch": 1.8984326018808777, "grad_norm": 10.527555465698242, "learning_rate": 3.0280000000000005e-06, "loss": 2.0826, "step": 6056 }, { "epoch": 1.8987460815047021, "grad_norm": 81.97634887695312, "learning_rate": 3.0285000000000003e-06, "loss": 2.5342, "step": 6057 }, { "epoch": 1.8990595611285266, "grad_norm": 24.821269989013672, "learning_rate": 3.029e-06, "loss": 2.2356, "step": 6058 }, { "epoch": 1.899373040752351, "grad_norm": 20.83104133605957, "learning_rate": 3.0295000000000004e-06, "loss": 2.2116, "step": 6059 }, { "epoch": 1.8996865203761755, "grad_norm": 25.46314811706543, "learning_rate": 3.0300000000000002e-06, "loss": 2.5997, "step": 6060 }, { "epoch": 1.9, "grad_norm": 9.644169807434082, "learning_rate": 3.0305e-06, "loss": 2.0949, "step": 6061 }, { "epoch": 1.9003134796238244, "grad_norm": 35.5372200012207, "learning_rate": 3.031e-06, "loss": 2.5674, "step": 6062 }, { "epoch": 1.9006269592476488, "grad_norm": 16.273813247680664, "learning_rate": 3.0315e-06, "loss": 2.465, "step": 6063 }, { "epoch": 1.9009404388714732, "grad_norm": 31.08625030517578, "learning_rate": 3.0320000000000004e-06, "loss": 2.3942, "step": 6064 }, { "epoch": 1.9012539184952977, "grad_norm": 45.639808654785156, "learning_rate": 3.0325000000000006e-06, "loss": 2.5761, "step": 6065 }, { "epoch": 1.9015673981191221, "grad_norm": 34.586517333984375, "learning_rate": 3.0330000000000004e-06, "loss": 2.297, "step": 6066 }, { "epoch": 1.9018808777429466, "grad_norm": 14.389083862304688, "learning_rate": 3.0335000000000003e-06, "loss": 2.3685, "step": 6067 }, { "epoch": 1.902194357366771, "grad_norm": 54.00825881958008, "learning_rate": 3.034e-06, "loss": 2.7093, "step": 6068 }, { "epoch": 1.9025078369905954, "grad_norm": 8.659297943115234, "learning_rate": 3.0345000000000003e-06, "loss": 2.2077, "step": 6069 }, { "epoch": 1.90282131661442, "grad_norm": 40.629173278808594, "learning_rate": 3.035e-06, "loss": 2.1629, "step": 6070 }, { "epoch": 1.9031347962382446, "grad_norm": 13.120607376098633, "learning_rate": 3.0355e-06, "loss": 2.2908, "step": 6071 }, { "epoch": 1.903448275862069, "grad_norm": 15.2529878616333, "learning_rate": 3.0360000000000002e-06, "loss": 2.109, "step": 6072 }, { "epoch": 1.9037617554858934, "grad_norm": 19.432376861572266, "learning_rate": 3.0365e-06, "loss": 2.3507, "step": 6073 }, { "epoch": 1.9040752351097179, "grad_norm": 11.499388694763184, "learning_rate": 3.0370000000000003e-06, "loss": 2.3288, "step": 6074 }, { "epoch": 1.9043887147335423, "grad_norm": 13.910191535949707, "learning_rate": 3.0375000000000006e-06, "loss": 2.5768, "step": 6075 }, { "epoch": 1.9047021943573668, "grad_norm": 33.155765533447266, "learning_rate": 3.0380000000000004e-06, "loss": 2.279, "step": 6076 }, { "epoch": 1.9050156739811912, "grad_norm": 17.429643630981445, "learning_rate": 3.0385000000000002e-06, "loss": 2.4015, "step": 6077 }, { "epoch": 1.9053291536050156, "grad_norm": 9.968888282775879, "learning_rate": 3.0390000000000005e-06, "loss": 2.2738, "step": 6078 }, { "epoch": 1.90564263322884, "grad_norm": 19.311132431030273, "learning_rate": 3.0395000000000003e-06, "loss": 2.3267, "step": 6079 }, { "epoch": 1.9059561128526645, "grad_norm": 15.164823532104492, "learning_rate": 3.04e-06, "loss": 2.9306, "step": 6080 }, { "epoch": 1.9062695924764892, "grad_norm": 20.48821449279785, "learning_rate": 3.0405e-06, "loss": 2.3529, "step": 6081 }, { "epoch": 1.9065830721003136, "grad_norm": 42.66050338745117, "learning_rate": 3.041e-06, "loss": 2.5175, "step": 6082 }, { "epoch": 1.906896551724138, "grad_norm": 15.134031295776367, "learning_rate": 3.0415e-06, "loss": 2.2948, "step": 6083 }, { "epoch": 1.9072100313479625, "grad_norm": 12.295634269714355, "learning_rate": 3.0420000000000007e-06, "loss": 2.1349, "step": 6084 }, { "epoch": 1.907523510971787, "grad_norm": 11.567246437072754, "learning_rate": 3.0425000000000005e-06, "loss": 2.1183, "step": 6085 }, { "epoch": 1.9078369905956114, "grad_norm": 16.417787551879883, "learning_rate": 3.0430000000000003e-06, "loss": 2.1569, "step": 6086 }, { "epoch": 1.9081504702194358, "grad_norm": 21.135101318359375, "learning_rate": 3.0435e-06, "loss": 2.3614, "step": 6087 }, { "epoch": 1.9084639498432603, "grad_norm": 23.375978469848633, "learning_rate": 3.0440000000000004e-06, "loss": 2.5555, "step": 6088 }, { "epoch": 1.9087774294670847, "grad_norm": 11.53591537475586, "learning_rate": 3.0445000000000002e-06, "loss": 2.3772, "step": 6089 }, { "epoch": 1.9090909090909092, "grad_norm": 15.471573829650879, "learning_rate": 3.045e-06, "loss": 2.3107, "step": 6090 }, { "epoch": 1.9094043887147336, "grad_norm": 70.57958984375, "learning_rate": 3.0455000000000003e-06, "loss": 2.1541, "step": 6091 }, { "epoch": 1.909717868338558, "grad_norm": 24.529329299926758, "learning_rate": 3.046e-06, "loss": 2.9641, "step": 6092 }, { "epoch": 1.9100313479623825, "grad_norm": 9.5241117477417, "learning_rate": 3.0465e-06, "loss": 2.1275, "step": 6093 }, { "epoch": 1.910344827586207, "grad_norm": 14.434778213500977, "learning_rate": 3.0470000000000006e-06, "loss": 2.2802, "step": 6094 }, { "epoch": 1.9106583072100314, "grad_norm": 11.293213844299316, "learning_rate": 3.0475000000000005e-06, "loss": 2.3955, "step": 6095 }, { "epoch": 1.9109717868338558, "grad_norm": 53.23150634765625, "learning_rate": 3.0480000000000003e-06, "loss": 2.4416, "step": 6096 }, { "epoch": 1.9112852664576803, "grad_norm": 11.527998924255371, "learning_rate": 3.0485000000000005e-06, "loss": 2.2852, "step": 6097 }, { "epoch": 1.9115987460815047, "grad_norm": 15.593719482421875, "learning_rate": 3.0490000000000004e-06, "loss": 2.3357, "step": 6098 }, { "epoch": 1.9119122257053291, "grad_norm": 15.388204574584961, "learning_rate": 3.0495e-06, "loss": 2.2743, "step": 6099 }, { "epoch": 1.9122257053291536, "grad_norm": 9.700482368469238, "learning_rate": 3.05e-06, "loss": 2.3766, "step": 6100 }, { "epoch": 1.912539184952978, "grad_norm": 22.43489646911621, "learning_rate": 3.0505000000000003e-06, "loss": 2.169, "step": 6101 }, { "epoch": 1.9128526645768025, "grad_norm": 32.98069763183594, "learning_rate": 3.051e-06, "loss": 2.221, "step": 6102 }, { "epoch": 1.913166144200627, "grad_norm": 13.035593032836914, "learning_rate": 3.0515e-06, "loss": 2.0808, "step": 6103 }, { "epoch": 1.9134796238244514, "grad_norm": 14.784707069396973, "learning_rate": 3.0520000000000006e-06, "loss": 2.6694, "step": 6104 }, { "epoch": 1.9137931034482758, "grad_norm": 19.96974754333496, "learning_rate": 3.0525000000000004e-06, "loss": 2.2853, "step": 6105 }, { "epoch": 1.9141065830721002, "grad_norm": 21.09632682800293, "learning_rate": 3.0530000000000002e-06, "loss": 2.2353, "step": 6106 }, { "epoch": 1.9144200626959247, "grad_norm": 12.733104705810547, "learning_rate": 3.0535000000000005e-06, "loss": 2.1974, "step": 6107 }, { "epoch": 1.9147335423197491, "grad_norm": 21.158296585083008, "learning_rate": 3.0540000000000003e-06, "loss": 2.349, "step": 6108 }, { "epoch": 1.9150470219435736, "grad_norm": 12.683104515075684, "learning_rate": 3.0545e-06, "loss": 2.1925, "step": 6109 }, { "epoch": 1.915360501567398, "grad_norm": 21.3099422454834, "learning_rate": 3.0550000000000004e-06, "loss": 2.3814, "step": 6110 }, { "epoch": 1.9156739811912225, "grad_norm": 31.082923889160156, "learning_rate": 3.0555000000000002e-06, "loss": 2.4238, "step": 6111 }, { "epoch": 1.915987460815047, "grad_norm": 10.956926345825195, "learning_rate": 3.056e-06, "loss": 2.2932, "step": 6112 }, { "epoch": 1.9163009404388713, "grad_norm": 18.108543395996094, "learning_rate": 3.0565e-06, "loss": 2.3098, "step": 6113 }, { "epoch": 1.9166144200626958, "grad_norm": 8.127020835876465, "learning_rate": 3.0570000000000005e-06, "loss": 2.2654, "step": 6114 }, { "epoch": 1.9169278996865202, "grad_norm": 9.577781677246094, "learning_rate": 3.0575000000000004e-06, "loss": 2.1266, "step": 6115 }, { "epoch": 1.9172413793103447, "grad_norm": 34.873382568359375, "learning_rate": 3.0580000000000006e-06, "loss": 1.9902, "step": 6116 }, { "epoch": 1.9175548589341693, "grad_norm": 18.865312576293945, "learning_rate": 3.0585000000000004e-06, "loss": 2.2634, "step": 6117 }, { "epoch": 1.9178683385579938, "grad_norm": 59.097164154052734, "learning_rate": 3.0590000000000003e-06, "loss": 2.5326, "step": 6118 }, { "epoch": 1.9181818181818182, "grad_norm": 7.89798641204834, "learning_rate": 3.0595e-06, "loss": 2.201, "step": 6119 }, { "epoch": 1.9184952978056427, "grad_norm": 9.707578659057617, "learning_rate": 3.0600000000000003e-06, "loss": 2.2415, "step": 6120 }, { "epoch": 1.918808777429467, "grad_norm": 9.2268705368042, "learning_rate": 3.0605e-06, "loss": 2.0483, "step": 6121 }, { "epoch": 1.9191222570532915, "grad_norm": 30.342716217041016, "learning_rate": 3.061e-06, "loss": 2.4259, "step": 6122 }, { "epoch": 1.919435736677116, "grad_norm": 23.01906967163086, "learning_rate": 3.0615e-06, "loss": 2.2752, "step": 6123 }, { "epoch": 1.9197492163009404, "grad_norm": 9.22539234161377, "learning_rate": 3.0620000000000005e-06, "loss": 2.0471, "step": 6124 }, { "epoch": 1.9200626959247649, "grad_norm": 21.07687759399414, "learning_rate": 3.0625000000000003e-06, "loss": 2.4733, "step": 6125 }, { "epoch": 1.9203761755485893, "grad_norm": 19.99814224243164, "learning_rate": 3.0630000000000006e-06, "loss": 2.7217, "step": 6126 }, { "epoch": 1.9206896551724137, "grad_norm": 49.313392639160156, "learning_rate": 3.0635000000000004e-06, "loss": 3.131, "step": 6127 }, { "epoch": 1.9210031347962384, "grad_norm": 8.287384986877441, "learning_rate": 3.0640000000000002e-06, "loss": 2.1604, "step": 6128 }, { "epoch": 1.9213166144200629, "grad_norm": 16.254018783569336, "learning_rate": 3.0645e-06, "loss": 2.1469, "step": 6129 }, { "epoch": 1.9216300940438873, "grad_norm": 13.05605411529541, "learning_rate": 3.0650000000000003e-06, "loss": 2.2322, "step": 6130 }, { "epoch": 1.9219435736677117, "grad_norm": 13.967427253723145, "learning_rate": 3.0655e-06, "loss": 2.1692, "step": 6131 }, { "epoch": 1.9222570532915362, "grad_norm": 18.97654151916504, "learning_rate": 3.066e-06, "loss": 2.1141, "step": 6132 }, { "epoch": 1.9225705329153606, "grad_norm": 17.071155548095703, "learning_rate": 3.0665e-06, "loss": 2.0225, "step": 6133 }, { "epoch": 1.922884012539185, "grad_norm": 44.785980224609375, "learning_rate": 3.067e-06, "loss": 2.3243, "step": 6134 }, { "epoch": 1.9231974921630095, "grad_norm": 12.6852445602417, "learning_rate": 3.0675000000000003e-06, "loss": 2.3113, "step": 6135 }, { "epoch": 1.923510971786834, "grad_norm": 11.303502082824707, "learning_rate": 3.0680000000000005e-06, "loss": 1.9648, "step": 6136 }, { "epoch": 1.9238244514106584, "grad_norm": 9.669048309326172, "learning_rate": 3.0685000000000004e-06, "loss": 2.2378, "step": 6137 }, { "epoch": 1.9241379310344828, "grad_norm": 67.32134246826172, "learning_rate": 3.069e-06, "loss": 2.2429, "step": 6138 }, { "epoch": 1.9244514106583073, "grad_norm": 14.467636108398438, "learning_rate": 3.0695000000000004e-06, "loss": 2.3443, "step": 6139 }, { "epoch": 1.9247648902821317, "grad_norm": 26.40109634399414, "learning_rate": 3.0700000000000003e-06, "loss": 2.2199, "step": 6140 }, { "epoch": 1.9250783699059562, "grad_norm": 32.24627685546875, "learning_rate": 3.0705e-06, "loss": 2.0711, "step": 6141 }, { "epoch": 1.9253918495297806, "grad_norm": 43.16203689575195, "learning_rate": 3.071e-06, "loss": 2.0225, "step": 6142 }, { "epoch": 1.925705329153605, "grad_norm": 15.031079292297363, "learning_rate": 3.0715e-06, "loss": 2.7297, "step": 6143 }, { "epoch": 1.9260188087774295, "grad_norm": 11.947735786437988, "learning_rate": 3.072e-06, "loss": 2.2531, "step": 6144 }, { "epoch": 1.926332288401254, "grad_norm": 11.026966094970703, "learning_rate": 3.0725000000000007e-06, "loss": 2.1997, "step": 6145 }, { "epoch": 1.9266457680250784, "grad_norm": 12.493368148803711, "learning_rate": 3.0730000000000005e-06, "loss": 2.2338, "step": 6146 }, { "epoch": 1.9269592476489028, "grad_norm": 7.645673751831055, "learning_rate": 3.0735000000000003e-06, "loss": 2.2303, "step": 6147 }, { "epoch": 1.9272727272727272, "grad_norm": 42.979488372802734, "learning_rate": 3.074e-06, "loss": 2.0703, "step": 6148 }, { "epoch": 1.9275862068965517, "grad_norm": 9.205596923828125, "learning_rate": 3.0745000000000004e-06, "loss": 2.1779, "step": 6149 }, { "epoch": 1.9278996865203761, "grad_norm": 12.70852279663086, "learning_rate": 3.075e-06, "loss": 2.2296, "step": 6150 }, { "epoch": 1.9282131661442006, "grad_norm": 12.988320350646973, "learning_rate": 3.0755e-06, "loss": 2.1411, "step": 6151 }, { "epoch": 1.928526645768025, "grad_norm": 9.5177583694458, "learning_rate": 3.0760000000000003e-06, "loss": 2.3787, "step": 6152 }, { "epoch": 1.9288401253918495, "grad_norm": 16.0618839263916, "learning_rate": 3.0765e-06, "loss": 1.8497, "step": 6153 }, { "epoch": 1.929153605015674, "grad_norm": 51.27670669555664, "learning_rate": 3.077e-06, "loss": 2.3243, "step": 6154 }, { "epoch": 1.9294670846394983, "grad_norm": 19.865938186645508, "learning_rate": 3.0775000000000006e-06, "loss": 2.0024, "step": 6155 }, { "epoch": 1.9297805642633228, "grad_norm": 65.7314682006836, "learning_rate": 3.0780000000000004e-06, "loss": 2.3256, "step": 6156 }, { "epoch": 1.9300940438871472, "grad_norm": 55.74231719970703, "learning_rate": 3.0785000000000003e-06, "loss": 2.622, "step": 6157 }, { "epoch": 1.9304075235109717, "grad_norm": 15.176788330078125, "learning_rate": 3.0790000000000005e-06, "loss": 2.3372, "step": 6158 }, { "epoch": 1.930721003134796, "grad_norm": 12.638873100280762, "learning_rate": 3.0795000000000003e-06, "loss": 2.5779, "step": 6159 }, { "epoch": 1.9310344827586206, "grad_norm": 12.51217269897461, "learning_rate": 3.08e-06, "loss": 2.2006, "step": 6160 }, { "epoch": 1.931347962382445, "grad_norm": 16.80646514892578, "learning_rate": 3.0805e-06, "loss": 2.208, "step": 6161 }, { "epoch": 1.9316614420062694, "grad_norm": 19.71490478515625, "learning_rate": 3.0810000000000002e-06, "loss": 2.3028, "step": 6162 }, { "epoch": 1.9319749216300939, "grad_norm": 15.717697143554688, "learning_rate": 3.0815e-06, "loss": 2.3063, "step": 6163 }, { "epoch": 1.9322884012539185, "grad_norm": 31.971899032592773, "learning_rate": 3.082e-06, "loss": 1.9921, "step": 6164 }, { "epoch": 1.932601880877743, "grad_norm": 34.533424377441406, "learning_rate": 3.0825000000000006e-06, "loss": 2.0271, "step": 6165 }, { "epoch": 1.9329153605015674, "grad_norm": 81.34234619140625, "learning_rate": 3.0830000000000004e-06, "loss": 2.3326, "step": 6166 }, { "epoch": 1.9332288401253919, "grad_norm": 23.47624969482422, "learning_rate": 3.0835e-06, "loss": 2.111, "step": 6167 }, { "epoch": 1.9335423197492163, "grad_norm": 15.721684455871582, "learning_rate": 3.0840000000000005e-06, "loss": 2.1075, "step": 6168 }, { "epoch": 1.9338557993730408, "grad_norm": 41.16612243652344, "learning_rate": 3.0845000000000003e-06, "loss": 2.0034, "step": 6169 }, { "epoch": 1.9341692789968652, "grad_norm": 8.464729309082031, "learning_rate": 3.085e-06, "loss": 2.0479, "step": 6170 }, { "epoch": 1.9344827586206896, "grad_norm": 23.668128967285156, "learning_rate": 3.0855000000000004e-06, "loss": 2.283, "step": 6171 }, { "epoch": 1.934796238244514, "grad_norm": 18.2292537689209, "learning_rate": 3.086e-06, "loss": 2.5008, "step": 6172 }, { "epoch": 1.9351097178683385, "grad_norm": 15.80252456665039, "learning_rate": 3.0865e-06, "loss": 2.059, "step": 6173 }, { "epoch": 1.935423197492163, "grad_norm": 14.27904224395752, "learning_rate": 3.087e-06, "loss": 2.488, "step": 6174 }, { "epoch": 1.9357366771159876, "grad_norm": 21.141990661621094, "learning_rate": 3.0875000000000005e-06, "loss": 2.1918, "step": 6175 }, { "epoch": 1.936050156739812, "grad_norm": 13.165213584899902, "learning_rate": 3.0880000000000003e-06, "loss": 2.1132, "step": 6176 }, { "epoch": 1.9363636363636365, "grad_norm": 11.983834266662598, "learning_rate": 3.0885000000000006e-06, "loss": 2.1287, "step": 6177 }, { "epoch": 1.936677115987461, "grad_norm": 12.687564849853516, "learning_rate": 3.0890000000000004e-06, "loss": 2.433, "step": 6178 }, { "epoch": 1.9369905956112854, "grad_norm": 53.66118621826172, "learning_rate": 3.0895000000000002e-06, "loss": 2.3676, "step": 6179 }, { "epoch": 1.9373040752351098, "grad_norm": 28.613327026367188, "learning_rate": 3.09e-06, "loss": 2.297, "step": 6180 }, { "epoch": 1.9376175548589343, "grad_norm": 13.394174575805664, "learning_rate": 3.0905000000000003e-06, "loss": 2.0543, "step": 6181 }, { "epoch": 1.9379310344827587, "grad_norm": 15.933115005493164, "learning_rate": 3.091e-06, "loss": 2.2325, "step": 6182 }, { "epoch": 1.9382445141065832, "grad_norm": 26.931055068969727, "learning_rate": 3.0915e-06, "loss": 2.072, "step": 6183 }, { "epoch": 1.9385579937304076, "grad_norm": 21.357282638549805, "learning_rate": 3.092e-06, "loss": 2.1702, "step": 6184 }, { "epoch": 1.938871473354232, "grad_norm": 16.33864974975586, "learning_rate": 3.0925000000000005e-06, "loss": 2.5012, "step": 6185 }, { "epoch": 1.9391849529780565, "grad_norm": 17.6795711517334, "learning_rate": 3.0930000000000003e-06, "loss": 2.4467, "step": 6186 }, { "epoch": 1.939498432601881, "grad_norm": 83.75550079345703, "learning_rate": 3.0935000000000005e-06, "loss": 3.2387, "step": 6187 }, { "epoch": 1.9398119122257054, "grad_norm": 23.691688537597656, "learning_rate": 3.0940000000000004e-06, "loss": 2.1602, "step": 6188 }, { "epoch": 1.9401253918495298, "grad_norm": 15.03283977508545, "learning_rate": 3.0945e-06, "loss": 2.265, "step": 6189 }, { "epoch": 1.9404388714733543, "grad_norm": 19.019834518432617, "learning_rate": 3.0950000000000004e-06, "loss": 2.1407, "step": 6190 }, { "epoch": 1.9407523510971787, "grad_norm": 17.266340255737305, "learning_rate": 3.0955000000000003e-06, "loss": 2.2096, "step": 6191 }, { "epoch": 1.9410658307210031, "grad_norm": 19.639511108398438, "learning_rate": 3.096e-06, "loss": 2.2862, "step": 6192 }, { "epoch": 1.9413793103448276, "grad_norm": 32.09061813354492, "learning_rate": 3.0965e-06, "loss": 2.0854, "step": 6193 }, { "epoch": 1.941692789968652, "grad_norm": 12.486015319824219, "learning_rate": 3.097e-06, "loss": 2.2919, "step": 6194 }, { "epoch": 1.9420062695924765, "grad_norm": 15.492584228515625, "learning_rate": 3.0975000000000004e-06, "loss": 2.2623, "step": 6195 }, { "epoch": 1.942319749216301, "grad_norm": 9.039558410644531, "learning_rate": 3.0980000000000007e-06, "loss": 2.2022, "step": 6196 }, { "epoch": 1.9426332288401253, "grad_norm": 14.141021728515625, "learning_rate": 3.0985000000000005e-06, "loss": 2.7238, "step": 6197 }, { "epoch": 1.9429467084639498, "grad_norm": 7.6339945793151855, "learning_rate": 3.0990000000000003e-06, "loss": 2.6758, "step": 6198 }, { "epoch": 1.9432601880877742, "grad_norm": 10.205277442932129, "learning_rate": 3.0995e-06, "loss": 2.2798, "step": 6199 }, { "epoch": 1.9435736677115987, "grad_norm": 15.242185592651367, "learning_rate": 3.1000000000000004e-06, "loss": 2.6331, "step": 6200 }, { "epoch": 1.9438871473354231, "grad_norm": 25.186603546142578, "learning_rate": 3.1005000000000002e-06, "loss": 2.1404, "step": 6201 }, { "epoch": 1.9442006269592476, "grad_norm": 51.08753204345703, "learning_rate": 3.101e-06, "loss": 2.3513, "step": 6202 }, { "epoch": 1.944514106583072, "grad_norm": 9.865594863891602, "learning_rate": 3.1015e-06, "loss": 2.2377, "step": 6203 }, { "epoch": 1.9448275862068964, "grad_norm": 9.022784233093262, "learning_rate": 3.102e-06, "loss": 2.0802, "step": 6204 }, { "epoch": 1.9451410658307209, "grad_norm": 38.94915771484375, "learning_rate": 3.1025000000000004e-06, "loss": 2.7855, "step": 6205 }, { "epoch": 1.9454545454545453, "grad_norm": 16.95958137512207, "learning_rate": 3.1030000000000006e-06, "loss": 1.9462, "step": 6206 }, { "epoch": 1.9457680250783698, "grad_norm": 103.67754364013672, "learning_rate": 3.1035000000000004e-06, "loss": 2.4162, "step": 6207 }, { "epoch": 1.9460815047021942, "grad_norm": 20.15214729309082, "learning_rate": 3.1040000000000003e-06, "loss": 2.2354, "step": 6208 }, { "epoch": 1.9463949843260187, "grad_norm": 29.8538875579834, "learning_rate": 3.1045e-06, "loss": 2.1404, "step": 6209 }, { "epoch": 1.9467084639498433, "grad_norm": 18.50629997253418, "learning_rate": 3.1050000000000003e-06, "loss": 2.2808, "step": 6210 }, { "epoch": 1.9470219435736678, "grad_norm": 13.072752952575684, "learning_rate": 3.1055e-06, "loss": 2.1819, "step": 6211 }, { "epoch": 1.9473354231974922, "grad_norm": 12.113188743591309, "learning_rate": 3.106e-06, "loss": 2.2628, "step": 6212 }, { "epoch": 1.9476489028213166, "grad_norm": 19.383434295654297, "learning_rate": 3.1065000000000002e-06, "loss": 2.2511, "step": 6213 }, { "epoch": 1.947962382445141, "grad_norm": 13.1512451171875, "learning_rate": 3.107e-06, "loss": 1.8007, "step": 6214 }, { "epoch": 1.9482758620689655, "grad_norm": 29.93698501586914, "learning_rate": 3.1075000000000003e-06, "loss": 2.9913, "step": 6215 }, { "epoch": 1.94858934169279, "grad_norm": 51.03324890136719, "learning_rate": 3.1080000000000006e-06, "loss": 2.4474, "step": 6216 }, { "epoch": 1.9489028213166144, "grad_norm": 21.86216163635254, "learning_rate": 3.1085000000000004e-06, "loss": 2.3511, "step": 6217 }, { "epoch": 1.9492163009404389, "grad_norm": 44.02346420288086, "learning_rate": 3.1090000000000002e-06, "loss": 2.2321, "step": 6218 }, { "epoch": 1.9495297805642633, "grad_norm": 10.752887725830078, "learning_rate": 3.1095000000000005e-06, "loss": 2.1363, "step": 6219 }, { "epoch": 1.9498432601880877, "grad_norm": 9.247302055358887, "learning_rate": 3.1100000000000003e-06, "loss": 2.2588, "step": 6220 }, { "epoch": 1.9501567398119122, "grad_norm": 15.984376907348633, "learning_rate": 3.1105e-06, "loss": 2.384, "step": 6221 }, { "epoch": 1.9504702194357368, "grad_norm": 33.194580078125, "learning_rate": 3.111e-06, "loss": 2.8689, "step": 6222 }, { "epoch": 1.9507836990595613, "grad_norm": 10.841005325317383, "learning_rate": 3.1115e-06, "loss": 1.9113, "step": 6223 }, { "epoch": 1.9510971786833857, "grad_norm": 9.533475875854492, "learning_rate": 3.112e-06, "loss": 2.1351, "step": 6224 }, { "epoch": 1.9514106583072102, "grad_norm": 18.278017044067383, "learning_rate": 3.1125000000000007e-06, "loss": 2.4297, "step": 6225 }, { "epoch": 1.9517241379310346, "grad_norm": 14.732829093933105, "learning_rate": 3.1130000000000005e-06, "loss": 2.0193, "step": 6226 }, { "epoch": 1.952037617554859, "grad_norm": 14.808735847473145, "learning_rate": 3.1135000000000003e-06, "loss": 2.1229, "step": 6227 }, { "epoch": 1.9523510971786835, "grad_norm": 40.86724090576172, "learning_rate": 3.114e-06, "loss": 2.3927, "step": 6228 }, { "epoch": 1.952664576802508, "grad_norm": 16.639362335205078, "learning_rate": 3.1145000000000004e-06, "loss": 2.5477, "step": 6229 }, { "epoch": 1.9529780564263324, "grad_norm": 24.524995803833008, "learning_rate": 3.1150000000000002e-06, "loss": 2.2817, "step": 6230 }, { "epoch": 1.9532915360501568, "grad_norm": 32.46340560913086, "learning_rate": 3.1155e-06, "loss": 2.3405, "step": 6231 }, { "epoch": 1.9536050156739813, "grad_norm": 15.452544212341309, "learning_rate": 3.1160000000000003e-06, "loss": 2.4422, "step": 6232 }, { "epoch": 1.9539184952978057, "grad_norm": 15.222606658935547, "learning_rate": 3.1165e-06, "loss": 2.336, "step": 6233 }, { "epoch": 1.9542319749216301, "grad_norm": 25.828502655029297, "learning_rate": 3.117e-06, "loss": 2.2059, "step": 6234 }, { "epoch": 1.9545454545454546, "grad_norm": 15.925555229187012, "learning_rate": 3.1175000000000006e-06, "loss": 2.2686, "step": 6235 }, { "epoch": 1.954858934169279, "grad_norm": 13.170594215393066, "learning_rate": 3.1180000000000005e-06, "loss": 2.5014, "step": 6236 }, { "epoch": 1.9551724137931035, "grad_norm": 12.45872688293457, "learning_rate": 3.1185000000000003e-06, "loss": 2.0243, "step": 6237 }, { "epoch": 1.955485893416928, "grad_norm": 15.75046443939209, "learning_rate": 3.1190000000000005e-06, "loss": 2.0471, "step": 6238 }, { "epoch": 1.9557993730407524, "grad_norm": 92.76771545410156, "learning_rate": 3.1195000000000004e-06, "loss": 2.7066, "step": 6239 }, { "epoch": 1.9561128526645768, "grad_norm": 47.9924201965332, "learning_rate": 3.12e-06, "loss": 2.211, "step": 6240 }, { "epoch": 1.9564263322884012, "grad_norm": 13.54965591430664, "learning_rate": 3.1205e-06, "loss": 2.274, "step": 6241 }, { "epoch": 1.9567398119122257, "grad_norm": 11.754216194152832, "learning_rate": 3.1210000000000003e-06, "loss": 2.4049, "step": 6242 }, { "epoch": 1.9570532915360501, "grad_norm": 12.851913452148438, "learning_rate": 3.1215e-06, "loss": 2.3721, "step": 6243 }, { "epoch": 1.9573667711598746, "grad_norm": 11.860594749450684, "learning_rate": 3.122e-06, "loss": 2.1425, "step": 6244 }, { "epoch": 1.957680250783699, "grad_norm": 29.15382957458496, "learning_rate": 3.1225000000000006e-06, "loss": 2.2544, "step": 6245 }, { "epoch": 1.9579937304075234, "grad_norm": 32.79502487182617, "learning_rate": 3.1230000000000004e-06, "loss": 2.195, "step": 6246 }, { "epoch": 1.958307210031348, "grad_norm": 8.380058288574219, "learning_rate": 3.1235000000000002e-06, "loss": 2.2671, "step": 6247 }, { "epoch": 1.9586206896551723, "grad_norm": 41.89897918701172, "learning_rate": 3.1240000000000005e-06, "loss": 2.641, "step": 6248 }, { "epoch": 1.9589341692789968, "grad_norm": 10.770977020263672, "learning_rate": 3.1245000000000003e-06, "loss": 2.4666, "step": 6249 }, { "epoch": 1.9592476489028212, "grad_norm": 21.347427368164062, "learning_rate": 3.125e-06, "loss": 2.3943, "step": 6250 }, { "epoch": 1.9595611285266457, "grad_norm": 20.523212432861328, "learning_rate": 3.1255000000000004e-06, "loss": 2.3104, "step": 6251 }, { "epoch": 1.95987460815047, "grad_norm": 45.21733474731445, "learning_rate": 3.1260000000000002e-06, "loss": 2.0047, "step": 6252 }, { "epoch": 1.9601880877742945, "grad_norm": 19.431230545043945, "learning_rate": 3.1265e-06, "loss": 2.1989, "step": 6253 }, { "epoch": 1.960501567398119, "grad_norm": 43.393760681152344, "learning_rate": 3.127e-06, "loss": 2.1272, "step": 6254 }, { "epoch": 1.9608150470219434, "grad_norm": 19.50667381286621, "learning_rate": 3.1275e-06, "loss": 2.1979, "step": 6255 }, { "epoch": 1.9611285266457679, "grad_norm": 24.86729621887207, "learning_rate": 3.1280000000000004e-06, "loss": 2.296, "step": 6256 }, { "epoch": 1.9614420062695925, "grad_norm": 10.600141525268555, "learning_rate": 3.1285000000000006e-06, "loss": 2.0203, "step": 6257 }, { "epoch": 1.961755485893417, "grad_norm": 15.078958511352539, "learning_rate": 3.1290000000000005e-06, "loss": 2.7117, "step": 6258 }, { "epoch": 1.9620689655172414, "grad_norm": 31.340831756591797, "learning_rate": 3.1295000000000003e-06, "loss": 2.2201, "step": 6259 }, { "epoch": 1.9623824451410659, "grad_norm": 11.659078598022461, "learning_rate": 3.13e-06, "loss": 2.2339, "step": 6260 }, { "epoch": 1.9626959247648903, "grad_norm": 21.408418655395508, "learning_rate": 3.1305000000000004e-06, "loss": 2.2118, "step": 6261 }, { "epoch": 1.9630094043887147, "grad_norm": 32.824012756347656, "learning_rate": 3.131e-06, "loss": 2.5834, "step": 6262 }, { "epoch": 1.9633228840125392, "grad_norm": 22.480697631835938, "learning_rate": 3.1315e-06, "loss": 2.8845, "step": 6263 }, { "epoch": 1.9636363636363636, "grad_norm": 14.943920135498047, "learning_rate": 3.132e-06, "loss": 2.5693, "step": 6264 }, { "epoch": 1.963949843260188, "grad_norm": 21.227256774902344, "learning_rate": 3.1325e-06, "loss": 2.3999, "step": 6265 }, { "epoch": 1.9642633228840125, "grad_norm": 10.880897521972656, "learning_rate": 3.1330000000000003e-06, "loss": 2.2018, "step": 6266 }, { "epoch": 1.964576802507837, "grad_norm": 63.30581283569336, "learning_rate": 3.1335000000000006e-06, "loss": 1.9168, "step": 6267 }, { "epoch": 1.9648902821316614, "grad_norm": 10.682960510253906, "learning_rate": 3.1340000000000004e-06, "loss": 2.054, "step": 6268 }, { "epoch": 1.965203761755486, "grad_norm": 11.94248104095459, "learning_rate": 3.1345000000000002e-06, "loss": 2.3176, "step": 6269 }, { "epoch": 1.9655172413793105, "grad_norm": 16.106670379638672, "learning_rate": 3.135e-06, "loss": 2.0857, "step": 6270 }, { "epoch": 1.965830721003135, "grad_norm": 6.923719882965088, "learning_rate": 3.1355000000000003e-06, "loss": 2.1489, "step": 6271 }, { "epoch": 1.9661442006269594, "grad_norm": 12.050187110900879, "learning_rate": 3.136e-06, "loss": 2.2787, "step": 6272 }, { "epoch": 1.9664576802507838, "grad_norm": 33.00879669189453, "learning_rate": 3.1365e-06, "loss": 2.3985, "step": 6273 }, { "epoch": 1.9667711598746083, "grad_norm": 12.105003356933594, "learning_rate": 3.137e-06, "loss": 2.3161, "step": 6274 }, { "epoch": 1.9670846394984327, "grad_norm": 13.618671417236328, "learning_rate": 3.1375e-06, "loss": 2.2129, "step": 6275 }, { "epoch": 1.9673981191222571, "grad_norm": 12.591639518737793, "learning_rate": 3.1380000000000003e-06, "loss": 2.4553, "step": 6276 }, { "epoch": 1.9677115987460816, "grad_norm": 17.883012771606445, "learning_rate": 3.1385000000000005e-06, "loss": 2.1222, "step": 6277 }, { "epoch": 1.968025078369906, "grad_norm": 16.255971908569336, "learning_rate": 3.1390000000000004e-06, "loss": 2.2342, "step": 6278 }, { "epoch": 1.9683385579937305, "grad_norm": 16.84834861755371, "learning_rate": 3.1395e-06, "loss": 2.473, "step": 6279 }, { "epoch": 1.968652037617555, "grad_norm": 19.836959838867188, "learning_rate": 3.1400000000000004e-06, "loss": 2.2049, "step": 6280 }, { "epoch": 1.9689655172413794, "grad_norm": 46.876102447509766, "learning_rate": 3.1405000000000003e-06, "loss": 2.7783, "step": 6281 }, { "epoch": 1.9692789968652038, "grad_norm": 19.187061309814453, "learning_rate": 3.141e-06, "loss": 2.2508, "step": 6282 }, { "epoch": 1.9695924764890282, "grad_norm": 23.875375747680664, "learning_rate": 3.1415e-06, "loss": 2.7126, "step": 6283 }, { "epoch": 1.9699059561128527, "grad_norm": 15.11329460144043, "learning_rate": 3.142e-06, "loss": 2.1725, "step": 6284 }, { "epoch": 1.9702194357366771, "grad_norm": 64.44976806640625, "learning_rate": 3.1425e-06, "loss": 2.3958, "step": 6285 }, { "epoch": 1.9705329153605016, "grad_norm": 57.60660171508789, "learning_rate": 3.1430000000000007e-06, "loss": 2.4208, "step": 6286 }, { "epoch": 1.970846394984326, "grad_norm": 34.09165954589844, "learning_rate": 3.1435000000000005e-06, "loss": 2.2666, "step": 6287 }, { "epoch": 1.9711598746081505, "grad_norm": 32.35491180419922, "learning_rate": 3.1440000000000003e-06, "loss": 2.2841, "step": 6288 }, { "epoch": 1.971473354231975, "grad_norm": 11.348264694213867, "learning_rate": 3.1445e-06, "loss": 2.0545, "step": 6289 }, { "epoch": 1.9717868338557993, "grad_norm": 12.587067604064941, "learning_rate": 3.1450000000000004e-06, "loss": 2.0804, "step": 6290 }, { "epoch": 1.9721003134796238, "grad_norm": 19.97517967224121, "learning_rate": 3.1455e-06, "loss": 2.5052, "step": 6291 }, { "epoch": 1.9724137931034482, "grad_norm": 15.031810760498047, "learning_rate": 3.146e-06, "loss": 2.192, "step": 6292 }, { "epoch": 1.9727272727272727, "grad_norm": 10.090429306030273, "learning_rate": 3.1465000000000003e-06, "loss": 2.3494, "step": 6293 }, { "epoch": 1.973040752351097, "grad_norm": 16.087080001831055, "learning_rate": 3.147e-06, "loss": 2.1995, "step": 6294 }, { "epoch": 1.9733542319749215, "grad_norm": 50.80659866333008, "learning_rate": 3.1475e-06, "loss": 2.3741, "step": 6295 }, { "epoch": 1.973667711598746, "grad_norm": 8.694010734558105, "learning_rate": 3.1480000000000006e-06, "loss": 2.1153, "step": 6296 }, { "epoch": 1.9739811912225704, "grad_norm": 22.7263240814209, "learning_rate": 3.1485000000000004e-06, "loss": 2.2891, "step": 6297 }, { "epoch": 1.9742946708463949, "grad_norm": 6.88250207901001, "learning_rate": 3.1490000000000003e-06, "loss": 2.2056, "step": 6298 }, { "epoch": 1.9746081504702193, "grad_norm": 13.31670093536377, "learning_rate": 3.1495000000000005e-06, "loss": 2.0936, "step": 6299 }, { "epoch": 1.9749216300940438, "grad_norm": 11.524837493896484, "learning_rate": 3.1500000000000003e-06, "loss": 2.08, "step": 6300 }, { "epoch": 1.9752351097178682, "grad_norm": 8.468381881713867, "learning_rate": 3.1505e-06, "loss": 2.1082, "step": 6301 }, { "epoch": 1.9755485893416926, "grad_norm": 22.814212799072266, "learning_rate": 3.151e-06, "loss": 2.2338, "step": 6302 }, { "epoch": 1.975862068965517, "grad_norm": 15.636788368225098, "learning_rate": 3.1515000000000002e-06, "loss": 2.7083, "step": 6303 }, { "epoch": 1.9761755485893417, "grad_norm": 14.776278495788574, "learning_rate": 3.152e-06, "loss": 2.2016, "step": 6304 }, { "epoch": 1.9764890282131662, "grad_norm": 15.003888130187988, "learning_rate": 3.1525e-06, "loss": 2.2169, "step": 6305 }, { "epoch": 1.9768025078369906, "grad_norm": 17.022552490234375, "learning_rate": 3.1530000000000006e-06, "loss": 2.2239, "step": 6306 }, { "epoch": 1.977115987460815, "grad_norm": 12.406381607055664, "learning_rate": 3.1535000000000004e-06, "loss": 2.2029, "step": 6307 }, { "epoch": 1.9774294670846395, "grad_norm": 10.545319557189941, "learning_rate": 3.154e-06, "loss": 1.9441, "step": 6308 }, { "epoch": 1.977742946708464, "grad_norm": 11.90626049041748, "learning_rate": 3.1545000000000005e-06, "loss": 2.2987, "step": 6309 }, { "epoch": 1.9780564263322884, "grad_norm": 15.618658065795898, "learning_rate": 3.1550000000000003e-06, "loss": 1.8933, "step": 6310 }, { "epoch": 1.9783699059561128, "grad_norm": 9.833617210388184, "learning_rate": 3.1555e-06, "loss": 1.979, "step": 6311 }, { "epoch": 1.9786833855799373, "grad_norm": 13.40846061706543, "learning_rate": 3.1560000000000004e-06, "loss": 2.1384, "step": 6312 }, { "epoch": 1.9789968652037617, "grad_norm": 16.884504318237305, "learning_rate": 3.1565e-06, "loss": 2.3817, "step": 6313 }, { "epoch": 1.9793103448275862, "grad_norm": 30.242950439453125, "learning_rate": 3.157e-06, "loss": 2.4918, "step": 6314 }, { "epoch": 1.9796238244514106, "grad_norm": 12.3598051071167, "learning_rate": 3.1575e-06, "loss": 1.9898, "step": 6315 }, { "epoch": 1.9799373040752353, "grad_norm": 23.940189361572266, "learning_rate": 3.1580000000000005e-06, "loss": 2.2078, "step": 6316 }, { "epoch": 1.9802507836990597, "grad_norm": 6.898366928100586, "learning_rate": 3.1585000000000003e-06, "loss": 2.1173, "step": 6317 }, { "epoch": 1.9805642633228842, "grad_norm": 21.28299903869629, "learning_rate": 3.1590000000000006e-06, "loss": 2.7358, "step": 6318 }, { "epoch": 1.9808777429467086, "grad_norm": 14.182652473449707, "learning_rate": 3.1595000000000004e-06, "loss": 2.1299, "step": 6319 }, { "epoch": 1.981191222570533, "grad_norm": 9.222768783569336, "learning_rate": 3.1600000000000002e-06, "loss": 2.1796, "step": 6320 }, { "epoch": 1.9815047021943575, "grad_norm": 87.75921630859375, "learning_rate": 3.1605e-06, "loss": 2.3708, "step": 6321 }, { "epoch": 1.981818181818182, "grad_norm": 18.29620361328125, "learning_rate": 3.1610000000000003e-06, "loss": 2.4295, "step": 6322 }, { "epoch": 1.9821316614420064, "grad_norm": 14.226387977600098, "learning_rate": 3.1615e-06, "loss": 2.1621, "step": 6323 }, { "epoch": 1.9824451410658308, "grad_norm": 6.915779113769531, "learning_rate": 3.162e-06, "loss": 2.3139, "step": 6324 }, { "epoch": 1.9827586206896552, "grad_norm": 31.989805221557617, "learning_rate": 3.1625000000000002e-06, "loss": 2.4353, "step": 6325 }, { "epoch": 1.9830721003134797, "grad_norm": 10.334614753723145, "learning_rate": 3.1630000000000005e-06, "loss": 2.3891, "step": 6326 }, { "epoch": 1.9833855799373041, "grad_norm": 32.01905822753906, "learning_rate": 3.1635000000000003e-06, "loss": 2.1868, "step": 6327 }, { "epoch": 1.9836990595611286, "grad_norm": 19.62434959411621, "learning_rate": 3.1640000000000005e-06, "loss": 2.3079, "step": 6328 }, { "epoch": 1.984012539184953, "grad_norm": 12.64426040649414, "learning_rate": 3.1645000000000004e-06, "loss": 2.0242, "step": 6329 }, { "epoch": 1.9843260188087775, "grad_norm": 28.206012725830078, "learning_rate": 3.165e-06, "loss": 2.1971, "step": 6330 }, { "epoch": 1.984639498432602, "grad_norm": 17.226451873779297, "learning_rate": 3.1655000000000004e-06, "loss": 2.1305, "step": 6331 }, { "epoch": 1.9849529780564263, "grad_norm": 13.701321601867676, "learning_rate": 3.1660000000000003e-06, "loss": 2.0846, "step": 6332 }, { "epoch": 1.9852664576802508, "grad_norm": 53.87226104736328, "learning_rate": 3.1665e-06, "loss": 2.5442, "step": 6333 }, { "epoch": 1.9855799373040752, "grad_norm": 57.23225402832031, "learning_rate": 3.167e-06, "loss": 2.3749, "step": 6334 }, { "epoch": 1.9858934169278997, "grad_norm": 21.677019119262695, "learning_rate": 3.1675e-06, "loss": 2.0964, "step": 6335 }, { "epoch": 1.986206896551724, "grad_norm": 53.68826675415039, "learning_rate": 3.1680000000000004e-06, "loss": 3.0364, "step": 6336 }, { "epoch": 1.9865203761755486, "grad_norm": 33.148406982421875, "learning_rate": 3.1685000000000007e-06, "loss": 2.275, "step": 6337 }, { "epoch": 1.986833855799373, "grad_norm": 7.798027038574219, "learning_rate": 3.1690000000000005e-06, "loss": 1.9571, "step": 6338 }, { "epoch": 1.9871473354231974, "grad_norm": 9.131011962890625, "learning_rate": 3.1695000000000003e-06, "loss": 2.3012, "step": 6339 }, { "epoch": 1.9874608150470219, "grad_norm": 14.6558256149292, "learning_rate": 3.17e-06, "loss": 2.1853, "step": 6340 }, { "epoch": 1.9877742946708463, "grad_norm": 25.692052841186523, "learning_rate": 3.1705000000000004e-06, "loss": 2.2872, "step": 6341 }, { "epoch": 1.9880877742946708, "grad_norm": 43.00083541870117, "learning_rate": 3.1710000000000002e-06, "loss": 2.8004, "step": 6342 }, { "epoch": 1.9884012539184952, "grad_norm": 17.228382110595703, "learning_rate": 3.1715e-06, "loss": 2.2165, "step": 6343 }, { "epoch": 1.9887147335423196, "grad_norm": 19.439098358154297, "learning_rate": 3.172e-06, "loss": 2.3464, "step": 6344 }, { "epoch": 1.989028213166144, "grad_norm": 49.01555633544922, "learning_rate": 3.1725e-06, "loss": 2.4074, "step": 6345 }, { "epoch": 1.9893416927899685, "grad_norm": 14.276300430297852, "learning_rate": 3.1730000000000004e-06, "loss": 2.1991, "step": 6346 }, { "epoch": 1.989655172413793, "grad_norm": 23.116601943969727, "learning_rate": 3.1735000000000006e-06, "loss": 2.1692, "step": 6347 }, { "epoch": 1.9899686520376174, "grad_norm": 15.478215217590332, "learning_rate": 3.1740000000000004e-06, "loss": 2.0933, "step": 6348 }, { "epoch": 1.9902821316614419, "grad_norm": 17.975017547607422, "learning_rate": 3.1745000000000003e-06, "loss": 2.1837, "step": 6349 }, { "epoch": 1.9905956112852663, "grad_norm": 6.596172332763672, "learning_rate": 3.175e-06, "loss": 2.0905, "step": 6350 }, { "epoch": 1.990909090909091, "grad_norm": 13.028762817382812, "learning_rate": 3.1755000000000003e-06, "loss": 2.2581, "step": 6351 }, { "epoch": 1.9912225705329154, "grad_norm": 10.148426055908203, "learning_rate": 3.176e-06, "loss": 2.3209, "step": 6352 }, { "epoch": 1.9915360501567398, "grad_norm": 7.7970476150512695, "learning_rate": 3.1765e-06, "loss": 2.0534, "step": 6353 }, { "epoch": 1.9918495297805643, "grad_norm": 47.781227111816406, "learning_rate": 3.1770000000000002e-06, "loss": 2.2884, "step": 6354 }, { "epoch": 1.9921630094043887, "grad_norm": 13.419988632202148, "learning_rate": 3.1775e-06, "loss": 2.347, "step": 6355 }, { "epoch": 1.9924764890282132, "grad_norm": 14.526352882385254, "learning_rate": 3.1780000000000003e-06, "loss": 2.3624, "step": 6356 }, { "epoch": 1.9927899686520376, "grad_norm": 51.24506759643555, "learning_rate": 3.1785000000000006e-06, "loss": 2.3981, "step": 6357 }, { "epoch": 1.993103448275862, "grad_norm": 46.352943420410156, "learning_rate": 3.1790000000000004e-06, "loss": 2.729, "step": 6358 }, { "epoch": 1.9934169278996865, "grad_norm": 11.26982307434082, "learning_rate": 3.1795000000000002e-06, "loss": 2.1656, "step": 6359 }, { "epoch": 1.993730407523511, "grad_norm": 12.368547439575195, "learning_rate": 3.1800000000000005e-06, "loss": 2.3287, "step": 6360 }, { "epoch": 1.9940438871473354, "grad_norm": 12.606298446655273, "learning_rate": 3.1805000000000003e-06, "loss": 2.1878, "step": 6361 }, { "epoch": 1.9943573667711598, "grad_norm": 6.567707538604736, "learning_rate": 3.181e-06, "loss": 2.2835, "step": 6362 }, { "epoch": 1.9946708463949845, "grad_norm": 25.984466552734375, "learning_rate": 3.1815e-06, "loss": 2.4004, "step": 6363 }, { "epoch": 1.994984326018809, "grad_norm": 17.47072410583496, "learning_rate": 3.182e-06, "loss": 2.3559, "step": 6364 }, { "epoch": 1.9952978056426334, "grad_norm": 15.98811149597168, "learning_rate": 3.1825e-06, "loss": 2.1197, "step": 6365 }, { "epoch": 1.9956112852664578, "grad_norm": 6.573192119598389, "learning_rate": 3.1830000000000007e-06, "loss": 2.4837, "step": 6366 }, { "epoch": 1.9959247648902823, "grad_norm": 18.482707977294922, "learning_rate": 3.1835000000000005e-06, "loss": 2.2992, "step": 6367 }, { "epoch": 1.9962382445141067, "grad_norm": 19.308034896850586, "learning_rate": 3.1840000000000003e-06, "loss": 2.3165, "step": 6368 }, { "epoch": 1.9965517241379311, "grad_norm": 6.105698585510254, "learning_rate": 3.1845e-06, "loss": 1.9594, "step": 6369 }, { "epoch": 1.9968652037617556, "grad_norm": 32.73683547973633, "learning_rate": 3.1850000000000004e-06, "loss": 2.3862, "step": 6370 }, { "epoch": 1.99717868338558, "grad_norm": 13.382736206054688, "learning_rate": 3.1855000000000002e-06, "loss": 2.3056, "step": 6371 }, { "epoch": 1.9974921630094045, "grad_norm": 50.91844177246094, "learning_rate": 3.186e-06, "loss": 2.1562, "step": 6372 }, { "epoch": 1.997805642633229, "grad_norm": 16.6110897064209, "learning_rate": 3.1865000000000003e-06, "loss": 2.1552, "step": 6373 }, { "epoch": 1.9981191222570533, "grad_norm": 12.067429542541504, "learning_rate": 3.187e-06, "loss": 1.9695, "step": 6374 }, { "epoch": 1.9984326018808778, "grad_norm": 47.696414947509766, "learning_rate": 3.1875e-06, "loss": 2.3831, "step": 6375 }, { "epoch": 1.9987460815047022, "grad_norm": 22.687246322631836, "learning_rate": 3.188e-06, "loss": 2.4178, "step": 6376 }, { "epoch": 1.9990595611285267, "grad_norm": 16.281429290771484, "learning_rate": 3.1885000000000005e-06, "loss": 2.5671, "step": 6377 }, { "epoch": 1.9993730407523511, "grad_norm": 12.381708145141602, "learning_rate": 3.1890000000000003e-06, "loss": 2.4764, "step": 6378 }, { "epoch": 1.9996865203761756, "grad_norm": 12.664398193359375, "learning_rate": 3.1895000000000005e-06, "loss": 1.9393, "step": 6379 }, { "epoch": 2.0, "grad_norm": 11.222203254699707, "learning_rate": 3.1900000000000004e-06, "loss": 2.177, "step": 6380 } ], "logging_steps": 1, "max_steps": 12760, "num_input_tokens_seen": 0, "num_train_epochs": 4, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 3.4373825752006656e+16, "train_batch_size": 16, "trial_name": null, "trial_params": null }