{ "best_metric": null, "best_model_checkpoint": null, "epoch": 29.940119760479043, "eval_steps": 500, "global_step": 20000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0014970059880239522, "grad_norm": 12.25, "learning_rate": 0.0008, "loss": 3.8621, "step": 1 }, { "epoch": 0.0029940119760479044, "grad_norm": 147.0, "learning_rate": 0.0008, "loss": 17.8731, "step": 2 }, { "epoch": 0.004491017964071856, "grad_norm": 19.375, "learning_rate": 0.0008, "loss": 3.9524, "step": 3 }, { "epoch": 0.005988023952095809, "grad_norm": 39.25, "learning_rate": 0.0008, "loss": 9.6708, "step": 4 }, { "epoch": 0.0074850299401197605, "grad_norm": 8.5, "learning_rate": 0.0008, "loss": 2.3872, "step": 5 }, { "epoch": 0.008982035928143712, "grad_norm": 25.0, "learning_rate": 0.0008, "loss": 3.4757, "step": 6 }, { "epoch": 0.010479041916167664, "grad_norm": 6.6875, "learning_rate": 0.0008, "loss": 2.1171, "step": 7 }, { "epoch": 0.011976047904191617, "grad_norm": 29.0, "learning_rate": 0.0008, "loss": 3.6517, "step": 8 }, { "epoch": 0.01347305389221557, "grad_norm": 33.5, "learning_rate": 0.0008, "loss": 4.0679, "step": 9 }, { "epoch": 0.014970059880239521, "grad_norm": 15.1875, "learning_rate": 0.0008, "loss": 2.6297, "step": 10 }, { "epoch": 0.016467065868263474, "grad_norm": 5.4375, "learning_rate": 0.0008, "loss": 2.1564, "step": 11 }, { "epoch": 0.017964071856287425, "grad_norm": 6.5, "learning_rate": 0.0008, "loss": 2.2034, "step": 12 }, { "epoch": 0.019461077844311378, "grad_norm": 0.98828125, "learning_rate": 0.0008, "loss": 2.0095, "step": 13 }, { "epoch": 0.020958083832335328, "grad_norm": 5.5, "learning_rate": 0.0008, "loss": 2.2219, "step": 14 }, { "epoch": 0.02245508982035928, "grad_norm": 3.359375, "learning_rate": 0.0008, "loss": 2.1824, "step": 15 }, { "epoch": 0.023952095808383235, "grad_norm": 2.765625, "learning_rate": 0.0008, "loss": 2.0574, "step": 16 }, { "epoch": 0.025449101796407185, "grad_norm": 1.9296875, "learning_rate": 0.0008, "loss": 2.0102, "step": 17 }, { "epoch": 0.02694610778443114, "grad_norm": 2.171875, "learning_rate": 0.0008, "loss": 2.0261, "step": 18 }, { "epoch": 0.02844311377245509, "grad_norm": 1.3671875, "learning_rate": 0.0008, "loss": 1.9611, "step": 19 }, { "epoch": 0.029940119760479042, "grad_norm": 2.078125, "learning_rate": 0.0008, "loss": 1.9503, "step": 20 }, { "epoch": 0.03143712574850299, "grad_norm": 1.4765625, "learning_rate": 0.0008, "loss": 1.9254, "step": 21 }, { "epoch": 0.03293413173652695, "grad_norm": 1.671875, "learning_rate": 0.0008, "loss": 1.9061, "step": 22 }, { "epoch": 0.0344311377245509, "grad_norm": 1.578125, "learning_rate": 0.0008, "loss": 1.875, "step": 23 }, { "epoch": 0.03592814371257485, "grad_norm": 0.6171875, "learning_rate": 0.0008, "loss": 1.8203, "step": 24 }, { "epoch": 0.0374251497005988, "grad_norm": 0.98828125, "learning_rate": 0.0008, "loss": 1.7867, "step": 25 }, { "epoch": 0.038922155688622756, "grad_norm": 1.296875, "learning_rate": 0.0008, "loss": 1.7805, "step": 26 }, { "epoch": 0.040419161676646706, "grad_norm": 1.25, "learning_rate": 0.0008, "loss": 1.7473, "step": 27 }, { "epoch": 0.041916167664670656, "grad_norm": 1.1875, "learning_rate": 0.0008, "loss": 1.7456, "step": 28 }, { "epoch": 0.04341317365269461, "grad_norm": 1.125, "learning_rate": 0.0008, "loss": 1.776, "step": 29 }, { "epoch": 0.04491017964071856, "grad_norm": 0.859375, "learning_rate": 0.0008, "loss": 1.722, "step": 30 }, { "epoch": 0.04640718562874251, "grad_norm": 0.85546875, "learning_rate": 0.0008, "loss": 1.6853, "step": 31 }, { "epoch": 0.04790419161676647, "grad_norm": 0.52734375, "learning_rate": 0.0008, "loss": 1.649, "step": 32 }, { "epoch": 0.04940119760479042, "grad_norm": 0.482421875, "learning_rate": 0.0008, "loss": 1.6499, "step": 33 }, { "epoch": 0.05089820359281437, "grad_norm": 0.41015625, "learning_rate": 0.0008, "loss": 1.6225, "step": 34 }, { "epoch": 0.05239520958083832, "grad_norm": 0.419921875, "learning_rate": 0.0008, "loss": 1.6077, "step": 35 }, { "epoch": 0.05389221556886228, "grad_norm": 0.4296875, "learning_rate": 0.0008, "loss": 1.5821, "step": 36 }, { "epoch": 0.05538922155688623, "grad_norm": 0.39453125, "learning_rate": 0.0008, "loss": 1.5864, "step": 37 }, { "epoch": 0.05688622754491018, "grad_norm": 0.482421875, "learning_rate": 0.0008, "loss": 1.5684, "step": 38 }, { "epoch": 0.058383233532934134, "grad_norm": 0.326171875, "learning_rate": 0.0008, "loss": 1.5676, "step": 39 }, { "epoch": 0.059880239520958084, "grad_norm": 0.27734375, "learning_rate": 0.0008, "loss": 1.554, "step": 40 }, { "epoch": 0.061377245508982034, "grad_norm": 0.279296875, "learning_rate": 0.0008, "loss": 1.5503, "step": 41 }, { "epoch": 0.06287425149700598, "grad_norm": 0.26953125, "learning_rate": 0.0008, "loss": 1.5388, "step": 42 }, { "epoch": 0.06437125748502993, "grad_norm": 0.232421875, "learning_rate": 0.0008, "loss": 1.547, "step": 43 }, { "epoch": 0.0658682634730539, "grad_norm": 0.2578125, "learning_rate": 0.0008, "loss": 1.5204, "step": 44 }, { "epoch": 0.06736526946107785, "grad_norm": 0.279296875, "learning_rate": 0.0008, "loss": 1.511, "step": 45 }, { "epoch": 0.0688622754491018, "grad_norm": 0.423828125, "learning_rate": 0.0008, "loss": 1.5279, "step": 46 }, { "epoch": 0.07035928143712575, "grad_norm": 0.330078125, "learning_rate": 0.0008, "loss": 1.5328, "step": 47 }, { "epoch": 0.0718562874251497, "grad_norm": 0.29296875, "learning_rate": 0.0008, "loss": 1.5189, "step": 48 }, { "epoch": 0.07335329341317365, "grad_norm": 0.2158203125, "learning_rate": 0.0008, "loss": 1.5183, "step": 49 }, { "epoch": 0.0748502994011976, "grad_norm": 0.2470703125, "learning_rate": 0.0008, "loss": 1.5107, "step": 50 }, { "epoch": 0.07634730538922156, "grad_norm": 0.2216796875, "learning_rate": 0.0008, "loss": 1.507, "step": 51 }, { "epoch": 0.07784431137724551, "grad_norm": 0.1884765625, "learning_rate": 0.0008, "loss": 1.483, "step": 52 }, { "epoch": 0.07934131736526946, "grad_norm": 0.2470703125, "learning_rate": 0.0008, "loss": 1.4858, "step": 53 }, { "epoch": 0.08083832335329341, "grad_norm": 0.275390625, "learning_rate": 0.0008, "loss": 1.4907, "step": 54 }, { "epoch": 0.08233532934131736, "grad_norm": 0.32421875, "learning_rate": 0.0008, "loss": 1.4804, "step": 55 }, { "epoch": 0.08383233532934131, "grad_norm": 0.376953125, "learning_rate": 0.0008, "loss": 1.4833, "step": 56 }, { "epoch": 0.08532934131736528, "grad_norm": 0.462890625, "learning_rate": 0.0008, "loss": 1.4743, "step": 57 }, { "epoch": 0.08682634730538923, "grad_norm": 0.51171875, "learning_rate": 0.0008, "loss": 1.4808, "step": 58 }, { "epoch": 0.08832335329341318, "grad_norm": 0.578125, "learning_rate": 0.0008, "loss": 1.4926, "step": 59 }, { "epoch": 0.08982035928143713, "grad_norm": 0.65234375, "learning_rate": 0.0008, "loss": 1.4869, "step": 60 }, { "epoch": 0.09131736526946108, "grad_norm": 0.423828125, "learning_rate": 0.0008, "loss": 1.4629, "step": 61 }, { "epoch": 0.09281437125748503, "grad_norm": 0.29296875, "learning_rate": 0.0008, "loss": 1.4634, "step": 62 }, { "epoch": 0.09431137724550898, "grad_norm": 0.455078125, "learning_rate": 0.0008, "loss": 1.4597, "step": 63 }, { "epoch": 0.09580838323353294, "grad_norm": 0.1767578125, "learning_rate": 0.0008, "loss": 1.4483, "step": 64 }, { "epoch": 0.09730538922155689, "grad_norm": 0.48046875, "learning_rate": 0.0008, "loss": 1.452, "step": 65 }, { "epoch": 0.09880239520958084, "grad_norm": 0.361328125, "learning_rate": 0.0008, "loss": 1.4557, "step": 66 }, { "epoch": 0.10029940119760479, "grad_norm": 0.212890625, "learning_rate": 0.0008, "loss": 1.4346, "step": 67 }, { "epoch": 0.10179640718562874, "grad_norm": 0.447265625, "learning_rate": 0.0008, "loss": 1.4549, "step": 68 }, { "epoch": 0.10329341317365269, "grad_norm": 0.287109375, "learning_rate": 0.0008, "loss": 1.4367, "step": 69 }, { "epoch": 0.10479041916167664, "grad_norm": 0.267578125, "learning_rate": 0.0008, "loss": 1.4298, "step": 70 }, { "epoch": 0.1062874251497006, "grad_norm": 0.453125, "learning_rate": 0.0008, "loss": 1.4434, "step": 71 }, { "epoch": 0.10778443113772455, "grad_norm": 0.23046875, "learning_rate": 0.0008, "loss": 1.4333, "step": 72 }, { "epoch": 0.1092814371257485, "grad_norm": 0.2353515625, "learning_rate": 0.0008, "loss": 1.4331, "step": 73 }, { "epoch": 0.11077844311377245, "grad_norm": 0.1923828125, "learning_rate": 0.0008, "loss": 1.4263, "step": 74 }, { "epoch": 0.1122754491017964, "grad_norm": 0.220703125, "learning_rate": 0.0008, "loss": 1.4307, "step": 75 }, { "epoch": 0.11377245508982035, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.4165, "step": 76 }, { "epoch": 0.11526946107784432, "grad_norm": 0.2041015625, "learning_rate": 0.0008, "loss": 1.4138, "step": 77 }, { "epoch": 0.11676646706586827, "grad_norm": 0.166015625, "learning_rate": 0.0008, "loss": 1.4204, "step": 78 }, { "epoch": 0.11826347305389222, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.4122, "step": 79 }, { "epoch": 0.11976047904191617, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.4143, "step": 80 }, { "epoch": 0.12125748502994012, "grad_norm": 0.18359375, "learning_rate": 0.0008, "loss": 1.4032, "step": 81 }, { "epoch": 0.12275449101796407, "grad_norm": 0.1806640625, "learning_rate": 0.0008, "loss": 1.4069, "step": 82 }, { "epoch": 0.12425149700598802, "grad_norm": 0.1923828125, "learning_rate": 0.0008, "loss": 1.4059, "step": 83 }, { "epoch": 0.12574850299401197, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.4065, "step": 84 }, { "epoch": 0.12724550898203593, "grad_norm": 0.15234375, "learning_rate": 0.0008, "loss": 1.3974, "step": 85 }, { "epoch": 0.12874251497005987, "grad_norm": 0.1728515625, "learning_rate": 0.0008, "loss": 1.3896, "step": 86 }, { "epoch": 0.13023952095808383, "grad_norm": 0.23046875, "learning_rate": 0.0008, "loss": 1.4011, "step": 87 }, { "epoch": 0.1317365269461078, "grad_norm": 0.376953125, "learning_rate": 0.0008, "loss": 1.391, "step": 88 }, { "epoch": 0.13323353293413173, "grad_norm": 0.498046875, "learning_rate": 0.0008, "loss": 1.3908, "step": 89 }, { "epoch": 0.1347305389221557, "grad_norm": 1.3046875, "learning_rate": 0.0008, "loss": 1.4263, "step": 90 }, { "epoch": 0.13622754491017963, "grad_norm": 5.34375, "learning_rate": 0.0008, "loss": 1.5911, "step": 91 }, { "epoch": 0.1377245508982036, "grad_norm": 0.86328125, "learning_rate": 0.0008, "loss": 1.4479, "step": 92 }, { "epoch": 0.13922155688622753, "grad_norm": 0.90625, "learning_rate": 0.0008, "loss": 1.5098, "step": 93 }, { "epoch": 0.1407185628742515, "grad_norm": 0.51953125, "learning_rate": 0.0008, "loss": 1.4637, "step": 94 }, { "epoch": 0.14221556886227546, "grad_norm": 0.52734375, "learning_rate": 0.0008, "loss": 1.4494, "step": 95 }, { "epoch": 0.1437125748502994, "grad_norm": 0.71875, "learning_rate": 0.0008, "loss": 1.4602, "step": 96 }, { "epoch": 0.14520958083832336, "grad_norm": 1.7109375, "learning_rate": 0.0008, "loss": 1.4749, "step": 97 }, { "epoch": 0.1467065868263473, "grad_norm": 0.9765625, "learning_rate": 0.0008, "loss": 1.4621, "step": 98 }, { "epoch": 0.14820359281437126, "grad_norm": 0.466796875, "learning_rate": 0.0008, "loss": 1.4306, "step": 99 }, { "epoch": 0.1497005988023952, "grad_norm": 0.46484375, "learning_rate": 0.0008, "loss": 1.4282, "step": 100 }, { "epoch": 0.15119760479041916, "grad_norm": 0.41015625, "learning_rate": 0.0008, "loss": 1.4251, "step": 101 }, { "epoch": 0.15269461077844312, "grad_norm": 0.30078125, "learning_rate": 0.0008, "loss": 1.4319, "step": 102 }, { "epoch": 0.15419161676646706, "grad_norm": 0.2890625, "learning_rate": 0.0008, "loss": 1.4123, "step": 103 }, { "epoch": 0.15568862275449102, "grad_norm": 0.265625, "learning_rate": 0.0008, "loss": 1.4124, "step": 104 }, { "epoch": 0.15718562874251496, "grad_norm": 0.2890625, "learning_rate": 0.0008, "loss": 1.4222, "step": 105 }, { "epoch": 0.15868263473053892, "grad_norm": 0.25390625, "learning_rate": 0.0008, "loss": 1.4084, "step": 106 }, { "epoch": 0.1601796407185629, "grad_norm": 0.20703125, "learning_rate": 0.0008, "loss": 1.4113, "step": 107 }, { "epoch": 0.16167664670658682, "grad_norm": 0.283203125, "learning_rate": 0.0008, "loss": 1.403, "step": 108 }, { "epoch": 0.1631736526946108, "grad_norm": 0.185546875, "learning_rate": 0.0008, "loss": 1.3906, "step": 109 }, { "epoch": 0.16467065868263472, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.3873, "step": 110 }, { "epoch": 0.1661676646706587, "grad_norm": 0.1962890625, "learning_rate": 0.0008, "loss": 1.3932, "step": 111 }, { "epoch": 0.16766467065868262, "grad_norm": 0.169921875, "learning_rate": 0.0008, "loss": 1.397, "step": 112 }, { "epoch": 0.1691616766467066, "grad_norm": 0.185546875, "learning_rate": 0.0008, "loss": 1.388, "step": 113 }, { "epoch": 0.17065868263473055, "grad_norm": 0.1865234375, "learning_rate": 0.0008, "loss": 1.3821, "step": 114 }, { "epoch": 0.1721556886227545, "grad_norm": 0.212890625, "learning_rate": 0.0008, "loss": 1.3754, "step": 115 }, { "epoch": 0.17365269461077845, "grad_norm": 0.365234375, "learning_rate": 0.0008, "loss": 1.3932, "step": 116 }, { "epoch": 0.1751497005988024, "grad_norm": 0.82421875, "learning_rate": 0.0008, "loss": 1.3986, "step": 117 }, { "epoch": 0.17664670658682635, "grad_norm": 1.1953125, "learning_rate": 0.0008, "loss": 1.4088, "step": 118 }, { "epoch": 0.1781437125748503, "grad_norm": 1.2890625, "learning_rate": 0.0008, "loss": 1.4281, "step": 119 }, { "epoch": 0.17964071856287425, "grad_norm": 0.37890625, "learning_rate": 0.0008, "loss": 1.3933, "step": 120 }, { "epoch": 0.18113772455089822, "grad_norm": 1.015625, "learning_rate": 0.0008, "loss": 1.4178, "step": 121 }, { "epoch": 0.18263473053892215, "grad_norm": 0.578125, "learning_rate": 0.0008, "loss": 1.3884, "step": 122 }, { "epoch": 0.18413173652694612, "grad_norm": 0.4140625, "learning_rate": 0.0008, "loss": 1.3867, "step": 123 }, { "epoch": 0.18562874251497005, "grad_norm": 0.34765625, "learning_rate": 0.0008, "loss": 1.3795, "step": 124 }, { "epoch": 0.18712574850299402, "grad_norm": 0.33203125, "learning_rate": 0.0008, "loss": 1.3872, "step": 125 }, { "epoch": 0.18862275449101795, "grad_norm": 0.267578125, "learning_rate": 0.0008, "loss": 1.3764, "step": 126 }, { "epoch": 0.19011976047904192, "grad_norm": 0.2734375, "learning_rate": 0.0008, "loss": 1.3773, "step": 127 }, { "epoch": 0.19161676646706588, "grad_norm": 0.26171875, "learning_rate": 0.0008, "loss": 1.377, "step": 128 }, { "epoch": 0.19311377245508982, "grad_norm": 0.2001953125, "learning_rate": 0.0008, "loss": 1.3746, "step": 129 }, { "epoch": 0.19461077844311378, "grad_norm": 0.236328125, "learning_rate": 0.0008, "loss": 1.3691, "step": 130 }, { "epoch": 0.19610778443113772, "grad_norm": 0.2236328125, "learning_rate": 0.0008, "loss": 1.373, "step": 131 }, { "epoch": 0.19760479041916168, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.3726, "step": 132 }, { "epoch": 0.19910179640718562, "grad_norm": 0.197265625, "learning_rate": 0.0008, "loss": 1.3653, "step": 133 }, { "epoch": 0.20059880239520958, "grad_norm": 0.173828125, "learning_rate": 0.0008, "loss": 1.3572, "step": 134 }, { "epoch": 0.20209580838323354, "grad_norm": 0.162109375, "learning_rate": 0.0008, "loss": 1.3518, "step": 135 }, { "epoch": 0.20359281437125748, "grad_norm": 0.1708984375, "learning_rate": 0.0008, "loss": 1.3573, "step": 136 }, { "epoch": 0.20508982035928144, "grad_norm": 0.150390625, "learning_rate": 0.0008, "loss": 1.3523, "step": 137 }, { "epoch": 0.20658682634730538, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.3393, "step": 138 }, { "epoch": 0.20808383233532934, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.3482, "step": 139 }, { "epoch": 0.20958083832335328, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.3397, "step": 140 }, { "epoch": 0.21107784431137724, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.3446, "step": 141 }, { "epoch": 0.2125748502994012, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.3432, "step": 142 }, { "epoch": 0.21407185628742514, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.3407, "step": 143 }, { "epoch": 0.2155688622754491, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.334, "step": 144 }, { "epoch": 0.21706586826347304, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.3358, "step": 145 }, { "epoch": 0.218562874251497, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.3459, "step": 146 }, { "epoch": 0.22005988023952097, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.3315, "step": 147 }, { "epoch": 0.2215568862275449, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.3374, "step": 148 }, { "epoch": 0.22305389221556887, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.3344, "step": 149 }, { "epoch": 0.2245508982035928, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.345, "step": 150 }, { "epoch": 0.22604790419161677, "grad_norm": 0.169921875, "learning_rate": 0.0008, "loss": 1.3301, "step": 151 }, { "epoch": 0.2275449101796407, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.331, "step": 152 }, { "epoch": 0.22904191616766467, "grad_norm": 0.1767578125, "learning_rate": 0.0008, "loss": 1.322, "step": 153 }, { "epoch": 0.23053892215568864, "grad_norm": 0.259765625, "learning_rate": 0.0008, "loss": 1.3297, "step": 154 }, { "epoch": 0.23203592814371257, "grad_norm": 0.31640625, "learning_rate": 0.0008, "loss": 1.3389, "step": 155 }, { "epoch": 0.23353293413173654, "grad_norm": 0.19921875, "learning_rate": 0.0008, "loss": 1.3189, "step": 156 }, { "epoch": 0.23502994011976047, "grad_norm": 0.197265625, "learning_rate": 0.0008, "loss": 1.3325, "step": 157 }, { "epoch": 0.23652694610778444, "grad_norm": 0.28515625, "learning_rate": 0.0008, "loss": 1.3364, "step": 158 }, { "epoch": 0.23802395209580837, "grad_norm": 0.318359375, "learning_rate": 0.0008, "loss": 1.3401, "step": 159 }, { "epoch": 0.23952095808383234, "grad_norm": 0.3515625, "learning_rate": 0.0008, "loss": 1.3334, "step": 160 }, { "epoch": 0.2410179640718563, "grad_norm": 0.75390625, "learning_rate": 0.0008, "loss": 1.3468, "step": 161 }, { "epoch": 0.24251497005988024, "grad_norm": 1.921875, "learning_rate": 0.0008, "loss": 1.407, "step": 162 }, { "epoch": 0.2440119760479042, "grad_norm": 0.439453125, "learning_rate": 0.0008, "loss": 1.3459, "step": 163 }, { "epoch": 0.24550898203592814, "grad_norm": 1.03125, "learning_rate": 0.0008, "loss": 1.3732, "step": 164 }, { "epoch": 0.2470059880239521, "grad_norm": 1.1640625, "learning_rate": 0.0008, "loss": 1.3869, "step": 165 }, { "epoch": 0.24850299401197604, "grad_norm": 0.515625, "learning_rate": 0.0008, "loss": 1.3686, "step": 166 }, { "epoch": 0.25, "grad_norm": 0.365234375, "learning_rate": 0.0008, "loss": 1.3486, "step": 167 }, { "epoch": 0.25149700598802394, "grad_norm": 0.431640625, "learning_rate": 0.0008, "loss": 1.357, "step": 168 }, { "epoch": 0.25299401197604793, "grad_norm": 0.310546875, "learning_rate": 0.0008, "loss": 1.3466, "step": 169 }, { "epoch": 0.25449101796407186, "grad_norm": 0.271484375, "learning_rate": 0.0008, "loss": 1.3512, "step": 170 }, { "epoch": 0.2559880239520958, "grad_norm": 0.34765625, "learning_rate": 0.0008, "loss": 1.3607, "step": 171 }, { "epoch": 0.25748502994011974, "grad_norm": 0.19921875, "learning_rate": 0.0008, "loss": 1.3398, "step": 172 }, { "epoch": 0.25898203592814373, "grad_norm": 0.2080078125, "learning_rate": 0.0008, "loss": 1.3372, "step": 173 }, { "epoch": 0.26047904191616766, "grad_norm": 0.25390625, "learning_rate": 0.0008, "loss": 1.3342, "step": 174 }, { "epoch": 0.2619760479041916, "grad_norm": 0.203125, "learning_rate": 0.0008, "loss": 1.3536, "step": 175 }, { "epoch": 0.2634730538922156, "grad_norm": 0.14453125, "learning_rate": 0.0008, "loss": 1.3416, "step": 176 }, { "epoch": 0.26497005988023953, "grad_norm": 0.17578125, "learning_rate": 0.0008, "loss": 1.3383, "step": 177 }, { "epoch": 0.26646706586826346, "grad_norm": 0.1865234375, "learning_rate": 0.0008, "loss": 1.3285, "step": 178 }, { "epoch": 0.2679640718562874, "grad_norm": 0.150390625, "learning_rate": 0.0008, "loss": 1.3252, "step": 179 }, { "epoch": 0.2694610778443114, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.3281, "step": 180 }, { "epoch": 0.27095808383233533, "grad_norm": 0.1943359375, "learning_rate": 0.0008, "loss": 1.3373, "step": 181 }, { "epoch": 0.27245508982035926, "grad_norm": 0.1533203125, "learning_rate": 0.0008, "loss": 1.3247, "step": 182 }, { "epoch": 0.27395209580838326, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.3321, "step": 183 }, { "epoch": 0.2754491017964072, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.3243, "step": 184 }, { "epoch": 0.27694610778443113, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.323, "step": 185 }, { "epoch": 0.27844311377245506, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.3149, "step": 186 }, { "epoch": 0.27994011976047906, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.3184, "step": 187 }, { "epoch": 0.281437125748503, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.3214, "step": 188 }, { "epoch": 0.28293413173652693, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.3092, "step": 189 }, { "epoch": 0.2844311377245509, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.3186, "step": 190 }, { "epoch": 0.28592814371257486, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.3194, "step": 191 }, { "epoch": 0.2874251497005988, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.3081, "step": 192 }, { "epoch": 0.28892215568862273, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.3161, "step": 193 }, { "epoch": 0.2904191616766467, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.3115, "step": 194 }, { "epoch": 0.29191616766467066, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.3152, "step": 195 }, { "epoch": 0.2934131736526946, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.3126, "step": 196 }, { "epoch": 0.2949101796407186, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.3046, "step": 197 }, { "epoch": 0.2964071856287425, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.3053, "step": 198 }, { "epoch": 0.29790419161676646, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.3166, "step": 199 }, { "epoch": 0.2994011976047904, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.3115, "step": 200 }, { "epoch": 0.3008982035928144, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.3097, "step": 201 }, { "epoch": 0.3023952095808383, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.3051, "step": 202 }, { "epoch": 0.30389221556886226, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.3034, "step": 203 }, { "epoch": 0.30538922155688625, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.305, "step": 204 }, { "epoch": 0.3068862275449102, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2913, "step": 205 }, { "epoch": 0.3083832335329341, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.3031, "step": 206 }, { "epoch": 0.30988023952095806, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.3113, "step": 207 }, { "epoch": 0.31137724550898205, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.3012, "step": 208 }, { "epoch": 0.312874251497006, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.2968, "step": 209 }, { "epoch": 0.3143712574850299, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.2994, "step": 210 }, { "epoch": 0.3158682634730539, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.3021, "step": 211 }, { "epoch": 0.31736526946107785, "grad_norm": 0.1806640625, "learning_rate": 0.0008, "loss": 1.3061, "step": 212 }, { "epoch": 0.3188622754491018, "grad_norm": 0.44140625, "learning_rate": 0.0008, "loss": 1.3072, "step": 213 }, { "epoch": 0.3203592814371258, "grad_norm": 0.73046875, "learning_rate": 0.0008, "loss": 1.3489, "step": 214 }, { "epoch": 0.3218562874251497, "grad_norm": 0.47265625, "learning_rate": 0.0008, "loss": 1.3112, "step": 215 }, { "epoch": 0.32335329341317365, "grad_norm": 1.4765625, "learning_rate": 0.0008, "loss": 1.3582, "step": 216 }, { "epoch": 0.3248502994011976, "grad_norm": 1.2890625, "learning_rate": 0.0008, "loss": 1.3626, "step": 217 }, { "epoch": 0.3263473053892216, "grad_norm": 0.328125, "learning_rate": 0.0008, "loss": 1.3273, "step": 218 }, { "epoch": 0.3278443113772455, "grad_norm": 0.77734375, "learning_rate": 0.0008, "loss": 1.3443, "step": 219 }, { "epoch": 0.32934131736526945, "grad_norm": 0.69921875, "learning_rate": 0.0008, "loss": 1.3579, "step": 220 }, { "epoch": 0.33083832335329344, "grad_norm": 0.390625, "learning_rate": 0.0008, "loss": 1.3535, "step": 221 }, { "epoch": 0.3323353293413174, "grad_norm": 0.4140625, "learning_rate": 0.0008, "loss": 1.3344, "step": 222 }, { "epoch": 0.3338323353293413, "grad_norm": 0.341796875, "learning_rate": 0.0008, "loss": 1.3299, "step": 223 }, { "epoch": 0.33532934131736525, "grad_norm": 0.298828125, "learning_rate": 0.0008, "loss": 1.3354, "step": 224 }, { "epoch": 0.33682634730538924, "grad_norm": 0.2216796875, "learning_rate": 0.0008, "loss": 1.3313, "step": 225 }, { "epoch": 0.3383233532934132, "grad_norm": 0.28125, "learning_rate": 0.0008, "loss": 1.3231, "step": 226 }, { "epoch": 0.3398203592814371, "grad_norm": 0.220703125, "learning_rate": 0.0008, "loss": 1.3302, "step": 227 }, { "epoch": 0.3413173652694611, "grad_norm": 0.216796875, "learning_rate": 0.0008, "loss": 1.3175, "step": 228 }, { "epoch": 0.34281437125748504, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.3205, "step": 229 }, { "epoch": 0.344311377245509, "grad_norm": 0.1875, "learning_rate": 0.0008, "loss": 1.3124, "step": 230 }, { "epoch": 0.3458083832335329, "grad_norm": 0.16015625, "learning_rate": 0.0008, "loss": 1.3172, "step": 231 }, { "epoch": 0.3473053892215569, "grad_norm": 0.1591796875, "learning_rate": 0.0008, "loss": 1.3148, "step": 232 }, { "epoch": 0.34880239520958084, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.3087, "step": 233 }, { "epoch": 0.3502994011976048, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.3094, "step": 234 }, { "epoch": 0.35179640718562877, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.3113, "step": 235 }, { "epoch": 0.3532934131736527, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.3054, "step": 236 }, { "epoch": 0.35479041916167664, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.3145, "step": 237 }, { "epoch": 0.3562874251497006, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.304, "step": 238 }, { "epoch": 0.35778443113772457, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.3063, "step": 239 }, { "epoch": 0.3592814371257485, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.3018, "step": 240 }, { "epoch": 0.36077844311377244, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.3051, "step": 241 }, { "epoch": 0.36227544910179643, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2974, "step": 242 }, { "epoch": 0.36377245508982037, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.303, "step": 243 }, { "epoch": 0.3652694610778443, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.3004, "step": 244 }, { "epoch": 0.36676646706586824, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.2908, "step": 245 }, { "epoch": 0.36826347305389223, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.2944, "step": 246 }, { "epoch": 0.36976047904191617, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.3007, "step": 247 }, { "epoch": 0.3712574850299401, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.2976, "step": 248 }, { "epoch": 0.3727544910179641, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.293, "step": 249 }, { "epoch": 0.37425149700598803, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.2994, "step": 250 }, { "epoch": 0.37574850299401197, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.2936, "step": 251 }, { "epoch": 0.3772455089820359, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.2927, "step": 252 }, { "epoch": 0.3787425149700599, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.2853, "step": 253 }, { "epoch": 0.38023952095808383, "grad_norm": 0.15234375, "learning_rate": 0.0008, "loss": 1.285, "step": 254 }, { "epoch": 0.38173652694610777, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.2995, "step": 255 }, { "epoch": 0.38323353293413176, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.2804, "step": 256 }, { "epoch": 0.3847305389221557, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.2919, "step": 257 }, { "epoch": 0.38622754491017963, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.2924, "step": 258 }, { "epoch": 0.38772455089820357, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2885, "step": 259 }, { "epoch": 0.38922155688622756, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.2904, "step": 260 }, { "epoch": 0.3907185628742515, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.2872, "step": 261 }, { "epoch": 0.39221556886227543, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2919, "step": 262 }, { "epoch": 0.3937125748502994, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.2957, "step": 263 }, { "epoch": 0.39520958083832336, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2855, "step": 264 }, { "epoch": 0.3967065868263473, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.2825, "step": 265 }, { "epoch": 0.39820359281437123, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.2848, "step": 266 }, { "epoch": 0.3997005988023952, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.2872, "step": 267 }, { "epoch": 0.40119760479041916, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.2917, "step": 268 }, { "epoch": 0.4026946107784431, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.2817, "step": 269 }, { "epoch": 0.4041916167664671, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.2888, "step": 270 }, { "epoch": 0.405688622754491, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.287, "step": 271 }, { "epoch": 0.40718562874251496, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2871, "step": 272 }, { "epoch": 0.4086826347305389, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.2878, "step": 273 }, { "epoch": 0.4101796407185629, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.2845, "step": 274 }, { "epoch": 0.4116766467065868, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.274, "step": 275 }, { "epoch": 0.41317365269461076, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.2769, "step": 276 }, { "epoch": 0.41467065868263475, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.284, "step": 277 }, { "epoch": 0.4161676646706587, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.2755, "step": 278 }, { "epoch": 0.4176646706586826, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.278, "step": 279 }, { "epoch": 0.41916167664670656, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.2831, "step": 280 }, { "epoch": 0.42065868263473055, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.2768, "step": 281 }, { "epoch": 0.4221556886227545, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.2748, "step": 282 }, { "epoch": 0.4236526946107784, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.2761, "step": 283 }, { "epoch": 0.4251497005988024, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.2679, "step": 284 }, { "epoch": 0.42664670658682635, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.276, "step": 285 }, { "epoch": 0.4281437125748503, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.2812, "step": 286 }, { "epoch": 0.4296407185628742, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.2695, "step": 287 }, { "epoch": 0.4311377245508982, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.2785, "step": 288 }, { "epoch": 0.43263473053892215, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.2744, "step": 289 }, { "epoch": 0.4341317365269461, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.2805, "step": 290 }, { "epoch": 0.4356287425149701, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.2771, "step": 291 }, { "epoch": 0.437125748502994, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.2746, "step": 292 }, { "epoch": 0.43862275449101795, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.284, "step": 293 }, { "epoch": 0.44011976047904194, "grad_norm": 0.1982421875, "learning_rate": 0.0008, "loss": 1.2814, "step": 294 }, { "epoch": 0.4416167664670659, "grad_norm": 0.1982421875, "learning_rate": 0.0008, "loss": 1.2742, "step": 295 }, { "epoch": 0.4431137724550898, "grad_norm": 0.302734375, "learning_rate": 0.0008, "loss": 1.2815, "step": 296 }, { "epoch": 0.44461077844311375, "grad_norm": 0.625, "learning_rate": 0.0008, "loss": 1.3037, "step": 297 }, { "epoch": 0.44610778443113774, "grad_norm": 0.97265625, "learning_rate": 0.0008, "loss": 1.3257, "step": 298 }, { "epoch": 0.4476047904191617, "grad_norm": 1.640625, "learning_rate": 0.0008, "loss": 1.3587, "step": 299 }, { "epoch": 0.4491017964071856, "grad_norm": 0.75390625, "learning_rate": 0.0008, "loss": 1.3257, "step": 300 }, { "epoch": 0.4505988023952096, "grad_norm": 0.455078125, "learning_rate": 0.0008, "loss": 1.3205, "step": 301 }, { "epoch": 0.45209580838323354, "grad_norm": 0.384765625, "learning_rate": 0.0008, "loss": 1.3186, "step": 302 }, { "epoch": 0.4535928143712575, "grad_norm": 0.4375, "learning_rate": 0.0008, "loss": 1.3042, "step": 303 }, { "epoch": 0.4550898203592814, "grad_norm": 0.248046875, "learning_rate": 0.0008, "loss": 1.299, "step": 304 }, { "epoch": 0.4565868263473054, "grad_norm": 0.259765625, "learning_rate": 0.0008, "loss": 1.2962, "step": 305 }, { "epoch": 0.45808383233532934, "grad_norm": 0.25, "learning_rate": 0.0008, "loss": 1.2882, "step": 306 }, { "epoch": 0.4595808383233533, "grad_norm": 0.1806640625, "learning_rate": 0.0008, "loss": 1.2876, "step": 307 }, { "epoch": 0.46107784431137727, "grad_norm": 0.181640625, "learning_rate": 0.0008, "loss": 1.2961, "step": 308 }, { "epoch": 0.4625748502994012, "grad_norm": 0.2216796875, "learning_rate": 0.0008, "loss": 1.295, "step": 309 }, { "epoch": 0.46407185628742514, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.2868, "step": 310 }, { "epoch": 0.4655688622754491, "grad_norm": 0.17578125, "learning_rate": 0.0008, "loss": 1.2954, "step": 311 }, { "epoch": 0.46706586826347307, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.2825, "step": 312 }, { "epoch": 0.468562874251497, "grad_norm": 0.2021484375, "learning_rate": 0.0008, "loss": 1.2827, "step": 313 }, { "epoch": 0.47005988023952094, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.2897, "step": 314 }, { "epoch": 0.47155688622754494, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.2906, "step": 315 }, { "epoch": 0.47305389221556887, "grad_norm": 0.201171875, "learning_rate": 0.0008, "loss": 1.2799, "step": 316 }, { "epoch": 0.4745508982035928, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.2748, "step": 317 }, { "epoch": 0.47604790419161674, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.2831, "step": 318 }, { "epoch": 0.47754491017964074, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.2808, "step": 319 }, { "epoch": 0.47904191616766467, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.2816, "step": 320 }, { "epoch": 0.4805389221556886, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2835, "step": 321 }, { "epoch": 0.4820359281437126, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.2818, "step": 322 }, { "epoch": 0.48353293413173654, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.2763, "step": 323 }, { "epoch": 0.48502994011976047, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.2805, "step": 324 }, { "epoch": 0.4865269461077844, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.2805, "step": 325 }, { "epoch": 0.4880239520958084, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.276, "step": 326 }, { "epoch": 0.48952095808383234, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.2827, "step": 327 }, { "epoch": 0.49101796407185627, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.2658, "step": 328 }, { "epoch": 0.49251497005988026, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.2749, "step": 329 }, { "epoch": 0.4940119760479042, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.2715, "step": 330 }, { "epoch": 0.49550898203592814, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.2661, "step": 331 }, { "epoch": 0.49700598802395207, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.2714, "step": 332 }, { "epoch": 0.49850299401197606, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.2666, "step": 333 }, { "epoch": 0.5, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.2837, "step": 334 }, { "epoch": 0.5014970059880239, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.2692, "step": 335 }, { "epoch": 0.5029940119760479, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.268, "step": 336 }, { "epoch": 0.5044910179640718, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2758, "step": 337 }, { "epoch": 0.5059880239520959, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.2652, "step": 338 }, { "epoch": 0.5074850299401198, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.2729, "step": 339 }, { "epoch": 0.5089820359281437, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.2636, "step": 340 }, { "epoch": 0.5104790419161677, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2692, "step": 341 }, { "epoch": 0.5119760479041916, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.2694, "step": 342 }, { "epoch": 0.5134730538922155, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2742, "step": 343 }, { "epoch": 0.5149700598802395, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.267, "step": 344 }, { "epoch": 0.5164670658682635, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.2677, "step": 345 }, { "epoch": 0.5179640718562875, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.2671, "step": 346 }, { "epoch": 0.5194610778443114, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.2737, "step": 347 }, { "epoch": 0.5209580838323353, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.2797, "step": 348 }, { "epoch": 0.5224550898203593, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.2638, "step": 349 }, { "epoch": 0.5239520958083832, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2663, "step": 350 }, { "epoch": 0.5254491017964071, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.2715, "step": 351 }, { "epoch": 0.5269461077844312, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2662, "step": 352 }, { "epoch": 0.5284431137724551, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.2748, "step": 353 }, { "epoch": 0.5299401197604791, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.2716, "step": 354 }, { "epoch": 0.531437125748503, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.2581, "step": 355 }, { "epoch": 0.5329341317365269, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2679, "step": 356 }, { "epoch": 0.5344311377245509, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.2615, "step": 357 }, { "epoch": 0.5359281437125748, "grad_norm": 0.1796875, "learning_rate": 0.0008, "loss": 1.2642, "step": 358 }, { "epoch": 0.5374251497005988, "grad_norm": 0.166015625, "learning_rate": 0.0008, "loss": 1.267, "step": 359 }, { "epoch": 0.5389221556886228, "grad_norm": 0.1904296875, "learning_rate": 0.0008, "loss": 1.2599, "step": 360 }, { "epoch": 0.5404191616766467, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.2684, "step": 361 }, { "epoch": 0.5419161676646707, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.2617, "step": 362 }, { "epoch": 0.5434131736526946, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.2601, "step": 363 }, { "epoch": 0.5449101796407185, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.2675, "step": 364 }, { "epoch": 0.5464071856287425, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.269, "step": 365 }, { "epoch": 0.5479041916167665, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.264, "step": 366 }, { "epoch": 0.5494011976047904, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.2626, "step": 367 }, { "epoch": 0.5508982035928144, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.2651, "step": 368 }, { "epoch": 0.5523952095808383, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.2659, "step": 369 }, { "epoch": 0.5538922155688623, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.2623, "step": 370 }, { "epoch": 0.5553892215568862, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.2614, "step": 371 }, { "epoch": 0.5568862275449101, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2689, "step": 372 }, { "epoch": 0.5583832335329342, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2632, "step": 373 }, { "epoch": 0.5598802395209581, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.2591, "step": 374 }, { "epoch": 0.561377245508982, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.2601, "step": 375 }, { "epoch": 0.562874251497006, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.2628, "step": 376 }, { "epoch": 0.5643712574850299, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.2577, "step": 377 }, { "epoch": 0.5658682634730539, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.2571, "step": 378 }, { "epoch": 0.5673652694610778, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.2581, "step": 379 }, { "epoch": 0.5688622754491018, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.2596, "step": 380 }, { "epoch": 0.5703592814371258, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.2629, "step": 381 }, { "epoch": 0.5718562874251497, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.2597, "step": 382 }, { "epoch": 0.5733532934131736, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.2626, "step": 383 }, { "epoch": 0.5748502994011976, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.2557, "step": 384 }, { "epoch": 0.5763473053892215, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.2703, "step": 385 }, { "epoch": 0.5778443113772455, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.2595, "step": 386 }, { "epoch": 0.5793413173652695, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.2534, "step": 387 }, { "epoch": 0.5808383233532934, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.2662, "step": 388 }, { "epoch": 0.5823353293413174, "grad_norm": 0.158203125, "learning_rate": 0.0008, "loss": 1.2494, "step": 389 }, { "epoch": 0.5838323353293413, "grad_norm": 0.14453125, "learning_rate": 0.0008, "loss": 1.2616, "step": 390 }, { "epoch": 0.5853293413173652, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.2516, "step": 391 }, { "epoch": 0.5868263473053892, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.2502, "step": 392 }, { "epoch": 0.5883233532934131, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.2538, "step": 393 }, { "epoch": 0.5898203592814372, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.2571, "step": 394 }, { "epoch": 0.5913173652694611, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.2501, "step": 395 }, { "epoch": 0.592814371257485, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.2572, "step": 396 }, { "epoch": 0.594311377245509, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.2573, "step": 397 }, { "epoch": 0.5958083832335329, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.2487, "step": 398 }, { "epoch": 0.5973053892215568, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.2464, "step": 399 }, { "epoch": 0.5988023952095808, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.251, "step": 400 }, { "epoch": 0.6002994011976048, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.2577, "step": 401 }, { "epoch": 0.6017964071856288, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2532, "step": 402 }, { "epoch": 0.6032934131736527, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.2586, "step": 403 }, { "epoch": 0.6047904191616766, "grad_norm": 0.216796875, "learning_rate": 0.0008, "loss": 1.2574, "step": 404 }, { "epoch": 0.6062874251497006, "grad_norm": 0.365234375, "learning_rate": 0.0008, "loss": 1.2516, "step": 405 }, { "epoch": 0.6077844311377245, "grad_norm": 1.171875, "learning_rate": 0.0008, "loss": 1.2707, "step": 406 }, { "epoch": 0.6092814371257484, "grad_norm": 1.890625, "learning_rate": 0.0008, "loss": 1.3835, "step": 407 }, { "epoch": 0.6107784431137725, "grad_norm": 0.81640625, "learning_rate": 0.0008, "loss": 1.3279, "step": 408 }, { "epoch": 0.6122754491017964, "grad_norm": 2.703125, "learning_rate": 0.0008, "loss": 1.4153, "step": 409 }, { "epoch": 0.6137724550898204, "grad_norm": 0.76953125, "learning_rate": 0.0008, "loss": 1.3757, "step": 410 }, { "epoch": 0.6152694610778443, "grad_norm": 1.2265625, "learning_rate": 0.0008, "loss": 1.4161, "step": 411 }, { "epoch": 0.6167664670658682, "grad_norm": 0.96875, "learning_rate": 0.0008, "loss": 1.3939, "step": 412 }, { "epoch": 0.6182634730538922, "grad_norm": 0.64453125, "learning_rate": 0.0008, "loss": 1.3756, "step": 413 }, { "epoch": 0.6197604790419161, "grad_norm": 0.515625, "learning_rate": 0.0008, "loss": 1.373, "step": 414 }, { "epoch": 0.6212574850299402, "grad_norm": 0.423828125, "learning_rate": 0.0008, "loss": 1.3487, "step": 415 }, { "epoch": 0.6227544910179641, "grad_norm": 0.373046875, "learning_rate": 0.0008, "loss": 1.3556, "step": 416 }, { "epoch": 0.624251497005988, "grad_norm": 0.3203125, "learning_rate": 0.0008, "loss": 1.3364, "step": 417 }, { "epoch": 0.625748502994012, "grad_norm": 0.3125, "learning_rate": 0.0008, "loss": 1.3407, "step": 418 }, { "epoch": 0.6272455089820359, "grad_norm": 0.26171875, "learning_rate": 0.0008, "loss": 1.3387, "step": 419 }, { "epoch": 0.6287425149700598, "grad_norm": 0.2216796875, "learning_rate": 0.0008, "loss": 1.3285, "step": 420 }, { "epoch": 0.6302395209580839, "grad_norm": 0.2451171875, "learning_rate": 0.0008, "loss": 1.3225, "step": 421 }, { "epoch": 0.6317365269461078, "grad_norm": 0.2099609375, "learning_rate": 0.0008, "loss": 1.3073, "step": 422 }, { "epoch": 0.6332335329341318, "grad_norm": 0.2041015625, "learning_rate": 0.0008, "loss": 1.3177, "step": 423 }, { "epoch": 0.6347305389221557, "grad_norm": 0.2216796875, "learning_rate": 0.0008, "loss": 1.3074, "step": 424 }, { "epoch": 0.6362275449101796, "grad_norm": 0.21875, "learning_rate": 0.0008, "loss": 1.2996, "step": 425 }, { "epoch": 0.6377245508982036, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.3035, "step": 426 }, { "epoch": 0.6392215568862275, "grad_norm": 0.2197265625, "learning_rate": 0.0008, "loss": 1.3063, "step": 427 }, { "epoch": 0.6407185628742516, "grad_norm": 0.1708984375, "learning_rate": 0.0008, "loss": 1.3045, "step": 428 }, { "epoch": 0.6422155688622755, "grad_norm": 0.1572265625, "learning_rate": 0.0008, "loss": 1.2989, "step": 429 }, { "epoch": 0.6437125748502994, "grad_norm": 0.2041015625, "learning_rate": 0.0008, "loss": 1.293, "step": 430 }, { "epoch": 0.6452095808383234, "grad_norm": 0.216796875, "learning_rate": 0.0008, "loss": 1.2885, "step": 431 }, { "epoch": 0.6467065868263473, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.2915, "step": 432 }, { "epoch": 0.6482035928143712, "grad_norm": 0.15625, "learning_rate": 0.0008, "loss": 1.2892, "step": 433 }, { "epoch": 0.6497005988023952, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.2859, "step": 434 }, { "epoch": 0.6511976047904192, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.2829, "step": 435 }, { "epoch": 0.6526946107784432, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.2888, "step": 436 }, { "epoch": 0.6541916167664671, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.278, "step": 437 }, { "epoch": 0.655688622754491, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.2785, "step": 438 }, { "epoch": 0.657185628742515, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.2866, "step": 439 }, { "epoch": 0.6586826347305389, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.2802, "step": 440 }, { "epoch": 0.6601796407185628, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.2791, "step": 441 }, { "epoch": 0.6616766467065869, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.2747, "step": 442 }, { "epoch": 0.6631736526946108, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.2801, "step": 443 }, { "epoch": 0.6646706586826348, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.2771, "step": 444 }, { "epoch": 0.6661676646706587, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.2773, "step": 445 }, { "epoch": 0.6676646706586826, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2719, "step": 446 }, { "epoch": 0.6691616766467066, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.2724, "step": 447 }, { "epoch": 0.6706586826347305, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.2643, "step": 448 }, { "epoch": 0.6721556886227545, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.2621, "step": 449 }, { "epoch": 0.6736526946107785, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.272, "step": 450 }, { "epoch": 0.6751497005988024, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.2579, "step": 451 }, { "epoch": 0.6766467065868264, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.2697, "step": 452 }, { "epoch": 0.6781437125748503, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.2567, "step": 453 }, { "epoch": 0.6796407185628742, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.2649, "step": 454 }, { "epoch": 0.6811377245508982, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2707, "step": 455 }, { "epoch": 0.6826347305389222, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.2685, "step": 456 }, { "epoch": 0.6841317365269461, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2598, "step": 457 }, { "epoch": 0.6856287425149701, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.2675, "step": 458 }, { "epoch": 0.687125748502994, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2612, "step": 459 }, { "epoch": 0.688622754491018, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.2551, "step": 460 }, { "epoch": 0.6901197604790419, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2577, "step": 461 }, { "epoch": 0.6916167664670658, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.2606, "step": 462 }, { "epoch": 0.6931137724550899, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2602, "step": 463 }, { "epoch": 0.6946107784431138, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.2544, "step": 464 }, { "epoch": 0.6961077844311377, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.2547, "step": 465 }, { "epoch": 0.6976047904191617, "grad_norm": 0.337890625, "learning_rate": 0.0008, "loss": 1.2575, "step": 466 }, { "epoch": 0.6991017964071856, "grad_norm": 0.71484375, "learning_rate": 0.0008, "loss": 1.2726, "step": 467 }, { "epoch": 0.7005988023952096, "grad_norm": 1.34375, "learning_rate": 0.0008, "loss": 1.3157, "step": 468 }, { "epoch": 0.7020958083832335, "grad_norm": 0.609375, "learning_rate": 0.0008, "loss": 1.2837, "step": 469 }, { "epoch": 0.7035928143712575, "grad_norm": 0.640625, "learning_rate": 0.0008, "loss": 1.3059, "step": 470 }, { "epoch": 0.7050898203592815, "grad_norm": 0.52734375, "learning_rate": 0.0008, "loss": 1.3061, "step": 471 }, { "epoch": 0.7065868263473054, "grad_norm": 0.4921875, "learning_rate": 0.0008, "loss": 1.2931, "step": 472 }, { "epoch": 0.7080838323353293, "grad_norm": 0.48046875, "learning_rate": 0.0008, "loss": 1.2869, "step": 473 }, { "epoch": 0.7095808383233533, "grad_norm": 0.353515625, "learning_rate": 0.0008, "loss": 1.2761, "step": 474 }, { "epoch": 0.7110778443113772, "grad_norm": 0.333984375, "learning_rate": 0.0008, "loss": 1.2764, "step": 475 }, { "epoch": 0.7125748502994012, "grad_norm": 0.294921875, "learning_rate": 0.0008, "loss": 1.2857, "step": 476 }, { "epoch": 0.7140718562874252, "grad_norm": 0.2431640625, "learning_rate": 0.0008, "loss": 1.2777, "step": 477 }, { "epoch": 0.7155688622754491, "grad_norm": 0.1982421875, "learning_rate": 0.0008, "loss": 1.2747, "step": 478 }, { "epoch": 0.7170658682634731, "grad_norm": 0.2255859375, "learning_rate": 0.0008, "loss": 1.2732, "step": 479 }, { "epoch": 0.718562874251497, "grad_norm": 0.1826171875, "learning_rate": 0.0008, "loss": 1.2695, "step": 480 }, { "epoch": 0.7200598802395209, "grad_norm": 0.193359375, "learning_rate": 0.0008, "loss": 1.2706, "step": 481 }, { "epoch": 0.7215568862275449, "grad_norm": 0.1826171875, "learning_rate": 0.0008, "loss": 1.2783, "step": 482 }, { "epoch": 0.7230538922155688, "grad_norm": 0.1728515625, "learning_rate": 0.0008, "loss": 1.2682, "step": 483 }, { "epoch": 0.7245508982035929, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.2582, "step": 484 }, { "epoch": 0.7260479041916168, "grad_norm": 0.1484375, "learning_rate": 0.0008, "loss": 1.275, "step": 485 }, { "epoch": 0.7275449101796407, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.2671, "step": 486 }, { "epoch": 0.7290419161676647, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.2655, "step": 487 }, { "epoch": 0.7305389221556886, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.2631, "step": 488 }, { "epoch": 0.7320359281437125, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.2566, "step": 489 }, { "epoch": 0.7335329341317365, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.2566, "step": 490 }, { "epoch": 0.7350299401197605, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.2554, "step": 491 }, { "epoch": 0.7365269461077845, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.2556, "step": 492 }, { "epoch": 0.7380239520958084, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.2574, "step": 493 }, { "epoch": 0.7395209580838323, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.257, "step": 494 }, { "epoch": 0.7410179640718563, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2622, "step": 495 }, { "epoch": 0.7425149700598802, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.2635, "step": 496 }, { "epoch": 0.7440119760479041, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2536, "step": 497 }, { "epoch": 0.7455089820359282, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.2551, "step": 498 }, { "epoch": 0.7470059880239521, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.2509, "step": 499 }, { "epoch": 0.7485029940119761, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.2617, "step": 500 }, { "epoch": 0.75, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.2542, "step": 501 }, { "epoch": 0.7514970059880239, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.254, "step": 502 }, { "epoch": 0.7529940119760479, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.2516, "step": 503 }, { "epoch": 0.7544910179640718, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.2534, "step": 504 }, { "epoch": 0.7559880239520959, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.2536, "step": 505 }, { "epoch": 0.7574850299401198, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.2457, "step": 506 }, { "epoch": 0.7589820359281437, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.2576, "step": 507 }, { "epoch": 0.7604790419161677, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.2521, "step": 508 }, { "epoch": 0.7619760479041916, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.2629, "step": 509 }, { "epoch": 0.7634730538922155, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.2444, "step": 510 }, { "epoch": 0.7649700598802395, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.2592, "step": 511 }, { "epoch": 0.7664670658682635, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2509, "step": 512 }, { "epoch": 0.7679640718562875, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.2494, "step": 513 }, { "epoch": 0.7694610778443114, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2442, "step": 514 }, { "epoch": 0.7709580838323353, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.2492, "step": 515 }, { "epoch": 0.7724550898203593, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.2487, "step": 516 }, { "epoch": 0.7739520958083832, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.2504, "step": 517 }, { "epoch": 0.7754491017964071, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.2428, "step": 518 }, { "epoch": 0.7769461077844312, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.2555, "step": 519 }, { "epoch": 0.7784431137724551, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.2431, "step": 520 }, { "epoch": 0.7799401197604791, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.2415, "step": 521 }, { "epoch": 0.781437125748503, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.2507, "step": 522 }, { "epoch": 0.7829341317365269, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.2531, "step": 523 }, { "epoch": 0.7844311377245509, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2447, "step": 524 }, { "epoch": 0.7859281437125748, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.2391, "step": 525 }, { "epoch": 0.7874251497005988, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2424, "step": 526 }, { "epoch": 0.7889221556886228, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.2469, "step": 527 }, { "epoch": 0.7904191616766467, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.2487, "step": 528 }, { "epoch": 0.7919161676646707, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.2405, "step": 529 }, { "epoch": 0.7934131736526946, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.2493, "step": 530 }, { "epoch": 0.7949101796407185, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.2407, "step": 531 }, { "epoch": 0.7964071856287425, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.2494, "step": 532 }, { "epoch": 0.7979041916167665, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.2468, "step": 533 }, { "epoch": 0.7994011976047904, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.2456, "step": 534 }, { "epoch": 0.8008982035928144, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2483, "step": 535 }, { "epoch": 0.8023952095808383, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.2478, "step": 536 }, { "epoch": 0.8038922155688623, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2463, "step": 537 }, { "epoch": 0.8053892215568862, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.2434, "step": 538 }, { "epoch": 0.8068862275449101, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.2472, "step": 539 }, { "epoch": 0.8083832335329342, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.2514, "step": 540 }, { "epoch": 0.8098802395209581, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.2483, "step": 541 }, { "epoch": 0.811377245508982, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.2396, "step": 542 }, { "epoch": 0.812874251497006, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2467, "step": 543 }, { "epoch": 0.8143712574850299, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.2467, "step": 544 }, { "epoch": 0.8158682634730539, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.2465, "step": 545 }, { "epoch": 0.8173652694610778, "grad_norm": 0.2099609375, "learning_rate": 0.0008, "loss": 1.2495, "step": 546 }, { "epoch": 0.8188622754491018, "grad_norm": 0.1962890625, "learning_rate": 0.0008, "loss": 1.2516, "step": 547 }, { "epoch": 0.8203592814371258, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.2468, "step": 548 }, { "epoch": 0.8218562874251497, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.2523, "step": 549 }, { "epoch": 0.8233532934131736, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.2377, "step": 550 }, { "epoch": 0.8248502994011976, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.2487, "step": 551 }, { "epoch": 0.8263473053892215, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.2451, "step": 552 }, { "epoch": 0.8278443113772455, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.2477, "step": 553 }, { "epoch": 0.8293413173652695, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.2495, "step": 554 }, { "epoch": 0.8308383233532934, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.2438, "step": 555 }, { "epoch": 0.8323353293413174, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.2469, "step": 556 }, { "epoch": 0.8338323353293413, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.2444, "step": 557 }, { "epoch": 0.8353293413173652, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.2592, "step": 558 }, { "epoch": 0.8368263473053892, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.2508, "step": 559 }, { "epoch": 0.8383233532934131, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.2474, "step": 560 }, { "epoch": 0.8398203592814372, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.2452, "step": 561 }, { "epoch": 0.8413173652694611, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.2421, "step": 562 }, { "epoch": 0.842814371257485, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.2452, "step": 563 }, { "epoch": 0.844311377245509, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.2469, "step": 564 }, { "epoch": 0.8458083832335329, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.2491, "step": 565 }, { "epoch": 0.8473053892215568, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.2406, "step": 566 }, { "epoch": 0.8488023952095808, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.2382, "step": 567 }, { "epoch": 0.8502994011976048, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2385, "step": 568 }, { "epoch": 0.8517964071856288, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.2477, "step": 569 }, { "epoch": 0.8532934131736527, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.2428, "step": 570 }, { "epoch": 0.8547904191616766, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2465, "step": 571 }, { "epoch": 0.8562874251497006, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.2548, "step": 572 }, { "epoch": 0.8577844311377245, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.2468, "step": 573 }, { "epoch": 0.8592814371257484, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.2416, "step": 574 }, { "epoch": 0.8607784431137725, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.2437, "step": 575 }, { "epoch": 0.8622754491017964, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.2291, "step": 576 }, { "epoch": 0.8637724550898204, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.2391, "step": 577 }, { "epoch": 0.8652694610778443, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.2451, "step": 578 }, { "epoch": 0.8667664670658682, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.2325, "step": 579 }, { "epoch": 0.8682634730538922, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.2394, "step": 580 }, { "epoch": 0.8697604790419161, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2332, "step": 581 }, { "epoch": 0.8712574850299402, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.239, "step": 582 }, { "epoch": 0.8727544910179641, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.2363, "step": 583 }, { "epoch": 0.874251497005988, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.2463, "step": 584 }, { "epoch": 0.875748502994012, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.2459, "step": 585 }, { "epoch": 0.8772455089820359, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.2403, "step": 586 }, { "epoch": 0.8787425149700598, "grad_norm": 0.14453125, "learning_rate": 0.0008, "loss": 1.249, "step": 587 }, { "epoch": 0.8802395209580839, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.2359, "step": 588 }, { "epoch": 0.8817365269461078, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.2448, "step": 589 }, { "epoch": 0.8832335329341318, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.2473, "step": 590 }, { "epoch": 0.8847305389221557, "grad_norm": 0.1533203125, "learning_rate": 0.0008, "loss": 1.2437, "step": 591 }, { "epoch": 0.8862275449101796, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.243, "step": 592 }, { "epoch": 0.8877245508982036, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.2377, "step": 593 }, { "epoch": 0.8892215568862275, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.2365, "step": 594 }, { "epoch": 0.8907185628742516, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.2386, "step": 595 }, { "epoch": 0.8922155688622755, "grad_norm": 0.158203125, "learning_rate": 0.0008, "loss": 1.2347, "step": 596 }, { "epoch": 0.8937125748502994, "grad_norm": 0.16015625, "learning_rate": 0.0008, "loss": 1.2459, "step": 597 }, { "epoch": 0.8952095808383234, "grad_norm": 0.1845703125, "learning_rate": 0.0008, "loss": 1.2417, "step": 598 }, { "epoch": 0.8967065868263473, "grad_norm": 0.197265625, "learning_rate": 0.0008, "loss": 1.2378, "step": 599 }, { "epoch": 0.8982035928143712, "grad_norm": 0.1572265625, "learning_rate": 0.0008, "loss": 1.2402, "step": 600 }, { "epoch": 0.8997005988023952, "grad_norm": 0.177734375, "learning_rate": 0.0008, "loss": 1.2324, "step": 601 }, { "epoch": 0.9011976047904192, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.2395, "step": 602 }, { "epoch": 0.9026946107784432, "grad_norm": 0.150390625, "learning_rate": 0.0008, "loss": 1.2374, "step": 603 }, { "epoch": 0.9041916167664671, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.2418, "step": 604 }, { "epoch": 0.905688622754491, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.2431, "step": 605 }, { "epoch": 0.907185628742515, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.2416, "step": 606 }, { "epoch": 0.9086826347305389, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.2369, "step": 607 }, { "epoch": 0.9101796407185628, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.2399, "step": 608 }, { "epoch": 0.9116766467065869, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.2445, "step": 609 }, { "epoch": 0.9131736526946108, "grad_norm": 0.1845703125, "learning_rate": 0.0008, "loss": 1.2343, "step": 610 }, { "epoch": 0.9146706586826348, "grad_norm": 0.2041015625, "learning_rate": 0.0008, "loss": 1.2317, "step": 611 }, { "epoch": 0.9161676646706587, "grad_norm": 0.3203125, "learning_rate": 0.0008, "loss": 1.2381, "step": 612 }, { "epoch": 0.9176646706586826, "grad_norm": 0.7421875, "learning_rate": 0.0008, "loss": 1.2429, "step": 613 }, { "epoch": 0.9191616766467066, "grad_norm": 1.625, "learning_rate": 0.0008, "loss": 1.3109, "step": 614 }, { "epoch": 0.9206586826347305, "grad_norm": 0.8125, "learning_rate": 0.0008, "loss": 1.2647, "step": 615 }, { "epoch": 0.9221556886227545, "grad_norm": 0.71484375, "learning_rate": 0.0008, "loss": 1.2848, "step": 616 }, { "epoch": 0.9236526946107785, "grad_norm": 0.484375, "learning_rate": 0.0008, "loss": 1.2868, "step": 617 }, { "epoch": 0.9251497005988024, "grad_norm": 0.353515625, "learning_rate": 0.0008, "loss": 1.2641, "step": 618 }, { "epoch": 0.9266467065868264, "grad_norm": 0.298828125, "learning_rate": 0.0008, "loss": 1.261, "step": 619 }, { "epoch": 0.9281437125748503, "grad_norm": 0.349609375, "learning_rate": 0.0008, "loss": 1.2685, "step": 620 }, { "epoch": 0.9296407185628742, "grad_norm": 0.2431640625, "learning_rate": 0.0008, "loss": 1.2581, "step": 621 }, { "epoch": 0.9311377245508982, "grad_norm": 0.267578125, "learning_rate": 0.0008, "loss": 1.2602, "step": 622 }, { "epoch": 0.9326347305389222, "grad_norm": 0.2333984375, "learning_rate": 0.0008, "loss": 1.2504, "step": 623 }, { "epoch": 0.9341317365269461, "grad_norm": 0.185546875, "learning_rate": 0.0008, "loss": 1.2536, "step": 624 }, { "epoch": 0.9356287425149701, "grad_norm": 0.21875, "learning_rate": 0.0008, "loss": 1.2571, "step": 625 }, { "epoch": 0.937125748502994, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.2459, "step": 626 }, { "epoch": 0.938622754491018, "grad_norm": 0.169921875, "learning_rate": 0.0008, "loss": 1.2506, "step": 627 }, { "epoch": 0.9401197604790419, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.2524, "step": 628 }, { "epoch": 0.9416167664670658, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.2505, "step": 629 }, { "epoch": 0.9431137724550899, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.2487, "step": 630 }, { "epoch": 0.9446107784431138, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.2488, "step": 631 }, { "epoch": 0.9461077844311377, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.2459, "step": 632 }, { "epoch": 0.9476047904191617, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.2413, "step": 633 }, { "epoch": 0.9491017964071856, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.2407, "step": 634 }, { "epoch": 0.9505988023952096, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.2363, "step": 635 }, { "epoch": 0.9520958083832335, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.2352, "step": 636 }, { "epoch": 0.9535928143712575, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.2365, "step": 637 }, { "epoch": 0.9550898203592815, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.2447, "step": 638 }, { "epoch": 0.9565868263473054, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.245, "step": 639 }, { "epoch": 0.9580838323353293, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2441, "step": 640 }, { "epoch": 0.9595808383233533, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.2406, "step": 641 }, { "epoch": 0.9610778443113772, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2385, "step": 642 }, { "epoch": 0.9625748502994012, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.2329, "step": 643 }, { "epoch": 0.9640718562874252, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.2412, "step": 644 }, { "epoch": 0.9655688622754491, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.2366, "step": 645 }, { "epoch": 0.9670658682634731, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.2435, "step": 646 }, { "epoch": 0.968562874251497, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.2438, "step": 647 }, { "epoch": 0.9700598802395209, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.2343, "step": 648 }, { "epoch": 0.9715568862275449, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.2347, "step": 649 }, { "epoch": 0.9730538922155688, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.2401, "step": 650 }, { "epoch": 0.9745508982035929, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.235, "step": 651 }, { "epoch": 0.9760479041916168, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.2344, "step": 652 }, { "epoch": 0.9775449101796407, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.2315, "step": 653 }, { "epoch": 0.9790419161676647, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2327, "step": 654 }, { "epoch": 0.9805389221556886, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.236, "step": 655 }, { "epoch": 0.9820359281437125, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.2296, "step": 656 }, { "epoch": 0.9835329341317365, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.2338, "step": 657 }, { "epoch": 0.9850299401197605, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.2435, "step": 658 }, { "epoch": 0.9865269461077845, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.2294, "step": 659 }, { "epoch": 0.9880239520958084, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.232, "step": 660 }, { "epoch": 0.9895209580838323, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.2238, "step": 661 }, { "epoch": 0.9910179640718563, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.2328, "step": 662 }, { "epoch": 0.9925149700598802, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.2334, "step": 663 }, { "epoch": 0.9940119760479041, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2371, "step": 664 }, { "epoch": 0.9955089820359282, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.2422, "step": 665 }, { "epoch": 0.9970059880239521, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.2314, "step": 666 }, { "epoch": 0.9985029940119761, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.2306, "step": 667 }, { "epoch": 1.0, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.2275, "step": 668 }, { "epoch": 1.001497005988024, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.2372, "step": 669 }, { "epoch": 1.0029940119760479, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2366, "step": 670 }, { "epoch": 1.0044910179640718, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.2293, "step": 671 }, { "epoch": 1.0059880239520957, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.2365, "step": 672 }, { "epoch": 1.0074850299401197, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2288, "step": 673 }, { "epoch": 1.0089820359281436, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.2309, "step": 674 }, { "epoch": 1.0104790419161676, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.2281, "step": 675 }, { "epoch": 1.0119760479041917, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2286, "step": 676 }, { "epoch": 1.0134730538922156, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.2302, "step": 677 }, { "epoch": 1.0149700598802396, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.2238, "step": 678 }, { "epoch": 1.0164670658682635, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.2237, "step": 679 }, { "epoch": 1.0179640718562875, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.222, "step": 680 }, { "epoch": 1.0194610778443114, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.2323, "step": 681 }, { "epoch": 1.0209580838323353, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.2246, "step": 682 }, { "epoch": 1.0224550898203593, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.2249, "step": 683 }, { "epoch": 1.0239520958083832, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.2348, "step": 684 }, { "epoch": 1.0254491017964071, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2309, "step": 685 }, { "epoch": 1.026946107784431, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.2295, "step": 686 }, { "epoch": 1.028443113772455, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2291, "step": 687 }, { "epoch": 1.029940119760479, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.2245, "step": 688 }, { "epoch": 1.031437125748503, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.2247, "step": 689 }, { "epoch": 1.032934131736527, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.2322, "step": 690 }, { "epoch": 1.034431137724551, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.2273, "step": 691 }, { "epoch": 1.035928143712575, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.223, "step": 692 }, { "epoch": 1.0374251497005988, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.2276, "step": 693 }, { "epoch": 1.0389221556886228, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.221, "step": 694 }, { "epoch": 1.0404191616766467, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.2212, "step": 695 }, { "epoch": 1.0419161676646707, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.222, "step": 696 }, { "epoch": 1.0434131736526946, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.2294, "step": 697 }, { "epoch": 1.0449101796407185, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.2231, "step": 698 }, { "epoch": 1.0464071856287425, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.2274, "step": 699 }, { "epoch": 1.0479041916167664, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.2232, "step": 700 }, { "epoch": 1.0494011976047903, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2189, "step": 701 }, { "epoch": 1.0508982035928143, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.2336, "step": 702 }, { "epoch": 1.0523952095808382, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.23, "step": 703 }, { "epoch": 1.0538922155688624, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.2317, "step": 704 }, { "epoch": 1.0553892215568863, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2261, "step": 705 }, { "epoch": 1.0568862275449102, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.2241, "step": 706 }, { "epoch": 1.0583832335329342, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.2223, "step": 707 }, { "epoch": 1.0598802395209581, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.2178, "step": 708 }, { "epoch": 1.061377245508982, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.2348, "step": 709 }, { "epoch": 1.062874251497006, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.2284, "step": 710 }, { "epoch": 1.06437125748503, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.2272, "step": 711 }, { "epoch": 1.0658682634730539, "grad_norm": 0.1728515625, "learning_rate": 0.0008, "loss": 1.2345, "step": 712 }, { "epoch": 1.0673652694610778, "grad_norm": 0.1953125, "learning_rate": 0.0008, "loss": 1.2282, "step": 713 }, { "epoch": 1.0688622754491017, "grad_norm": 0.1748046875, "learning_rate": 0.0008, "loss": 1.2278, "step": 714 }, { "epoch": 1.0703592814371257, "grad_norm": 0.1953125, "learning_rate": 0.0008, "loss": 1.2277, "step": 715 }, { "epoch": 1.0718562874251496, "grad_norm": 0.189453125, "learning_rate": 0.0008, "loss": 1.2264, "step": 716 }, { "epoch": 1.0733532934131738, "grad_norm": 0.267578125, "learning_rate": 0.0008, "loss": 1.2238, "step": 717 }, { "epoch": 1.0748502994011977, "grad_norm": 0.3671875, "learning_rate": 0.0008, "loss": 1.218, "step": 718 }, { "epoch": 1.0763473053892216, "grad_norm": 0.828125, "learning_rate": 0.0008, "loss": 1.2397, "step": 719 }, { "epoch": 1.0778443113772456, "grad_norm": 1.8203125, "learning_rate": 0.0008, "loss": 1.3052, "step": 720 }, { "epoch": 1.0793413173652695, "grad_norm": 0.2470703125, "learning_rate": 0.0008, "loss": 1.2408, "step": 721 }, { "epoch": 1.0808383233532934, "grad_norm": 1.1875, "learning_rate": 0.0008, "loss": 1.2891, "step": 722 }, { "epoch": 1.0823353293413174, "grad_norm": 0.53125, "learning_rate": 0.0008, "loss": 1.2595, "step": 723 }, { "epoch": 1.0838323353293413, "grad_norm": 0.5, "learning_rate": 0.0008, "loss": 1.2605, "step": 724 }, { "epoch": 1.0853293413173652, "grad_norm": 0.294921875, "learning_rate": 0.0008, "loss": 1.2537, "step": 725 }, { "epoch": 1.0868263473053892, "grad_norm": 0.287109375, "learning_rate": 0.0008, "loss": 1.2526, "step": 726 }, { "epoch": 1.0883233532934131, "grad_norm": 0.2373046875, "learning_rate": 0.0008, "loss": 1.2418, "step": 727 }, { "epoch": 1.089820359281437, "grad_norm": 0.177734375, "learning_rate": 0.0008, "loss": 1.2439, "step": 728 }, { "epoch": 1.091317365269461, "grad_norm": 0.2109375, "learning_rate": 0.0008, "loss": 1.2389, "step": 729 }, { "epoch": 1.092814371257485, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.2388, "step": 730 }, { "epoch": 1.0943113772455089, "grad_norm": 0.1845703125, "learning_rate": 0.0008, "loss": 1.2394, "step": 731 }, { "epoch": 1.095808383233533, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.2443, "step": 732 }, { "epoch": 1.097305389221557, "grad_norm": 0.15234375, "learning_rate": 0.0008, "loss": 1.2308, "step": 733 }, { "epoch": 1.098802395209581, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.23, "step": 734 }, { "epoch": 1.1002994011976048, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.2342, "step": 735 }, { "epoch": 1.1017964071856288, "grad_norm": 0.1689453125, "learning_rate": 0.0008, "loss": 1.2372, "step": 736 }, { "epoch": 1.1032934131736527, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.2339, "step": 737 }, { "epoch": 1.1047904191616766, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.2252, "step": 738 }, { "epoch": 1.1062874251497006, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.2298, "step": 739 }, { "epoch": 1.1077844311377245, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.2372, "step": 740 }, { "epoch": 1.1092814371257484, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.2223, "step": 741 }, { "epoch": 1.1107784431137724, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.2322, "step": 742 }, { "epoch": 1.1122754491017963, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.2261, "step": 743 }, { "epoch": 1.1137724550898203, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.2329, "step": 744 }, { "epoch": 1.1152694610778444, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.2268, "step": 745 }, { "epoch": 1.1167664670658684, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.2296, "step": 746 }, { "epoch": 1.1182634730538923, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.2252, "step": 747 }, { "epoch": 1.1197604790419162, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2342, "step": 748 }, { "epoch": 1.1212574850299402, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.2307, "step": 749 }, { "epoch": 1.122754491017964, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.224, "step": 750 }, { "epoch": 1.124251497005988, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.2255, "step": 751 }, { "epoch": 1.125748502994012, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.2305, "step": 752 }, { "epoch": 1.127245508982036, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.228, "step": 753 }, { "epoch": 1.1287425149700598, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.2216, "step": 754 }, { "epoch": 1.1302395209580838, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.2284, "step": 755 }, { "epoch": 1.1317365269461077, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.2271, "step": 756 }, { "epoch": 1.1332335329341316, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.2271, "step": 757 }, { "epoch": 1.1347305389221556, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.2278, "step": 758 }, { "epoch": 1.1362275449101795, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.2244, "step": 759 }, { "epoch": 1.1377245508982037, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.2322, "step": 760 }, { "epoch": 1.1392215568862276, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.2234, "step": 761 }, { "epoch": 1.1407185628742516, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.2225, "step": 762 }, { "epoch": 1.1422155688622755, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.2199, "step": 763 }, { "epoch": 1.1437125748502994, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.224, "step": 764 }, { "epoch": 1.1452095808383234, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.2217, "step": 765 }, { "epoch": 1.1467065868263473, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.2292, "step": 766 }, { "epoch": 1.1482035928143712, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.2188, "step": 767 }, { "epoch": 1.1497005988023952, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.2298, "step": 768 }, { "epoch": 1.151197604790419, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2207, "step": 769 }, { "epoch": 1.152694610778443, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.2295, "step": 770 }, { "epoch": 1.154191616766467, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.2203, "step": 771 }, { "epoch": 1.1556886227544911, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.2198, "step": 772 }, { "epoch": 1.157185628742515, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.2228, "step": 773 }, { "epoch": 1.158682634730539, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.2141, "step": 774 }, { "epoch": 1.160179640718563, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.22, "step": 775 }, { "epoch": 1.1616766467065869, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.2225, "step": 776 }, { "epoch": 1.1631736526946108, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.2187, "step": 777 }, { "epoch": 1.1646706586826348, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.2214, "step": 778 }, { "epoch": 1.1661676646706587, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.2203, "step": 779 }, { "epoch": 1.1676646706586826, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.2223, "step": 780 }, { "epoch": 1.1691616766467066, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.219, "step": 781 }, { "epoch": 1.1706586826347305, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.2217, "step": 782 }, { "epoch": 1.1721556886227544, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.2189, "step": 783 }, { "epoch": 1.1736526946107784, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.2181, "step": 784 }, { "epoch": 1.1751497005988023, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2276, "step": 785 }, { "epoch": 1.1766467065868262, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.2178, "step": 786 }, { "epoch": 1.1781437125748502, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2254, "step": 787 }, { "epoch": 1.1796407185628743, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.2179, "step": 788 }, { "epoch": 1.1811377245508983, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.219, "step": 789 }, { "epoch": 1.1826347305389222, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.2253, "step": 790 }, { "epoch": 1.1841317365269461, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2195, "step": 791 }, { "epoch": 1.18562874251497, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.2154, "step": 792 }, { "epoch": 1.187125748502994, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.2129, "step": 793 }, { "epoch": 1.188622754491018, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2196, "step": 794 }, { "epoch": 1.1901197604790419, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.2216, "step": 795 }, { "epoch": 1.1916167664670658, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2218, "step": 796 }, { "epoch": 1.1931137724550898, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.2223, "step": 797 }, { "epoch": 1.1946107784431137, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.2176, "step": 798 }, { "epoch": 1.1961077844311376, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.2178, "step": 799 }, { "epoch": 1.1976047904191618, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.2196, "step": 800 }, { "epoch": 1.1991017964071857, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.222, "step": 801 }, { "epoch": 1.2005988023952097, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.2207, "step": 802 }, { "epoch": 1.2020958083832336, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.2174, "step": 803 }, { "epoch": 1.2035928143712575, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.2221, "step": 804 }, { "epoch": 1.2050898203592815, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.2173, "step": 805 }, { "epoch": 1.2065868263473054, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2134, "step": 806 }, { "epoch": 1.2080838323353293, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2161, "step": 807 }, { "epoch": 1.2095808383233533, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.2128, "step": 808 }, { "epoch": 1.2110778443113772, "grad_norm": 0.2734375, "learning_rate": 0.0008, "loss": 1.2226, "step": 809 }, { "epoch": 1.2125748502994012, "grad_norm": 0.9140625, "learning_rate": 0.0008, "loss": 1.2246, "step": 810 }, { "epoch": 1.214071856287425, "grad_norm": 1.9140625, "learning_rate": 0.0008, "loss": 1.3315, "step": 811 }, { "epoch": 1.215568862275449, "grad_norm": 0.4453125, "learning_rate": 0.0008, "loss": 1.2585, "step": 812 }, { "epoch": 1.217065868263473, "grad_norm": 0.91015625, "learning_rate": 0.0008, "loss": 1.2749, "step": 813 }, { "epoch": 1.218562874251497, "grad_norm": 1.4375, "learning_rate": 0.0008, "loss": 1.2944, "step": 814 }, { "epoch": 1.220059880239521, "grad_norm": 0.4609375, "learning_rate": 0.0008, "loss": 1.2678, "step": 815 }, { "epoch": 1.221556886227545, "grad_norm": 0.3984375, "learning_rate": 0.0008, "loss": 1.2712, "step": 816 }, { "epoch": 1.223053892215569, "grad_norm": 0.359375, "learning_rate": 0.0008, "loss": 1.2548, "step": 817 }, { "epoch": 1.2245508982035929, "grad_norm": 0.318359375, "learning_rate": 0.0008, "loss": 1.2598, "step": 818 }, { "epoch": 1.2260479041916168, "grad_norm": 0.259765625, "learning_rate": 0.0008, "loss": 1.2494, "step": 819 }, { "epoch": 1.2275449101796407, "grad_norm": 0.2265625, "learning_rate": 0.0008, "loss": 1.258, "step": 820 }, { "epoch": 1.2290419161676647, "grad_norm": 0.1845703125, "learning_rate": 0.0008, "loss": 1.2467, "step": 821 }, { "epoch": 1.2305389221556886, "grad_norm": 0.19140625, "learning_rate": 0.0008, "loss": 1.238, "step": 822 }, { "epoch": 1.2320359281437125, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.2271, "step": 823 }, { "epoch": 1.2335329341317365, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.2371, "step": 824 }, { "epoch": 1.2350299401197604, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.2355, "step": 825 }, { "epoch": 1.2365269461077844, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.2355, "step": 826 }, { "epoch": 1.2380239520958083, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.2356, "step": 827 }, { "epoch": 1.2395209580838324, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.2343, "step": 828 }, { "epoch": 1.2410179640718564, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.2363, "step": 829 }, { "epoch": 1.2425149700598803, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.2363, "step": 830 }, { "epoch": 1.2440119760479043, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.22, "step": 831 }, { "epoch": 1.2455089820359282, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.2313, "step": 832 }, { "epoch": 1.2470059880239521, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.2283, "step": 833 }, { "epoch": 1.248502994011976, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.232, "step": 834 }, { "epoch": 1.25, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2278, "step": 835 }, { "epoch": 1.251497005988024, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.2336, "step": 836 }, { "epoch": 1.2529940119760479, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.218, "step": 837 }, { "epoch": 1.2544910179640718, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.2327, "step": 838 }, { "epoch": 1.2559880239520957, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.2223, "step": 839 }, { "epoch": 1.2574850299401197, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.2284, "step": 840 }, { "epoch": 1.2589820359281436, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.2221, "step": 841 }, { "epoch": 1.2604790419161676, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.2212, "step": 842 }, { "epoch": 1.2619760479041915, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.2208, "step": 843 }, { "epoch": 1.2634730538922156, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.2243, "step": 844 }, { "epoch": 1.2649700598802396, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.2203, "step": 845 }, { "epoch": 1.2664670658682635, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.2163, "step": 846 }, { "epoch": 1.2679640718562875, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.2256, "step": 847 }, { "epoch": 1.2694610778443114, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.2152, "step": 848 }, { "epoch": 1.2709580838323353, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.2169, "step": 849 }, { "epoch": 1.2724550898203593, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.2325, "step": 850 }, { "epoch": 1.2739520958083832, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2178, "step": 851 }, { "epoch": 1.2754491017964071, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.2087, "step": 852 }, { "epoch": 1.276946107784431, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.2163, "step": 853 }, { "epoch": 1.278443113772455, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.2198, "step": 854 }, { "epoch": 1.2799401197604792, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.2201, "step": 855 }, { "epoch": 1.281437125748503, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.2171, "step": 856 }, { "epoch": 1.282934131736527, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.213, "step": 857 }, { "epoch": 1.284431137724551, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.2188, "step": 858 }, { "epoch": 1.285928143712575, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.2155, "step": 859 }, { "epoch": 1.2874251497005988, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.2163, "step": 860 }, { "epoch": 1.2889221556886228, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.2242, "step": 861 }, { "epoch": 1.2904191616766467, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.2176, "step": 862 }, { "epoch": 1.2919161676646707, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.2127, "step": 863 }, { "epoch": 1.2934131736526946, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.2145, "step": 864 }, { "epoch": 1.2949101796407185, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.2258, "step": 865 }, { "epoch": 1.2964071856287425, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.219, "step": 866 }, { "epoch": 1.2979041916167664, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2148, "step": 867 }, { "epoch": 1.2994011976047903, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.215, "step": 868 }, { "epoch": 1.3008982035928143, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.207, "step": 869 }, { "epoch": 1.3023952095808382, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2146, "step": 870 }, { "epoch": 1.3038922155688621, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.2173, "step": 871 }, { "epoch": 1.3053892215568863, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.2191, "step": 872 }, { "epoch": 1.3068862275449102, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.2156, "step": 873 }, { "epoch": 1.3083832335329342, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.2124, "step": 874 }, { "epoch": 1.3098802395209581, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.2191, "step": 875 }, { "epoch": 1.311377245508982, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.2193, "step": 876 }, { "epoch": 1.312874251497006, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2146, "step": 877 }, { "epoch": 1.31437125748503, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.2121, "step": 878 }, { "epoch": 1.3158682634730539, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.2162, "step": 879 }, { "epoch": 1.3173652694610778, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.2154, "step": 880 }, { "epoch": 1.3188622754491017, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.2199, "step": 881 }, { "epoch": 1.3203592814371259, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.2234, "step": 882 }, { "epoch": 1.3218562874251498, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.222, "step": 883 }, { "epoch": 1.3233532934131738, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.2121, "step": 884 }, { "epoch": 1.3248502994011977, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.2114, "step": 885 }, { "epoch": 1.3263473053892216, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.2124, "step": 886 }, { "epoch": 1.3278443113772456, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2143, "step": 887 }, { "epoch": 1.3293413173652695, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.2189, "step": 888 }, { "epoch": 1.3308383233532934, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.2093, "step": 889 }, { "epoch": 1.3323353293413174, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.2117, "step": 890 }, { "epoch": 1.3338323353293413, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.2157, "step": 891 }, { "epoch": 1.3353293413173652, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.2195, "step": 892 }, { "epoch": 1.3368263473053892, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1979, "step": 893 }, { "epoch": 1.3383233532934131, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.215, "step": 894 }, { "epoch": 1.339820359281437, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2084, "step": 895 }, { "epoch": 1.341317365269461, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.2044, "step": 896 }, { "epoch": 1.342814371257485, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.2095, "step": 897 }, { "epoch": 1.3443113772455089, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.2104, "step": 898 }, { "epoch": 1.3458083832335328, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.2132, "step": 899 }, { "epoch": 1.347305389221557, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.2128, "step": 900 }, { "epoch": 1.348802395209581, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.2178, "step": 901 }, { "epoch": 1.3502994011976048, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.2073, "step": 902 }, { "epoch": 1.3517964071856288, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.2028, "step": 903 }, { "epoch": 1.3532934131736527, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.208, "step": 904 }, { "epoch": 1.3547904191616766, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.2178, "step": 905 }, { "epoch": 1.3562874251497006, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.2105, "step": 906 }, { "epoch": 1.3577844311377245, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.213, "step": 907 }, { "epoch": 1.3592814371257484, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.2116, "step": 908 }, { "epoch": 1.3607784431137724, "grad_norm": 0.1982421875, "learning_rate": 0.0008, "loss": 1.2187, "step": 909 }, { "epoch": 1.3622754491017965, "grad_norm": 0.400390625, "learning_rate": 0.0008, "loss": 1.219, "step": 910 }, { "epoch": 1.3637724550898205, "grad_norm": 0.9140625, "learning_rate": 0.0008, "loss": 1.2323, "step": 911 }, { "epoch": 1.3652694610778444, "grad_norm": 1.8671875, "learning_rate": 0.0008, "loss": 1.2959, "step": 912 }, { "epoch": 1.3667664670658684, "grad_norm": 0.259765625, "learning_rate": 0.0008, "loss": 1.2367, "step": 913 }, { "epoch": 1.3682634730538923, "grad_norm": 1.0234375, "learning_rate": 0.0008, "loss": 1.2636, "step": 914 }, { "epoch": 1.3697604790419162, "grad_norm": 0.50390625, "learning_rate": 0.0008, "loss": 1.2372, "step": 915 }, { "epoch": 1.3712574850299402, "grad_norm": 0.5390625, "learning_rate": 0.0008, "loss": 1.2443, "step": 916 }, { "epoch": 1.372754491017964, "grad_norm": 0.30859375, "learning_rate": 0.0008, "loss": 1.2336, "step": 917 }, { "epoch": 1.374251497005988, "grad_norm": 0.392578125, "learning_rate": 0.0008, "loss": 1.2485, "step": 918 }, { "epoch": 1.375748502994012, "grad_norm": 0.287109375, "learning_rate": 0.0008, "loss": 1.2288, "step": 919 }, { "epoch": 1.377245508982036, "grad_norm": 0.28125, "learning_rate": 0.0008, "loss": 1.2385, "step": 920 }, { "epoch": 1.3787425149700598, "grad_norm": 0.232421875, "learning_rate": 0.0008, "loss": 1.2306, "step": 921 }, { "epoch": 1.3802395209580838, "grad_norm": 0.171875, "learning_rate": 0.0008, "loss": 1.2302, "step": 922 }, { "epoch": 1.3817365269461077, "grad_norm": 0.169921875, "learning_rate": 0.0008, "loss": 1.2309, "step": 923 }, { "epoch": 1.3832335329341316, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.2244, "step": 924 }, { "epoch": 1.3847305389221556, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.2262, "step": 925 }, { "epoch": 1.3862275449101795, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.208, "step": 926 }, { "epoch": 1.3877245508982035, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.2209, "step": 927 }, { "epoch": 1.3892215568862276, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.2219, "step": 928 }, { "epoch": 1.3907185628742516, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.216, "step": 929 }, { "epoch": 1.3922155688622755, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.2252, "step": 930 }, { "epoch": 1.3937125748502994, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.2221, "step": 931 }, { "epoch": 1.3952095808383234, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.2244, "step": 932 }, { "epoch": 1.3967065868263473, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.2251, "step": 933 }, { "epoch": 1.3982035928143712, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.217, "step": 934 }, { "epoch": 1.3997005988023952, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.2183, "step": 935 }, { "epoch": 1.401197604790419, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.213, "step": 936 }, { "epoch": 1.402694610778443, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.2141, "step": 937 }, { "epoch": 1.4041916167664672, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.2129, "step": 938 }, { "epoch": 1.4056886227544911, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.2105, "step": 939 }, { "epoch": 1.407185628742515, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2148, "step": 940 }, { "epoch": 1.408682634730539, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.2099, "step": 941 }, { "epoch": 1.410179640718563, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.2095, "step": 942 }, { "epoch": 1.4116766467065869, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2154, "step": 943 }, { "epoch": 1.4131736526946108, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.217, "step": 944 }, { "epoch": 1.4146706586826348, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.2108, "step": 945 }, { "epoch": 1.4161676646706587, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.2119, "step": 946 }, { "epoch": 1.4176646706586826, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.2122, "step": 947 }, { "epoch": 1.4191616766467066, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.2096, "step": 948 }, { "epoch": 1.4206586826347305, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.2052, "step": 949 }, { "epoch": 1.4221556886227544, "grad_norm": 0.051025390625, "learning_rate": 0.0008, "loss": 1.2113, "step": 950 }, { "epoch": 1.4236526946107784, "grad_norm": 0.052734375, "learning_rate": 0.0008, "loss": 1.2125, "step": 951 }, { "epoch": 1.4251497005988023, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.2084, "step": 952 }, { "epoch": 1.4266467065868262, "grad_norm": 0.049072265625, "learning_rate": 0.0008, "loss": 1.2112, "step": 953 }, { "epoch": 1.4281437125748502, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2088, "step": 954 }, { "epoch": 1.4296407185628741, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.2017, "step": 955 }, { "epoch": 1.4311377245508983, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.2095, "step": 956 }, { "epoch": 1.4326347305389222, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.2201, "step": 957 }, { "epoch": 1.4341317365269461, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.2119, "step": 958 }, { "epoch": 1.43562874251497, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.2162, "step": 959 }, { "epoch": 1.437125748502994, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.2035, "step": 960 }, { "epoch": 1.438622754491018, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.2084, "step": 961 }, { "epoch": 1.4401197604790419, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.2118, "step": 962 }, { "epoch": 1.4416167664670658, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.2041, "step": 963 }, { "epoch": 1.4431137724550898, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.21, "step": 964 }, { "epoch": 1.4446107784431137, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.2105, "step": 965 }, { "epoch": 1.4461077844311379, "grad_norm": 0.046875, "learning_rate": 0.0008, "loss": 1.2103, "step": 966 }, { "epoch": 1.4476047904191618, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.2033, "step": 967 }, { "epoch": 1.4491017964071857, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.2092, "step": 968 }, { "epoch": 1.4505988023952097, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.2062, "step": 969 }, { "epoch": 1.4520958083832336, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.2027, "step": 970 }, { "epoch": 1.4535928143712575, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.2035, "step": 971 }, { "epoch": 1.4550898203592815, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.2088, "step": 972 }, { "epoch": 1.4565868263473054, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.2028, "step": 973 }, { "epoch": 1.4580838323353293, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.2031, "step": 974 }, { "epoch": 1.4595808383233533, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.2023, "step": 975 }, { "epoch": 1.4610778443113772, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.205, "step": 976 }, { "epoch": 1.4625748502994012, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.2101, "step": 977 }, { "epoch": 1.464071856287425, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.2114, "step": 978 }, { "epoch": 1.465568862275449, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1998, "step": 979 }, { "epoch": 1.467065868263473, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2041, "step": 980 }, { "epoch": 1.468562874251497, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.2183, "step": 981 }, { "epoch": 1.4700598802395208, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.2202, "step": 982 }, { "epoch": 1.471556886227545, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.2125, "step": 983 }, { "epoch": 1.473053892215569, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.2027, "step": 984 }, { "epoch": 1.4745508982035929, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.2034, "step": 985 }, { "epoch": 1.4760479041916168, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.2009, "step": 986 }, { "epoch": 1.4775449101796407, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.2107, "step": 987 }, { "epoch": 1.4790419161676647, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.2079, "step": 988 }, { "epoch": 1.4805389221556886, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.2113, "step": 989 }, { "epoch": 1.4820359281437125, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.2029, "step": 990 }, { "epoch": 1.4835329341317365, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.2102, "step": 991 }, { "epoch": 1.4850299401197604, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.2034, "step": 992 }, { "epoch": 1.4865269461077844, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.2087, "step": 993 }, { "epoch": 1.4880239520958085, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.2089, "step": 994 }, { "epoch": 1.4895209580838324, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.2062, "step": 995 }, { "epoch": 1.4910179640718564, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.2015, "step": 996 }, { "epoch": 1.4925149700598803, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2039, "step": 997 }, { "epoch": 1.4940119760479043, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.2031, "step": 998 }, { "epoch": 1.4955089820359282, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.2136, "step": 999 }, { "epoch": 1.4970059880239521, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.2024, "step": 1000 }, { "epoch": 1.498502994011976, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.2109, "step": 1001 }, { "epoch": 1.5, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.2178, "step": 1002 }, { "epoch": 1.501497005988024, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.2105, "step": 1003 }, { "epoch": 1.5029940119760479, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2054, "step": 1004 }, { "epoch": 1.5044910179640718, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.2028, "step": 1005 }, { "epoch": 1.5059880239520957, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.2002, "step": 1006 }, { "epoch": 1.5074850299401197, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2092, "step": 1007 }, { "epoch": 1.5089820359281436, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2041, "step": 1008 }, { "epoch": 1.5104790419161676, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.2079, "step": 1009 }, { "epoch": 1.5119760479041915, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1992, "step": 1010 }, { "epoch": 1.5134730538922154, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.2069, "step": 1011 }, { "epoch": 1.5149700598802394, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.2013, "step": 1012 }, { "epoch": 1.5164670658682635, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.2034, "step": 1013 }, { "epoch": 1.5179640718562875, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.2027, "step": 1014 }, { "epoch": 1.5194610778443114, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.2092, "step": 1015 }, { "epoch": 1.5209580838323353, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.2086, "step": 1016 }, { "epoch": 1.5224550898203593, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2071, "step": 1017 }, { "epoch": 1.5239520958083832, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.2035, "step": 1018 }, { "epoch": 1.5254491017964071, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.203, "step": 1019 }, { "epoch": 1.5269461077844313, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.2158, "step": 1020 }, { "epoch": 1.5284431137724552, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2044, "step": 1021 }, { "epoch": 1.5299401197604792, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.2039, "step": 1022 }, { "epoch": 1.531437125748503, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1994, "step": 1023 }, { "epoch": 1.532934131736527, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2034, "step": 1024 }, { "epoch": 1.534431137724551, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.2066, "step": 1025 }, { "epoch": 1.535928143712575, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2026, "step": 1026 }, { "epoch": 1.5374251497005988, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2077, "step": 1027 }, { "epoch": 1.5389221556886228, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.2121, "step": 1028 }, { "epoch": 1.5404191616766467, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.2023, "step": 1029 }, { "epoch": 1.5419161676646707, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.2043, "step": 1030 }, { "epoch": 1.5434131736526946, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.2075, "step": 1031 }, { "epoch": 1.5449101796407185, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2064, "step": 1032 }, { "epoch": 1.5464071856287425, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.2031, "step": 1033 }, { "epoch": 1.5479041916167664, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.2031, "step": 1034 }, { "epoch": 1.5494011976047903, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.2081, "step": 1035 }, { "epoch": 1.5508982035928143, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.2075, "step": 1036 }, { "epoch": 1.5523952095808382, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.2064, "step": 1037 }, { "epoch": 1.5538922155688621, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.2106, "step": 1038 }, { "epoch": 1.555389221556886, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.2055, "step": 1039 }, { "epoch": 1.55688622754491, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.2047, "step": 1040 }, { "epoch": 1.5583832335329342, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.2051, "step": 1041 }, { "epoch": 1.5598802395209581, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.2072, "step": 1042 }, { "epoch": 1.561377245508982, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.2055, "step": 1043 }, { "epoch": 1.562874251497006, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.2028, "step": 1044 }, { "epoch": 1.56437125748503, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.204, "step": 1045 }, { "epoch": 1.5658682634730539, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.2117, "step": 1046 }, { "epoch": 1.5673652694610778, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.2005, "step": 1047 }, { "epoch": 1.568862275449102, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.2032, "step": 1048 }, { "epoch": 1.5703592814371259, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.2109, "step": 1049 }, { "epoch": 1.5718562874251498, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.2036, "step": 1050 }, { "epoch": 1.5733532934131738, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1996, "step": 1051 }, { "epoch": 1.5748502994011977, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2076, "step": 1052 }, { "epoch": 1.5763473053892216, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.2064, "step": 1053 }, { "epoch": 1.5778443113772456, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2028, "step": 1054 }, { "epoch": 1.5793413173652695, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.2081, "step": 1055 }, { "epoch": 1.5808383233532934, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1996, "step": 1056 }, { "epoch": 1.5823353293413174, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.2042, "step": 1057 }, { "epoch": 1.5838323353293413, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1969, "step": 1058 }, { "epoch": 1.5853293413173652, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.2051, "step": 1059 }, { "epoch": 1.5868263473053892, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.2038, "step": 1060 }, { "epoch": 1.5883233532934131, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.2081, "step": 1061 }, { "epoch": 1.589820359281437, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.2055, "step": 1062 }, { "epoch": 1.591317365269461, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.202, "step": 1063 }, { "epoch": 1.592814371257485, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.2008, "step": 1064 }, { "epoch": 1.5943113772455089, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.2024, "step": 1065 }, { "epoch": 1.5958083832335328, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.2078, "step": 1066 }, { "epoch": 1.5973053892215567, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.2053, "step": 1067 }, { "epoch": 1.5988023952095807, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.2041, "step": 1068 }, { "epoch": 1.6002994011976048, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.2056, "step": 1069 }, { "epoch": 1.6017964071856288, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.2009, "step": 1070 }, { "epoch": 1.6032934131736527, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.211, "step": 1071 }, { "epoch": 1.6047904191616766, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.2129, "step": 1072 }, { "epoch": 1.6062874251497006, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.2025, "step": 1073 }, { "epoch": 1.6077844311377245, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.2019, "step": 1074 }, { "epoch": 1.6092814371257484, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.2059, "step": 1075 }, { "epoch": 1.6107784431137726, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.2036, "step": 1076 }, { "epoch": 1.6122754491017965, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2065, "step": 1077 }, { "epoch": 1.6137724550898205, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2114, "step": 1078 }, { "epoch": 1.6152694610778444, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2037, "step": 1079 }, { "epoch": 1.6167664670658684, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.2015, "step": 1080 }, { "epoch": 1.6182634730538923, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.199, "step": 1081 }, { "epoch": 1.6197604790419162, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2123, "step": 1082 }, { "epoch": 1.6212574850299402, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.2005, "step": 1083 }, { "epoch": 1.622754491017964, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1993, "step": 1084 }, { "epoch": 1.624251497005988, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.204, "step": 1085 }, { "epoch": 1.625748502994012, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.2042, "step": 1086 }, { "epoch": 1.627245508982036, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1993, "step": 1087 }, { "epoch": 1.6287425149700598, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.2052, "step": 1088 }, { "epoch": 1.6302395209580838, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2036, "step": 1089 }, { "epoch": 1.6317365269461077, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1921, "step": 1090 }, { "epoch": 1.6332335329341316, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2045, "step": 1091 }, { "epoch": 1.6347305389221556, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.2008, "step": 1092 }, { "epoch": 1.6362275449101795, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.204, "step": 1093 }, { "epoch": 1.6377245508982035, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1952, "step": 1094 }, { "epoch": 1.6392215568862274, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.201, "step": 1095 }, { "epoch": 1.6407185628742516, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1998, "step": 1096 }, { "epoch": 1.6422155688622755, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.2058, "step": 1097 }, { "epoch": 1.6437125748502994, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.207, "step": 1098 }, { "epoch": 1.6452095808383234, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.2026, "step": 1099 }, { "epoch": 1.6467065868263473, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1981, "step": 1100 }, { "epoch": 1.6482035928143712, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2022, "step": 1101 }, { "epoch": 1.6497005988023952, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.2003, "step": 1102 }, { "epoch": 1.6511976047904193, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1976, "step": 1103 }, { "epoch": 1.6526946107784433, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.2066, "step": 1104 }, { "epoch": 1.6541916167664672, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.2024, "step": 1105 }, { "epoch": 1.6556886227544911, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.208, "step": 1106 }, { "epoch": 1.657185628742515, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1998, "step": 1107 }, { "epoch": 1.658682634730539, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.2023, "step": 1108 }, { "epoch": 1.660179640718563, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.2033, "step": 1109 }, { "epoch": 1.6616766467065869, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1999, "step": 1110 }, { "epoch": 1.6631736526946108, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.2061, "step": 1111 }, { "epoch": 1.6646706586826348, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.2035, "step": 1112 }, { "epoch": 1.6661676646706587, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1901, "step": 1113 }, { "epoch": 1.6676646706586826, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1957, "step": 1114 }, { "epoch": 1.6691616766467066, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1969, "step": 1115 }, { "epoch": 1.6706586826347305, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.2045, "step": 1116 }, { "epoch": 1.6721556886227544, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2036, "step": 1117 }, { "epoch": 1.6736526946107784, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.2022, "step": 1118 }, { "epoch": 1.6751497005988023, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2028, "step": 1119 }, { "epoch": 1.6766467065868262, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.2063, "step": 1120 }, { "epoch": 1.6781437125748502, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.203, "step": 1121 }, { "epoch": 1.6796407185628741, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.2046, "step": 1122 }, { "epoch": 1.681137724550898, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.2002, "step": 1123 }, { "epoch": 1.6826347305389222, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.2078, "step": 1124 }, { "epoch": 1.6841317365269461, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2104, "step": 1125 }, { "epoch": 1.68562874251497, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.2006, "step": 1126 }, { "epoch": 1.687125748502994, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.207, "step": 1127 }, { "epoch": 1.688622754491018, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.2039, "step": 1128 }, { "epoch": 1.6901197604790419, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.2034, "step": 1129 }, { "epoch": 1.6916167664670658, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.2091, "step": 1130 }, { "epoch": 1.69311377245509, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1965, "step": 1131 }, { "epoch": 1.694610778443114, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.2039, "step": 1132 }, { "epoch": 1.6961077844311379, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1979, "step": 1133 }, { "epoch": 1.6976047904191618, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.2006, "step": 1134 }, { "epoch": 1.6991017964071857, "grad_norm": 0.236328125, "learning_rate": 0.0008, "loss": 1.1998, "step": 1135 }, { "epoch": 1.7005988023952097, "grad_norm": 0.482421875, "learning_rate": 0.0008, "loss": 1.2086, "step": 1136 }, { "epoch": 1.7020958083832336, "grad_norm": 1.328125, "learning_rate": 0.0008, "loss": 1.2406, "step": 1137 }, { "epoch": 1.7035928143712575, "grad_norm": 1.28125, "learning_rate": 0.0008, "loss": 1.2445, "step": 1138 }, { "epoch": 1.7050898203592815, "grad_norm": 0.32421875, "learning_rate": 0.0008, "loss": 1.2307, "step": 1139 }, { "epoch": 1.7065868263473054, "grad_norm": 0.50390625, "learning_rate": 0.0008, "loss": 1.2324, "step": 1140 }, { "epoch": 1.7080838323353293, "grad_norm": 0.400390625, "learning_rate": 0.0008, "loss": 1.2266, "step": 1141 }, { "epoch": 1.7095808383233533, "grad_norm": 0.31640625, "learning_rate": 0.0008, "loss": 1.2242, "step": 1142 }, { "epoch": 1.7110778443113772, "grad_norm": 0.30078125, "learning_rate": 0.0008, "loss": 1.2131, "step": 1143 }, { "epoch": 1.7125748502994012, "grad_norm": 0.25390625, "learning_rate": 0.0008, "loss": 1.2149, "step": 1144 }, { "epoch": 1.714071856287425, "grad_norm": 0.2041015625, "learning_rate": 0.0008, "loss": 1.2205, "step": 1145 }, { "epoch": 1.715568862275449, "grad_norm": 0.1728515625, "learning_rate": 0.0008, "loss": 1.2154, "step": 1146 }, { "epoch": 1.717065868263473, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.2186, "step": 1147 }, { "epoch": 1.718562874251497, "grad_norm": 0.1591796875, "learning_rate": 0.0008, "loss": 1.2107, "step": 1148 }, { "epoch": 1.7200598802395208, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.2122, "step": 1149 }, { "epoch": 1.7215568862275448, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.2155, "step": 1150 }, { "epoch": 1.7230538922155687, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.2003, "step": 1151 }, { "epoch": 1.7245508982035929, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.2131, "step": 1152 }, { "epoch": 1.7260479041916168, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.2108, "step": 1153 }, { "epoch": 1.7275449101796407, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.2048, "step": 1154 }, { "epoch": 1.7290419161676647, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.2034, "step": 1155 }, { "epoch": 1.7305389221556886, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.2099, "step": 1156 }, { "epoch": 1.7320359281437125, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.2067, "step": 1157 }, { "epoch": 1.7335329341317365, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1989, "step": 1158 }, { "epoch": 1.7350299401197606, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.2037, "step": 1159 }, { "epoch": 1.7365269461077846, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1999, "step": 1160 }, { "epoch": 1.7380239520958085, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.2002, "step": 1161 }, { "epoch": 1.7395209580838324, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.2082, "step": 1162 }, { "epoch": 1.7410179640718564, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.2056, "step": 1163 }, { "epoch": 1.7425149700598803, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.204, "step": 1164 }, { "epoch": 1.7440119760479043, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.2007, "step": 1165 }, { "epoch": 1.7455089820359282, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.2006, "step": 1166 }, { "epoch": 1.7470059880239521, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.2037, "step": 1167 }, { "epoch": 1.748502994011976, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1902, "step": 1168 }, { "epoch": 1.75, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.2009, "step": 1169 }, { "epoch": 1.751497005988024, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1987, "step": 1170 }, { "epoch": 1.7529940119760479, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.2076, "step": 1171 }, { "epoch": 1.7544910179640718, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.2014, "step": 1172 }, { "epoch": 1.7559880239520957, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1927, "step": 1173 }, { "epoch": 1.7574850299401197, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.2002, "step": 1174 }, { "epoch": 1.7589820359281436, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.2018, "step": 1175 }, { "epoch": 1.7604790419161676, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1954, "step": 1176 }, { "epoch": 1.7619760479041915, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1957, "step": 1177 }, { "epoch": 1.7634730538922154, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.201, "step": 1178 }, { "epoch": 1.7649700598802394, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1982, "step": 1179 }, { "epoch": 1.7664670658682635, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1993, "step": 1180 }, { "epoch": 1.7679640718562875, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1968, "step": 1181 }, { "epoch": 1.7694610778443114, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.198, "step": 1182 }, { "epoch": 1.7709580838323353, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1968, "step": 1183 }, { "epoch": 1.7724550898203593, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.2056, "step": 1184 }, { "epoch": 1.7739520958083832, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1906, "step": 1185 }, { "epoch": 1.7754491017964071, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.206, "step": 1186 }, { "epoch": 1.7769461077844313, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1957, "step": 1187 }, { "epoch": 1.7784431137724552, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1938, "step": 1188 }, { "epoch": 1.7799401197604792, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1938, "step": 1189 }, { "epoch": 1.781437125748503, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.2022, "step": 1190 }, { "epoch": 1.782934131736527, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.2001, "step": 1191 }, { "epoch": 1.784431137724551, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.2002, "step": 1192 }, { "epoch": 1.785928143712575, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.2004, "step": 1193 }, { "epoch": 1.7874251497005988, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1993, "step": 1194 }, { "epoch": 1.7889221556886228, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2006, "step": 1195 }, { "epoch": 1.7904191616766467, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.2053, "step": 1196 }, { "epoch": 1.7919161676646707, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.2009, "step": 1197 }, { "epoch": 1.7934131736526946, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1989, "step": 1198 }, { "epoch": 1.7949101796407185, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.199, "step": 1199 }, { "epoch": 1.7964071856287425, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1943, "step": 1200 }, { "epoch": 1.7979041916167664, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1942, "step": 1201 }, { "epoch": 1.7994011976047903, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1922, "step": 1202 }, { "epoch": 1.8008982035928143, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1993, "step": 1203 }, { "epoch": 1.8023952095808382, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.197, "step": 1204 }, { "epoch": 1.8038922155688621, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.2087, "step": 1205 }, { "epoch": 1.805389221556886, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1982, "step": 1206 }, { "epoch": 1.80688622754491, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1941, "step": 1207 }, { "epoch": 1.8083832335329342, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1965, "step": 1208 }, { "epoch": 1.8098802395209581, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.2009, "step": 1209 }, { "epoch": 1.811377245508982, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1945, "step": 1210 }, { "epoch": 1.812874251497006, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1946, "step": 1211 }, { "epoch": 1.81437125748503, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1955, "step": 1212 }, { "epoch": 1.8158682634730539, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.2027, "step": 1213 }, { "epoch": 1.8173652694610778, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.2017, "step": 1214 }, { "epoch": 1.818862275449102, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.2033, "step": 1215 }, { "epoch": 1.8203592814371259, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.2018, "step": 1216 }, { "epoch": 1.8218562874251498, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.2031, "step": 1217 }, { "epoch": 1.8233532934131738, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1977, "step": 1218 }, { "epoch": 1.8248502994011977, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1978, "step": 1219 }, { "epoch": 1.8263473053892216, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.2027, "step": 1220 }, { "epoch": 1.8278443113772456, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1979, "step": 1221 }, { "epoch": 1.8293413173652695, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.2007, "step": 1222 }, { "epoch": 1.8308383233532934, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.2038, "step": 1223 }, { "epoch": 1.8323353293413174, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.197, "step": 1224 }, { "epoch": 1.8338323353293413, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1969, "step": 1225 }, { "epoch": 1.8353293413173652, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.2045, "step": 1226 }, { "epoch": 1.8368263473053892, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1982, "step": 1227 }, { "epoch": 1.8383233532934131, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1956, "step": 1228 }, { "epoch": 1.839820359281437, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1997, "step": 1229 }, { "epoch": 1.841317365269461, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1935, "step": 1230 }, { "epoch": 1.842814371257485, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1979, "step": 1231 }, { "epoch": 1.8443113772455089, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1963, "step": 1232 }, { "epoch": 1.8458083832335328, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2048, "step": 1233 }, { "epoch": 1.8473053892215567, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1986, "step": 1234 }, { "epoch": 1.8488023952095807, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1992, "step": 1235 }, { "epoch": 1.8502994011976048, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1962, "step": 1236 }, { "epoch": 1.8517964071856288, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1967, "step": 1237 }, { "epoch": 1.8532934131736527, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1933, "step": 1238 }, { "epoch": 1.8547904191616766, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1951, "step": 1239 }, { "epoch": 1.8562874251497006, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1929, "step": 1240 }, { "epoch": 1.8577844311377245, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1963, "step": 1241 }, { "epoch": 1.8592814371257484, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1939, "step": 1242 }, { "epoch": 1.8607784431137726, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1955, "step": 1243 }, { "epoch": 1.8622754491017965, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1972, "step": 1244 }, { "epoch": 1.8637724550898205, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.2039, "step": 1245 }, { "epoch": 1.8652694610778444, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1959, "step": 1246 }, { "epoch": 1.8667664670658684, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1993, "step": 1247 }, { "epoch": 1.8682634730538923, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1956, "step": 1248 }, { "epoch": 1.8697604790419162, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1985, "step": 1249 }, { "epoch": 1.8712574850299402, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1916, "step": 1250 }, { "epoch": 1.872754491017964, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1957, "step": 1251 }, { "epoch": 1.874251497005988, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1984, "step": 1252 }, { "epoch": 1.875748502994012, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1953, "step": 1253 }, { "epoch": 1.877245508982036, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.19, "step": 1254 }, { "epoch": 1.8787425149700598, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1997, "step": 1255 }, { "epoch": 1.8802395209580838, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1897, "step": 1256 }, { "epoch": 1.8817365269461077, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1943, "step": 1257 }, { "epoch": 1.8832335329341316, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1961, "step": 1258 }, { "epoch": 1.8847305389221556, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1967, "step": 1259 }, { "epoch": 1.8862275449101795, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1975, "step": 1260 }, { "epoch": 1.8877245508982035, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1958, "step": 1261 }, { "epoch": 1.8892215568862274, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1961, "step": 1262 }, { "epoch": 1.8907185628742516, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1865, "step": 1263 }, { "epoch": 1.8922155688622755, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1995, "step": 1264 }, { "epoch": 1.8937125748502994, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.195, "step": 1265 }, { "epoch": 1.8952095808383234, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1946, "step": 1266 }, { "epoch": 1.8967065868263473, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.2047, "step": 1267 }, { "epoch": 1.8982035928143712, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1946, "step": 1268 }, { "epoch": 1.8997005988023952, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1983, "step": 1269 }, { "epoch": 1.9011976047904193, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.197, "step": 1270 }, { "epoch": 1.9026946107784433, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.193, "step": 1271 }, { "epoch": 1.9041916167664672, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1948, "step": 1272 }, { "epoch": 1.9056886227544911, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1926, "step": 1273 }, { "epoch": 1.907185628742515, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1966, "step": 1274 }, { "epoch": 1.908682634730539, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.2019, "step": 1275 }, { "epoch": 1.910179640718563, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1955, "step": 1276 }, { "epoch": 1.9116766467065869, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1993, "step": 1277 }, { "epoch": 1.9131736526946108, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1926, "step": 1278 }, { "epoch": 1.9146706586826348, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1943, "step": 1279 }, { "epoch": 1.9161676646706587, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.2037, "step": 1280 }, { "epoch": 1.9176646706586826, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.194, "step": 1281 }, { "epoch": 1.9191616766467066, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1899, "step": 1282 }, { "epoch": 1.9206586826347305, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.2013, "step": 1283 }, { "epoch": 1.9221556886227544, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1933, "step": 1284 }, { "epoch": 1.9236526946107784, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1984, "step": 1285 }, { "epoch": 1.9251497005988023, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1964, "step": 1286 }, { "epoch": 1.9266467065868262, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.2036, "step": 1287 }, { "epoch": 1.9281437125748502, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1956, "step": 1288 }, { "epoch": 1.9296407185628741, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1956, "step": 1289 }, { "epoch": 1.931137724550898, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1928, "step": 1290 }, { "epoch": 1.9326347305389222, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1901, "step": 1291 }, { "epoch": 1.9341317365269461, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.2008, "step": 1292 }, { "epoch": 1.93562874251497, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1991, "step": 1293 }, { "epoch": 1.937125748502994, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1963, "step": 1294 }, { "epoch": 1.938622754491018, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1948, "step": 1295 }, { "epoch": 1.9401197604790419, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1986, "step": 1296 }, { "epoch": 1.9416167664670658, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1952, "step": 1297 }, { "epoch": 1.94311377245509, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.195, "step": 1298 }, { "epoch": 1.944610778443114, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.2005, "step": 1299 }, { "epoch": 1.9461077844311379, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1994, "step": 1300 }, { "epoch": 1.9476047904191618, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1986, "step": 1301 }, { "epoch": 1.9491017964071857, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1969, "step": 1302 }, { "epoch": 1.9505988023952097, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1933, "step": 1303 }, { "epoch": 1.9520958083832336, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1958, "step": 1304 }, { "epoch": 1.9535928143712575, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1988, "step": 1305 }, { "epoch": 1.9550898203592815, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1983, "step": 1306 }, { "epoch": 1.9565868263473054, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.2006, "step": 1307 }, { "epoch": 1.9580838323353293, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1964, "step": 1308 }, { "epoch": 1.9595808383233533, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1933, "step": 1309 }, { "epoch": 1.9610778443113772, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.2014, "step": 1310 }, { "epoch": 1.9625748502994012, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.195, "step": 1311 }, { "epoch": 1.964071856287425, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1968, "step": 1312 }, { "epoch": 1.965568862275449, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1899, "step": 1313 }, { "epoch": 1.967065868263473, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1916, "step": 1314 }, { "epoch": 1.968562874251497, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1935, "step": 1315 }, { "epoch": 1.9700598802395208, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1894, "step": 1316 }, { "epoch": 1.9715568862275448, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.194, "step": 1317 }, { "epoch": 1.9730538922155687, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1992, "step": 1318 }, { "epoch": 1.9745508982035929, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1932, "step": 1319 }, { "epoch": 1.9760479041916168, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1972, "step": 1320 }, { "epoch": 1.9775449101796407, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1878, "step": 1321 }, { "epoch": 1.9790419161676647, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1934, "step": 1322 }, { "epoch": 1.9805389221556886, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1848, "step": 1323 }, { "epoch": 1.9820359281437125, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1922, "step": 1324 }, { "epoch": 1.9835329341317365, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.2001, "step": 1325 }, { "epoch": 1.9850299401197606, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1906, "step": 1326 }, { "epoch": 1.9865269461077846, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1969, "step": 1327 }, { "epoch": 1.9880239520958085, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1859, "step": 1328 }, { "epoch": 1.9895209580838324, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1982, "step": 1329 }, { "epoch": 1.9910179640718564, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1932, "step": 1330 }, { "epoch": 1.9925149700598803, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1922, "step": 1331 }, { "epoch": 1.9940119760479043, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1961, "step": 1332 }, { "epoch": 1.9955089820359282, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.195, "step": 1333 }, { "epoch": 1.9970059880239521, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1998, "step": 1334 }, { "epoch": 1.998502994011976, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.2005, "step": 1335 }, { "epoch": 2.0, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.2005, "step": 1336 }, { "epoch": 2.001497005988024, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1958, "step": 1337 }, { "epoch": 2.002994011976048, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1954, "step": 1338 }, { "epoch": 2.004491017964072, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1946, "step": 1339 }, { "epoch": 2.0059880239520957, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.199, "step": 1340 }, { "epoch": 2.0074850299401197, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1871, "step": 1341 }, { "epoch": 2.0089820359281436, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1966, "step": 1342 }, { "epoch": 2.0104790419161676, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1905, "step": 1343 }, { "epoch": 2.0119760479041915, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.199, "step": 1344 }, { "epoch": 2.0134730538922154, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1907, "step": 1345 }, { "epoch": 2.0149700598802394, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.2011, "step": 1346 }, { "epoch": 2.0164670658682633, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1897, "step": 1347 }, { "epoch": 2.0179640718562872, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1983, "step": 1348 }, { "epoch": 2.019461077844311, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1901, "step": 1349 }, { "epoch": 2.020958083832335, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1962, "step": 1350 }, { "epoch": 2.0224550898203595, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.194, "step": 1351 }, { "epoch": 2.0239520958083834, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1909, "step": 1352 }, { "epoch": 2.0254491017964074, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.2032, "step": 1353 }, { "epoch": 2.0269461077844313, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1925, "step": 1354 }, { "epoch": 2.0284431137724552, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1884, "step": 1355 }, { "epoch": 2.029940119760479, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1977, "step": 1356 }, { "epoch": 2.031437125748503, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1943, "step": 1357 }, { "epoch": 2.032934131736527, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1926, "step": 1358 }, { "epoch": 2.034431137724551, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1863, "step": 1359 }, { "epoch": 2.035928143712575, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1936, "step": 1360 }, { "epoch": 2.037425149700599, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1934, "step": 1361 }, { "epoch": 2.038922155688623, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1982, "step": 1362 }, { "epoch": 2.0404191616766467, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1959, "step": 1363 }, { "epoch": 2.0419161676646707, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1966, "step": 1364 }, { "epoch": 2.0434131736526946, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1923, "step": 1365 }, { "epoch": 2.0449101796407185, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.194, "step": 1366 }, { "epoch": 2.0464071856287425, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1953, "step": 1367 }, { "epoch": 2.0479041916167664, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.193, "step": 1368 }, { "epoch": 2.0494011976047903, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.194, "step": 1369 }, { "epoch": 2.0508982035928143, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1945, "step": 1370 }, { "epoch": 2.052395209580838, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1969, "step": 1371 }, { "epoch": 2.053892215568862, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1943, "step": 1372 }, { "epoch": 2.055389221556886, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1939, "step": 1373 }, { "epoch": 2.05688622754491, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1909, "step": 1374 }, { "epoch": 2.058383233532934, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1855, "step": 1375 }, { "epoch": 2.059880239520958, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1896, "step": 1376 }, { "epoch": 2.061377245508982, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.188, "step": 1377 }, { "epoch": 2.062874251497006, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1959, "step": 1378 }, { "epoch": 2.06437125748503, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1956, "step": 1379 }, { "epoch": 2.065868263473054, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1839, "step": 1380 }, { "epoch": 2.067365269461078, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1939, "step": 1381 }, { "epoch": 2.068862275449102, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1906, "step": 1382 }, { "epoch": 2.070359281437126, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1945, "step": 1383 }, { "epoch": 2.07185628742515, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1945, "step": 1384 }, { "epoch": 2.0733532934131738, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1956, "step": 1385 }, { "epoch": 2.0748502994011977, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1963, "step": 1386 }, { "epoch": 2.0763473053892216, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1867, "step": 1387 }, { "epoch": 2.0778443113772456, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1957, "step": 1388 }, { "epoch": 2.0793413173652695, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1934, "step": 1389 }, { "epoch": 2.0808383233532934, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.193, "step": 1390 }, { "epoch": 2.0823353293413174, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1868, "step": 1391 }, { "epoch": 2.0838323353293413, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1901, "step": 1392 }, { "epoch": 2.0853293413173652, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1895, "step": 1393 }, { "epoch": 2.086826347305389, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1976, "step": 1394 }, { "epoch": 2.088323353293413, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1947, "step": 1395 }, { "epoch": 2.089820359281437, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1936, "step": 1396 }, { "epoch": 2.091317365269461, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1937, "step": 1397 }, { "epoch": 2.092814371257485, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.186, "step": 1398 }, { "epoch": 2.094311377245509, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1921, "step": 1399 }, { "epoch": 2.095808383233533, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1882, "step": 1400 }, { "epoch": 2.0973053892215567, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1893, "step": 1401 }, { "epoch": 2.0988023952095807, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1941, "step": 1402 }, { "epoch": 2.1002994011976046, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1909, "step": 1403 }, { "epoch": 2.1017964071856285, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1901, "step": 1404 }, { "epoch": 2.1032934131736525, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1978, "step": 1405 }, { "epoch": 2.1047904191616764, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1931, "step": 1406 }, { "epoch": 2.106287425149701, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1882, "step": 1407 }, { "epoch": 2.1077844311377247, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.186, "step": 1408 }, { "epoch": 2.1092814371257487, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1917, "step": 1409 }, { "epoch": 2.1107784431137726, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1915, "step": 1410 }, { "epoch": 2.1122754491017965, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1943, "step": 1411 }, { "epoch": 2.1137724550898205, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1916, "step": 1412 }, { "epoch": 2.1152694610778444, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1839, "step": 1413 }, { "epoch": 2.1167664670658684, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1864, "step": 1414 }, { "epoch": 2.1182634730538923, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1986, "step": 1415 }, { "epoch": 2.1197604790419162, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1941, "step": 1416 }, { "epoch": 2.12125748502994, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1936, "step": 1417 }, { "epoch": 2.122754491017964, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1944, "step": 1418 }, { "epoch": 2.124251497005988, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1981, "step": 1419 }, { "epoch": 2.125748502994012, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1893, "step": 1420 }, { "epoch": 2.127245508982036, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1952, "step": 1421 }, { "epoch": 2.12874251497006, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1851, "step": 1422 }, { "epoch": 2.1302395209580838, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1933, "step": 1423 }, { "epoch": 2.1317365269461077, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1958, "step": 1424 }, { "epoch": 2.1332335329341316, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1903, "step": 1425 }, { "epoch": 2.1347305389221556, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.191, "step": 1426 }, { "epoch": 2.1362275449101795, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.1894, "step": 1427 }, { "epoch": 2.1377245508982035, "grad_norm": 0.228515625, "learning_rate": 0.0008, "loss": 1.1844, "step": 1428 }, { "epoch": 2.1392215568862274, "grad_norm": 0.419921875, "learning_rate": 0.0008, "loss": 1.1981, "step": 1429 }, { "epoch": 2.1407185628742513, "grad_norm": 0.99609375, "learning_rate": 0.0008, "loss": 1.205, "step": 1430 }, { "epoch": 2.1422155688622753, "grad_norm": 1.875, "learning_rate": 0.0008, "loss": 1.2759, "step": 1431 }, { "epoch": 2.143712574850299, "grad_norm": 0.490234375, "learning_rate": 0.0008, "loss": 1.2266, "step": 1432 }, { "epoch": 2.1452095808383236, "grad_norm": 1.2109375, "learning_rate": 0.0008, "loss": 1.2652, "step": 1433 }, { "epoch": 2.1467065868263475, "grad_norm": 0.423828125, "learning_rate": 0.0008, "loss": 1.2499, "step": 1434 }, { "epoch": 2.1482035928143715, "grad_norm": 0.6171875, "learning_rate": 0.0008, "loss": 1.2517, "step": 1435 }, { "epoch": 2.1497005988023954, "grad_norm": 0.388671875, "learning_rate": 0.0008, "loss": 1.2465, "step": 1436 }, { "epoch": 2.1511976047904193, "grad_norm": 0.341796875, "learning_rate": 0.0008, "loss": 1.2361, "step": 1437 }, { "epoch": 2.1526946107784433, "grad_norm": 0.365234375, "learning_rate": 0.0008, "loss": 1.2366, "step": 1438 }, { "epoch": 2.154191616766467, "grad_norm": 0.322265625, "learning_rate": 0.0008, "loss": 1.2336, "step": 1439 }, { "epoch": 2.155688622754491, "grad_norm": 0.248046875, "learning_rate": 0.0008, "loss": 1.23, "step": 1440 }, { "epoch": 2.157185628742515, "grad_norm": 0.2490234375, "learning_rate": 0.0008, "loss": 1.2179, "step": 1441 }, { "epoch": 2.158682634730539, "grad_norm": 0.265625, "learning_rate": 0.0008, "loss": 1.2181, "step": 1442 }, { "epoch": 2.160179640718563, "grad_norm": 0.1943359375, "learning_rate": 0.0008, "loss": 1.2198, "step": 1443 }, { "epoch": 2.161676646706587, "grad_norm": 0.193359375, "learning_rate": 0.0008, "loss": 1.2163, "step": 1444 }, { "epoch": 2.163173652694611, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.212, "step": 1445 }, { "epoch": 2.1646706586826348, "grad_norm": 0.1630859375, "learning_rate": 0.0008, "loss": 1.2124, "step": 1446 }, { "epoch": 2.1661676646706587, "grad_norm": 0.162109375, "learning_rate": 0.0008, "loss": 1.2083, "step": 1447 }, { "epoch": 2.1676646706586826, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.2074, "step": 1448 }, { "epoch": 2.1691616766467066, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.2028, "step": 1449 }, { "epoch": 2.1706586826347305, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.2135, "step": 1450 }, { "epoch": 2.1721556886227544, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.2023, "step": 1451 }, { "epoch": 2.1736526946107784, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.2095, "step": 1452 }, { "epoch": 2.1751497005988023, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.209, "step": 1453 }, { "epoch": 2.1766467065868262, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.2018, "step": 1454 }, { "epoch": 2.17814371257485, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1958, "step": 1455 }, { "epoch": 2.179640718562874, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1998, "step": 1456 }, { "epoch": 2.181137724550898, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.2, "step": 1457 }, { "epoch": 2.182634730538922, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.2027, "step": 1458 }, { "epoch": 2.184131736526946, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1967, "step": 1459 }, { "epoch": 2.18562874251497, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.2036, "step": 1460 }, { "epoch": 2.187125748502994, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1984, "step": 1461 }, { "epoch": 2.1886227544910177, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.2036, "step": 1462 }, { "epoch": 2.190119760479042, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.205, "step": 1463 }, { "epoch": 2.191616766467066, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1921, "step": 1464 }, { "epoch": 2.19311377245509, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.199, "step": 1465 }, { "epoch": 2.194610778443114, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1914, "step": 1466 }, { "epoch": 2.196107784431138, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1985, "step": 1467 }, { "epoch": 2.197604790419162, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.2006, "step": 1468 }, { "epoch": 2.1991017964071857, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1987, "step": 1469 }, { "epoch": 2.2005988023952097, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1985, "step": 1470 }, { "epoch": 2.2020958083832336, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1966, "step": 1471 }, { "epoch": 2.2035928143712575, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1943, "step": 1472 }, { "epoch": 2.2050898203592815, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.2011, "step": 1473 }, { "epoch": 2.2065868263473054, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1975, "step": 1474 }, { "epoch": 2.2080838323353293, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.2034, "step": 1475 }, { "epoch": 2.2095808383233533, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1873, "step": 1476 }, { "epoch": 2.211077844311377, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.2029, "step": 1477 }, { "epoch": 2.212574850299401, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1927, "step": 1478 }, { "epoch": 2.214071856287425, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.188, "step": 1479 }, { "epoch": 2.215568862275449, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1931, "step": 1480 }, { "epoch": 2.217065868263473, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1884, "step": 1481 }, { "epoch": 2.218562874251497, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1996, "step": 1482 }, { "epoch": 2.220059880239521, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1932, "step": 1483 }, { "epoch": 2.2215568862275448, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.197, "step": 1484 }, { "epoch": 2.2230538922155687, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1909, "step": 1485 }, { "epoch": 2.2245508982035926, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1947, "step": 1486 }, { "epoch": 2.2260479041916166, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1912, "step": 1487 }, { "epoch": 2.2275449101796405, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1913, "step": 1488 }, { "epoch": 2.229041916167665, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1919, "step": 1489 }, { "epoch": 2.230538922155689, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.187, "step": 1490 }, { "epoch": 2.2320359281437128, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1928, "step": 1491 }, { "epoch": 2.2335329341317367, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1785, "step": 1492 }, { "epoch": 2.2350299401197606, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1893, "step": 1493 }, { "epoch": 2.2365269461077846, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1801, "step": 1494 }, { "epoch": 2.2380239520958085, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1915, "step": 1495 }, { "epoch": 2.2395209580838324, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1852, "step": 1496 }, { "epoch": 2.2410179640718564, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1882, "step": 1497 }, { "epoch": 2.2425149700598803, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1948, "step": 1498 }, { "epoch": 2.2440119760479043, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1855, "step": 1499 }, { "epoch": 2.245508982035928, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1905, "step": 1500 }, { "epoch": 2.247005988023952, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1941, "step": 1501 }, { "epoch": 2.248502994011976, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1944, "step": 1502 }, { "epoch": 2.25, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1907, "step": 1503 }, { "epoch": 2.251497005988024, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1924, "step": 1504 }, { "epoch": 2.252994011976048, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.189, "step": 1505 }, { "epoch": 2.254491017964072, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1919, "step": 1506 }, { "epoch": 2.2559880239520957, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1871, "step": 1507 }, { "epoch": 2.2574850299401197, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.196, "step": 1508 }, { "epoch": 2.2589820359281436, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1979, "step": 1509 }, { "epoch": 2.2604790419161676, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1822, "step": 1510 }, { "epoch": 2.2619760479041915, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1863, "step": 1511 }, { "epoch": 2.2634730538922154, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1856, "step": 1512 }, { "epoch": 2.2649700598802394, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1891, "step": 1513 }, { "epoch": 2.2664670658682633, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1848, "step": 1514 }, { "epoch": 2.2679640718562872, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1944, "step": 1515 }, { "epoch": 2.269461077844311, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1957, "step": 1516 }, { "epoch": 2.270958083832335, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1876, "step": 1517 }, { "epoch": 2.272455089820359, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.19, "step": 1518 }, { "epoch": 2.2739520958083834, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1802, "step": 1519 }, { "epoch": 2.2754491017964074, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1838, "step": 1520 }, { "epoch": 2.2769461077844313, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1959, "step": 1521 }, { "epoch": 2.2784431137724552, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1993, "step": 1522 }, { "epoch": 2.279940119760479, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1901, "step": 1523 }, { "epoch": 2.281437125748503, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1815, "step": 1524 }, { "epoch": 2.282934131736527, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1943, "step": 1525 }, { "epoch": 2.284431137724551, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.2008, "step": 1526 }, { "epoch": 2.285928143712575, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1897, "step": 1527 }, { "epoch": 2.287425149700599, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.192, "step": 1528 }, { "epoch": 2.288922155688623, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.192, "step": 1529 }, { "epoch": 2.2904191616766467, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1904, "step": 1530 }, { "epoch": 2.2919161676646707, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1796, "step": 1531 }, { "epoch": 2.2934131736526946, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1839, "step": 1532 }, { "epoch": 2.2949101796407185, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1858, "step": 1533 }, { "epoch": 2.2964071856287425, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1899, "step": 1534 }, { "epoch": 2.2979041916167664, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.2006, "step": 1535 }, { "epoch": 2.2994011976047903, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1873, "step": 1536 }, { "epoch": 2.3008982035928143, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.187, "step": 1537 }, { "epoch": 2.302395209580838, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1934, "step": 1538 }, { "epoch": 2.303892215568862, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1892, "step": 1539 }, { "epoch": 2.305389221556886, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1885, "step": 1540 }, { "epoch": 2.30688622754491, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.197, "step": 1541 }, { "epoch": 2.308383233532934, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1912, "step": 1542 }, { "epoch": 2.309880239520958, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1864, "step": 1543 }, { "epoch": 2.3113772455089823, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1952, "step": 1544 }, { "epoch": 2.312874251497006, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1903, "step": 1545 }, { "epoch": 2.31437125748503, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1918, "step": 1546 }, { "epoch": 2.315868263473054, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1836, "step": 1547 }, { "epoch": 2.317365269461078, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.187, "step": 1548 }, { "epoch": 2.318862275449102, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1908, "step": 1549 }, { "epoch": 2.320359281437126, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1937, "step": 1550 }, { "epoch": 2.32185628742515, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1874, "step": 1551 }, { "epoch": 2.3233532934131738, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1874, "step": 1552 }, { "epoch": 2.3248502994011977, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1883, "step": 1553 }, { "epoch": 2.3263473053892216, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1901, "step": 1554 }, { "epoch": 2.3278443113772456, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1942, "step": 1555 }, { "epoch": 2.3293413173652695, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1838, "step": 1556 }, { "epoch": 2.3308383233532934, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.187, "step": 1557 }, { "epoch": 2.3323353293413174, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1858, "step": 1558 }, { "epoch": 2.3338323353293413, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1824, "step": 1559 }, { "epoch": 2.3353293413173652, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1842, "step": 1560 }, { "epoch": 2.336826347305389, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1931, "step": 1561 }, { "epoch": 2.338323353293413, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1898, "step": 1562 }, { "epoch": 2.339820359281437, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.185, "step": 1563 }, { "epoch": 2.341317365269461, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1816, "step": 1564 }, { "epoch": 2.342814371257485, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1845, "step": 1565 }, { "epoch": 2.344311377245509, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1906, "step": 1566 }, { "epoch": 2.345808383233533, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1877, "step": 1567 }, { "epoch": 2.3473053892215567, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1931, "step": 1568 }, { "epoch": 2.3488023952095807, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1833, "step": 1569 }, { "epoch": 2.3502994011976046, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1893, "step": 1570 }, { "epoch": 2.3517964071856285, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1978, "step": 1571 }, { "epoch": 2.3532934131736525, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1813, "step": 1572 }, { "epoch": 2.3547904191616764, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1875, "step": 1573 }, { "epoch": 2.3562874251497004, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1831, "step": 1574 }, { "epoch": 2.3577844311377247, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1852, "step": 1575 }, { "epoch": 2.3592814371257487, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1816, "step": 1576 }, { "epoch": 2.3607784431137726, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1909, "step": 1577 }, { "epoch": 2.3622754491017965, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1883, "step": 1578 }, { "epoch": 2.3637724550898205, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1891, "step": 1579 }, { "epoch": 2.3652694610778444, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.189, "step": 1580 }, { "epoch": 2.3667664670658684, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1867, "step": 1581 }, { "epoch": 2.3682634730538923, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1893, "step": 1582 }, { "epoch": 2.3697604790419162, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1884, "step": 1583 }, { "epoch": 2.37125748502994, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.182, "step": 1584 }, { "epoch": 2.372754491017964, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.187, "step": 1585 }, { "epoch": 2.374251497005988, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1887, "step": 1586 }, { "epoch": 2.375748502994012, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1878, "step": 1587 }, { "epoch": 2.377245508982036, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1938, "step": 1588 }, { "epoch": 2.37874251497006, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1822, "step": 1589 }, { "epoch": 2.3802395209580838, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1834, "step": 1590 }, { "epoch": 2.3817365269461077, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1851, "step": 1591 }, { "epoch": 2.3832335329341316, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1846, "step": 1592 }, { "epoch": 2.3847305389221556, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1839, "step": 1593 }, { "epoch": 2.3862275449101795, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.182, "step": 1594 }, { "epoch": 2.3877245508982035, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1845, "step": 1595 }, { "epoch": 2.3892215568862274, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1937, "step": 1596 }, { "epoch": 2.3907185628742513, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1867, "step": 1597 }, { "epoch": 2.3922155688622753, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1821, "step": 1598 }, { "epoch": 2.3937125748502996, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1908, "step": 1599 }, { "epoch": 2.3952095808383236, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1937, "step": 1600 }, { "epoch": 2.3967065868263475, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1865, "step": 1601 }, { "epoch": 2.3982035928143715, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1889, "step": 1602 }, { "epoch": 2.3997005988023954, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1921, "step": 1603 }, { "epoch": 2.4011976047904193, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1867, "step": 1604 }, { "epoch": 2.4026946107784433, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1819, "step": 1605 }, { "epoch": 2.404191616766467, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1875, "step": 1606 }, { "epoch": 2.405688622754491, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1847, "step": 1607 }, { "epoch": 2.407185628742515, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1846, "step": 1608 }, { "epoch": 2.408682634730539, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1948, "step": 1609 }, { "epoch": 2.410179640718563, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.186, "step": 1610 }, { "epoch": 2.411676646706587, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1886, "step": 1611 }, { "epoch": 2.413173652694611, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1887, "step": 1612 }, { "epoch": 2.4146706586826348, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1831, "step": 1613 }, { "epoch": 2.4161676646706587, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.183, "step": 1614 }, { "epoch": 2.4176646706586826, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1863, "step": 1615 }, { "epoch": 2.4191616766467066, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.187, "step": 1616 }, { "epoch": 2.4206586826347305, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1896, "step": 1617 }, { "epoch": 2.4221556886227544, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1889, "step": 1618 }, { "epoch": 2.4236526946107784, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1882, "step": 1619 }, { "epoch": 2.4251497005988023, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1835, "step": 1620 }, { "epoch": 2.4266467065868262, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1882, "step": 1621 }, { "epoch": 2.42814371257485, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1873, "step": 1622 }, { "epoch": 2.429640718562874, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1948, "step": 1623 }, { "epoch": 2.431137724550898, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1879, "step": 1624 }, { "epoch": 2.432634730538922, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1909, "step": 1625 }, { "epoch": 2.434131736526946, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.178, "step": 1626 }, { "epoch": 2.43562874251497, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1953, "step": 1627 }, { "epoch": 2.437125748502994, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1878, "step": 1628 }, { "epoch": 2.4386227544910177, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1837, "step": 1629 }, { "epoch": 2.440119760479042, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1779, "step": 1630 }, { "epoch": 2.441616766467066, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1887, "step": 1631 }, { "epoch": 2.44311377245509, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1849, "step": 1632 }, { "epoch": 2.444610778443114, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1887, "step": 1633 }, { "epoch": 2.446107784431138, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1871, "step": 1634 }, { "epoch": 2.447604790419162, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1879, "step": 1635 }, { "epoch": 2.4491017964071857, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1874, "step": 1636 }, { "epoch": 2.4505988023952097, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1824, "step": 1637 }, { "epoch": 2.4520958083832336, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1778, "step": 1638 }, { "epoch": 2.4535928143712575, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1872, "step": 1639 }, { "epoch": 2.4550898203592815, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1913, "step": 1640 }, { "epoch": 2.4565868263473054, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1838, "step": 1641 }, { "epoch": 2.4580838323353293, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1818, "step": 1642 }, { "epoch": 2.4595808383233533, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1771, "step": 1643 }, { "epoch": 2.461077844311377, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1915, "step": 1644 }, { "epoch": 2.462574850299401, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1867, "step": 1645 }, { "epoch": 2.464071856287425, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1905, "step": 1646 }, { "epoch": 2.465568862275449, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1843, "step": 1647 }, { "epoch": 2.467065868263473, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1892, "step": 1648 }, { "epoch": 2.468562874251497, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.185, "step": 1649 }, { "epoch": 2.470059880239521, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1863, "step": 1650 }, { "epoch": 2.4715568862275448, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1839, "step": 1651 }, { "epoch": 2.4730538922155687, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1917, "step": 1652 }, { "epoch": 2.4745508982035926, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1821, "step": 1653 }, { "epoch": 2.4760479041916166, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1858, "step": 1654 }, { "epoch": 2.477544910179641, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1864, "step": 1655 }, { "epoch": 2.479041916167665, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1803, "step": 1656 }, { "epoch": 2.480538922155689, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1925, "step": 1657 }, { "epoch": 2.4820359281437128, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1813, "step": 1658 }, { "epoch": 2.4835329341317367, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1904, "step": 1659 }, { "epoch": 2.4850299401197606, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1896, "step": 1660 }, { "epoch": 2.4865269461077846, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.188, "step": 1661 }, { "epoch": 2.4880239520958085, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1813, "step": 1662 }, { "epoch": 2.4895209580838324, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1888, "step": 1663 }, { "epoch": 2.4910179640718564, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1751, "step": 1664 }, { "epoch": 2.4925149700598803, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1868, "step": 1665 }, { "epoch": 2.4940119760479043, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1874, "step": 1666 }, { "epoch": 2.495508982035928, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1818, "step": 1667 }, { "epoch": 2.497005988023952, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1792, "step": 1668 }, { "epoch": 2.498502994011976, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1833, "step": 1669 }, { "epoch": 2.5, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1899, "step": 1670 }, { "epoch": 2.501497005988024, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1855, "step": 1671 }, { "epoch": 2.502994011976048, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1812, "step": 1672 }, { "epoch": 2.504491017964072, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1922, "step": 1673 }, { "epoch": 2.5059880239520957, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.187, "step": 1674 }, { "epoch": 2.5074850299401197, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1827, "step": 1675 }, { "epoch": 2.5089820359281436, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1886, "step": 1676 }, { "epoch": 2.5104790419161676, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1913, "step": 1677 }, { "epoch": 2.5119760479041915, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1848, "step": 1678 }, { "epoch": 2.5134730538922154, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1834, "step": 1679 }, { "epoch": 2.5149700598802394, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1853, "step": 1680 }, { "epoch": 2.5164670658682633, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.187, "step": 1681 }, { "epoch": 2.5179640718562872, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.177, "step": 1682 }, { "epoch": 2.519461077844311, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1822, "step": 1683 }, { "epoch": 2.520958083832335, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1879, "step": 1684 }, { "epoch": 2.522455089820359, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1877, "step": 1685 }, { "epoch": 2.523952095808383, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1873, "step": 1686 }, { "epoch": 2.525449101796407, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1849, "step": 1687 }, { "epoch": 2.5269461077844313, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1831, "step": 1688 }, { "epoch": 2.5284431137724552, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1909, "step": 1689 }, { "epoch": 2.529940119760479, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.18, "step": 1690 }, { "epoch": 2.531437125748503, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1869, "step": 1691 }, { "epoch": 2.532934131736527, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.19, "step": 1692 }, { "epoch": 2.534431137724551, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1832, "step": 1693 }, { "epoch": 2.535928143712575, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1862, "step": 1694 }, { "epoch": 2.537425149700599, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1863, "step": 1695 }, { "epoch": 2.538922155688623, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.192, "step": 1696 }, { "epoch": 2.5404191616766467, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1827, "step": 1697 }, { "epoch": 2.5419161676646707, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1887, "step": 1698 }, { "epoch": 2.5434131736526946, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1739, "step": 1699 }, { "epoch": 2.5449101796407185, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1846, "step": 1700 }, { "epoch": 2.5464071856287425, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1784, "step": 1701 }, { "epoch": 2.5479041916167664, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.187, "step": 1702 }, { "epoch": 2.5494011976047903, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1772, "step": 1703 }, { "epoch": 2.5508982035928143, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1907, "step": 1704 }, { "epoch": 2.552395209580838, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1876, "step": 1705 }, { "epoch": 2.553892215568862, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1839, "step": 1706 }, { "epoch": 2.555389221556886, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1796, "step": 1707 }, { "epoch": 2.55688622754491, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1888, "step": 1708 }, { "epoch": 2.5583832335329344, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1837, "step": 1709 }, { "epoch": 2.5598802395209583, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1887, "step": 1710 }, { "epoch": 2.5613772455089823, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1904, "step": 1711 }, { "epoch": 2.562874251497006, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1858, "step": 1712 }, { "epoch": 2.56437125748503, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1877, "step": 1713 }, { "epoch": 2.565868263473054, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1824, "step": 1714 }, { "epoch": 2.567365269461078, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1892, "step": 1715 }, { "epoch": 2.568862275449102, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1901, "step": 1716 }, { "epoch": 2.570359281437126, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1904, "step": 1717 }, { "epoch": 2.57185628742515, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1868, "step": 1718 }, { "epoch": 2.5733532934131738, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1862, "step": 1719 }, { "epoch": 2.5748502994011977, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1782, "step": 1720 }, { "epoch": 2.5763473053892216, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1874, "step": 1721 }, { "epoch": 2.5778443113772456, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.186, "step": 1722 }, { "epoch": 2.5793413173652695, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1846, "step": 1723 }, { "epoch": 2.5808383233532934, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1858, "step": 1724 }, { "epoch": 2.5823353293413174, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1868, "step": 1725 }, { "epoch": 2.5838323353293413, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1804, "step": 1726 }, { "epoch": 2.5853293413173652, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.182, "step": 1727 }, { "epoch": 2.586826347305389, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1889, "step": 1728 }, { "epoch": 2.588323353293413, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1837, "step": 1729 }, { "epoch": 2.589820359281437, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1867, "step": 1730 }, { "epoch": 2.591317365269461, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1822, "step": 1731 }, { "epoch": 2.592814371257485, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.181, "step": 1732 }, { "epoch": 2.594311377245509, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1816, "step": 1733 }, { "epoch": 2.595808383233533, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.191, "step": 1734 }, { "epoch": 2.5973053892215567, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1893, "step": 1735 }, { "epoch": 2.5988023952095807, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1843, "step": 1736 }, { "epoch": 2.6002994011976046, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1866, "step": 1737 }, { "epoch": 2.6017964071856285, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1834, "step": 1738 }, { "epoch": 2.6032934131736525, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1893, "step": 1739 }, { "epoch": 2.6047904191616764, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1811, "step": 1740 }, { "epoch": 2.6062874251497004, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.189, "step": 1741 }, { "epoch": 2.6077844311377243, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1893, "step": 1742 }, { "epoch": 2.6092814371257482, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1838, "step": 1743 }, { "epoch": 2.6107784431137726, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.181, "step": 1744 }, { "epoch": 2.6122754491017965, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1785, "step": 1745 }, { "epoch": 2.6137724550898205, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1845, "step": 1746 }, { "epoch": 2.6152694610778444, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1825, "step": 1747 }, { "epoch": 2.6167664670658684, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1812, "step": 1748 }, { "epoch": 2.6182634730538923, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1811, "step": 1749 }, { "epoch": 2.6197604790419162, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1755, "step": 1750 }, { "epoch": 2.62125748502994, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1808, "step": 1751 }, { "epoch": 2.622754491017964, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1755, "step": 1752 }, { "epoch": 2.624251497005988, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1854, "step": 1753 }, { "epoch": 2.625748502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1842, "step": 1754 }, { "epoch": 2.627245508982036, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.187, "step": 1755 }, { "epoch": 2.62874251497006, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1825, "step": 1756 }, { "epoch": 2.6302395209580838, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1924, "step": 1757 }, { "epoch": 2.6317365269461077, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1828, "step": 1758 }, { "epoch": 2.6332335329341316, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1905, "step": 1759 }, { "epoch": 2.6347305389221556, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1835, "step": 1760 }, { "epoch": 2.6362275449101795, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1801, "step": 1761 }, { "epoch": 2.6377245508982035, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1871, "step": 1762 }, { "epoch": 2.6392215568862274, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1849, "step": 1763 }, { "epoch": 2.6407185628742518, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1822, "step": 1764 }, { "epoch": 2.6422155688622757, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1874, "step": 1765 }, { "epoch": 2.6437125748502996, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1794, "step": 1766 }, { "epoch": 2.6452095808383236, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1843, "step": 1767 }, { "epoch": 2.6467065868263475, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1862, "step": 1768 }, { "epoch": 2.6482035928143715, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1827, "step": 1769 }, { "epoch": 2.6497005988023954, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.184, "step": 1770 }, { "epoch": 2.6511976047904193, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1893, "step": 1771 }, { "epoch": 2.6526946107784433, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1823, "step": 1772 }, { "epoch": 2.654191616766467, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1814, "step": 1773 }, { "epoch": 2.655688622754491, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.185, "step": 1774 }, { "epoch": 2.657185628742515, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1882, "step": 1775 }, { "epoch": 2.658682634730539, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1881, "step": 1776 }, { "epoch": 2.660179640718563, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1912, "step": 1777 }, { "epoch": 2.661676646706587, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1795, "step": 1778 }, { "epoch": 2.663173652694611, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1891, "step": 1779 }, { "epoch": 2.6646706586826348, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1874, "step": 1780 }, { "epoch": 2.6661676646706587, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1862, "step": 1781 }, { "epoch": 2.6676646706586826, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1875, "step": 1782 }, { "epoch": 2.6691616766467066, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1809, "step": 1783 }, { "epoch": 2.6706586826347305, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1803, "step": 1784 }, { "epoch": 2.6721556886227544, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1839, "step": 1785 }, { "epoch": 2.6736526946107784, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1915, "step": 1786 }, { "epoch": 2.6751497005988023, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1808, "step": 1787 }, { "epoch": 2.6766467065868262, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1849, "step": 1788 }, { "epoch": 2.67814371257485, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1855, "step": 1789 }, { "epoch": 2.679640718562874, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1898, "step": 1790 }, { "epoch": 2.681137724550898, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1809, "step": 1791 }, { "epoch": 2.682634730538922, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1851, "step": 1792 }, { "epoch": 2.684131736526946, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1938, "step": 1793 }, { "epoch": 2.68562874251497, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1843, "step": 1794 }, { "epoch": 2.687125748502994, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1805, "step": 1795 }, { "epoch": 2.6886227544910177, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1851, "step": 1796 }, { "epoch": 2.6901197604790417, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1797, "step": 1797 }, { "epoch": 2.6916167664670656, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1814, "step": 1798 }, { "epoch": 2.69311377245509, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1902, "step": 1799 }, { "epoch": 2.694610778443114, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.179, "step": 1800 }, { "epoch": 2.696107784431138, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1783, "step": 1801 }, { "epoch": 2.697604790419162, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1787, "step": 1802 }, { "epoch": 2.6991017964071857, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1894, "step": 1803 }, { "epoch": 2.7005988023952097, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1855, "step": 1804 }, { "epoch": 2.7020958083832336, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1845, "step": 1805 }, { "epoch": 2.7035928143712575, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1827, "step": 1806 }, { "epoch": 2.7050898203592815, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1903, "step": 1807 }, { "epoch": 2.7065868263473054, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1753, "step": 1808 }, { "epoch": 2.7080838323353293, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.187, "step": 1809 }, { "epoch": 2.7095808383233533, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1876, "step": 1810 }, { "epoch": 2.711077844311377, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1766, "step": 1811 }, { "epoch": 2.712574850299401, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1816, "step": 1812 }, { "epoch": 2.714071856287425, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1825, "step": 1813 }, { "epoch": 2.715568862275449, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1904, "step": 1814 }, { "epoch": 2.717065868263473, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1805, "step": 1815 }, { "epoch": 2.718562874251497, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1896, "step": 1816 }, { "epoch": 2.720059880239521, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1796, "step": 1817 }, { "epoch": 2.7215568862275448, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1883, "step": 1818 }, { "epoch": 2.7230538922155687, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1784, "step": 1819 }, { "epoch": 2.724550898203593, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1865, "step": 1820 }, { "epoch": 2.726047904191617, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1855, "step": 1821 }, { "epoch": 2.727544910179641, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.185, "step": 1822 }, { "epoch": 2.729041916167665, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.185, "step": 1823 }, { "epoch": 2.730538922155689, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1836, "step": 1824 }, { "epoch": 2.7320359281437128, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1859, "step": 1825 }, { "epoch": 2.7335329341317367, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1846, "step": 1826 }, { "epoch": 2.7350299401197606, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1845, "step": 1827 }, { "epoch": 2.7365269461077846, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1817, "step": 1828 }, { "epoch": 2.7380239520958085, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1821, "step": 1829 }, { "epoch": 2.7395209580838324, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.183, "step": 1830 }, { "epoch": 2.7410179640718564, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1794, "step": 1831 }, { "epoch": 2.7425149700598803, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1798, "step": 1832 }, { "epoch": 2.7440119760479043, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1797, "step": 1833 }, { "epoch": 2.745508982035928, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1787, "step": 1834 }, { "epoch": 2.747005988023952, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1872, "step": 1835 }, { "epoch": 2.748502994011976, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1893, "step": 1836 }, { "epoch": 2.75, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1863, "step": 1837 }, { "epoch": 2.751497005988024, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1857, "step": 1838 }, { "epoch": 2.752994011976048, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1812, "step": 1839 }, { "epoch": 2.754491017964072, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1798, "step": 1840 }, { "epoch": 2.7559880239520957, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1804, "step": 1841 }, { "epoch": 2.7574850299401197, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1831, "step": 1842 }, { "epoch": 2.7589820359281436, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1801, "step": 1843 }, { "epoch": 2.7604790419161676, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1858, "step": 1844 }, { "epoch": 2.7619760479041915, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1807, "step": 1845 }, { "epoch": 2.7634730538922154, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1874, "step": 1846 }, { "epoch": 2.7649700598802394, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1818, "step": 1847 }, { "epoch": 2.7664670658682633, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1781, "step": 1848 }, { "epoch": 2.7679640718562872, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1867, "step": 1849 }, { "epoch": 2.769461077844311, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1813, "step": 1850 }, { "epoch": 2.770958083832335, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1781, "step": 1851 }, { "epoch": 2.772455089820359, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1795, "step": 1852 }, { "epoch": 2.773952095808383, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1887, "step": 1853 }, { "epoch": 2.775449101796407, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.18, "step": 1854 }, { "epoch": 2.7769461077844313, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1788, "step": 1855 }, { "epoch": 2.7784431137724552, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1769, "step": 1856 }, { "epoch": 2.779940119760479, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1805, "step": 1857 }, { "epoch": 2.781437125748503, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1799, "step": 1858 }, { "epoch": 2.782934131736527, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1787, "step": 1859 }, { "epoch": 2.784431137724551, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1841, "step": 1860 }, { "epoch": 2.785928143712575, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1733, "step": 1861 }, { "epoch": 2.787425149700599, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1865, "step": 1862 }, { "epoch": 2.788922155688623, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1776, "step": 1863 }, { "epoch": 2.7904191616766467, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1818, "step": 1864 }, { "epoch": 2.7919161676646707, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1833, "step": 1865 }, { "epoch": 2.7934131736526946, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1837, "step": 1866 }, { "epoch": 2.7949101796407185, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1839, "step": 1867 }, { "epoch": 2.7964071856287425, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1821, "step": 1868 }, { "epoch": 2.7979041916167664, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1764, "step": 1869 }, { "epoch": 2.7994011976047903, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1805, "step": 1870 }, { "epoch": 2.8008982035928143, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1782, "step": 1871 }, { "epoch": 2.802395209580838, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1757, "step": 1872 }, { "epoch": 2.803892215568862, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.19, "step": 1873 }, { "epoch": 2.805389221556886, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1822, "step": 1874 }, { "epoch": 2.80688622754491, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1874, "step": 1875 }, { "epoch": 2.8083832335329344, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1769, "step": 1876 }, { "epoch": 2.8098802395209583, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1805, "step": 1877 }, { "epoch": 2.8113772455089823, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1895, "step": 1878 }, { "epoch": 2.812874251497006, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1847, "step": 1879 }, { "epoch": 2.81437125748503, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1874, "step": 1880 }, { "epoch": 2.815868263473054, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.182, "step": 1881 }, { "epoch": 2.817365269461078, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1909, "step": 1882 }, { "epoch": 2.818862275449102, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1779, "step": 1883 }, { "epoch": 2.820359281437126, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1803, "step": 1884 }, { "epoch": 2.82185628742515, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1854, "step": 1885 }, { "epoch": 2.8233532934131738, "grad_norm": 0.169921875, "learning_rate": 0.0008, "loss": 1.1795, "step": 1886 }, { "epoch": 2.8248502994011977, "grad_norm": 0.201171875, "learning_rate": 0.0008, "loss": 1.1766, "step": 1887 }, { "epoch": 2.8263473053892216, "grad_norm": 0.322265625, "learning_rate": 0.0008, "loss": 1.1869, "step": 1888 }, { "epoch": 2.8278443113772456, "grad_norm": 0.71875, "learning_rate": 0.0008, "loss": 1.1896, "step": 1889 }, { "epoch": 2.8293413173652695, "grad_norm": 1.6328125, "learning_rate": 0.0008, "loss": 1.2451, "step": 1890 }, { "epoch": 2.8308383233532934, "grad_norm": 0.31640625, "learning_rate": 0.0008, "loss": 1.1999, "step": 1891 }, { "epoch": 2.8323353293413174, "grad_norm": 0.70703125, "learning_rate": 0.0008, "loss": 1.2046, "step": 1892 }, { "epoch": 2.8338323353293413, "grad_norm": 0.6015625, "learning_rate": 0.0008, "loss": 1.2139, "step": 1893 }, { "epoch": 2.8353293413173652, "grad_norm": 0.390625, "learning_rate": 0.0008, "loss": 1.2034, "step": 1894 }, { "epoch": 2.836826347305389, "grad_norm": 0.80078125, "learning_rate": 0.0008, "loss": 1.2113, "step": 1895 }, { "epoch": 2.838323353293413, "grad_norm": 0.6484375, "learning_rate": 0.0008, "loss": 1.2191, "step": 1896 }, { "epoch": 2.839820359281437, "grad_norm": 0.494140625, "learning_rate": 0.0008, "loss": 1.2157, "step": 1897 }, { "epoch": 2.841317365269461, "grad_norm": 0.44140625, "learning_rate": 0.0008, "loss": 1.2156, "step": 1898 }, { "epoch": 2.842814371257485, "grad_norm": 0.283203125, "learning_rate": 0.0008, "loss": 1.2043, "step": 1899 }, { "epoch": 2.844311377245509, "grad_norm": 0.29296875, "learning_rate": 0.0008, "loss": 1.2027, "step": 1900 }, { "epoch": 2.845808383233533, "grad_norm": 0.2255859375, "learning_rate": 0.0008, "loss": 1.2037, "step": 1901 }, { "epoch": 2.8473053892215567, "grad_norm": 0.197265625, "learning_rate": 0.0008, "loss": 1.1945, "step": 1902 }, { "epoch": 2.8488023952095807, "grad_norm": 0.171875, "learning_rate": 0.0008, "loss": 1.1947, "step": 1903 }, { "epoch": 2.8502994011976046, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1927, "step": 1904 }, { "epoch": 2.8517964071856285, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1911, "step": 1905 }, { "epoch": 2.8532934131736525, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1888, "step": 1906 }, { "epoch": 2.8547904191616764, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1883, "step": 1907 }, { "epoch": 2.8562874251497004, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1916, "step": 1908 }, { "epoch": 2.8577844311377243, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1928, "step": 1909 }, { "epoch": 2.8592814371257482, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1925, "step": 1910 }, { "epoch": 2.8607784431137726, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1893, "step": 1911 }, { "epoch": 2.8622754491017965, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1854, "step": 1912 }, { "epoch": 2.8637724550898205, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1877, "step": 1913 }, { "epoch": 2.8652694610778444, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1842, "step": 1914 }, { "epoch": 2.8667664670658684, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1872, "step": 1915 }, { "epoch": 2.8682634730538923, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1879, "step": 1916 }, { "epoch": 2.8697604790419162, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1901, "step": 1917 }, { "epoch": 2.87125748502994, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1848, "step": 1918 }, { "epoch": 2.872754491017964, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1843, "step": 1919 }, { "epoch": 2.874251497005988, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1872, "step": 1920 }, { "epoch": 2.875748502994012, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1794, "step": 1921 }, { "epoch": 2.877245508982036, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1844, "step": 1922 }, { "epoch": 2.87874251497006, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1907, "step": 1923 }, { "epoch": 2.8802395209580838, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1878, "step": 1924 }, { "epoch": 2.8817365269461077, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1843, "step": 1925 }, { "epoch": 2.8832335329341316, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.178, "step": 1926 }, { "epoch": 2.8847305389221556, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1822, "step": 1927 }, { "epoch": 2.8862275449101795, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1834, "step": 1928 }, { "epoch": 2.8877245508982035, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.186, "step": 1929 }, { "epoch": 2.8892215568862274, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1909, "step": 1930 }, { "epoch": 2.8907185628742518, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1883, "step": 1931 }, { "epoch": 2.8922155688622757, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1848, "step": 1932 }, { "epoch": 2.8937125748502996, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.184, "step": 1933 }, { "epoch": 2.8952095808383236, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1867, "step": 1934 }, { "epoch": 2.8967065868263475, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1754, "step": 1935 }, { "epoch": 2.8982035928143715, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1803, "step": 1936 }, { "epoch": 2.8997005988023954, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1808, "step": 1937 }, { "epoch": 2.9011976047904193, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1799, "step": 1938 }, { "epoch": 2.9026946107784433, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.181, "step": 1939 }, { "epoch": 2.904191616766467, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1814, "step": 1940 }, { "epoch": 2.905688622754491, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1832, "step": 1941 }, { "epoch": 2.907185628742515, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1739, "step": 1942 }, { "epoch": 2.908682634730539, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1809, "step": 1943 }, { "epoch": 2.910179640718563, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1867, "step": 1944 }, { "epoch": 2.911676646706587, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1907, "step": 1945 }, { "epoch": 2.913173652694611, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1828, "step": 1946 }, { "epoch": 2.9146706586826348, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1797, "step": 1947 }, { "epoch": 2.9161676646706587, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1848, "step": 1948 }, { "epoch": 2.9176646706586826, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1761, "step": 1949 }, { "epoch": 2.9191616766467066, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1902, "step": 1950 }, { "epoch": 2.9206586826347305, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1825, "step": 1951 }, { "epoch": 2.9221556886227544, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1784, "step": 1952 }, { "epoch": 2.9236526946107784, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1792, "step": 1953 }, { "epoch": 2.9251497005988023, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.175, "step": 1954 }, { "epoch": 2.9266467065868262, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1842, "step": 1955 }, { "epoch": 2.92814371257485, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1811, "step": 1956 }, { "epoch": 2.929640718562874, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1818, "step": 1957 }, { "epoch": 2.931137724550898, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1782, "step": 1958 }, { "epoch": 2.932634730538922, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1768, "step": 1959 }, { "epoch": 2.934131736526946, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1846, "step": 1960 }, { "epoch": 2.93562874251497, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1858, "step": 1961 }, { "epoch": 2.937125748502994, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1895, "step": 1962 }, { "epoch": 2.9386227544910177, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1838, "step": 1963 }, { "epoch": 2.9401197604790417, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1748, "step": 1964 }, { "epoch": 2.9416167664670656, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1831, "step": 1965 }, { "epoch": 2.94311377245509, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1788, "step": 1966 }, { "epoch": 2.944610778443114, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1804, "step": 1967 }, { "epoch": 2.946107784431138, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1756, "step": 1968 }, { "epoch": 2.947604790419162, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1852, "step": 1969 }, { "epoch": 2.9491017964071857, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1794, "step": 1970 }, { "epoch": 2.9505988023952097, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1787, "step": 1971 }, { "epoch": 2.9520958083832336, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1844, "step": 1972 }, { "epoch": 2.9535928143712575, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.177, "step": 1973 }, { "epoch": 2.9550898203592815, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1811, "step": 1974 }, { "epoch": 2.9565868263473054, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1809, "step": 1975 }, { "epoch": 2.9580838323353293, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.183, "step": 1976 }, { "epoch": 2.9595808383233533, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1787, "step": 1977 }, { "epoch": 2.961077844311377, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1835, "step": 1978 }, { "epoch": 2.962574850299401, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1846, "step": 1979 }, { "epoch": 2.964071856287425, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1834, "step": 1980 }, { "epoch": 2.965568862275449, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1786, "step": 1981 }, { "epoch": 2.967065868263473, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1827, "step": 1982 }, { "epoch": 2.968562874251497, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1846, "step": 1983 }, { "epoch": 2.970059880239521, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1791, "step": 1984 }, { "epoch": 2.9715568862275448, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.179, "step": 1985 }, { "epoch": 2.9730538922155687, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1833, "step": 1986 }, { "epoch": 2.974550898203593, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1856, "step": 1987 }, { "epoch": 2.976047904191617, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1822, "step": 1988 }, { "epoch": 2.977544910179641, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1882, "step": 1989 }, { "epoch": 2.979041916167665, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.184, "step": 1990 }, { "epoch": 2.980538922155689, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1791, "step": 1991 }, { "epoch": 2.9820359281437128, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1781, "step": 1992 }, { "epoch": 2.9835329341317367, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1777, "step": 1993 }, { "epoch": 2.9850299401197606, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1788, "step": 1994 }, { "epoch": 2.9865269461077846, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.179, "step": 1995 }, { "epoch": 2.9880239520958085, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1746, "step": 1996 }, { "epoch": 2.9895209580838324, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1774, "step": 1997 }, { "epoch": 2.9910179640718564, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1793, "step": 1998 }, { "epoch": 2.9925149700598803, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1913, "step": 1999 }, { "epoch": 2.9940119760479043, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1788, "step": 2000 }, { "epoch": 2.995508982035928, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1812, "step": 2001 }, { "epoch": 2.997005988023952, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1774, "step": 2002 }, { "epoch": 2.998502994011976, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1888, "step": 2003 }, { "epoch": 3.0, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1848, "step": 2004 }, { "epoch": 3.001497005988024, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1775, "step": 2005 }, { "epoch": 3.002994011976048, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1796, "step": 2006 }, { "epoch": 3.004491017964072, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1843, "step": 2007 }, { "epoch": 3.0059880239520957, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1879, "step": 2008 }, { "epoch": 3.0074850299401197, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1827, "step": 2009 }, { "epoch": 3.0089820359281436, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1779, "step": 2010 }, { "epoch": 3.0104790419161676, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1726, "step": 2011 }, { "epoch": 3.0119760479041915, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1801, "step": 2012 }, { "epoch": 3.0134730538922154, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1811, "step": 2013 }, { "epoch": 3.0149700598802394, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1767, "step": 2014 }, { "epoch": 3.0164670658682633, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1843, "step": 2015 }, { "epoch": 3.0179640718562872, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1747, "step": 2016 }, { "epoch": 3.019461077844311, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1788, "step": 2017 }, { "epoch": 3.020958083832335, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1785, "step": 2018 }, { "epoch": 3.0224550898203595, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.171, "step": 2019 }, { "epoch": 3.0239520958083834, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1845, "step": 2020 }, { "epoch": 3.0254491017964074, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1837, "step": 2021 }, { "epoch": 3.0269461077844313, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1806, "step": 2022 }, { "epoch": 3.0284431137724552, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1809, "step": 2023 }, { "epoch": 3.029940119760479, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1848, "step": 2024 }, { "epoch": 3.031437125748503, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1782, "step": 2025 }, { "epoch": 3.032934131736527, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1786, "step": 2026 }, { "epoch": 3.034431137724551, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1764, "step": 2027 }, { "epoch": 3.035928143712575, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1857, "step": 2028 }, { "epoch": 3.037425149700599, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1756, "step": 2029 }, { "epoch": 3.038922155688623, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1794, "step": 2030 }, { "epoch": 3.0404191616766467, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.177, "step": 2031 }, { "epoch": 3.0419161676646707, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1807, "step": 2032 }, { "epoch": 3.0434131736526946, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1807, "step": 2033 }, { "epoch": 3.0449101796407185, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1773, "step": 2034 }, { "epoch": 3.0464071856287425, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1882, "step": 2035 }, { "epoch": 3.0479041916167664, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1763, "step": 2036 }, { "epoch": 3.0494011976047903, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1784, "step": 2037 }, { "epoch": 3.0508982035928143, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1741, "step": 2038 }, { "epoch": 3.052395209580838, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1798, "step": 2039 }, { "epoch": 3.053892215568862, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1907, "step": 2040 }, { "epoch": 3.055389221556886, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1814, "step": 2041 }, { "epoch": 3.05688622754491, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1848, "step": 2042 }, { "epoch": 3.058383233532934, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1797, "step": 2043 }, { "epoch": 3.059880239520958, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1734, "step": 2044 }, { "epoch": 3.061377245508982, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1807, "step": 2045 }, { "epoch": 3.062874251497006, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1824, "step": 2046 }, { "epoch": 3.06437125748503, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1758, "step": 2047 }, { "epoch": 3.065868263473054, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1851, "step": 2048 }, { "epoch": 3.067365269461078, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1798, "step": 2049 }, { "epoch": 3.068862275449102, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1795, "step": 2050 }, { "epoch": 3.070359281437126, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1779, "step": 2051 }, { "epoch": 3.07185628742515, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1715, "step": 2052 }, { "epoch": 3.0733532934131738, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1796, "step": 2053 }, { "epoch": 3.0748502994011977, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1826, "step": 2054 }, { "epoch": 3.0763473053892216, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1817, "step": 2055 }, { "epoch": 3.0778443113772456, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1742, "step": 2056 }, { "epoch": 3.0793413173652695, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1801, "step": 2057 }, { "epoch": 3.0808383233532934, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1813, "step": 2058 }, { "epoch": 3.0823353293413174, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1811, "step": 2059 }, { "epoch": 3.0838323353293413, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1775, "step": 2060 }, { "epoch": 3.0853293413173652, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1868, "step": 2061 }, { "epoch": 3.086826347305389, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1824, "step": 2062 }, { "epoch": 3.088323353293413, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1796, "step": 2063 }, { "epoch": 3.089820359281437, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1829, "step": 2064 }, { "epoch": 3.091317365269461, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1758, "step": 2065 }, { "epoch": 3.092814371257485, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1802, "step": 2066 }, { "epoch": 3.094311377245509, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1823, "step": 2067 }, { "epoch": 3.095808383233533, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1808, "step": 2068 }, { "epoch": 3.0973053892215567, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1794, "step": 2069 }, { "epoch": 3.0988023952095807, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1798, "step": 2070 }, { "epoch": 3.1002994011976046, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1805, "step": 2071 }, { "epoch": 3.1017964071856285, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.1762, "step": 2072 }, { "epoch": 3.1032934131736525, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1714, "step": 2073 }, { "epoch": 3.1047904191616764, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1789, "step": 2074 }, { "epoch": 3.106287425149701, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1812, "step": 2075 }, { "epoch": 3.1077844311377247, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1792, "step": 2076 }, { "epoch": 3.1092814371257487, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.182, "step": 2077 }, { "epoch": 3.1107784431137726, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1872, "step": 2078 }, { "epoch": 3.1122754491017965, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1803, "step": 2079 }, { "epoch": 3.1137724550898205, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1795, "step": 2080 }, { "epoch": 3.1152694610778444, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1806, "step": 2081 }, { "epoch": 3.1167664670658684, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1792, "step": 2082 }, { "epoch": 3.1182634730538923, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1825, "step": 2083 }, { "epoch": 3.1197604790419162, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1732, "step": 2084 }, { "epoch": 3.12125748502994, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.182, "step": 2085 }, { "epoch": 3.122754491017964, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1848, "step": 2086 }, { "epoch": 3.124251497005988, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1771, "step": 2087 }, { "epoch": 3.125748502994012, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1805, "step": 2088 }, { "epoch": 3.127245508982036, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1764, "step": 2089 }, { "epoch": 3.12874251497006, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1797, "step": 2090 }, { "epoch": 3.1302395209580838, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1759, "step": 2091 }, { "epoch": 3.1317365269461077, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1763, "step": 2092 }, { "epoch": 3.1332335329341316, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1781, "step": 2093 }, { "epoch": 3.1347305389221556, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1765, "step": 2094 }, { "epoch": 3.1362275449101795, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1793, "step": 2095 }, { "epoch": 3.1377245508982035, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1753, "step": 2096 }, { "epoch": 3.1392215568862274, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1779, "step": 2097 }, { "epoch": 3.1407185628742513, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1788, "step": 2098 }, { "epoch": 3.1422155688622753, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1756, "step": 2099 }, { "epoch": 3.143712574850299, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1834, "step": 2100 }, { "epoch": 3.1452095808383236, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1811, "step": 2101 }, { "epoch": 3.1467065868263475, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1784, "step": 2102 }, { "epoch": 3.1482035928143715, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1851, "step": 2103 }, { "epoch": 3.1497005988023954, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1792, "step": 2104 }, { "epoch": 3.1511976047904193, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1804, "step": 2105 }, { "epoch": 3.1526946107784433, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1754, "step": 2106 }, { "epoch": 3.154191616766467, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1767, "step": 2107 }, { "epoch": 3.155688622754491, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.176, "step": 2108 }, { "epoch": 3.157185628742515, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1814, "step": 2109 }, { "epoch": 3.158682634730539, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1724, "step": 2110 }, { "epoch": 3.160179640718563, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1801, "step": 2111 }, { "epoch": 3.161676646706587, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1784, "step": 2112 }, { "epoch": 3.163173652694611, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1851, "step": 2113 }, { "epoch": 3.1646706586826348, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1815, "step": 2114 }, { "epoch": 3.1661676646706587, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1734, "step": 2115 }, { "epoch": 3.1676646706586826, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1785, "step": 2116 }, { "epoch": 3.1691616766467066, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1772, "step": 2117 }, { "epoch": 3.1706586826347305, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1814, "step": 2118 }, { "epoch": 3.1721556886227544, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1816, "step": 2119 }, { "epoch": 3.1736526946107784, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1771, "step": 2120 }, { "epoch": 3.1751497005988023, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1777, "step": 2121 }, { "epoch": 3.1766467065868262, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1839, "step": 2122 }, { "epoch": 3.17814371257485, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1765, "step": 2123 }, { "epoch": 3.179640718562874, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1825, "step": 2124 }, { "epoch": 3.181137724550898, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1825, "step": 2125 }, { "epoch": 3.182634730538922, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1772, "step": 2126 }, { "epoch": 3.184131736526946, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1787, "step": 2127 }, { "epoch": 3.18562874251497, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1821, "step": 2128 }, { "epoch": 3.187125748502994, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1792, "step": 2129 }, { "epoch": 3.1886227544910177, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1773, "step": 2130 }, { "epoch": 3.190119760479042, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.178, "step": 2131 }, { "epoch": 3.191616766467066, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1744, "step": 2132 }, { "epoch": 3.19311377245509, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1752, "step": 2133 }, { "epoch": 3.194610778443114, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1762, "step": 2134 }, { "epoch": 3.196107784431138, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1749, "step": 2135 }, { "epoch": 3.197604790419162, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1742, "step": 2136 }, { "epoch": 3.1991017964071857, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1725, "step": 2137 }, { "epoch": 3.2005988023952097, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1854, "step": 2138 }, { "epoch": 3.2020958083832336, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1817, "step": 2139 }, { "epoch": 3.2035928143712575, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1702, "step": 2140 }, { "epoch": 3.2050898203592815, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1785, "step": 2141 }, { "epoch": 3.2065868263473054, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1799, "step": 2142 }, { "epoch": 3.2080838323353293, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1773, "step": 2143 }, { "epoch": 3.2095808383233533, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1805, "step": 2144 }, { "epoch": 3.211077844311377, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1805, "step": 2145 }, { "epoch": 3.212574850299401, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.183, "step": 2146 }, { "epoch": 3.214071856287425, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.181, "step": 2147 }, { "epoch": 3.215568862275449, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1748, "step": 2148 }, { "epoch": 3.217065868263473, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1759, "step": 2149 }, { "epoch": 3.218562874251497, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1768, "step": 2150 }, { "epoch": 3.220059880239521, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1803, "step": 2151 }, { "epoch": 3.2215568862275448, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1827, "step": 2152 }, { "epoch": 3.2230538922155687, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1733, "step": 2153 }, { "epoch": 3.2245508982035926, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1779, "step": 2154 }, { "epoch": 3.2260479041916166, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1827, "step": 2155 }, { "epoch": 3.2275449101796405, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1818, "step": 2156 }, { "epoch": 3.229041916167665, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.178, "step": 2157 }, { "epoch": 3.230538922155689, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1727, "step": 2158 }, { "epoch": 3.2320359281437128, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1732, "step": 2159 }, { "epoch": 3.2335329341317367, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1828, "step": 2160 }, { "epoch": 3.2350299401197606, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1781, "step": 2161 }, { "epoch": 3.2365269461077846, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1762, "step": 2162 }, { "epoch": 3.2380239520958085, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1736, "step": 2163 }, { "epoch": 3.2395209580838324, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1843, "step": 2164 }, { "epoch": 3.2410179640718564, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1785, "step": 2165 }, { "epoch": 3.2425149700598803, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1746, "step": 2166 }, { "epoch": 3.2440119760479043, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1779, "step": 2167 }, { "epoch": 3.245508982035928, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1755, "step": 2168 }, { "epoch": 3.247005988023952, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1766, "step": 2169 }, { "epoch": 3.248502994011976, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1796, "step": 2170 }, { "epoch": 3.25, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1746, "step": 2171 }, { "epoch": 3.251497005988024, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1758, "step": 2172 }, { "epoch": 3.252994011976048, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1778, "step": 2173 }, { "epoch": 3.254491017964072, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1766, "step": 2174 }, { "epoch": 3.2559880239520957, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1801, "step": 2175 }, { "epoch": 3.2574850299401197, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1742, "step": 2176 }, { "epoch": 3.2589820359281436, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1758, "step": 2177 }, { "epoch": 3.2604790419161676, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1743, "step": 2178 }, { "epoch": 3.2619760479041915, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1859, "step": 2179 }, { "epoch": 3.2634730538922154, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1767, "step": 2180 }, { "epoch": 3.2649700598802394, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1796, "step": 2181 }, { "epoch": 3.2664670658682633, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1756, "step": 2182 }, { "epoch": 3.2679640718562872, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1789, "step": 2183 }, { "epoch": 3.269461077844311, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1796, "step": 2184 }, { "epoch": 3.270958083832335, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1811, "step": 2185 }, { "epoch": 3.272455089820359, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1732, "step": 2186 }, { "epoch": 3.2739520958083834, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1761, "step": 2187 }, { "epoch": 3.2754491017964074, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1772, "step": 2188 }, { "epoch": 3.2769461077844313, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1784, "step": 2189 }, { "epoch": 3.2784431137724552, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1743, "step": 2190 }, { "epoch": 3.279940119760479, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1662, "step": 2191 }, { "epoch": 3.281437125748503, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1671, "step": 2192 }, { "epoch": 3.282934131736527, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1824, "step": 2193 }, { "epoch": 3.284431137724551, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1787, "step": 2194 }, { "epoch": 3.285928143712575, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1704, "step": 2195 }, { "epoch": 3.287425149700599, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1758, "step": 2196 }, { "epoch": 3.288922155688623, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.173, "step": 2197 }, { "epoch": 3.2904191616766467, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1809, "step": 2198 }, { "epoch": 3.2919161676646707, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1722, "step": 2199 }, { "epoch": 3.2934131736526946, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1825, "step": 2200 }, { "epoch": 3.2949101796407185, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1796, "step": 2201 }, { "epoch": 3.2964071856287425, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1778, "step": 2202 }, { "epoch": 3.2979041916167664, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1735, "step": 2203 }, { "epoch": 3.2994011976047903, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.179, "step": 2204 }, { "epoch": 3.3008982035928143, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1829, "step": 2205 }, { "epoch": 3.302395209580838, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1733, "step": 2206 }, { "epoch": 3.303892215568862, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.177, "step": 2207 }, { "epoch": 3.305389221556886, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1741, "step": 2208 }, { "epoch": 3.30688622754491, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1789, "step": 2209 }, { "epoch": 3.308383233532934, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1835, "step": 2210 }, { "epoch": 3.309880239520958, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1797, "step": 2211 }, { "epoch": 3.3113772455089823, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1758, "step": 2212 }, { "epoch": 3.312874251497006, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1719, "step": 2213 }, { "epoch": 3.31437125748503, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1754, "step": 2214 }, { "epoch": 3.315868263473054, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1757, "step": 2215 }, { "epoch": 3.317365269461078, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1843, "step": 2216 }, { "epoch": 3.318862275449102, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1759, "step": 2217 }, { "epoch": 3.320359281437126, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1732, "step": 2218 }, { "epoch": 3.32185628742515, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1723, "step": 2219 }, { "epoch": 3.3233532934131738, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1794, "step": 2220 }, { "epoch": 3.3248502994011977, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1812, "step": 2221 }, { "epoch": 3.3263473053892216, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.18, "step": 2222 }, { "epoch": 3.3278443113772456, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1728, "step": 2223 }, { "epoch": 3.3293413173652695, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1739, "step": 2224 }, { "epoch": 3.3308383233532934, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1791, "step": 2225 }, { "epoch": 3.3323353293413174, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1732, "step": 2226 }, { "epoch": 3.3338323353293413, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1792, "step": 2227 }, { "epoch": 3.3353293413173652, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.175, "step": 2228 }, { "epoch": 3.336826347305389, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1789, "step": 2229 }, { "epoch": 3.338323353293413, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1707, "step": 2230 }, { "epoch": 3.339820359281437, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1766, "step": 2231 }, { "epoch": 3.341317365269461, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1738, "step": 2232 }, { "epoch": 3.342814371257485, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1719, "step": 2233 }, { "epoch": 3.344311377245509, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1797, "step": 2234 }, { "epoch": 3.345808383233533, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.178, "step": 2235 }, { "epoch": 3.3473053892215567, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1732, "step": 2236 }, { "epoch": 3.3488023952095807, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1717, "step": 2237 }, { "epoch": 3.3502994011976046, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1818, "step": 2238 }, { "epoch": 3.3517964071856285, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1789, "step": 2239 }, { "epoch": 3.3532934131736525, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1745, "step": 2240 }, { "epoch": 3.3547904191616764, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1743, "step": 2241 }, { "epoch": 3.3562874251497004, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1751, "step": 2242 }, { "epoch": 3.3577844311377247, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1774, "step": 2243 }, { "epoch": 3.3592814371257487, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1812, "step": 2244 }, { "epoch": 3.3607784431137726, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.172, "step": 2245 }, { "epoch": 3.3622754491017965, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1708, "step": 2246 }, { "epoch": 3.3637724550898205, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1784, "step": 2247 }, { "epoch": 3.3652694610778444, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1759, "step": 2248 }, { "epoch": 3.3667664670658684, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1759, "step": 2249 }, { "epoch": 3.3682634730538923, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.183, "step": 2250 }, { "epoch": 3.3697604790419162, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1802, "step": 2251 }, { "epoch": 3.37125748502994, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1726, "step": 2252 }, { "epoch": 3.372754491017964, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.1756, "step": 2253 }, { "epoch": 3.374251497005988, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1774, "step": 2254 }, { "epoch": 3.375748502994012, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1728, "step": 2255 }, { "epoch": 3.377245508982036, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1756, "step": 2256 }, { "epoch": 3.37874251497006, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1718, "step": 2257 }, { "epoch": 3.3802395209580838, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1773, "step": 2258 }, { "epoch": 3.3817365269461077, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1834, "step": 2259 }, { "epoch": 3.3832335329341316, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1754, "step": 2260 }, { "epoch": 3.3847305389221556, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1675, "step": 2261 }, { "epoch": 3.3862275449101795, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.166, "step": 2262 }, { "epoch": 3.3877245508982035, "grad_norm": 0.15234375, "learning_rate": 0.0008, "loss": 1.1785, "step": 2263 }, { "epoch": 3.3892215568862274, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1747, "step": 2264 }, { "epoch": 3.3907185628742513, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1714, "step": 2265 }, { "epoch": 3.3922155688622753, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.181, "step": 2266 }, { "epoch": 3.3937125748502996, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1781, "step": 2267 }, { "epoch": 3.3952095808383236, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1779, "step": 2268 }, { "epoch": 3.3967065868263475, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1732, "step": 2269 }, { "epoch": 3.3982035928143715, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1754, "step": 2270 }, { "epoch": 3.3997005988023954, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1756, "step": 2271 }, { "epoch": 3.4011976047904193, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1793, "step": 2272 }, { "epoch": 3.4026946107784433, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1767, "step": 2273 }, { "epoch": 3.404191616766467, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1695, "step": 2274 }, { "epoch": 3.405688622754491, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1709, "step": 2275 }, { "epoch": 3.407185628742515, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1738, "step": 2276 }, { "epoch": 3.408682634730539, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1665, "step": 2277 }, { "epoch": 3.410179640718563, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1753, "step": 2278 }, { "epoch": 3.411676646706587, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1796, "step": 2279 }, { "epoch": 3.413173652694611, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1796, "step": 2280 }, { "epoch": 3.4146706586826348, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1808, "step": 2281 }, { "epoch": 3.4161676646706587, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1722, "step": 2282 }, { "epoch": 3.4176646706586826, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1758, "step": 2283 }, { "epoch": 3.4191616766467066, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.173, "step": 2284 }, { "epoch": 3.4206586826347305, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1709, "step": 2285 }, { "epoch": 3.4221556886227544, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.178, "step": 2286 }, { "epoch": 3.4236526946107784, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1745, "step": 2287 }, { "epoch": 3.4251497005988023, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1794, "step": 2288 }, { "epoch": 3.4266467065868262, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1739, "step": 2289 }, { "epoch": 3.42814371257485, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1737, "step": 2290 }, { "epoch": 3.429640718562874, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1778, "step": 2291 }, { "epoch": 3.431137724550898, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1774, "step": 2292 }, { "epoch": 3.432634730538922, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1737, "step": 2293 }, { "epoch": 3.434131736526946, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1736, "step": 2294 }, { "epoch": 3.43562874251497, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1722, "step": 2295 }, { "epoch": 3.437125748502994, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1791, "step": 2296 }, { "epoch": 3.4386227544910177, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1789, "step": 2297 }, { "epoch": 3.440119760479042, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1798, "step": 2298 }, { "epoch": 3.441616766467066, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.176, "step": 2299 }, { "epoch": 3.44311377245509, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1806, "step": 2300 }, { "epoch": 3.444610778443114, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1718, "step": 2301 }, { "epoch": 3.446107784431138, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1687, "step": 2302 }, { "epoch": 3.447604790419162, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1742, "step": 2303 }, { "epoch": 3.4491017964071857, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1784, "step": 2304 }, { "epoch": 3.4505988023952097, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1756, "step": 2305 }, { "epoch": 3.4520958083832336, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1715, "step": 2306 }, { "epoch": 3.4535928143712575, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1709, "step": 2307 }, { "epoch": 3.4550898203592815, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1721, "step": 2308 }, { "epoch": 3.4565868263473054, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1842, "step": 2309 }, { "epoch": 3.4580838323353293, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1835, "step": 2310 }, { "epoch": 3.4595808383233533, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1761, "step": 2311 }, { "epoch": 3.461077844311377, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1756, "step": 2312 }, { "epoch": 3.462574850299401, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1753, "step": 2313 }, { "epoch": 3.464071856287425, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1773, "step": 2314 }, { "epoch": 3.465568862275449, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1696, "step": 2315 }, { "epoch": 3.467065868263473, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1746, "step": 2316 }, { "epoch": 3.468562874251497, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1728, "step": 2317 }, { "epoch": 3.470059880239521, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1763, "step": 2318 }, { "epoch": 3.4715568862275448, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1686, "step": 2319 }, { "epoch": 3.4730538922155687, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1728, "step": 2320 }, { "epoch": 3.4745508982035926, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1736, "step": 2321 }, { "epoch": 3.4760479041916166, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1753, "step": 2322 }, { "epoch": 3.477544910179641, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1808, "step": 2323 }, { "epoch": 3.479041916167665, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1715, "step": 2324 }, { "epoch": 3.480538922155689, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1839, "step": 2325 }, { "epoch": 3.4820359281437128, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1747, "step": 2326 }, { "epoch": 3.4835329341317367, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1778, "step": 2327 }, { "epoch": 3.4850299401197606, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1673, "step": 2328 }, { "epoch": 3.4865269461077846, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1832, "step": 2329 }, { "epoch": 3.4880239520958085, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.17, "step": 2330 }, { "epoch": 3.4895209580838324, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1751, "step": 2331 }, { "epoch": 3.4910179640718564, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.172, "step": 2332 }, { "epoch": 3.4925149700598803, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.172, "step": 2333 }, { "epoch": 3.4940119760479043, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1814, "step": 2334 }, { "epoch": 3.495508982035928, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1769, "step": 2335 }, { "epoch": 3.497005988023952, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1684, "step": 2336 }, { "epoch": 3.498502994011976, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1831, "step": 2337 }, { "epoch": 3.5, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1738, "step": 2338 }, { "epoch": 3.501497005988024, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1823, "step": 2339 }, { "epoch": 3.502994011976048, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1778, "step": 2340 }, { "epoch": 3.504491017964072, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1802, "step": 2341 }, { "epoch": 3.5059880239520957, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1745, "step": 2342 }, { "epoch": 3.5074850299401197, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1705, "step": 2343 }, { "epoch": 3.5089820359281436, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1705, "step": 2344 }, { "epoch": 3.5104790419161676, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1762, "step": 2345 }, { "epoch": 3.5119760479041915, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1699, "step": 2346 }, { "epoch": 3.5134730538922154, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1743, "step": 2347 }, { "epoch": 3.5149700598802394, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1783, "step": 2348 }, { "epoch": 3.5164670658682633, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1728, "step": 2349 }, { "epoch": 3.5179640718562872, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1739, "step": 2350 }, { "epoch": 3.519461077844311, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.178, "step": 2351 }, { "epoch": 3.520958083832335, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.184, "step": 2352 }, { "epoch": 3.522455089820359, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1738, "step": 2353 }, { "epoch": 3.523952095808383, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1737, "step": 2354 }, { "epoch": 3.525449101796407, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1819, "step": 2355 }, { "epoch": 3.5269461077844313, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1711, "step": 2356 }, { "epoch": 3.5284431137724552, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1685, "step": 2357 }, { "epoch": 3.529940119760479, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1759, "step": 2358 }, { "epoch": 3.531437125748503, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1716, "step": 2359 }, { "epoch": 3.532934131736527, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1773, "step": 2360 }, { "epoch": 3.534431137724551, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1782, "step": 2361 }, { "epoch": 3.535928143712575, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1748, "step": 2362 }, { "epoch": 3.537425149700599, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1752, "step": 2363 }, { "epoch": 3.538922155688623, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1691, "step": 2364 }, { "epoch": 3.5404191616766467, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1711, "step": 2365 }, { "epoch": 3.5419161676646707, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1825, "step": 2366 }, { "epoch": 3.5434131736526946, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1831, "step": 2367 }, { "epoch": 3.5449101796407185, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1745, "step": 2368 }, { "epoch": 3.5464071856287425, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1772, "step": 2369 }, { "epoch": 3.5479041916167664, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1674, "step": 2370 }, { "epoch": 3.5494011976047903, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1781, "step": 2371 }, { "epoch": 3.5508982035928143, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1764, "step": 2372 }, { "epoch": 3.552395209580838, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1799, "step": 2373 }, { "epoch": 3.553892215568862, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1786, "step": 2374 }, { "epoch": 3.555389221556886, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.178, "step": 2375 }, { "epoch": 3.55688622754491, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1824, "step": 2376 }, { "epoch": 3.5583832335329344, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1739, "step": 2377 }, { "epoch": 3.5598802395209583, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1771, "step": 2378 }, { "epoch": 3.5613772455089823, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1802, "step": 2379 }, { "epoch": 3.562874251497006, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1759, "step": 2380 }, { "epoch": 3.56437125748503, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1756, "step": 2381 }, { "epoch": 3.565868263473054, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.178, "step": 2382 }, { "epoch": 3.567365269461078, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1741, "step": 2383 }, { "epoch": 3.568862275449102, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1812, "step": 2384 }, { "epoch": 3.570359281437126, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.178, "step": 2385 }, { "epoch": 3.57185628742515, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1743, "step": 2386 }, { "epoch": 3.5733532934131738, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1732, "step": 2387 }, { "epoch": 3.5748502994011977, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1734, "step": 2388 }, { "epoch": 3.5763473053892216, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1696, "step": 2389 }, { "epoch": 3.5778443113772456, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1779, "step": 2390 }, { "epoch": 3.5793413173652695, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1778, "step": 2391 }, { "epoch": 3.5808383233532934, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1757, "step": 2392 }, { "epoch": 3.5823353293413174, "grad_norm": 0.1533203125, "learning_rate": 0.0008, "loss": 1.1754, "step": 2393 }, { "epoch": 3.5838323353293413, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.1786, "step": 2394 }, { "epoch": 3.5853293413173652, "grad_norm": 0.16015625, "learning_rate": 0.0008, "loss": 1.1803, "step": 2395 }, { "epoch": 3.586826347305389, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1738, "step": 2396 }, { "epoch": 3.588323353293413, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.165, "step": 2397 }, { "epoch": 3.589820359281437, "grad_norm": 0.17578125, "learning_rate": 0.0008, "loss": 1.1724, "step": 2398 }, { "epoch": 3.591317365269461, "grad_norm": 0.224609375, "learning_rate": 0.0008, "loss": 1.1817, "step": 2399 }, { "epoch": 3.592814371257485, "grad_norm": 0.287109375, "learning_rate": 0.0008, "loss": 1.1725, "step": 2400 }, { "epoch": 3.594311377245509, "grad_norm": 0.486328125, "learning_rate": 0.0008, "loss": 1.1851, "step": 2401 }, { "epoch": 3.595808383233533, "grad_norm": 1.0, "learning_rate": 0.0008, "loss": 1.1945, "step": 2402 }, { "epoch": 3.5973053892215567, "grad_norm": 1.6484375, "learning_rate": 0.0008, "loss": 1.24, "step": 2403 }, { "epoch": 3.5988023952095807, "grad_norm": 0.37890625, "learning_rate": 0.0008, "loss": 1.1938, "step": 2404 }, { "epoch": 3.6002994011976046, "grad_norm": 1.1484375, "learning_rate": 0.0008, "loss": 1.2344, "step": 2405 }, { "epoch": 3.6017964071856285, "grad_norm": 0.39453125, "learning_rate": 0.0008, "loss": 1.2008, "step": 2406 }, { "epoch": 3.6032934131736525, "grad_norm": 0.546875, "learning_rate": 0.0008, "loss": 1.2244, "step": 2407 }, { "epoch": 3.6047904191616764, "grad_norm": 0.314453125, "learning_rate": 0.0008, "loss": 1.2083, "step": 2408 }, { "epoch": 3.6062874251497004, "grad_norm": 0.24609375, "learning_rate": 0.0008, "loss": 1.198, "step": 2409 }, { "epoch": 3.6077844311377243, "grad_norm": 0.2333984375, "learning_rate": 0.0008, "loss": 1.2014, "step": 2410 }, { "epoch": 3.6092814371257482, "grad_norm": 0.1982421875, "learning_rate": 0.0008, "loss": 1.1989, "step": 2411 }, { "epoch": 3.6107784431137726, "grad_norm": 0.1767578125, "learning_rate": 0.0008, "loss": 1.1992, "step": 2412 }, { "epoch": 3.6122754491017965, "grad_norm": 0.171875, "learning_rate": 0.0008, "loss": 1.1956, "step": 2413 }, { "epoch": 3.6137724550898205, "grad_norm": 0.2421875, "learning_rate": 0.0008, "loss": 1.1881, "step": 2414 }, { "epoch": 3.6152694610778444, "grad_norm": 0.189453125, "learning_rate": 0.0008, "loss": 1.195, "step": 2415 }, { "epoch": 3.6167664670658684, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1901, "step": 2416 }, { "epoch": 3.6182634730538923, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1778, "step": 2417 }, { "epoch": 3.6197604790419162, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.1927, "step": 2418 }, { "epoch": 3.62125748502994, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.184, "step": 2419 }, { "epoch": 3.622754491017964, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1833, "step": 2420 }, { "epoch": 3.624251497005988, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1809, "step": 2421 }, { "epoch": 3.625748502994012, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1795, "step": 2422 }, { "epoch": 3.627245508982036, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1864, "step": 2423 }, { "epoch": 3.62874251497006, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1775, "step": 2424 }, { "epoch": 3.6302395209580838, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.185, "step": 2425 }, { "epoch": 3.6317365269461077, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.185, "step": 2426 }, { "epoch": 3.6332335329341316, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1895, "step": 2427 }, { "epoch": 3.6347305389221556, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1788, "step": 2428 }, { "epoch": 3.6362275449101795, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1795, "step": 2429 }, { "epoch": 3.6377245508982035, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1752, "step": 2430 }, { "epoch": 3.6392215568862274, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1735, "step": 2431 }, { "epoch": 3.6407185628742518, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1855, "step": 2432 }, { "epoch": 3.6422155688622757, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1738, "step": 2433 }, { "epoch": 3.6437125748502996, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.175, "step": 2434 }, { "epoch": 3.6452095808383236, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1782, "step": 2435 }, { "epoch": 3.6467065868263475, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1755, "step": 2436 }, { "epoch": 3.6482035928143715, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1661, "step": 2437 }, { "epoch": 3.6497005988023954, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1763, "step": 2438 }, { "epoch": 3.6511976047904193, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.171, "step": 2439 }, { "epoch": 3.6526946107784433, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1772, "step": 2440 }, { "epoch": 3.654191616766467, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1738, "step": 2441 }, { "epoch": 3.655688622754491, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1767, "step": 2442 }, { "epoch": 3.657185628742515, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1754, "step": 2443 }, { "epoch": 3.658682634730539, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1735, "step": 2444 }, { "epoch": 3.660179640718563, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1733, "step": 2445 }, { "epoch": 3.661676646706587, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1792, "step": 2446 }, { "epoch": 3.663173652694611, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1738, "step": 2447 }, { "epoch": 3.6646706586826348, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1787, "step": 2448 }, { "epoch": 3.6661676646706587, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1723, "step": 2449 }, { "epoch": 3.6676646706586826, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1706, "step": 2450 }, { "epoch": 3.6691616766467066, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1735, "step": 2451 }, { "epoch": 3.6706586826347305, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1801, "step": 2452 }, { "epoch": 3.6721556886227544, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1686, "step": 2453 }, { "epoch": 3.6736526946107784, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1707, "step": 2454 }, { "epoch": 3.6751497005988023, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1685, "step": 2455 }, { "epoch": 3.6766467065868262, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1617, "step": 2456 }, { "epoch": 3.67814371257485, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1713, "step": 2457 }, { "epoch": 3.679640718562874, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.181, "step": 2458 }, { "epoch": 3.681137724550898, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1706, "step": 2459 }, { "epoch": 3.682634730538922, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1726, "step": 2460 }, { "epoch": 3.684131736526946, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1736, "step": 2461 }, { "epoch": 3.68562874251497, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1752, "step": 2462 }, { "epoch": 3.687125748502994, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1707, "step": 2463 }, { "epoch": 3.6886227544910177, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1779, "step": 2464 }, { "epoch": 3.6901197604790417, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1761, "step": 2465 }, { "epoch": 3.6916167664670656, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1791, "step": 2466 }, { "epoch": 3.69311377245509, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1753, "step": 2467 }, { "epoch": 3.694610778443114, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1748, "step": 2468 }, { "epoch": 3.696107784431138, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1829, "step": 2469 }, { "epoch": 3.697604790419162, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1736, "step": 2470 }, { "epoch": 3.6991017964071857, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1752, "step": 2471 }, { "epoch": 3.7005988023952097, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1702, "step": 2472 }, { "epoch": 3.7020958083832336, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.172, "step": 2473 }, { "epoch": 3.7035928143712575, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1714, "step": 2474 }, { "epoch": 3.7050898203592815, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1745, "step": 2475 }, { "epoch": 3.7065868263473054, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1675, "step": 2476 }, { "epoch": 3.7080838323353293, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1662, "step": 2477 }, { "epoch": 3.7095808383233533, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1772, "step": 2478 }, { "epoch": 3.711077844311377, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1769, "step": 2479 }, { "epoch": 3.712574850299401, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1733, "step": 2480 }, { "epoch": 3.714071856287425, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.176, "step": 2481 }, { "epoch": 3.715568862275449, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1835, "step": 2482 }, { "epoch": 3.717065868263473, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.182, "step": 2483 }, { "epoch": 3.718562874251497, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.173, "step": 2484 }, { "epoch": 3.720059880239521, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1762, "step": 2485 }, { "epoch": 3.7215568862275448, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1737, "step": 2486 }, { "epoch": 3.7230538922155687, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1685, "step": 2487 }, { "epoch": 3.724550898203593, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1731, "step": 2488 }, { "epoch": 3.726047904191617, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1713, "step": 2489 }, { "epoch": 3.727544910179641, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.168, "step": 2490 }, { "epoch": 3.729041916167665, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1633, "step": 2491 }, { "epoch": 3.730538922155689, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1759, "step": 2492 }, { "epoch": 3.7320359281437128, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.169, "step": 2493 }, { "epoch": 3.7335329341317367, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1729, "step": 2494 }, { "epoch": 3.7350299401197606, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1702, "step": 2495 }, { "epoch": 3.7365269461077846, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1731, "step": 2496 }, { "epoch": 3.7380239520958085, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1754, "step": 2497 }, { "epoch": 3.7395209580838324, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1674, "step": 2498 }, { "epoch": 3.7410179640718564, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1773, "step": 2499 }, { "epoch": 3.7425149700598803, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1686, "step": 2500 }, { "epoch": 3.7440119760479043, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1743, "step": 2501 }, { "epoch": 3.745508982035928, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1712, "step": 2502 }, { "epoch": 3.747005988023952, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1732, "step": 2503 }, { "epoch": 3.748502994011976, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1741, "step": 2504 }, { "epoch": 3.75, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1766, "step": 2505 }, { "epoch": 3.751497005988024, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1733, "step": 2506 }, { "epoch": 3.752994011976048, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1707, "step": 2507 }, { "epoch": 3.754491017964072, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1724, "step": 2508 }, { "epoch": 3.7559880239520957, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1689, "step": 2509 }, { "epoch": 3.7574850299401197, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1753, "step": 2510 }, { "epoch": 3.7589820359281436, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1746, "step": 2511 }, { "epoch": 3.7604790419161676, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1692, "step": 2512 }, { "epoch": 3.7619760479041915, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1814, "step": 2513 }, { "epoch": 3.7634730538922154, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1742, "step": 2514 }, { "epoch": 3.7649700598802394, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1716, "step": 2515 }, { "epoch": 3.7664670658682633, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.1728, "step": 2516 }, { "epoch": 3.7679640718562872, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1756, "step": 2517 }, { "epoch": 3.769461077844311, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1734, "step": 2518 }, { "epoch": 3.770958083832335, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1751, "step": 2519 }, { "epoch": 3.772455089820359, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1694, "step": 2520 }, { "epoch": 3.773952095808383, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1784, "step": 2521 }, { "epoch": 3.775449101796407, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1704, "step": 2522 }, { "epoch": 3.7769461077844313, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1696, "step": 2523 }, { "epoch": 3.7784431137724552, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1694, "step": 2524 }, { "epoch": 3.779940119760479, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1702, "step": 2525 }, { "epoch": 3.781437125748503, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1701, "step": 2526 }, { "epoch": 3.782934131736527, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1714, "step": 2527 }, { "epoch": 3.784431137724551, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1746, "step": 2528 }, { "epoch": 3.785928143712575, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1776, "step": 2529 }, { "epoch": 3.787425149700599, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1674, "step": 2530 }, { "epoch": 3.788922155688623, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1682, "step": 2531 }, { "epoch": 3.7904191616766467, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1761, "step": 2532 }, { "epoch": 3.7919161676646707, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1727, "step": 2533 }, { "epoch": 3.7934131736526946, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1713, "step": 2534 }, { "epoch": 3.7949101796407185, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1706, "step": 2535 }, { "epoch": 3.7964071856287425, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1706, "step": 2536 }, { "epoch": 3.7979041916167664, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1829, "step": 2537 }, { "epoch": 3.7994011976047903, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.169, "step": 2538 }, { "epoch": 3.8008982035928143, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1814, "step": 2539 }, { "epoch": 3.802395209580838, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1663, "step": 2540 }, { "epoch": 3.803892215568862, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1709, "step": 2541 }, { "epoch": 3.805389221556886, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.167, "step": 2542 }, { "epoch": 3.80688622754491, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1723, "step": 2543 }, { "epoch": 3.8083832335329344, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1704, "step": 2544 }, { "epoch": 3.8098802395209583, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1757, "step": 2545 }, { "epoch": 3.8113772455089823, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1716, "step": 2546 }, { "epoch": 3.812874251497006, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.179, "step": 2547 }, { "epoch": 3.81437125748503, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1723, "step": 2548 }, { "epoch": 3.815868263473054, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1703, "step": 2549 }, { "epoch": 3.817365269461078, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1714, "step": 2550 }, { "epoch": 3.818862275449102, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.176, "step": 2551 }, { "epoch": 3.820359281437126, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1764, "step": 2552 }, { "epoch": 3.82185628742515, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.169, "step": 2553 }, { "epoch": 3.8233532934131738, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1763, "step": 2554 }, { "epoch": 3.8248502994011977, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1697, "step": 2555 }, { "epoch": 3.8263473053892216, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1669, "step": 2556 }, { "epoch": 3.8278443113772456, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1722, "step": 2557 }, { "epoch": 3.8293413173652695, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.165, "step": 2558 }, { "epoch": 3.8308383233532934, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1711, "step": 2559 }, { "epoch": 3.8323353293413174, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1667, "step": 2560 }, { "epoch": 3.8338323353293413, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1645, "step": 2561 }, { "epoch": 3.8353293413173652, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1736, "step": 2562 }, { "epoch": 3.836826347305389, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1743, "step": 2563 }, { "epoch": 3.838323353293413, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1659, "step": 2564 }, { "epoch": 3.839820359281437, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1734, "step": 2565 }, { "epoch": 3.841317365269461, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.176, "step": 2566 }, { "epoch": 3.842814371257485, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1731, "step": 2567 }, { "epoch": 3.844311377245509, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1751, "step": 2568 }, { "epoch": 3.845808383233533, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1704, "step": 2569 }, { "epoch": 3.8473053892215567, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1695, "step": 2570 }, { "epoch": 3.8488023952095807, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1724, "step": 2571 }, { "epoch": 3.8502994011976046, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1688, "step": 2572 }, { "epoch": 3.8517964071856285, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.171, "step": 2573 }, { "epoch": 3.8532934131736525, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1785, "step": 2574 }, { "epoch": 3.8547904191616764, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1703, "step": 2575 }, { "epoch": 3.8562874251497004, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.171, "step": 2576 }, { "epoch": 3.8577844311377243, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.172, "step": 2577 }, { "epoch": 3.8592814371257482, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1724, "step": 2578 }, { "epoch": 3.8607784431137726, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1699, "step": 2579 }, { "epoch": 3.8622754491017965, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1742, "step": 2580 }, { "epoch": 3.8637724550898205, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1762, "step": 2581 }, { "epoch": 3.8652694610778444, "grad_norm": 0.1904296875, "learning_rate": 0.0008, "loss": 1.1658, "step": 2582 }, { "epoch": 3.8667664670658684, "grad_norm": 0.265625, "learning_rate": 0.0008, "loss": 1.1766, "step": 2583 }, { "epoch": 3.8682634730538923, "grad_norm": 0.384765625, "learning_rate": 0.0008, "loss": 1.1762, "step": 2584 }, { "epoch": 3.8697604790419162, "grad_norm": 0.68359375, "learning_rate": 0.0008, "loss": 1.1842, "step": 2585 }, { "epoch": 3.87125748502994, "grad_norm": 1.140625, "learning_rate": 0.0008, "loss": 1.2016, "step": 2586 }, { "epoch": 3.872754491017964, "grad_norm": 0.91015625, "learning_rate": 0.0008, "loss": 1.195, "step": 2587 }, { "epoch": 3.874251497005988, "grad_norm": 0.328125, "learning_rate": 0.0008, "loss": 1.1825, "step": 2588 }, { "epoch": 3.875748502994012, "grad_norm": 0.48828125, "learning_rate": 0.0008, "loss": 1.195, "step": 2589 }, { "epoch": 3.877245508982036, "grad_norm": 0.2734375, "learning_rate": 0.0008, "loss": 1.1849, "step": 2590 }, { "epoch": 3.87874251497006, "grad_norm": 0.306640625, "learning_rate": 0.0008, "loss": 1.1929, "step": 2591 }, { "epoch": 3.8802395209580838, "grad_norm": 0.2421875, "learning_rate": 0.0008, "loss": 1.1809, "step": 2592 }, { "epoch": 3.8817365269461077, "grad_norm": 0.1962890625, "learning_rate": 0.0008, "loss": 1.1813, "step": 2593 }, { "epoch": 3.8832335329341316, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.1835, "step": 2594 }, { "epoch": 3.8847305389221556, "grad_norm": 0.1484375, "learning_rate": 0.0008, "loss": 1.1786, "step": 2595 }, { "epoch": 3.8862275449101795, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1772, "step": 2596 }, { "epoch": 3.8877245508982035, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1751, "step": 2597 }, { "epoch": 3.8892215568862274, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1808, "step": 2598 }, { "epoch": 3.8907185628742518, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1784, "step": 2599 }, { "epoch": 3.8922155688622757, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1724, "step": 2600 }, { "epoch": 3.8937125748502996, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1768, "step": 2601 }, { "epoch": 3.8952095808383236, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1761, "step": 2602 }, { "epoch": 3.8967065868263475, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1711, "step": 2603 }, { "epoch": 3.8982035928143715, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1731, "step": 2604 }, { "epoch": 3.8997005988023954, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1782, "step": 2605 }, { "epoch": 3.9011976047904193, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1786, "step": 2606 }, { "epoch": 3.9026946107784433, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1726, "step": 2607 }, { "epoch": 3.904191616766467, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1703, "step": 2608 }, { "epoch": 3.905688622754491, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1773, "step": 2609 }, { "epoch": 3.907185628742515, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1703, "step": 2610 }, { "epoch": 3.908682634730539, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1656, "step": 2611 }, { "epoch": 3.910179640718563, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1732, "step": 2612 }, { "epoch": 3.911676646706587, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1735, "step": 2613 }, { "epoch": 3.913173652694611, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1686, "step": 2614 }, { "epoch": 3.9146706586826348, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1787, "step": 2615 }, { "epoch": 3.9161676646706587, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1739, "step": 2616 }, { "epoch": 3.9176646706586826, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1768, "step": 2617 }, { "epoch": 3.9191616766467066, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1775, "step": 2618 }, { "epoch": 3.9206586826347305, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1748, "step": 2619 }, { "epoch": 3.9221556886227544, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1675, "step": 2620 }, { "epoch": 3.9236526946107784, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1699, "step": 2621 }, { "epoch": 3.9251497005988023, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1734, "step": 2622 }, { "epoch": 3.9266467065868262, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1738, "step": 2623 }, { "epoch": 3.92814371257485, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1694, "step": 2624 }, { "epoch": 3.929640718562874, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1726, "step": 2625 }, { "epoch": 3.931137724550898, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1719, "step": 2626 }, { "epoch": 3.932634730538922, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1727, "step": 2627 }, { "epoch": 3.934131736526946, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.17, "step": 2628 }, { "epoch": 3.93562874251497, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1735, "step": 2629 }, { "epoch": 3.937125748502994, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1718, "step": 2630 }, { "epoch": 3.9386227544910177, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1733, "step": 2631 }, { "epoch": 3.9401197604790417, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1728, "step": 2632 }, { "epoch": 3.9416167664670656, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1714, "step": 2633 }, { "epoch": 3.94311377245509, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1749, "step": 2634 }, { "epoch": 3.944610778443114, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1653, "step": 2635 }, { "epoch": 3.946107784431138, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1714, "step": 2636 }, { "epoch": 3.947604790419162, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1722, "step": 2637 }, { "epoch": 3.9491017964071857, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1649, "step": 2638 }, { "epoch": 3.9505988023952097, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1731, "step": 2639 }, { "epoch": 3.9520958083832336, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1622, "step": 2640 }, { "epoch": 3.9535928143712575, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1654, "step": 2641 }, { "epoch": 3.9550898203592815, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1689, "step": 2642 }, { "epoch": 3.9565868263473054, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.169, "step": 2643 }, { "epoch": 3.9580838323353293, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1737, "step": 2644 }, { "epoch": 3.9595808383233533, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.175, "step": 2645 }, { "epoch": 3.961077844311377, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1711, "step": 2646 }, { "epoch": 3.962574850299401, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1717, "step": 2647 }, { "epoch": 3.964071856287425, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1686, "step": 2648 }, { "epoch": 3.965568862275449, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1654, "step": 2649 }, { "epoch": 3.967065868263473, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.178, "step": 2650 }, { "epoch": 3.968562874251497, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1751, "step": 2651 }, { "epoch": 3.970059880239521, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1692, "step": 2652 }, { "epoch": 3.9715568862275448, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1743, "step": 2653 }, { "epoch": 3.9730538922155687, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1795, "step": 2654 }, { "epoch": 3.974550898203593, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1774, "step": 2655 }, { "epoch": 3.976047904191617, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1705, "step": 2656 }, { "epoch": 3.977544910179641, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1787, "step": 2657 }, { "epoch": 3.979041916167665, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1726, "step": 2658 }, { "epoch": 3.980538922155689, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.165, "step": 2659 }, { "epoch": 3.9820359281437128, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1711, "step": 2660 }, { "epoch": 3.9835329341317367, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1704, "step": 2661 }, { "epoch": 3.9850299401197606, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.17, "step": 2662 }, { "epoch": 3.9865269461077846, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1733, "step": 2663 }, { "epoch": 3.9880239520958085, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.169, "step": 2664 }, { "epoch": 3.9895209580838324, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1696, "step": 2665 }, { "epoch": 3.9910179640718564, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.18, "step": 2666 }, { "epoch": 3.9925149700598803, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1663, "step": 2667 }, { "epoch": 3.9940119760479043, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1693, "step": 2668 }, { "epoch": 3.995508982035928, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1683, "step": 2669 }, { "epoch": 3.997005988023952, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1745, "step": 2670 }, { "epoch": 3.998502994011976, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1649, "step": 2671 }, { "epoch": 4.0, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1702, "step": 2672 }, { "epoch": 4.001497005988024, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1743, "step": 2673 }, { "epoch": 4.002994011976048, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1709, "step": 2674 }, { "epoch": 4.004491017964072, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1735, "step": 2675 }, { "epoch": 4.005988023952096, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1728, "step": 2676 }, { "epoch": 4.00748502994012, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1692, "step": 2677 }, { "epoch": 4.008982035928144, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.171, "step": 2678 }, { "epoch": 4.0104790419161676, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1659, "step": 2679 }, { "epoch": 4.0119760479041915, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1722, "step": 2680 }, { "epoch": 4.013473053892215, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1771, "step": 2681 }, { "epoch": 4.014970059880239, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1743, "step": 2682 }, { "epoch": 4.016467065868263, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.176, "step": 2683 }, { "epoch": 4.017964071856287, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1657, "step": 2684 }, { "epoch": 4.019461077844311, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1695, "step": 2685 }, { "epoch": 4.020958083832335, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1688, "step": 2686 }, { "epoch": 4.022455089820359, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1737, "step": 2687 }, { "epoch": 4.023952095808383, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1599, "step": 2688 }, { "epoch": 4.025449101796407, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.17, "step": 2689 }, { "epoch": 4.026946107784431, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1746, "step": 2690 }, { "epoch": 4.028443113772455, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1708, "step": 2691 }, { "epoch": 4.029940119760479, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1705, "step": 2692 }, { "epoch": 4.031437125748503, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1654, "step": 2693 }, { "epoch": 4.032934131736527, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1732, "step": 2694 }, { "epoch": 4.0344311377245505, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1728, "step": 2695 }, { "epoch": 4.0359281437125745, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1683, "step": 2696 }, { "epoch": 4.037425149700598, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1724, "step": 2697 }, { "epoch": 4.038922155688622, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1634, "step": 2698 }, { "epoch": 4.040419161676646, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1735, "step": 2699 }, { "epoch": 4.04191616766467, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1717, "step": 2700 }, { "epoch": 4.043413173652695, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1666, "step": 2701 }, { "epoch": 4.044910179640719, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1659, "step": 2702 }, { "epoch": 4.046407185628743, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1679, "step": 2703 }, { "epoch": 4.047904191616767, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1803, "step": 2704 }, { "epoch": 4.049401197604791, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1744, "step": 2705 }, { "epoch": 4.050898203592815, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.172, "step": 2706 }, { "epoch": 4.052395209580839, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1713, "step": 2707 }, { "epoch": 4.053892215568863, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1728, "step": 2708 }, { "epoch": 4.0553892215568865, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1743, "step": 2709 }, { "epoch": 4.0568862275449105, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1618, "step": 2710 }, { "epoch": 4.058383233532934, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1694, "step": 2711 }, { "epoch": 4.059880239520958, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1634, "step": 2712 }, { "epoch": 4.061377245508982, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1679, "step": 2713 }, { "epoch": 4.062874251497006, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1737, "step": 2714 }, { "epoch": 4.06437125748503, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1684, "step": 2715 }, { "epoch": 4.065868263473054, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1723, "step": 2716 }, { "epoch": 4.067365269461078, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.173, "step": 2717 }, { "epoch": 4.068862275449102, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1679, "step": 2718 }, { "epoch": 4.070359281437126, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1663, "step": 2719 }, { "epoch": 4.07185628742515, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1788, "step": 2720 }, { "epoch": 4.073353293413174, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1678, "step": 2721 }, { "epoch": 4.074850299401198, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1627, "step": 2722 }, { "epoch": 4.076347305389222, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1725, "step": 2723 }, { "epoch": 4.077844311377246, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1735, "step": 2724 }, { "epoch": 4.0793413173652695, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1627, "step": 2725 }, { "epoch": 4.080838323353293, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1658, "step": 2726 }, { "epoch": 4.082335329341317, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1705, "step": 2727 }, { "epoch": 4.083832335329341, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1732, "step": 2728 }, { "epoch": 4.085329341317365, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.1729, "step": 2729 }, { "epoch": 4.086826347305389, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1642, "step": 2730 }, { "epoch": 4.088323353293413, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1716, "step": 2731 }, { "epoch": 4.089820359281437, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1649, "step": 2732 }, { "epoch": 4.091317365269461, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1698, "step": 2733 }, { "epoch": 4.092814371257485, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.17, "step": 2734 }, { "epoch": 4.094311377245509, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1727, "step": 2735 }, { "epoch": 4.095808383233533, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.17, "step": 2736 }, { "epoch": 4.097305389221557, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.164, "step": 2737 }, { "epoch": 4.098802395209581, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1721, "step": 2738 }, { "epoch": 4.100299401197605, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1679, "step": 2739 }, { "epoch": 4.1017964071856285, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1601, "step": 2740 }, { "epoch": 4.1032934131736525, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1654, "step": 2741 }, { "epoch": 4.104790419161676, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1728, "step": 2742 }, { "epoch": 4.1062874251497, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1683, "step": 2743 }, { "epoch": 4.107784431137724, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.169, "step": 2744 }, { "epoch": 4.109281437125748, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1686, "step": 2745 }, { "epoch": 4.110778443113772, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1742, "step": 2746 }, { "epoch": 4.112275449101796, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1683, "step": 2747 }, { "epoch": 4.11377245508982, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1677, "step": 2748 }, { "epoch": 4.115269461077844, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1623, "step": 2749 }, { "epoch": 4.116766467065868, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1598, "step": 2750 }, { "epoch": 4.118263473053892, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1741, "step": 2751 }, { "epoch": 4.119760479041916, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1774, "step": 2752 }, { "epoch": 4.12125748502994, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1661, "step": 2753 }, { "epoch": 4.122754491017964, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1627, "step": 2754 }, { "epoch": 4.124251497005988, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1728, "step": 2755 }, { "epoch": 4.125748502994012, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1683, "step": 2756 }, { "epoch": 4.127245508982036, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.168, "step": 2757 }, { "epoch": 4.12874251497006, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1746, "step": 2758 }, { "epoch": 4.130239520958084, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1751, "step": 2759 }, { "epoch": 4.131736526946108, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1747, "step": 2760 }, { "epoch": 4.133233532934132, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.162, "step": 2761 }, { "epoch": 4.134730538922156, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1757, "step": 2762 }, { "epoch": 4.13622754491018, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1781, "step": 2763 }, { "epoch": 4.137724550898204, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1706, "step": 2764 }, { "epoch": 4.139221556886228, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1634, "step": 2765 }, { "epoch": 4.140718562874252, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1711, "step": 2766 }, { "epoch": 4.142215568862276, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1717, "step": 2767 }, { "epoch": 4.1437125748503, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.175, "step": 2768 }, { "epoch": 4.145209580838324, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.172, "step": 2769 }, { "epoch": 4.1467065868263475, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.169, "step": 2770 }, { "epoch": 4.1482035928143715, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1679, "step": 2771 }, { "epoch": 4.149700598802395, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1721, "step": 2772 }, { "epoch": 4.151197604790419, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1694, "step": 2773 }, { "epoch": 4.152694610778443, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1631, "step": 2774 }, { "epoch": 4.154191616766467, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1742, "step": 2775 }, { "epoch": 4.155688622754491, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1761, "step": 2776 }, { "epoch": 4.157185628742515, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1733, "step": 2777 }, { "epoch": 4.158682634730539, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1678, "step": 2778 }, { "epoch": 4.160179640718563, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1673, "step": 2779 }, { "epoch": 4.161676646706587, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1646, "step": 2780 }, { "epoch": 4.163173652694611, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1773, "step": 2781 }, { "epoch": 4.164670658682635, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1657, "step": 2782 }, { "epoch": 4.166167664670659, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.166, "step": 2783 }, { "epoch": 4.167664670658683, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1711, "step": 2784 }, { "epoch": 4.169161676646707, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.17, "step": 2785 }, { "epoch": 4.1706586826347305, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.167, "step": 2786 }, { "epoch": 4.172155688622754, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1712, "step": 2787 }, { "epoch": 4.173652694610778, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1751, "step": 2788 }, { "epoch": 4.175149700598802, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1672, "step": 2789 }, { "epoch": 4.176646706586826, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1659, "step": 2790 }, { "epoch": 4.17814371257485, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1686, "step": 2791 }, { "epoch": 4.179640718562874, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1619, "step": 2792 }, { "epoch": 4.181137724550898, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1705, "step": 2793 }, { "epoch": 4.182634730538922, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1692, "step": 2794 }, { "epoch": 4.184131736526946, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1704, "step": 2795 }, { "epoch": 4.18562874251497, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1636, "step": 2796 }, { "epoch": 4.187125748502994, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.167, "step": 2797 }, { "epoch": 4.188622754491018, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1683, "step": 2798 }, { "epoch": 4.190119760479042, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1659, "step": 2799 }, { "epoch": 4.191616766467066, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1709, "step": 2800 }, { "epoch": 4.1931137724550895, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1673, "step": 2801 }, { "epoch": 4.1946107784431135, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1694, "step": 2802 }, { "epoch": 4.196107784431137, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1635, "step": 2803 }, { "epoch": 4.197604790419161, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1705, "step": 2804 }, { "epoch": 4.199101796407185, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1666, "step": 2805 }, { "epoch": 4.200598802395209, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1708, "step": 2806 }, { "epoch": 4.202095808383233, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1731, "step": 2807 }, { "epoch": 4.203592814371257, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1685, "step": 2808 }, { "epoch": 4.205089820359281, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1714, "step": 2809 }, { "epoch": 4.206586826347305, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1667, "step": 2810 }, { "epoch": 4.20808383233533, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1683, "step": 2811 }, { "epoch": 4.209580838323353, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1689, "step": 2812 }, { "epoch": 4.211077844311378, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1737, "step": 2813 }, { "epoch": 4.212574850299402, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1765, "step": 2814 }, { "epoch": 4.2140718562874255, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.163, "step": 2815 }, { "epoch": 4.2155688622754495, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1668, "step": 2816 }, { "epoch": 4.217065868263473, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1711, "step": 2817 }, { "epoch": 4.218562874251497, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1665, "step": 2818 }, { "epoch": 4.220059880239521, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1682, "step": 2819 }, { "epoch": 4.221556886227545, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1746, "step": 2820 }, { "epoch": 4.223053892215569, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1689, "step": 2821 }, { "epoch": 4.224550898203593, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1718, "step": 2822 }, { "epoch": 4.226047904191617, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1764, "step": 2823 }, { "epoch": 4.227544910179641, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1692, "step": 2824 }, { "epoch": 4.229041916167665, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1677, "step": 2825 }, { "epoch": 4.230538922155689, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1702, "step": 2826 }, { "epoch": 4.232035928143713, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1735, "step": 2827 }, { "epoch": 4.233532934131737, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1707, "step": 2828 }, { "epoch": 4.235029940119761, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1731, "step": 2829 }, { "epoch": 4.236526946107785, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1787, "step": 2830 }, { "epoch": 4.2380239520958085, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1683, "step": 2831 }, { "epoch": 4.2395209580838324, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1771, "step": 2832 }, { "epoch": 4.241017964071856, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1733, "step": 2833 }, { "epoch": 4.24251497005988, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1766, "step": 2834 }, { "epoch": 4.244011976047904, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1673, "step": 2835 }, { "epoch": 4.245508982035928, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1706, "step": 2836 }, { "epoch": 4.247005988023952, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1692, "step": 2837 }, { "epoch": 4.248502994011976, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1663, "step": 2838 }, { "epoch": 4.25, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1647, "step": 2839 }, { "epoch": 4.251497005988024, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1691, "step": 2840 }, { "epoch": 4.252994011976048, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1724, "step": 2841 }, { "epoch": 4.254491017964072, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1672, "step": 2842 }, { "epoch": 4.255988023952096, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1668, "step": 2843 }, { "epoch": 4.25748502994012, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1674, "step": 2844 }, { "epoch": 4.258982035928144, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1707, "step": 2845 }, { "epoch": 4.2604790419161676, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1698, "step": 2846 }, { "epoch": 4.2619760479041915, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1739, "step": 2847 }, { "epoch": 4.263473053892215, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1635, "step": 2848 }, { "epoch": 4.264970059880239, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1747, "step": 2849 }, { "epoch": 4.266467065868263, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1627, "step": 2850 }, { "epoch": 4.267964071856287, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.171, "step": 2851 }, { "epoch": 4.269461077844311, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1747, "step": 2852 }, { "epoch": 4.270958083832335, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1705, "step": 2853 }, { "epoch": 4.272455089820359, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1663, "step": 2854 }, { "epoch": 4.273952095808383, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1695, "step": 2855 }, { "epoch": 4.275449101796407, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1771, "step": 2856 }, { "epoch": 4.276946107784431, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1617, "step": 2857 }, { "epoch": 4.278443113772455, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1652, "step": 2858 }, { "epoch": 4.279940119760479, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1666, "step": 2859 }, { "epoch": 4.281437125748503, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1716, "step": 2860 }, { "epoch": 4.282934131736527, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1644, "step": 2861 }, { "epoch": 4.2844311377245505, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1652, "step": 2862 }, { "epoch": 4.2859281437125745, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1697, "step": 2863 }, { "epoch": 4.287425149700598, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1657, "step": 2864 }, { "epoch": 4.288922155688622, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1701, "step": 2865 }, { "epoch": 4.290419161676647, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1689, "step": 2866 }, { "epoch": 4.29191616766467, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1747, "step": 2867 }, { "epoch": 4.293413173652695, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1674, "step": 2868 }, { "epoch": 4.294910179640719, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1676, "step": 2869 }, { "epoch": 4.296407185628743, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1756, "step": 2870 }, { "epoch": 4.297904191616767, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1726, "step": 2871 }, { "epoch": 4.299401197604791, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.174, "step": 2872 }, { "epoch": 4.300898203592815, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1635, "step": 2873 }, { "epoch": 4.302395209580839, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1755, "step": 2874 }, { "epoch": 4.303892215568863, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.166, "step": 2875 }, { "epoch": 4.3053892215568865, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1719, "step": 2876 }, { "epoch": 4.3068862275449105, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1654, "step": 2877 }, { "epoch": 4.308383233532934, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.162, "step": 2878 }, { "epoch": 4.309880239520958, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1643, "step": 2879 }, { "epoch": 4.311377245508982, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1743, "step": 2880 }, { "epoch": 4.312874251497006, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1688, "step": 2881 }, { "epoch": 4.31437125748503, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1735, "step": 2882 }, { "epoch": 4.315868263473054, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1629, "step": 2883 }, { "epoch": 4.317365269461078, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1697, "step": 2884 }, { "epoch": 4.318862275449102, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1617, "step": 2885 }, { "epoch": 4.320359281437126, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1727, "step": 2886 }, { "epoch": 4.32185628742515, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1721, "step": 2887 }, { "epoch": 4.323353293413174, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1629, "step": 2888 }, { "epoch": 4.324850299401198, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1695, "step": 2889 }, { "epoch": 4.326347305389222, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.162, "step": 2890 }, { "epoch": 4.327844311377246, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1658, "step": 2891 }, { "epoch": 4.3293413173652695, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1608, "step": 2892 }, { "epoch": 4.330838323353293, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1706, "step": 2893 }, { "epoch": 4.332335329341317, "grad_norm": 0.16015625, "learning_rate": 0.0008, "loss": 1.1666, "step": 2894 }, { "epoch": 4.333832335329341, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1694, "step": 2895 }, { "epoch": 4.335329341317365, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1771, "step": 2896 }, { "epoch": 4.336826347305389, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1733, "step": 2897 }, { "epoch": 4.338323353293413, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1666, "step": 2898 }, { "epoch": 4.339820359281437, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1686, "step": 2899 }, { "epoch": 4.341317365269461, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1629, "step": 2900 }, { "epoch": 4.342814371257485, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1683, "step": 2901 }, { "epoch": 4.344311377245509, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1612, "step": 2902 }, { "epoch": 4.345808383233533, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1692, "step": 2903 }, { "epoch": 4.347305389221557, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1706, "step": 2904 }, { "epoch": 4.348802395209581, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1669, "step": 2905 }, { "epoch": 4.350299401197605, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1667, "step": 2906 }, { "epoch": 4.3517964071856285, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1667, "step": 2907 }, { "epoch": 4.3532934131736525, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1596, "step": 2908 }, { "epoch": 4.354790419161676, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1713, "step": 2909 }, { "epoch": 4.3562874251497, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1635, "step": 2910 }, { "epoch": 4.357784431137724, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1675, "step": 2911 }, { "epoch": 4.359281437125748, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.17, "step": 2912 }, { "epoch": 4.360778443113772, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1723, "step": 2913 }, { "epoch": 4.362275449101796, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1695, "step": 2914 }, { "epoch": 4.36377245508982, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.169, "step": 2915 }, { "epoch": 4.365269461077844, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1658, "step": 2916 }, { "epoch": 4.366766467065868, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1719, "step": 2917 }, { "epoch": 4.368263473053892, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.176, "step": 2918 }, { "epoch": 4.369760479041916, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.168, "step": 2919 }, { "epoch": 4.37125748502994, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1657, "step": 2920 }, { "epoch": 4.3727544910179645, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1696, "step": 2921 }, { "epoch": 4.374251497005988, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1687, "step": 2922 }, { "epoch": 4.375748502994012, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1703, "step": 2923 }, { "epoch": 4.3772455089820355, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1687, "step": 2924 }, { "epoch": 4.37874251497006, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1685, "step": 2925 }, { "epoch": 4.380239520958084, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1703, "step": 2926 }, { "epoch": 4.381736526946108, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1663, "step": 2927 }, { "epoch": 4.383233532934132, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1633, "step": 2928 }, { "epoch": 4.384730538922156, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1702, "step": 2929 }, { "epoch": 4.38622754491018, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1697, "step": 2930 }, { "epoch": 4.387724550898204, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1746, "step": 2931 }, { "epoch": 4.389221556886228, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1702, "step": 2932 }, { "epoch": 4.390718562874252, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1657, "step": 2933 }, { "epoch": 4.392215568862276, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1638, "step": 2934 }, { "epoch": 4.3937125748503, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1696, "step": 2935 }, { "epoch": 4.395209580838324, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1714, "step": 2936 }, { "epoch": 4.3967065868263475, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1649, "step": 2937 }, { "epoch": 4.3982035928143715, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.169, "step": 2938 }, { "epoch": 4.399700598802395, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1694, "step": 2939 }, { "epoch": 4.401197604790419, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1719, "step": 2940 }, { "epoch": 4.402694610778443, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1716, "step": 2941 }, { "epoch": 4.404191616766467, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1696, "step": 2942 }, { "epoch": 4.405688622754491, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1636, "step": 2943 }, { "epoch": 4.407185628742515, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1643, "step": 2944 }, { "epoch": 4.408682634730539, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1651, "step": 2945 }, { "epoch": 4.410179640718563, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1728, "step": 2946 }, { "epoch": 4.411676646706587, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1691, "step": 2947 }, { "epoch": 4.413173652694611, "grad_norm": 0.150390625, "learning_rate": 0.0008, "loss": 1.1686, "step": 2948 }, { "epoch": 4.414670658682635, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1662, "step": 2949 }, { "epoch": 4.416167664670659, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1645, "step": 2950 }, { "epoch": 4.417664670658683, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1752, "step": 2951 }, { "epoch": 4.419161676646707, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1684, "step": 2952 }, { "epoch": 4.4206586826347305, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1658, "step": 2953 }, { "epoch": 4.422155688622754, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1729, "step": 2954 }, { "epoch": 4.423652694610778, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1759, "step": 2955 }, { "epoch": 4.425149700598802, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1629, "step": 2956 }, { "epoch": 4.426646706586826, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1652, "step": 2957 }, { "epoch": 4.42814371257485, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1695, "step": 2958 }, { "epoch": 4.429640718562874, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1694, "step": 2959 }, { "epoch": 4.431137724550898, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1651, "step": 2960 }, { "epoch": 4.432634730538922, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1707, "step": 2961 }, { "epoch": 4.434131736526946, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1767, "step": 2962 }, { "epoch": 4.43562874251497, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1626, "step": 2963 }, { "epoch": 4.437125748502994, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1639, "step": 2964 }, { "epoch": 4.438622754491018, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1654, "step": 2965 }, { "epoch": 4.440119760479042, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1645, "step": 2966 }, { "epoch": 4.441616766467066, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1643, "step": 2967 }, { "epoch": 4.4431137724550895, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1705, "step": 2968 }, { "epoch": 4.4446107784431135, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1651, "step": 2969 }, { "epoch": 4.446107784431137, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1666, "step": 2970 }, { "epoch": 4.447604790419161, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1667, "step": 2971 }, { "epoch": 4.449101796407185, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1727, "step": 2972 }, { "epoch": 4.450598802395209, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.173, "step": 2973 }, { "epoch": 4.452095808383233, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.172, "step": 2974 }, { "epoch": 4.453592814371257, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1671, "step": 2975 }, { "epoch": 4.455089820359281, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1698, "step": 2976 }, { "epoch": 4.456586826347305, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1648, "step": 2977 }, { "epoch": 4.45808383233533, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.168, "step": 2978 }, { "epoch": 4.459580838323353, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1632, "step": 2979 }, { "epoch": 4.461077844311378, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1646, "step": 2980 }, { "epoch": 4.462574850299402, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1554, "step": 2981 }, { "epoch": 4.4640718562874255, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.164, "step": 2982 }, { "epoch": 4.4655688622754495, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1702, "step": 2983 }, { "epoch": 4.467065868263473, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1629, "step": 2984 }, { "epoch": 4.468562874251497, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1664, "step": 2985 }, { "epoch": 4.470059880239521, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.167, "step": 2986 }, { "epoch": 4.471556886227545, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1611, "step": 2987 }, { "epoch": 4.473053892215569, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1733, "step": 2988 }, { "epoch": 4.474550898203593, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1699, "step": 2989 }, { "epoch": 4.476047904191617, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1655, "step": 2990 }, { "epoch": 4.477544910179641, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1698, "step": 2991 }, { "epoch": 4.479041916167665, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1669, "step": 2992 }, { "epoch": 4.480538922155689, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1706, "step": 2993 }, { "epoch": 4.482035928143713, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1751, "step": 2994 }, { "epoch": 4.483532934131737, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1688, "step": 2995 }, { "epoch": 4.485029940119761, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1696, "step": 2996 }, { "epoch": 4.486526946107785, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1613, "step": 2997 }, { "epoch": 4.4880239520958085, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1698, "step": 2998 }, { "epoch": 4.4895209580838324, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1695, "step": 2999 }, { "epoch": 4.491017964071856, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1731, "step": 3000 }, { "epoch": 4.49251497005988, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1676, "step": 3001 }, { "epoch": 4.494011976047904, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1759, "step": 3002 }, { "epoch": 4.495508982035928, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1696, "step": 3003 }, { "epoch": 4.497005988023952, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1733, "step": 3004 }, { "epoch": 4.498502994011976, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1715, "step": 3005 }, { "epoch": 4.5, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1671, "step": 3006 }, { "epoch": 4.501497005988024, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1656, "step": 3007 }, { "epoch": 4.502994011976048, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1629, "step": 3008 }, { "epoch": 4.504491017964072, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1709, "step": 3009 }, { "epoch": 4.505988023952096, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1643, "step": 3010 }, { "epoch": 4.50748502994012, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1655, "step": 3011 }, { "epoch": 4.508982035928144, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1735, "step": 3012 }, { "epoch": 4.5104790419161676, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1668, "step": 3013 }, { "epoch": 4.5119760479041915, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1678, "step": 3014 }, { "epoch": 4.513473053892215, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1619, "step": 3015 }, { "epoch": 4.514970059880239, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.163, "step": 3016 }, { "epoch": 4.516467065868263, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1639, "step": 3017 }, { "epoch": 4.517964071856287, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1697, "step": 3018 }, { "epoch": 4.519461077844311, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1655, "step": 3019 }, { "epoch": 4.520958083832335, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1649, "step": 3020 }, { "epoch": 4.522455089820359, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.175, "step": 3021 }, { "epoch": 4.523952095808383, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1683, "step": 3022 }, { "epoch": 4.525449101796407, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1725, "step": 3023 }, { "epoch": 4.526946107784431, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1617, "step": 3024 }, { "epoch": 4.528443113772455, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1688, "step": 3025 }, { "epoch": 4.529940119760479, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.166, "step": 3026 }, { "epoch": 4.531437125748503, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1679, "step": 3027 }, { "epoch": 4.532934131736527, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.173, "step": 3028 }, { "epoch": 4.5344311377245505, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1705, "step": 3029 }, { "epoch": 4.5359281437125745, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.167, "step": 3030 }, { "epoch": 4.537425149700599, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1682, "step": 3031 }, { "epoch": 4.538922155688622, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1636, "step": 3032 }, { "epoch": 4.540419161676647, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1608, "step": 3033 }, { "epoch": 4.54191616766467, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1652, "step": 3034 }, { "epoch": 4.543413173652695, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.1697, "step": 3035 }, { "epoch": 4.544910179640718, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1692, "step": 3036 }, { "epoch": 4.546407185628743, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.169, "step": 3037 }, { "epoch": 4.547904191616767, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1739, "step": 3038 }, { "epoch": 4.549401197604791, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1746, "step": 3039 }, { "epoch": 4.550898203592815, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1607, "step": 3040 }, { "epoch": 4.552395209580839, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1612, "step": 3041 }, { "epoch": 4.553892215568863, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1604, "step": 3042 }, { "epoch": 4.5553892215568865, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1662, "step": 3043 }, { "epoch": 4.5568862275449105, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1652, "step": 3044 }, { "epoch": 4.558383233532934, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1689, "step": 3045 }, { "epoch": 4.559880239520958, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.174, "step": 3046 }, { "epoch": 4.561377245508982, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1716, "step": 3047 }, { "epoch": 4.562874251497006, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1642, "step": 3048 }, { "epoch": 4.56437125748503, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1681, "step": 3049 }, { "epoch": 4.565868263473054, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1729, "step": 3050 }, { "epoch": 4.567365269461078, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1577, "step": 3051 }, { "epoch": 4.568862275449102, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1674, "step": 3052 }, { "epoch": 4.570359281437126, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1718, "step": 3053 }, { "epoch": 4.57185628742515, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1728, "step": 3054 }, { "epoch": 4.573353293413174, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1677, "step": 3055 }, { "epoch": 4.574850299401198, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1677, "step": 3056 }, { "epoch": 4.576347305389222, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1714, "step": 3057 }, { "epoch": 4.577844311377246, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.173, "step": 3058 }, { "epoch": 4.5793413173652695, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1668, "step": 3059 }, { "epoch": 4.580838323353293, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1657, "step": 3060 }, { "epoch": 4.582335329341317, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1677, "step": 3061 }, { "epoch": 4.583832335329341, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1621, "step": 3062 }, { "epoch": 4.585329341317365, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1739, "step": 3063 }, { "epoch": 4.586826347305389, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1659, "step": 3064 }, { "epoch": 4.588323353293413, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1748, "step": 3065 }, { "epoch": 4.589820359281437, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1651, "step": 3066 }, { "epoch": 4.591317365269461, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.167, "step": 3067 }, { "epoch": 4.592814371257485, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1605, "step": 3068 }, { "epoch": 4.594311377245509, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1631, "step": 3069 }, { "epoch": 4.595808383233533, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1609, "step": 3070 }, { "epoch": 4.597305389221557, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1661, "step": 3071 }, { "epoch": 4.598802395209581, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1749, "step": 3072 }, { "epoch": 4.600299401197605, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1701, "step": 3073 }, { "epoch": 4.6017964071856285, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1682, "step": 3074 }, { "epoch": 4.6032934131736525, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1558, "step": 3075 }, { "epoch": 4.604790419161676, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1672, "step": 3076 }, { "epoch": 4.6062874251497, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1664, "step": 3077 }, { "epoch": 4.607784431137724, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1641, "step": 3078 }, { "epoch": 4.609281437125748, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1697, "step": 3079 }, { "epoch": 4.610778443113772, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1636, "step": 3080 }, { "epoch": 4.612275449101796, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1635, "step": 3081 }, { "epoch": 4.61377245508982, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1706, "step": 3082 }, { "epoch": 4.615269461077844, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1737, "step": 3083 }, { "epoch": 4.616766467065868, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1653, "step": 3084 }, { "epoch": 4.618263473053892, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1644, "step": 3085 }, { "epoch": 4.619760479041916, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1644, "step": 3086 }, { "epoch": 4.62125748502994, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1737, "step": 3087 }, { "epoch": 4.6227544910179645, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1691, "step": 3088 }, { "epoch": 4.624251497005988, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1714, "step": 3089 }, { "epoch": 4.625748502994012, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1664, "step": 3090 }, { "epoch": 4.6272455089820355, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1725, "step": 3091 }, { "epoch": 4.62874251497006, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1666, "step": 3092 }, { "epoch": 4.630239520958084, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1665, "step": 3093 }, { "epoch": 4.631736526946108, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.171, "step": 3094 }, { "epoch": 4.633233532934132, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1646, "step": 3095 }, { "epoch": 4.634730538922156, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1675, "step": 3096 }, { "epoch": 4.63622754491018, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1668, "step": 3097 }, { "epoch": 4.637724550898204, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1624, "step": 3098 }, { "epoch": 4.639221556886228, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1661, "step": 3099 }, { "epoch": 4.640718562874252, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1693, "step": 3100 }, { "epoch": 4.642215568862276, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1591, "step": 3101 }, { "epoch": 4.6437125748503, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1649, "step": 3102 }, { "epoch": 4.645209580838324, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1666, "step": 3103 }, { "epoch": 4.6467065868263475, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1674, "step": 3104 }, { "epoch": 4.6482035928143715, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1692, "step": 3105 }, { "epoch": 4.649700598802395, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1646, "step": 3106 }, { "epoch": 4.651197604790419, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.162, "step": 3107 }, { "epoch": 4.652694610778443, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1639, "step": 3108 }, { "epoch": 4.654191616766467, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1649, "step": 3109 }, { "epoch": 4.655688622754491, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1653, "step": 3110 }, { "epoch": 4.657185628742515, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1706, "step": 3111 }, { "epoch": 4.658682634730539, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1681, "step": 3112 }, { "epoch": 4.660179640718563, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1627, "step": 3113 }, { "epoch": 4.661676646706587, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1665, "step": 3114 }, { "epoch": 4.663173652694611, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1681, "step": 3115 }, { "epoch": 4.664670658682635, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.162, "step": 3116 }, { "epoch": 4.666167664670659, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1684, "step": 3117 }, { "epoch": 4.667664670658683, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1635, "step": 3118 }, { "epoch": 4.669161676646707, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1626, "step": 3119 }, { "epoch": 4.6706586826347305, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1644, "step": 3120 }, { "epoch": 4.672155688622754, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1724, "step": 3121 }, { "epoch": 4.673652694610778, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1717, "step": 3122 }, { "epoch": 4.675149700598802, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1676, "step": 3123 }, { "epoch": 4.676646706586826, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1738, "step": 3124 }, { "epoch": 4.67814371257485, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1696, "step": 3125 }, { "epoch": 4.679640718562874, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.169, "step": 3126 }, { "epoch": 4.681137724550898, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1619, "step": 3127 }, { "epoch": 4.682634730538922, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1707, "step": 3128 }, { "epoch": 4.684131736526946, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1588, "step": 3129 }, { "epoch": 4.68562874251497, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.168, "step": 3130 }, { "epoch": 4.687125748502994, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1681, "step": 3131 }, { "epoch": 4.688622754491018, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1706, "step": 3132 }, { "epoch": 4.690119760479042, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1674, "step": 3133 }, { "epoch": 4.691616766467066, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1641, "step": 3134 }, { "epoch": 4.6931137724550895, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1655, "step": 3135 }, { "epoch": 4.6946107784431135, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1663, "step": 3136 }, { "epoch": 4.696107784431137, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1751, "step": 3137 }, { "epoch": 4.697604790419161, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1675, "step": 3138 }, { "epoch": 4.699101796407185, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1548, "step": 3139 }, { "epoch": 4.700598802395209, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1629, "step": 3140 }, { "epoch": 4.702095808383233, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1635, "step": 3141 }, { "epoch": 4.703592814371257, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1649, "step": 3142 }, { "epoch": 4.705089820359282, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1619, "step": 3143 }, { "epoch": 4.706586826347305, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1636, "step": 3144 }, { "epoch": 4.70808383233533, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1755, "step": 3145 }, { "epoch": 4.709580838323353, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1597, "step": 3146 }, { "epoch": 4.711077844311378, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1711, "step": 3147 }, { "epoch": 4.712574850299401, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1664, "step": 3148 }, { "epoch": 4.7140718562874255, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1692, "step": 3149 }, { "epoch": 4.7155688622754495, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.161, "step": 3150 }, { "epoch": 4.717065868263473, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1669, "step": 3151 }, { "epoch": 4.718562874251497, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1654, "step": 3152 }, { "epoch": 4.720059880239521, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1579, "step": 3153 }, { "epoch": 4.721556886227545, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1623, "step": 3154 }, { "epoch": 4.723053892215569, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1585, "step": 3155 }, { "epoch": 4.724550898203593, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.165, "step": 3156 }, { "epoch": 4.726047904191617, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1652, "step": 3157 }, { "epoch": 4.727544910179641, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1679, "step": 3158 }, { "epoch": 4.729041916167665, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1629, "step": 3159 }, { "epoch": 4.730538922155689, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1618, "step": 3160 }, { "epoch": 4.732035928143713, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1634, "step": 3161 }, { "epoch": 4.733532934131737, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1709, "step": 3162 }, { "epoch": 4.735029940119761, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1709, "step": 3163 }, { "epoch": 4.736526946107785, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1647, "step": 3164 }, { "epoch": 4.7380239520958085, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1702, "step": 3165 }, { "epoch": 4.7395209580838324, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.168, "step": 3166 }, { "epoch": 4.741017964071856, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1602, "step": 3167 }, { "epoch": 4.74251497005988, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1724, "step": 3168 }, { "epoch": 4.744011976047904, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1743, "step": 3169 }, { "epoch": 4.745508982035928, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1657, "step": 3170 }, { "epoch": 4.747005988023952, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1656, "step": 3171 }, { "epoch": 4.748502994011976, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1647, "step": 3172 }, { "epoch": 4.75, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1656, "step": 3173 }, { "epoch": 4.751497005988024, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1684, "step": 3174 }, { "epoch": 4.752994011976048, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1633, "step": 3175 }, { "epoch": 4.754491017964072, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1675, "step": 3176 }, { "epoch": 4.755988023952096, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.168, "step": 3177 }, { "epoch": 4.75748502994012, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1658, "step": 3178 }, { "epoch": 4.758982035928144, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1692, "step": 3179 }, { "epoch": 4.7604790419161676, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1705, "step": 3180 }, { "epoch": 4.7619760479041915, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1725, "step": 3181 }, { "epoch": 4.763473053892215, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1713, "step": 3182 }, { "epoch": 4.764970059880239, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1662, "step": 3183 }, { "epoch": 4.766467065868263, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1702, "step": 3184 }, { "epoch": 4.767964071856287, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1666, "step": 3185 }, { "epoch": 4.769461077844311, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1624, "step": 3186 }, { "epoch": 4.770958083832335, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1699, "step": 3187 }, { "epoch": 4.772455089820359, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1688, "step": 3188 }, { "epoch": 4.773952095808383, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.164, "step": 3189 }, { "epoch": 4.775449101796407, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1676, "step": 3190 }, { "epoch": 4.776946107784431, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1743, "step": 3191 }, { "epoch": 4.778443113772455, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1705, "step": 3192 }, { "epoch": 4.779940119760479, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1645, "step": 3193 }, { "epoch": 4.781437125748503, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1635, "step": 3194 }, { "epoch": 4.782934131736527, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1635, "step": 3195 }, { "epoch": 4.7844311377245505, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.1697, "step": 3196 }, { "epoch": 4.7859281437125745, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1662, "step": 3197 }, { "epoch": 4.787425149700599, "grad_norm": 0.1630859375, "learning_rate": 0.0008, "loss": 1.1712, "step": 3198 }, { "epoch": 4.788922155688622, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1702, "step": 3199 }, { "epoch": 4.790419161676647, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1625, "step": 3200 }, { "epoch": 4.79191616766467, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1623, "step": 3201 }, { "epoch": 4.793413173652695, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1633, "step": 3202 }, { "epoch": 4.794910179640718, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.159, "step": 3203 }, { "epoch": 4.796407185628743, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1644, "step": 3204 }, { "epoch": 4.797904191616767, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1658, "step": 3205 }, { "epoch": 4.799401197604791, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1697, "step": 3206 }, { "epoch": 4.800898203592815, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.165, "step": 3207 }, { "epoch": 4.802395209580839, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1577, "step": 3208 }, { "epoch": 4.803892215568863, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.167, "step": 3209 }, { "epoch": 4.8053892215568865, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1696, "step": 3210 }, { "epoch": 4.8068862275449105, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1668, "step": 3211 }, { "epoch": 4.808383233532934, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1681, "step": 3212 }, { "epoch": 4.809880239520958, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1683, "step": 3213 }, { "epoch": 4.811377245508982, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1688, "step": 3214 }, { "epoch": 4.812874251497006, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1655, "step": 3215 }, { "epoch": 4.81437125748503, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1653, "step": 3216 }, { "epoch": 4.815868263473054, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1656, "step": 3217 }, { "epoch": 4.817365269461078, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1673, "step": 3218 }, { "epoch": 4.818862275449102, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1707, "step": 3219 }, { "epoch": 4.820359281437126, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1685, "step": 3220 }, { "epoch": 4.82185628742515, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1656, "step": 3221 }, { "epoch": 4.823353293413174, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1634, "step": 3222 }, { "epoch": 4.824850299401198, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1673, "step": 3223 }, { "epoch": 4.826347305389222, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.161, "step": 3224 }, { "epoch": 4.827844311377246, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.162, "step": 3225 }, { "epoch": 4.8293413173652695, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1656, "step": 3226 }, { "epoch": 4.830838323353293, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1593, "step": 3227 }, { "epoch": 4.832335329341317, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1696, "step": 3228 }, { "epoch": 4.833832335329341, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1686, "step": 3229 }, { "epoch": 4.835329341317365, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1619, "step": 3230 }, { "epoch": 4.836826347305389, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.162, "step": 3231 }, { "epoch": 4.838323353293413, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1608, "step": 3232 }, { "epoch": 4.839820359281437, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1682, "step": 3233 }, { "epoch": 4.841317365269461, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1605, "step": 3234 }, { "epoch": 4.842814371257485, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1587, "step": 3235 }, { "epoch": 4.844311377245509, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1689, "step": 3236 }, { "epoch": 4.845808383233533, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1647, "step": 3237 }, { "epoch": 4.847305389221557, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1666, "step": 3238 }, { "epoch": 4.848802395209581, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1671, "step": 3239 }, { "epoch": 4.850299401197605, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.165, "step": 3240 }, { "epoch": 4.8517964071856285, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1635, "step": 3241 }, { "epoch": 4.8532934131736525, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1659, "step": 3242 }, { "epoch": 4.854790419161676, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1631, "step": 3243 }, { "epoch": 4.8562874251497, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1643, "step": 3244 }, { "epoch": 4.857784431137724, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1673, "step": 3245 }, { "epoch": 4.859281437125748, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.156, "step": 3246 }, { "epoch": 4.860778443113772, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.161, "step": 3247 }, { "epoch": 4.862275449101796, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1672, "step": 3248 }, { "epoch": 4.86377245508982, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1671, "step": 3249 }, { "epoch": 4.865269461077844, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1615, "step": 3250 }, { "epoch": 4.866766467065868, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1596, "step": 3251 }, { "epoch": 4.868263473053892, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1671, "step": 3252 }, { "epoch": 4.869760479041916, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1658, "step": 3253 }, { "epoch": 4.87125748502994, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1654, "step": 3254 }, { "epoch": 4.8727544910179645, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1701, "step": 3255 }, { "epoch": 4.874251497005988, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.173, "step": 3256 }, { "epoch": 4.875748502994012, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1671, "step": 3257 }, { "epoch": 4.8772455089820355, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1647, "step": 3258 }, { "epoch": 4.87874251497006, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.166, "step": 3259 }, { "epoch": 4.880239520958084, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1652, "step": 3260 }, { "epoch": 4.881736526946108, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.168, "step": 3261 }, { "epoch": 4.883233532934132, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1606, "step": 3262 }, { "epoch": 4.884730538922156, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1633, "step": 3263 }, { "epoch": 4.88622754491018, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1619, "step": 3264 }, { "epoch": 4.887724550898204, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1683, "step": 3265 }, { "epoch": 4.889221556886228, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1594, "step": 3266 }, { "epoch": 4.890718562874252, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1672, "step": 3267 }, { "epoch": 4.892215568862276, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1696, "step": 3268 }, { "epoch": 4.8937125748503, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1656, "step": 3269 }, { "epoch": 4.895209580838324, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1642, "step": 3270 }, { "epoch": 4.8967065868263475, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1682, "step": 3271 }, { "epoch": 4.8982035928143715, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1652, "step": 3272 }, { "epoch": 4.899700598802395, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.159, "step": 3273 }, { "epoch": 4.901197604790419, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1628, "step": 3274 }, { "epoch": 4.902694610778443, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1656, "step": 3275 }, { "epoch": 4.904191616766467, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.162, "step": 3276 }, { "epoch": 4.905688622754491, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1629, "step": 3277 }, { "epoch": 4.907185628742515, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1627, "step": 3278 }, { "epoch": 4.908682634730539, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1672, "step": 3279 }, { "epoch": 4.910179640718563, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1626, "step": 3280 }, { "epoch": 4.911676646706587, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1627, "step": 3281 }, { "epoch": 4.913173652694611, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1646, "step": 3282 }, { "epoch": 4.914670658682635, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1653, "step": 3283 }, { "epoch": 4.916167664670659, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1558, "step": 3284 }, { "epoch": 4.917664670658683, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1644, "step": 3285 }, { "epoch": 4.919161676646707, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.174, "step": 3286 }, { "epoch": 4.9206586826347305, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1694, "step": 3287 }, { "epoch": 4.922155688622754, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1636, "step": 3288 }, { "epoch": 4.923652694610778, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1747, "step": 3289 }, { "epoch": 4.925149700598802, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1673, "step": 3290 }, { "epoch": 4.926646706586826, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1656, "step": 3291 }, { "epoch": 4.92814371257485, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.16, "step": 3292 }, { "epoch": 4.929640718562874, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1713, "step": 3293 }, { "epoch": 4.931137724550898, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1664, "step": 3294 }, { "epoch": 4.932634730538922, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1691, "step": 3295 }, { "epoch": 4.934131736526946, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1655, "step": 3296 }, { "epoch": 4.93562874251497, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1623, "step": 3297 }, { "epoch": 4.937125748502994, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1657, "step": 3298 }, { "epoch": 4.938622754491018, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1668, "step": 3299 }, { "epoch": 4.940119760479042, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1704, "step": 3300 }, { "epoch": 4.941616766467066, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.166, "step": 3301 }, { "epoch": 4.9431137724550895, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1714, "step": 3302 }, { "epoch": 4.9446107784431135, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1577, "step": 3303 }, { "epoch": 4.946107784431137, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1643, "step": 3304 }, { "epoch": 4.947604790419161, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1711, "step": 3305 }, { "epoch": 4.949101796407185, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.158, "step": 3306 }, { "epoch": 4.950598802395209, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1663, "step": 3307 }, { "epoch": 4.952095808383233, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1611, "step": 3308 }, { "epoch": 4.953592814371257, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.173, "step": 3309 }, { "epoch": 4.955089820359282, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1589, "step": 3310 }, { "epoch": 4.956586826347305, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.164, "step": 3311 }, { "epoch": 4.95808383233533, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1672, "step": 3312 }, { "epoch": 4.959580838323353, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1626, "step": 3313 }, { "epoch": 4.961077844311378, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1711, "step": 3314 }, { "epoch": 4.962574850299401, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1571, "step": 3315 }, { "epoch": 4.9640718562874255, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.169, "step": 3316 }, { "epoch": 4.9655688622754495, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.17, "step": 3317 }, { "epoch": 4.967065868263473, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1691, "step": 3318 }, { "epoch": 4.968562874251497, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1671, "step": 3319 }, { "epoch": 4.970059880239521, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1712, "step": 3320 }, { "epoch": 4.971556886227545, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1657, "step": 3321 }, { "epoch": 4.973053892215569, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1502, "step": 3322 }, { "epoch": 4.974550898203593, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1623, "step": 3323 }, { "epoch": 4.976047904191617, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1674, "step": 3324 }, { "epoch": 4.977544910179641, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1745, "step": 3325 }, { "epoch": 4.979041916167665, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1629, "step": 3326 }, { "epoch": 4.980538922155689, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1611, "step": 3327 }, { "epoch": 4.982035928143713, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.165, "step": 3328 }, { "epoch": 4.983532934131737, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1593, "step": 3329 }, { "epoch": 4.985029940119761, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.166, "step": 3330 }, { "epoch": 4.986526946107785, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1601, "step": 3331 }, { "epoch": 4.9880239520958085, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1646, "step": 3332 }, { "epoch": 4.9895209580838324, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1647, "step": 3333 }, { "epoch": 4.991017964071856, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.166, "step": 3334 }, { "epoch": 4.99251497005988, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1666, "step": 3335 }, { "epoch": 4.994011976047904, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.164, "step": 3336 }, { "epoch": 4.995508982035928, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1696, "step": 3337 }, { "epoch": 4.997005988023952, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1686, "step": 3338 }, { "epoch": 4.998502994011976, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1641, "step": 3339 }, { "epoch": 5.0, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1668, "step": 3340 }, { "epoch": 5.001497005988024, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1582, "step": 3341 }, { "epoch": 5.002994011976048, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1601, "step": 3342 }, { "epoch": 5.004491017964072, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1642, "step": 3343 }, { "epoch": 5.005988023952096, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1676, "step": 3344 }, { "epoch": 5.00748502994012, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1717, "step": 3345 }, { "epoch": 5.008982035928144, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1622, "step": 3346 }, { "epoch": 5.0104790419161676, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1643, "step": 3347 }, { "epoch": 5.0119760479041915, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1692, "step": 3348 }, { "epoch": 5.013473053892215, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1638, "step": 3349 }, { "epoch": 5.014970059880239, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1645, "step": 3350 }, { "epoch": 5.016467065868263, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1701, "step": 3351 }, { "epoch": 5.017964071856287, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.157, "step": 3352 }, { "epoch": 5.019461077844311, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1619, "step": 3353 }, { "epoch": 5.020958083832335, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1655, "step": 3354 }, { "epoch": 5.022455089820359, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1573, "step": 3355 }, { "epoch": 5.023952095808383, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1762, "step": 3356 }, { "epoch": 5.025449101796407, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1698, "step": 3357 }, { "epoch": 5.026946107784431, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.161, "step": 3358 }, { "epoch": 5.028443113772455, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1639, "step": 3359 }, { "epoch": 5.029940119760479, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.168, "step": 3360 }, { "epoch": 5.031437125748503, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1661, "step": 3361 }, { "epoch": 5.032934131736527, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1606, "step": 3362 }, { "epoch": 5.0344311377245505, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1598, "step": 3363 }, { "epoch": 5.0359281437125745, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1623, "step": 3364 }, { "epoch": 5.037425149700598, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1632, "step": 3365 }, { "epoch": 5.038922155688622, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1572, "step": 3366 }, { "epoch": 5.040419161676646, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1713, "step": 3367 }, { "epoch": 5.04191616766467, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.166, "step": 3368 }, { "epoch": 5.043413173652695, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1691, "step": 3369 }, { "epoch": 5.044910179640719, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1717, "step": 3370 }, { "epoch": 5.046407185628743, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1659, "step": 3371 }, { "epoch": 5.047904191616767, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1641, "step": 3372 }, { "epoch": 5.049401197604791, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1662, "step": 3373 }, { "epoch": 5.050898203592815, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1693, "step": 3374 }, { "epoch": 5.052395209580839, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1679, "step": 3375 }, { "epoch": 5.053892215568863, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1643, "step": 3376 }, { "epoch": 5.0553892215568865, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1663, "step": 3377 }, { "epoch": 5.0568862275449105, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1619, "step": 3378 }, { "epoch": 5.058383233532934, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1623, "step": 3379 }, { "epoch": 5.059880239520958, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1594, "step": 3380 }, { "epoch": 5.061377245508982, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1635, "step": 3381 }, { "epoch": 5.062874251497006, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1653, "step": 3382 }, { "epoch": 5.06437125748503, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1686, "step": 3383 }, { "epoch": 5.065868263473054, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1659, "step": 3384 }, { "epoch": 5.067365269461078, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1645, "step": 3385 }, { "epoch": 5.068862275449102, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1551, "step": 3386 }, { "epoch": 5.070359281437126, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1613, "step": 3387 }, { "epoch": 5.07185628742515, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1648, "step": 3388 }, { "epoch": 5.073353293413174, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1601, "step": 3389 }, { "epoch": 5.074850299401198, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1708, "step": 3390 }, { "epoch": 5.076347305389222, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1598, "step": 3391 }, { "epoch": 5.077844311377246, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1688, "step": 3392 }, { "epoch": 5.0793413173652695, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1632, "step": 3393 }, { "epoch": 5.080838323353293, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1649, "step": 3394 }, { "epoch": 5.082335329341317, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.159, "step": 3395 }, { "epoch": 5.083832335329341, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.164, "step": 3396 }, { "epoch": 5.085329341317365, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1611, "step": 3397 }, { "epoch": 5.086826347305389, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1651, "step": 3398 }, { "epoch": 5.088323353293413, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1558, "step": 3399 }, { "epoch": 5.089820359281437, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1615, "step": 3400 }, { "epoch": 5.091317365269461, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1693, "step": 3401 }, { "epoch": 5.092814371257485, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.168, "step": 3402 }, { "epoch": 5.094311377245509, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1744, "step": 3403 }, { "epoch": 5.095808383233533, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1562, "step": 3404 }, { "epoch": 5.097305389221557, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1656, "step": 3405 }, { "epoch": 5.098802395209581, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1655, "step": 3406 }, { "epoch": 5.100299401197605, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1549, "step": 3407 }, { "epoch": 5.1017964071856285, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.168, "step": 3408 }, { "epoch": 5.1032934131736525, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1604, "step": 3409 }, { "epoch": 5.104790419161676, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1647, "step": 3410 }, { "epoch": 5.1062874251497, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1709, "step": 3411 }, { "epoch": 5.107784431137724, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1618, "step": 3412 }, { "epoch": 5.109281437125748, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1612, "step": 3413 }, { "epoch": 5.110778443113772, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1657, "step": 3414 }, { "epoch": 5.112275449101796, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1679, "step": 3415 }, { "epoch": 5.11377245508982, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.165, "step": 3416 }, { "epoch": 5.115269461077844, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1633, "step": 3417 }, { "epoch": 5.116766467065868, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1663, "step": 3418 }, { "epoch": 5.118263473053892, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1645, "step": 3419 }, { "epoch": 5.119760479041916, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1636, "step": 3420 }, { "epoch": 5.12125748502994, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1633, "step": 3421 }, { "epoch": 5.122754491017964, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1624, "step": 3422 }, { "epoch": 5.124251497005988, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1629, "step": 3423 }, { "epoch": 5.125748502994012, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1685, "step": 3424 }, { "epoch": 5.127245508982036, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1576, "step": 3425 }, { "epoch": 5.12874251497006, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1643, "step": 3426 }, { "epoch": 5.130239520958084, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1566, "step": 3427 }, { "epoch": 5.131736526946108, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1652, "step": 3428 }, { "epoch": 5.133233532934132, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1653, "step": 3429 }, { "epoch": 5.134730538922156, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1596, "step": 3430 }, { "epoch": 5.13622754491018, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1574, "step": 3431 }, { "epoch": 5.137724550898204, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1614, "step": 3432 }, { "epoch": 5.139221556886228, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1617, "step": 3433 }, { "epoch": 5.140718562874252, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1699, "step": 3434 }, { "epoch": 5.142215568862276, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1679, "step": 3435 }, { "epoch": 5.1437125748503, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1617, "step": 3436 }, { "epoch": 5.145209580838324, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1602, "step": 3437 }, { "epoch": 5.1467065868263475, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1669, "step": 3438 }, { "epoch": 5.1482035928143715, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1605, "step": 3439 }, { "epoch": 5.149700598802395, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1646, "step": 3440 }, { "epoch": 5.151197604790419, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1663, "step": 3441 }, { "epoch": 5.152694610778443, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1648, "step": 3442 }, { "epoch": 5.154191616766467, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1627, "step": 3443 }, { "epoch": 5.155688622754491, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1548, "step": 3444 }, { "epoch": 5.157185628742515, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1664, "step": 3445 }, { "epoch": 5.158682634730539, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1631, "step": 3446 }, { "epoch": 5.160179640718563, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1651, "step": 3447 }, { "epoch": 5.161676646706587, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1661, "step": 3448 }, { "epoch": 5.163173652694611, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1638, "step": 3449 }, { "epoch": 5.164670658682635, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1748, "step": 3450 }, { "epoch": 5.166167664670659, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1658, "step": 3451 }, { "epoch": 5.167664670658683, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1654, "step": 3452 }, { "epoch": 5.169161676646707, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1674, "step": 3453 }, { "epoch": 5.1706586826347305, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1601, "step": 3454 }, { "epoch": 5.172155688622754, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1613, "step": 3455 }, { "epoch": 5.173652694610778, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.162, "step": 3456 }, { "epoch": 5.175149700598802, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.169, "step": 3457 }, { "epoch": 5.176646706586826, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1645, "step": 3458 }, { "epoch": 5.17814371257485, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.17, "step": 3459 }, { "epoch": 5.179640718562874, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1683, "step": 3460 }, { "epoch": 5.181137724550898, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1598, "step": 3461 }, { "epoch": 5.182634730538922, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1662, "step": 3462 }, { "epoch": 5.184131736526946, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1665, "step": 3463 }, { "epoch": 5.18562874251497, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1683, "step": 3464 }, { "epoch": 5.187125748502994, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1614, "step": 3465 }, { "epoch": 5.188622754491018, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1681, "step": 3466 }, { "epoch": 5.190119760479042, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1698, "step": 3467 }, { "epoch": 5.191616766467066, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1688, "step": 3468 }, { "epoch": 5.1931137724550895, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1646, "step": 3469 }, { "epoch": 5.1946107784431135, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1642, "step": 3470 }, { "epoch": 5.196107784431137, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1675, "step": 3471 }, { "epoch": 5.197604790419161, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1649, "step": 3472 }, { "epoch": 5.199101796407185, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1636, "step": 3473 }, { "epoch": 5.200598802395209, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.161, "step": 3474 }, { "epoch": 5.202095808383233, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1595, "step": 3475 }, { "epoch": 5.203592814371257, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1689, "step": 3476 }, { "epoch": 5.205089820359281, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1621, "step": 3477 }, { "epoch": 5.206586826347305, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1589, "step": 3478 }, { "epoch": 5.20808383233533, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1658, "step": 3479 }, { "epoch": 5.209580838323353, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1533, "step": 3480 }, { "epoch": 5.211077844311378, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1597, "step": 3481 }, { "epoch": 5.212574850299402, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1648, "step": 3482 }, { "epoch": 5.2140718562874255, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1689, "step": 3483 }, { "epoch": 5.2155688622754495, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1699, "step": 3484 }, { "epoch": 5.217065868263473, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1702, "step": 3485 }, { "epoch": 5.218562874251497, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.161, "step": 3486 }, { "epoch": 5.220059880239521, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.164, "step": 3487 }, { "epoch": 5.221556886227545, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1633, "step": 3488 }, { "epoch": 5.223053892215569, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1556, "step": 3489 }, { "epoch": 5.224550898203593, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1647, "step": 3490 }, { "epoch": 5.226047904191617, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1664, "step": 3491 }, { "epoch": 5.227544910179641, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1613, "step": 3492 }, { "epoch": 5.229041916167665, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1663, "step": 3493 }, { "epoch": 5.230538922155689, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1717, "step": 3494 }, { "epoch": 5.232035928143713, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1624, "step": 3495 }, { "epoch": 5.233532934131737, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1604, "step": 3496 }, { "epoch": 5.235029940119761, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1654, "step": 3497 }, { "epoch": 5.236526946107785, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1693, "step": 3498 }, { "epoch": 5.2380239520958085, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1629, "step": 3499 }, { "epoch": 5.2395209580838324, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1571, "step": 3500 }, { "epoch": 5.241017964071856, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1609, "step": 3501 }, { "epoch": 5.24251497005988, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1646, "step": 3502 }, { "epoch": 5.244011976047904, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.168, "step": 3503 }, { "epoch": 5.245508982035928, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1593, "step": 3504 }, { "epoch": 5.247005988023952, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1621, "step": 3505 }, { "epoch": 5.248502994011976, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1685, "step": 3506 }, { "epoch": 5.25, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1639, "step": 3507 }, { "epoch": 5.251497005988024, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.162, "step": 3508 }, { "epoch": 5.252994011976048, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1606, "step": 3509 }, { "epoch": 5.254491017964072, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1686, "step": 3510 }, { "epoch": 5.255988023952096, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1604, "step": 3511 }, { "epoch": 5.25748502994012, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1664, "step": 3512 }, { "epoch": 5.258982035928144, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1625, "step": 3513 }, { "epoch": 5.2604790419161676, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1613, "step": 3514 }, { "epoch": 5.2619760479041915, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1609, "step": 3515 }, { "epoch": 5.263473053892215, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.162, "step": 3516 }, { "epoch": 5.264970059880239, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1599, "step": 3517 }, { "epoch": 5.266467065868263, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1589, "step": 3518 }, { "epoch": 5.267964071856287, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1591, "step": 3519 }, { "epoch": 5.269461077844311, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.16, "step": 3520 }, { "epoch": 5.270958083832335, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1616, "step": 3521 }, { "epoch": 5.272455089820359, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1593, "step": 3522 }, { "epoch": 5.273952095808383, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1685, "step": 3523 }, { "epoch": 5.275449101796407, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1647, "step": 3524 }, { "epoch": 5.276946107784431, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.167, "step": 3525 }, { "epoch": 5.278443113772455, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1673, "step": 3526 }, { "epoch": 5.279940119760479, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.167, "step": 3527 }, { "epoch": 5.281437125748503, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1655, "step": 3528 }, { "epoch": 5.282934131736527, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1609, "step": 3529 }, { "epoch": 5.2844311377245505, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1625, "step": 3530 }, { "epoch": 5.2859281437125745, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1706, "step": 3531 }, { "epoch": 5.287425149700598, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1591, "step": 3532 }, { "epoch": 5.288922155688622, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1649, "step": 3533 }, { "epoch": 5.290419161676647, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1608, "step": 3534 }, { "epoch": 5.29191616766467, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1651, "step": 3535 }, { "epoch": 5.293413173652695, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1627, "step": 3536 }, { "epoch": 5.294910179640719, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1631, "step": 3537 }, { "epoch": 5.296407185628743, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1578, "step": 3538 }, { "epoch": 5.297904191616767, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1618, "step": 3539 }, { "epoch": 5.299401197604791, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1603, "step": 3540 }, { "epoch": 5.300898203592815, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1582, "step": 3541 }, { "epoch": 5.302395209580839, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1542, "step": 3542 }, { "epoch": 5.303892215568863, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1653, "step": 3543 }, { "epoch": 5.3053892215568865, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1629, "step": 3544 }, { "epoch": 5.3068862275449105, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1613, "step": 3545 }, { "epoch": 5.308383233532934, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1638, "step": 3546 }, { "epoch": 5.309880239520958, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1701, "step": 3547 }, { "epoch": 5.311377245508982, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1565, "step": 3548 }, { "epoch": 5.312874251497006, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1663, "step": 3549 }, { "epoch": 5.31437125748503, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1564, "step": 3550 }, { "epoch": 5.315868263473054, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1608, "step": 3551 }, { "epoch": 5.317365269461078, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1611, "step": 3552 }, { "epoch": 5.318862275449102, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.164, "step": 3553 }, { "epoch": 5.320359281437126, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.151, "step": 3554 }, { "epoch": 5.32185628742515, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.161, "step": 3555 }, { "epoch": 5.323353293413174, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1612, "step": 3556 }, { "epoch": 5.324850299401198, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1654, "step": 3557 }, { "epoch": 5.326347305389222, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1617, "step": 3558 }, { "epoch": 5.327844311377246, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1703, "step": 3559 }, { "epoch": 5.3293413173652695, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.165, "step": 3560 }, { "epoch": 5.330838323353293, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1632, "step": 3561 }, { "epoch": 5.332335329341317, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1688, "step": 3562 }, { "epoch": 5.333832335329341, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1617, "step": 3563 }, { "epoch": 5.335329341317365, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1612, "step": 3564 }, { "epoch": 5.336826347305389, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1638, "step": 3565 }, { "epoch": 5.338323353293413, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1617, "step": 3566 }, { "epoch": 5.339820359281437, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1609, "step": 3567 }, { "epoch": 5.341317365269461, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1656, "step": 3568 }, { "epoch": 5.342814371257485, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1664, "step": 3569 }, { "epoch": 5.344311377245509, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1604, "step": 3570 }, { "epoch": 5.345808383233533, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1661, "step": 3571 }, { "epoch": 5.347305389221557, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1591, "step": 3572 }, { "epoch": 5.348802395209581, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1593, "step": 3573 }, { "epoch": 5.350299401197605, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1634, "step": 3574 }, { "epoch": 5.3517964071856285, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1578, "step": 3575 }, { "epoch": 5.3532934131736525, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1638, "step": 3576 }, { "epoch": 5.354790419161676, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1669, "step": 3577 }, { "epoch": 5.3562874251497, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1682, "step": 3578 }, { "epoch": 5.357784431137724, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1599, "step": 3579 }, { "epoch": 5.359281437125748, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1619, "step": 3580 }, { "epoch": 5.360778443113772, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1601, "step": 3581 }, { "epoch": 5.362275449101796, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1634, "step": 3582 }, { "epoch": 5.36377245508982, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1597, "step": 3583 }, { "epoch": 5.365269461077844, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.164, "step": 3584 }, { "epoch": 5.366766467065868, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1637, "step": 3585 }, { "epoch": 5.368263473053892, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1577, "step": 3586 }, { "epoch": 5.369760479041916, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1649, "step": 3587 }, { "epoch": 5.37125748502994, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.16, "step": 3588 }, { "epoch": 5.3727544910179645, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1668, "step": 3589 }, { "epoch": 5.374251497005988, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1652, "step": 3590 }, { "epoch": 5.375748502994012, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1658, "step": 3591 }, { "epoch": 5.3772455089820355, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1648, "step": 3592 }, { "epoch": 5.37874251497006, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1591, "step": 3593 }, { "epoch": 5.380239520958084, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1663, "step": 3594 }, { "epoch": 5.381736526946108, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1676, "step": 3595 }, { "epoch": 5.383233532934132, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.166, "step": 3596 }, { "epoch": 5.384730538922156, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1707, "step": 3597 }, { "epoch": 5.38622754491018, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1614, "step": 3598 }, { "epoch": 5.387724550898204, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1671, "step": 3599 }, { "epoch": 5.389221556886228, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1648, "step": 3600 }, { "epoch": 5.390718562874252, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1592, "step": 3601 }, { "epoch": 5.392215568862276, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1697, "step": 3602 }, { "epoch": 5.3937125748503, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1602, "step": 3603 }, { "epoch": 5.395209580838324, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1581, "step": 3604 }, { "epoch": 5.3967065868263475, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1601, "step": 3605 }, { "epoch": 5.3982035928143715, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1603, "step": 3606 }, { "epoch": 5.399700598802395, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1604, "step": 3607 }, { "epoch": 5.401197604790419, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1654, "step": 3608 }, { "epoch": 5.402694610778443, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1605, "step": 3609 }, { "epoch": 5.404191616766467, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1627, "step": 3610 }, { "epoch": 5.405688622754491, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.161, "step": 3611 }, { "epoch": 5.407185628742515, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1615, "step": 3612 }, { "epoch": 5.408682634730539, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1618, "step": 3613 }, { "epoch": 5.410179640718563, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1614, "step": 3614 }, { "epoch": 5.411676646706587, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1614, "step": 3615 }, { "epoch": 5.413173652694611, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1651, "step": 3616 }, { "epoch": 5.414670658682635, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1678, "step": 3617 }, { "epoch": 5.416167664670659, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1681, "step": 3618 }, { "epoch": 5.417664670658683, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1568, "step": 3619 }, { "epoch": 5.419161676646707, "grad_norm": 0.14453125, "learning_rate": 0.0008, "loss": 1.1649, "step": 3620 }, { "epoch": 5.4206586826347305, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1609, "step": 3621 }, { "epoch": 5.422155688622754, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1628, "step": 3622 }, { "epoch": 5.423652694610778, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1583, "step": 3623 }, { "epoch": 5.425149700598802, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1661, "step": 3624 }, { "epoch": 5.426646706586826, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1665, "step": 3625 }, { "epoch": 5.42814371257485, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1563, "step": 3626 }, { "epoch": 5.429640718562874, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1615, "step": 3627 }, { "epoch": 5.431137724550898, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1607, "step": 3628 }, { "epoch": 5.432634730538922, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1594, "step": 3629 }, { "epoch": 5.434131736526946, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1639, "step": 3630 }, { "epoch": 5.43562874251497, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.164, "step": 3631 }, { "epoch": 5.437125748502994, "grad_norm": 0.1533203125, "learning_rate": 0.0008, "loss": 1.1699, "step": 3632 }, { "epoch": 5.438622754491018, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1591, "step": 3633 }, { "epoch": 5.440119760479042, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1636, "step": 3634 }, { "epoch": 5.441616766467066, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.1656, "step": 3635 }, { "epoch": 5.4431137724550895, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1653, "step": 3636 }, { "epoch": 5.4446107784431135, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1597, "step": 3637 }, { "epoch": 5.446107784431137, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1642, "step": 3638 }, { "epoch": 5.447604790419161, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1644, "step": 3639 }, { "epoch": 5.449101796407185, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1583, "step": 3640 }, { "epoch": 5.450598802395209, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1602, "step": 3641 }, { "epoch": 5.452095808383233, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1621, "step": 3642 }, { "epoch": 5.453592814371257, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1619, "step": 3643 }, { "epoch": 5.455089820359281, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1694, "step": 3644 }, { "epoch": 5.456586826347305, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.155, "step": 3645 }, { "epoch": 5.45808383233533, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1615, "step": 3646 }, { "epoch": 5.459580838323353, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1636, "step": 3647 }, { "epoch": 5.461077844311378, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1646, "step": 3648 }, { "epoch": 5.462574850299402, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1626, "step": 3649 }, { "epoch": 5.4640718562874255, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1647, "step": 3650 }, { "epoch": 5.4655688622754495, "grad_norm": 0.1728515625, "learning_rate": 0.0008, "loss": 1.1595, "step": 3651 }, { "epoch": 5.467065868263473, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1606, "step": 3652 }, { "epoch": 5.468562874251497, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1686, "step": 3653 }, { "epoch": 5.470059880239521, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1621, "step": 3654 }, { "epoch": 5.471556886227545, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1684, "step": 3655 }, { "epoch": 5.473053892215569, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1673, "step": 3656 }, { "epoch": 5.474550898203593, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1642, "step": 3657 }, { "epoch": 5.476047904191617, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1574, "step": 3658 }, { "epoch": 5.477544910179641, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.168, "step": 3659 }, { "epoch": 5.479041916167665, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1696, "step": 3660 }, { "epoch": 5.480538922155689, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1573, "step": 3661 }, { "epoch": 5.482035928143713, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1641, "step": 3662 }, { "epoch": 5.483532934131737, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1705, "step": 3663 }, { "epoch": 5.485029940119761, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1591, "step": 3664 }, { "epoch": 5.486526946107785, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1655, "step": 3665 }, { "epoch": 5.4880239520958085, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1506, "step": 3666 }, { "epoch": 5.4895209580838324, "grad_norm": 0.1572265625, "learning_rate": 0.0008, "loss": 1.1628, "step": 3667 }, { "epoch": 5.491017964071856, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1562, "step": 3668 }, { "epoch": 5.49251497005988, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1635, "step": 3669 }, { "epoch": 5.494011976047904, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1597, "step": 3670 }, { "epoch": 5.495508982035928, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1623, "step": 3671 }, { "epoch": 5.497005988023952, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1664, "step": 3672 }, { "epoch": 5.498502994011976, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.168, "step": 3673 }, { "epoch": 5.5, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1535, "step": 3674 }, { "epoch": 5.501497005988024, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1642, "step": 3675 }, { "epoch": 5.502994011976048, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1645, "step": 3676 }, { "epoch": 5.504491017964072, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1649, "step": 3677 }, { "epoch": 5.505988023952096, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1598, "step": 3678 }, { "epoch": 5.50748502994012, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1593, "step": 3679 }, { "epoch": 5.508982035928144, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1621, "step": 3680 }, { "epoch": 5.5104790419161676, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1632, "step": 3681 }, { "epoch": 5.5119760479041915, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.163, "step": 3682 }, { "epoch": 5.513473053892215, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1634, "step": 3683 }, { "epoch": 5.514970059880239, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1671, "step": 3684 }, { "epoch": 5.516467065868263, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1605, "step": 3685 }, { "epoch": 5.517964071856287, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1554, "step": 3686 }, { "epoch": 5.519461077844311, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1644, "step": 3687 }, { "epoch": 5.520958083832335, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1652, "step": 3688 }, { "epoch": 5.522455089820359, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1689, "step": 3689 }, { "epoch": 5.523952095808383, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1651, "step": 3690 }, { "epoch": 5.525449101796407, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1607, "step": 3691 }, { "epoch": 5.526946107784431, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1604, "step": 3692 }, { "epoch": 5.528443113772455, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1669, "step": 3693 }, { "epoch": 5.529940119760479, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1605, "step": 3694 }, { "epoch": 5.531437125748503, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1616, "step": 3695 }, { "epoch": 5.532934131736527, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1616, "step": 3696 }, { "epoch": 5.5344311377245505, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1579, "step": 3697 }, { "epoch": 5.5359281437125745, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1638, "step": 3698 }, { "epoch": 5.537425149700599, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1664, "step": 3699 }, { "epoch": 5.538922155688622, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1693, "step": 3700 }, { "epoch": 5.540419161676647, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1677, "step": 3701 }, { "epoch": 5.54191616766467, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1621, "step": 3702 }, { "epoch": 5.543413173652695, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1655, "step": 3703 }, { "epoch": 5.544910179640718, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1678, "step": 3704 }, { "epoch": 5.546407185628743, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1631, "step": 3705 }, { "epoch": 5.547904191616767, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1542, "step": 3706 }, { "epoch": 5.549401197604791, "grad_norm": 0.1875, "learning_rate": 0.0008, "loss": 1.1527, "step": 3707 }, { "epoch": 5.550898203592815, "grad_norm": 0.1767578125, "learning_rate": 0.0008, "loss": 1.1576, "step": 3708 }, { "epoch": 5.552395209580839, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1701, "step": 3709 }, { "epoch": 5.553892215568863, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.164, "step": 3710 }, { "epoch": 5.5553892215568865, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1592, "step": 3711 }, { "epoch": 5.5568862275449105, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1642, "step": 3712 }, { "epoch": 5.558383233532934, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.158, "step": 3713 }, { "epoch": 5.559880239520958, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1669, "step": 3714 }, { "epoch": 5.561377245508982, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1589, "step": 3715 }, { "epoch": 5.562874251497006, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1643, "step": 3716 }, { "epoch": 5.56437125748503, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1649, "step": 3717 }, { "epoch": 5.565868263473054, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1606, "step": 3718 }, { "epoch": 5.567365269461078, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1619, "step": 3719 }, { "epoch": 5.568862275449102, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1637, "step": 3720 }, { "epoch": 5.570359281437126, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1626, "step": 3721 }, { "epoch": 5.57185628742515, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1546, "step": 3722 }, { "epoch": 5.573353293413174, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1584, "step": 3723 }, { "epoch": 5.574850299401198, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1656, "step": 3724 }, { "epoch": 5.576347305389222, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1576, "step": 3725 }, { "epoch": 5.577844311377246, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1649, "step": 3726 }, { "epoch": 5.5793413173652695, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1622, "step": 3727 }, { "epoch": 5.580838323353293, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1682, "step": 3728 }, { "epoch": 5.582335329341317, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1683, "step": 3729 }, { "epoch": 5.583832335329341, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1581, "step": 3730 }, { "epoch": 5.585329341317365, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1588, "step": 3731 }, { "epoch": 5.586826347305389, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1562, "step": 3732 }, { "epoch": 5.588323353293413, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1583, "step": 3733 }, { "epoch": 5.589820359281437, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1668, "step": 3734 }, { "epoch": 5.591317365269461, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.156, "step": 3735 }, { "epoch": 5.592814371257485, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1615, "step": 3736 }, { "epoch": 5.594311377245509, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1632, "step": 3737 }, { "epoch": 5.595808383233533, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1571, "step": 3738 }, { "epoch": 5.597305389221557, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.156, "step": 3739 }, { "epoch": 5.598802395209581, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1601, "step": 3740 }, { "epoch": 5.600299401197605, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1663, "step": 3741 }, { "epoch": 5.6017964071856285, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.164, "step": 3742 }, { "epoch": 5.6032934131736525, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1562, "step": 3743 }, { "epoch": 5.604790419161676, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1632, "step": 3744 }, { "epoch": 5.6062874251497, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1627, "step": 3745 }, { "epoch": 5.607784431137724, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1606, "step": 3746 }, { "epoch": 5.609281437125748, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1651, "step": 3747 }, { "epoch": 5.610778443113772, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.165, "step": 3748 }, { "epoch": 5.612275449101796, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1647, "step": 3749 }, { "epoch": 5.61377245508982, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1639, "step": 3750 }, { "epoch": 5.615269461077844, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1606, "step": 3751 }, { "epoch": 5.616766467065868, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1589, "step": 3752 }, { "epoch": 5.618263473053892, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1644, "step": 3753 }, { "epoch": 5.619760479041916, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1664, "step": 3754 }, { "epoch": 5.62125748502994, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1665, "step": 3755 }, { "epoch": 5.6227544910179645, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1632, "step": 3756 }, { "epoch": 5.624251497005988, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1623, "step": 3757 }, { "epoch": 5.625748502994012, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1548, "step": 3758 }, { "epoch": 5.6272455089820355, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1658, "step": 3759 }, { "epoch": 5.62874251497006, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1544, "step": 3760 }, { "epoch": 5.630239520958084, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1625, "step": 3761 }, { "epoch": 5.631736526946108, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1593, "step": 3762 }, { "epoch": 5.633233532934132, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.159, "step": 3763 }, { "epoch": 5.634730538922156, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1634, "step": 3764 }, { "epoch": 5.63622754491018, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1559, "step": 3765 }, { "epoch": 5.637724550898204, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1667, "step": 3766 }, { "epoch": 5.639221556886228, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1633, "step": 3767 }, { "epoch": 5.640718562874252, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1593, "step": 3768 }, { "epoch": 5.642215568862276, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1652, "step": 3769 }, { "epoch": 5.6437125748503, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1628, "step": 3770 }, { "epoch": 5.645209580838324, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1647, "step": 3771 }, { "epoch": 5.6467065868263475, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1598, "step": 3772 }, { "epoch": 5.6482035928143715, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1643, "step": 3773 }, { "epoch": 5.649700598802395, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1568, "step": 3774 }, { "epoch": 5.651197604790419, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1623, "step": 3775 }, { "epoch": 5.652694610778443, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1624, "step": 3776 }, { "epoch": 5.654191616766467, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1646, "step": 3777 }, { "epoch": 5.655688622754491, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1565, "step": 3778 }, { "epoch": 5.657185628742515, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1603, "step": 3779 }, { "epoch": 5.658682634730539, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1607, "step": 3780 }, { "epoch": 5.660179640718563, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.159, "step": 3781 }, { "epoch": 5.661676646706587, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1696, "step": 3782 }, { "epoch": 5.663173652694611, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1634, "step": 3783 }, { "epoch": 5.664670658682635, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1617, "step": 3784 }, { "epoch": 5.666167664670659, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1559, "step": 3785 }, { "epoch": 5.667664670658683, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1617, "step": 3786 }, { "epoch": 5.669161676646707, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.156, "step": 3787 }, { "epoch": 5.6706586826347305, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1612, "step": 3788 }, { "epoch": 5.672155688622754, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1607, "step": 3789 }, { "epoch": 5.673652694610778, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1665, "step": 3790 }, { "epoch": 5.675149700598802, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1597, "step": 3791 }, { "epoch": 5.676646706586826, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1583, "step": 3792 }, { "epoch": 5.67814371257485, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1632, "step": 3793 }, { "epoch": 5.679640718562874, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1554, "step": 3794 }, { "epoch": 5.681137724550898, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1606, "step": 3795 }, { "epoch": 5.682634730538922, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1648, "step": 3796 }, { "epoch": 5.684131736526946, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1644, "step": 3797 }, { "epoch": 5.68562874251497, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1587, "step": 3798 }, { "epoch": 5.687125748502994, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1617, "step": 3799 }, { "epoch": 5.688622754491018, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.159, "step": 3800 }, { "epoch": 5.690119760479042, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1505, "step": 3801 }, { "epoch": 5.691616766467066, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1673, "step": 3802 }, { "epoch": 5.6931137724550895, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1621, "step": 3803 }, { "epoch": 5.6946107784431135, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.162, "step": 3804 }, { "epoch": 5.696107784431137, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1663, "step": 3805 }, { "epoch": 5.697604790419161, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1538, "step": 3806 }, { "epoch": 5.699101796407185, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1562, "step": 3807 }, { "epoch": 5.700598802395209, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1507, "step": 3808 }, { "epoch": 5.702095808383233, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1628, "step": 3809 }, { "epoch": 5.703592814371257, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1615, "step": 3810 }, { "epoch": 5.705089820359282, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1652, "step": 3811 }, { "epoch": 5.706586826347305, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1568, "step": 3812 }, { "epoch": 5.70808383233533, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1552, "step": 3813 }, { "epoch": 5.709580838323353, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1607, "step": 3814 }, { "epoch": 5.711077844311378, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.161, "step": 3815 }, { "epoch": 5.712574850299401, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.159, "step": 3816 }, { "epoch": 5.7140718562874255, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1604, "step": 3817 }, { "epoch": 5.7155688622754495, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1636, "step": 3818 }, { "epoch": 5.717065868263473, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1586, "step": 3819 }, { "epoch": 5.718562874251497, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1592, "step": 3820 }, { "epoch": 5.720059880239521, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1522, "step": 3821 }, { "epoch": 5.721556886227545, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1616, "step": 3822 }, { "epoch": 5.723053892215569, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1573, "step": 3823 }, { "epoch": 5.724550898203593, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1559, "step": 3824 }, { "epoch": 5.726047904191617, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1603, "step": 3825 }, { "epoch": 5.727544910179641, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1643, "step": 3826 }, { "epoch": 5.729041916167665, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1673, "step": 3827 }, { "epoch": 5.730538922155689, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1622, "step": 3828 }, { "epoch": 5.732035928143713, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1656, "step": 3829 }, { "epoch": 5.733532934131737, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1578, "step": 3830 }, { "epoch": 5.735029940119761, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1616, "step": 3831 }, { "epoch": 5.736526946107785, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1616, "step": 3832 }, { "epoch": 5.7380239520958085, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1657, "step": 3833 }, { "epoch": 5.7395209580838324, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1581, "step": 3834 }, { "epoch": 5.741017964071856, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1697, "step": 3835 }, { "epoch": 5.74251497005988, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1644, "step": 3836 }, { "epoch": 5.744011976047904, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.156, "step": 3837 }, { "epoch": 5.745508982035928, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1641, "step": 3838 }, { "epoch": 5.747005988023952, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1617, "step": 3839 }, { "epoch": 5.748502994011976, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.158, "step": 3840 }, { "epoch": 5.75, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1574, "step": 3841 }, { "epoch": 5.751497005988024, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1579, "step": 3842 }, { "epoch": 5.752994011976048, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1619, "step": 3843 }, { "epoch": 5.754491017964072, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1687, "step": 3844 }, { "epoch": 5.755988023952096, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1596, "step": 3845 }, { "epoch": 5.75748502994012, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1679, "step": 3846 }, { "epoch": 5.758982035928144, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1653, "step": 3847 }, { "epoch": 5.7604790419161676, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1633, "step": 3848 }, { "epoch": 5.7619760479041915, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1629, "step": 3849 }, { "epoch": 5.763473053892215, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1679, "step": 3850 }, { "epoch": 5.764970059880239, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.1606, "step": 3851 }, { "epoch": 5.766467065868263, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.164, "step": 3852 }, { "epoch": 5.767964071856287, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1652, "step": 3853 }, { "epoch": 5.769461077844311, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1572, "step": 3854 }, { "epoch": 5.770958083832335, "grad_norm": 0.1953125, "learning_rate": 0.0008, "loss": 1.1592, "step": 3855 }, { "epoch": 5.772455089820359, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.1532, "step": 3856 }, { "epoch": 5.773952095808383, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1672, "step": 3857 }, { "epoch": 5.775449101796407, "grad_norm": 0.1787109375, "learning_rate": 0.0008, "loss": 1.1631, "step": 3858 }, { "epoch": 5.776946107784431, "grad_norm": 0.201171875, "learning_rate": 0.0008, "loss": 1.158, "step": 3859 }, { "epoch": 5.778443113772455, "grad_norm": 0.1806640625, "learning_rate": 0.0008, "loss": 1.1605, "step": 3860 }, { "epoch": 5.779940119760479, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1592, "step": 3861 }, { "epoch": 5.781437125748503, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1546, "step": 3862 }, { "epoch": 5.782934131736527, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1582, "step": 3863 }, { "epoch": 5.7844311377245505, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1626, "step": 3864 }, { "epoch": 5.7859281437125745, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1553, "step": 3865 }, { "epoch": 5.787425149700599, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1667, "step": 3866 }, { "epoch": 5.788922155688622, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1556, "step": 3867 }, { "epoch": 5.790419161676647, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1645, "step": 3868 }, { "epoch": 5.79191616766467, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1613, "step": 3869 }, { "epoch": 5.793413173652695, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1553, "step": 3870 }, { "epoch": 5.794910179640718, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1563, "step": 3871 }, { "epoch": 5.796407185628743, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1603, "step": 3872 }, { "epoch": 5.797904191616767, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1715, "step": 3873 }, { "epoch": 5.799401197604791, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1677, "step": 3874 }, { "epoch": 5.800898203592815, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1706, "step": 3875 }, { "epoch": 5.802395209580839, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1504, "step": 3876 }, { "epoch": 5.803892215568863, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1596, "step": 3877 }, { "epoch": 5.8053892215568865, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1551, "step": 3878 }, { "epoch": 5.8068862275449105, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1641, "step": 3879 }, { "epoch": 5.808383233532934, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1584, "step": 3880 }, { "epoch": 5.809880239520958, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1567, "step": 3881 }, { "epoch": 5.811377245508982, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1616, "step": 3882 }, { "epoch": 5.812874251497006, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1648, "step": 3883 }, { "epoch": 5.81437125748503, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1638, "step": 3884 }, { "epoch": 5.815868263473054, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1607, "step": 3885 }, { "epoch": 5.817365269461078, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1576, "step": 3886 }, { "epoch": 5.818862275449102, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1538, "step": 3887 }, { "epoch": 5.820359281437126, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1573, "step": 3888 }, { "epoch": 5.82185628742515, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1583, "step": 3889 }, { "epoch": 5.823353293413174, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1534, "step": 3890 }, { "epoch": 5.824850299401198, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1625, "step": 3891 }, { "epoch": 5.826347305389222, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1598, "step": 3892 }, { "epoch": 5.827844311377246, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1574, "step": 3893 }, { "epoch": 5.8293413173652695, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1588, "step": 3894 }, { "epoch": 5.830838323353293, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1598, "step": 3895 }, { "epoch": 5.832335329341317, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1593, "step": 3896 }, { "epoch": 5.833832335329341, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1567, "step": 3897 }, { "epoch": 5.835329341317365, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1662, "step": 3898 }, { "epoch": 5.836826347305389, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1574, "step": 3899 }, { "epoch": 5.838323353293413, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1649, "step": 3900 }, { "epoch": 5.839820359281437, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1617, "step": 3901 }, { "epoch": 5.841317365269461, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1585, "step": 3902 }, { "epoch": 5.842814371257485, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1551, "step": 3903 }, { "epoch": 5.844311377245509, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1671, "step": 3904 }, { "epoch": 5.845808383233533, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1593, "step": 3905 }, { "epoch": 5.847305389221557, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.155, "step": 3906 }, { "epoch": 5.848802395209581, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1632, "step": 3907 }, { "epoch": 5.850299401197605, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1621, "step": 3908 }, { "epoch": 5.8517964071856285, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1519, "step": 3909 }, { "epoch": 5.8532934131736525, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1551, "step": 3910 }, { "epoch": 5.854790419161676, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.162, "step": 3911 }, { "epoch": 5.8562874251497, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1597, "step": 3912 }, { "epoch": 5.857784431137724, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1573, "step": 3913 }, { "epoch": 5.859281437125748, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1592, "step": 3914 }, { "epoch": 5.860778443113772, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1652, "step": 3915 }, { "epoch": 5.862275449101796, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.152, "step": 3916 }, { "epoch": 5.86377245508982, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1704, "step": 3917 }, { "epoch": 5.865269461077844, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1555, "step": 3918 }, { "epoch": 5.866766467065868, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.16, "step": 3919 }, { "epoch": 5.868263473053892, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1634, "step": 3920 }, { "epoch": 5.869760479041916, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1541, "step": 3921 }, { "epoch": 5.87125748502994, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1633, "step": 3922 }, { "epoch": 5.8727544910179645, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1583, "step": 3923 }, { "epoch": 5.874251497005988, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1674, "step": 3924 }, { "epoch": 5.875748502994012, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1681, "step": 3925 }, { "epoch": 5.8772455089820355, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1585, "step": 3926 }, { "epoch": 5.87874251497006, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1612, "step": 3927 }, { "epoch": 5.880239520958084, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1623, "step": 3928 }, { "epoch": 5.881736526946108, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.158, "step": 3929 }, { "epoch": 5.883233532934132, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1597, "step": 3930 }, { "epoch": 5.884730538922156, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1675, "step": 3931 }, { "epoch": 5.88622754491018, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1615, "step": 3932 }, { "epoch": 5.887724550898204, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1586, "step": 3933 }, { "epoch": 5.889221556886228, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1611, "step": 3934 }, { "epoch": 5.890718562874252, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1604, "step": 3935 }, { "epoch": 5.892215568862276, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1597, "step": 3936 }, { "epoch": 5.8937125748503, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1578, "step": 3937 }, { "epoch": 5.895209580838324, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1637, "step": 3938 }, { "epoch": 5.8967065868263475, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1591, "step": 3939 }, { "epoch": 5.8982035928143715, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1578, "step": 3940 }, { "epoch": 5.899700598802395, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1556, "step": 3941 }, { "epoch": 5.901197604790419, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1552, "step": 3942 }, { "epoch": 5.902694610778443, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1572, "step": 3943 }, { "epoch": 5.904191616766467, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1694, "step": 3944 }, { "epoch": 5.905688622754491, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1636, "step": 3945 }, { "epoch": 5.907185628742515, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1643, "step": 3946 }, { "epoch": 5.908682634730539, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1578, "step": 3947 }, { "epoch": 5.910179640718563, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.168, "step": 3948 }, { "epoch": 5.911676646706587, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1656, "step": 3949 }, { "epoch": 5.913173652694611, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1555, "step": 3950 }, { "epoch": 5.914670658682635, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1697, "step": 3951 }, { "epoch": 5.916167664670659, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1532, "step": 3952 }, { "epoch": 5.917664670658683, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1595, "step": 3953 }, { "epoch": 5.919161676646707, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1596, "step": 3954 }, { "epoch": 5.9206586826347305, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1592, "step": 3955 }, { "epoch": 5.922155688622754, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1684, "step": 3956 }, { "epoch": 5.923652694610778, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.163, "step": 3957 }, { "epoch": 5.925149700598802, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1562, "step": 3958 }, { "epoch": 5.926646706586826, "grad_norm": 0.1630859375, "learning_rate": 0.0008, "loss": 1.1629, "step": 3959 }, { "epoch": 5.92814371257485, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1519, "step": 3960 }, { "epoch": 5.929640718562874, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1534, "step": 3961 }, { "epoch": 5.931137724550898, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1634, "step": 3962 }, { "epoch": 5.932634730538922, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1582, "step": 3963 }, { "epoch": 5.934131736526946, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1585, "step": 3964 }, { "epoch": 5.93562874251497, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1548, "step": 3965 }, { "epoch": 5.937125748502994, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1657, "step": 3966 }, { "epoch": 5.938622754491018, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1551, "step": 3967 }, { "epoch": 5.940119760479042, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1623, "step": 3968 }, { "epoch": 5.941616766467066, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1614, "step": 3969 }, { "epoch": 5.9431137724550895, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.156, "step": 3970 }, { "epoch": 5.9446107784431135, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1635, "step": 3971 }, { "epoch": 5.946107784431137, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1616, "step": 3972 }, { "epoch": 5.947604790419161, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1511, "step": 3973 }, { "epoch": 5.949101796407185, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1573, "step": 3974 }, { "epoch": 5.950598802395209, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1654, "step": 3975 }, { "epoch": 5.952095808383233, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1559, "step": 3976 }, { "epoch": 5.953592814371257, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1572, "step": 3977 }, { "epoch": 5.955089820359282, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.169, "step": 3978 }, { "epoch": 5.956586826347305, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1567, "step": 3979 }, { "epoch": 5.95808383233533, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1571, "step": 3980 }, { "epoch": 5.959580838323353, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1643, "step": 3981 }, { "epoch": 5.961077844311378, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1559, "step": 3982 }, { "epoch": 5.962574850299401, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1634, "step": 3983 }, { "epoch": 5.9640718562874255, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1617, "step": 3984 }, { "epoch": 5.9655688622754495, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.152, "step": 3985 }, { "epoch": 5.967065868263473, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1487, "step": 3986 }, { "epoch": 5.968562874251497, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1566, "step": 3987 }, { "epoch": 5.970059880239521, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1679, "step": 3988 }, { "epoch": 5.971556886227545, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1588, "step": 3989 }, { "epoch": 5.973053892215569, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1609, "step": 3990 }, { "epoch": 5.974550898203593, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1604, "step": 3991 }, { "epoch": 5.976047904191617, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1558, "step": 3992 }, { "epoch": 5.977544910179641, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1585, "step": 3993 }, { "epoch": 5.979041916167665, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1647, "step": 3994 }, { "epoch": 5.980538922155689, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1654, "step": 3995 }, { "epoch": 5.982035928143713, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1579, "step": 3996 }, { "epoch": 5.983532934131737, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1494, "step": 3997 }, { "epoch": 5.985029940119761, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1647, "step": 3998 }, { "epoch": 5.986526946107785, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1605, "step": 3999 }, { "epoch": 5.9880239520958085, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1595, "step": 4000 }, { "epoch": 5.9895209580838324, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1553, "step": 4001 }, { "epoch": 5.991017964071856, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1653, "step": 4002 }, { "epoch": 5.99251497005988, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1578, "step": 4003 }, { "epoch": 5.994011976047904, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1634, "step": 4004 }, { "epoch": 5.995508982035928, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1597, "step": 4005 }, { "epoch": 5.997005988023952, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1563, "step": 4006 }, { "epoch": 5.998502994011976, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1578, "step": 4007 }, { "epoch": 6.0, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1599, "step": 4008 }, { "epoch": 6.001497005988024, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1583, "step": 4009 }, { "epoch": 6.002994011976048, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1622, "step": 4010 }, { "epoch": 6.004491017964072, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1628, "step": 4011 }, { "epoch": 6.005988023952096, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1581, "step": 4012 }, { "epoch": 6.00748502994012, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1613, "step": 4013 }, { "epoch": 6.008982035928144, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.157, "step": 4014 }, { "epoch": 6.0104790419161676, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1613, "step": 4015 }, { "epoch": 6.0119760479041915, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1556, "step": 4016 }, { "epoch": 6.013473053892215, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1612, "step": 4017 }, { "epoch": 6.014970059880239, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1658, "step": 4018 }, { "epoch": 6.016467065868263, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1598, "step": 4019 }, { "epoch": 6.017964071856287, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1642, "step": 4020 }, { "epoch": 6.019461077844311, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1587, "step": 4021 }, { "epoch": 6.020958083832335, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1517, "step": 4022 }, { "epoch": 6.022455089820359, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1482, "step": 4023 }, { "epoch": 6.023952095808383, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1591, "step": 4024 }, { "epoch": 6.025449101796407, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1593, "step": 4025 }, { "epoch": 6.026946107784431, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1562, "step": 4026 }, { "epoch": 6.028443113772455, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1614, "step": 4027 }, { "epoch": 6.029940119760479, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1642, "step": 4028 }, { "epoch": 6.031437125748503, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1615, "step": 4029 }, { "epoch": 6.032934131736527, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1633, "step": 4030 }, { "epoch": 6.0344311377245505, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1578, "step": 4031 }, { "epoch": 6.0359281437125745, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.163, "step": 4032 }, { "epoch": 6.037425149700598, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1579, "step": 4033 }, { "epoch": 6.038922155688622, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1572, "step": 4034 }, { "epoch": 6.040419161676646, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1578, "step": 4035 }, { "epoch": 6.04191616766467, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1609, "step": 4036 }, { "epoch": 6.043413173652695, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1549, "step": 4037 }, { "epoch": 6.044910179640719, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1605, "step": 4038 }, { "epoch": 6.046407185628743, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1623, "step": 4039 }, { "epoch": 6.047904191616767, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1644, "step": 4040 }, { "epoch": 6.049401197604791, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1594, "step": 4041 }, { "epoch": 6.050898203592815, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1622, "step": 4042 }, { "epoch": 6.052395209580839, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1634, "step": 4043 }, { "epoch": 6.053892215568863, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1616, "step": 4044 }, { "epoch": 6.0553892215568865, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1536, "step": 4045 }, { "epoch": 6.0568862275449105, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1619, "step": 4046 }, { "epoch": 6.058383233532934, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1552, "step": 4047 }, { "epoch": 6.059880239520958, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1615, "step": 4048 }, { "epoch": 6.061377245508982, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.156, "step": 4049 }, { "epoch": 6.062874251497006, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1609, "step": 4050 }, { "epoch": 6.06437125748503, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1583, "step": 4051 }, { "epoch": 6.065868263473054, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1619, "step": 4052 }, { "epoch": 6.067365269461078, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1634, "step": 4053 }, { "epoch": 6.068862275449102, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.155, "step": 4054 }, { "epoch": 6.070359281437126, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1533, "step": 4055 }, { "epoch": 6.07185628742515, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1556, "step": 4056 }, { "epoch": 6.073353293413174, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1571, "step": 4057 }, { "epoch": 6.074850299401198, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1549, "step": 4058 }, { "epoch": 6.076347305389222, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1566, "step": 4059 }, { "epoch": 6.077844311377246, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.153, "step": 4060 }, { "epoch": 6.0793413173652695, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1599, "step": 4061 }, { "epoch": 6.080838323353293, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1551, "step": 4062 }, { "epoch": 6.082335329341317, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1638, "step": 4063 }, { "epoch": 6.083832335329341, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1595, "step": 4064 }, { "epoch": 6.085329341317365, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1586, "step": 4065 }, { "epoch": 6.086826347305389, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1531, "step": 4066 }, { "epoch": 6.088323353293413, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1613, "step": 4067 }, { "epoch": 6.089820359281437, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1633, "step": 4068 }, { "epoch": 6.091317365269461, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1643, "step": 4069 }, { "epoch": 6.092814371257485, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1601, "step": 4070 }, { "epoch": 6.094311377245509, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1603, "step": 4071 }, { "epoch": 6.095808383233533, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1591, "step": 4072 }, { "epoch": 6.097305389221557, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1577, "step": 4073 }, { "epoch": 6.098802395209581, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1616, "step": 4074 }, { "epoch": 6.100299401197605, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1556, "step": 4075 }, { "epoch": 6.1017964071856285, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1556, "step": 4076 }, { "epoch": 6.1032934131736525, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1613, "step": 4077 }, { "epoch": 6.104790419161676, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.162, "step": 4078 }, { "epoch": 6.1062874251497, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1545, "step": 4079 }, { "epoch": 6.107784431137724, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1586, "step": 4080 }, { "epoch": 6.109281437125748, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1573, "step": 4081 }, { "epoch": 6.110778443113772, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1625, "step": 4082 }, { "epoch": 6.112275449101796, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.164, "step": 4083 }, { "epoch": 6.11377245508982, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1633, "step": 4084 }, { "epoch": 6.115269461077844, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.158, "step": 4085 }, { "epoch": 6.116766467065868, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1571, "step": 4086 }, { "epoch": 6.118263473053892, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1652, "step": 4087 }, { "epoch": 6.119760479041916, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1579, "step": 4088 }, { "epoch": 6.12125748502994, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1617, "step": 4089 }, { "epoch": 6.122754491017964, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1536, "step": 4090 }, { "epoch": 6.124251497005988, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1629, "step": 4091 }, { "epoch": 6.125748502994012, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1603, "step": 4092 }, { "epoch": 6.127245508982036, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.157, "step": 4093 }, { "epoch": 6.12874251497006, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1591, "step": 4094 }, { "epoch": 6.130239520958084, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1612, "step": 4095 }, { "epoch": 6.131736526946108, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1609, "step": 4096 }, { "epoch": 6.133233532934132, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1586, "step": 4097 }, { "epoch": 6.134730538922156, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1631, "step": 4098 }, { "epoch": 6.13622754491018, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1514, "step": 4099 }, { "epoch": 6.137724550898204, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1637, "step": 4100 }, { "epoch": 6.139221556886228, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1605, "step": 4101 }, { "epoch": 6.140718562874252, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1572, "step": 4102 }, { "epoch": 6.142215568862276, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.156, "step": 4103 }, { "epoch": 6.1437125748503, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1548, "step": 4104 }, { "epoch": 6.145209580838324, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1609, "step": 4105 }, { "epoch": 6.1467065868263475, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1599, "step": 4106 }, { "epoch": 6.1482035928143715, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1641, "step": 4107 }, { "epoch": 6.149700598802395, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1629, "step": 4108 }, { "epoch": 6.151197604790419, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1637, "step": 4109 }, { "epoch": 6.152694610778443, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1605, "step": 4110 }, { "epoch": 6.154191616766467, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1499, "step": 4111 }, { "epoch": 6.155688622754491, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1546, "step": 4112 }, { "epoch": 6.157185628742515, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1532, "step": 4113 }, { "epoch": 6.158682634730539, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.159, "step": 4114 }, { "epoch": 6.160179640718563, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1627, "step": 4115 }, { "epoch": 6.161676646706587, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1577, "step": 4116 }, { "epoch": 6.163173652694611, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1562, "step": 4117 }, { "epoch": 6.164670658682635, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1543, "step": 4118 }, { "epoch": 6.166167664670659, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1603, "step": 4119 }, { "epoch": 6.167664670658683, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1603, "step": 4120 }, { "epoch": 6.169161676646707, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1574, "step": 4121 }, { "epoch": 6.1706586826347305, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1577, "step": 4122 }, { "epoch": 6.172155688622754, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1594, "step": 4123 }, { "epoch": 6.173652694610778, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1538, "step": 4124 }, { "epoch": 6.175149700598802, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1553, "step": 4125 }, { "epoch": 6.176646706586826, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1514, "step": 4126 }, { "epoch": 6.17814371257485, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1688, "step": 4127 }, { "epoch": 6.179640718562874, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1612, "step": 4128 }, { "epoch": 6.181137724550898, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1569, "step": 4129 }, { "epoch": 6.182634730538922, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1568, "step": 4130 }, { "epoch": 6.184131736526946, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1585, "step": 4131 }, { "epoch": 6.18562874251497, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1519, "step": 4132 }, { "epoch": 6.187125748502994, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1607, "step": 4133 }, { "epoch": 6.188622754491018, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.163, "step": 4134 }, { "epoch": 6.190119760479042, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1564, "step": 4135 }, { "epoch": 6.191616766467066, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1604, "step": 4136 }, { "epoch": 6.1931137724550895, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1622, "step": 4137 }, { "epoch": 6.1946107784431135, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1555, "step": 4138 }, { "epoch": 6.196107784431137, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.16, "step": 4139 }, { "epoch": 6.197604790419161, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1532, "step": 4140 }, { "epoch": 6.199101796407185, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1596, "step": 4141 }, { "epoch": 6.200598802395209, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1615, "step": 4142 }, { "epoch": 6.202095808383233, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1552, "step": 4143 }, { "epoch": 6.203592814371257, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.158, "step": 4144 }, { "epoch": 6.205089820359281, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1635, "step": 4145 }, { "epoch": 6.206586826347305, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1614, "step": 4146 }, { "epoch": 6.20808383233533, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1643, "step": 4147 }, { "epoch": 6.209580838323353, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.154, "step": 4148 }, { "epoch": 6.211077844311378, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1586, "step": 4149 }, { "epoch": 6.212574850299402, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1617, "step": 4150 }, { "epoch": 6.2140718562874255, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.168, "step": 4151 }, { "epoch": 6.2155688622754495, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1502, "step": 4152 }, { "epoch": 6.217065868263473, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1571, "step": 4153 }, { "epoch": 6.218562874251497, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1568, "step": 4154 }, { "epoch": 6.220059880239521, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1536, "step": 4155 }, { "epoch": 6.221556886227545, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1604, "step": 4156 }, { "epoch": 6.223053892215569, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1538, "step": 4157 }, { "epoch": 6.224550898203593, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1544, "step": 4158 }, { "epoch": 6.226047904191617, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.16, "step": 4159 }, { "epoch": 6.227544910179641, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1569, "step": 4160 }, { "epoch": 6.229041916167665, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1669, "step": 4161 }, { "epoch": 6.230538922155689, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1599, "step": 4162 }, { "epoch": 6.232035928143713, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1568, "step": 4163 }, { "epoch": 6.233532934131737, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1545, "step": 4164 }, { "epoch": 6.235029940119761, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1507, "step": 4165 }, { "epoch": 6.236526946107785, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1606, "step": 4166 }, { "epoch": 6.2380239520958085, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.152, "step": 4167 }, { "epoch": 6.2395209580838324, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1574, "step": 4168 }, { "epoch": 6.241017964071856, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1648, "step": 4169 }, { "epoch": 6.24251497005988, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1556, "step": 4170 }, { "epoch": 6.244011976047904, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.157, "step": 4171 }, { "epoch": 6.245508982035928, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1563, "step": 4172 }, { "epoch": 6.247005988023952, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1592, "step": 4173 }, { "epoch": 6.248502994011976, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1605, "step": 4174 }, { "epoch": 6.25, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1616, "step": 4175 }, { "epoch": 6.251497005988024, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1542, "step": 4176 }, { "epoch": 6.252994011976048, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.163, "step": 4177 }, { "epoch": 6.254491017964072, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1551, "step": 4178 }, { "epoch": 6.255988023952096, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1624, "step": 4179 }, { "epoch": 6.25748502994012, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1611, "step": 4180 }, { "epoch": 6.258982035928144, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1646, "step": 4181 }, { "epoch": 6.2604790419161676, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1584, "step": 4182 }, { "epoch": 6.2619760479041915, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1599, "step": 4183 }, { "epoch": 6.263473053892215, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1591, "step": 4184 }, { "epoch": 6.264970059880239, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1612, "step": 4185 }, { "epoch": 6.266467065868263, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1562, "step": 4186 }, { "epoch": 6.267964071856287, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1572, "step": 4187 }, { "epoch": 6.269461077844311, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1636, "step": 4188 }, { "epoch": 6.270958083832335, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1574, "step": 4189 }, { "epoch": 6.272455089820359, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1566, "step": 4190 }, { "epoch": 6.273952095808383, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1634, "step": 4191 }, { "epoch": 6.275449101796407, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1664, "step": 4192 }, { "epoch": 6.276946107784431, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1589, "step": 4193 }, { "epoch": 6.278443113772455, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1503, "step": 4194 }, { "epoch": 6.279940119760479, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1613, "step": 4195 }, { "epoch": 6.281437125748503, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1525, "step": 4196 }, { "epoch": 6.282934131736527, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1668, "step": 4197 }, { "epoch": 6.2844311377245505, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.166, "step": 4198 }, { "epoch": 6.2859281437125745, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1543, "step": 4199 }, { "epoch": 6.287425149700598, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.1628, "step": 4200 }, { "epoch": 6.288922155688622, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1582, "step": 4201 }, { "epoch": 6.290419161676647, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.163, "step": 4202 }, { "epoch": 6.29191616766467, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1582, "step": 4203 }, { "epoch": 6.293413173652695, "grad_norm": 0.1767578125, "learning_rate": 0.0008, "loss": 1.1568, "step": 4204 }, { "epoch": 6.294910179640719, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.1603, "step": 4205 }, { "epoch": 6.296407185628743, "grad_norm": 0.1650390625, "learning_rate": 0.0008, "loss": 1.1553, "step": 4206 }, { "epoch": 6.297904191616767, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1602, "step": 4207 }, { "epoch": 6.299401197604791, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1601, "step": 4208 }, { "epoch": 6.300898203592815, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1569, "step": 4209 }, { "epoch": 6.302395209580839, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1573, "step": 4210 }, { "epoch": 6.303892215568863, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1609, "step": 4211 }, { "epoch": 6.3053892215568865, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1631, "step": 4212 }, { "epoch": 6.3068862275449105, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1578, "step": 4213 }, { "epoch": 6.308383233532934, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1635, "step": 4214 }, { "epoch": 6.309880239520958, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1605, "step": 4215 }, { "epoch": 6.311377245508982, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1609, "step": 4216 }, { "epoch": 6.312874251497006, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1568, "step": 4217 }, { "epoch": 6.31437125748503, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1547, "step": 4218 }, { "epoch": 6.315868263473054, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1561, "step": 4219 }, { "epoch": 6.317365269461078, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1547, "step": 4220 }, { "epoch": 6.318862275449102, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1526, "step": 4221 }, { "epoch": 6.320359281437126, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1536, "step": 4222 }, { "epoch": 6.32185628742515, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1523, "step": 4223 }, { "epoch": 6.323353293413174, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1604, "step": 4224 }, { "epoch": 6.324850299401198, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1599, "step": 4225 }, { "epoch": 6.326347305389222, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1569, "step": 4226 }, { "epoch": 6.327844311377246, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1596, "step": 4227 }, { "epoch": 6.3293413173652695, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.155, "step": 4228 }, { "epoch": 6.330838323353293, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1531, "step": 4229 }, { "epoch": 6.332335329341317, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1573, "step": 4230 }, { "epoch": 6.333832335329341, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1622, "step": 4231 }, { "epoch": 6.335329341317365, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1609, "step": 4232 }, { "epoch": 6.336826347305389, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1575, "step": 4233 }, { "epoch": 6.338323353293413, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1594, "step": 4234 }, { "epoch": 6.339820359281437, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1535, "step": 4235 }, { "epoch": 6.341317365269461, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1585, "step": 4236 }, { "epoch": 6.342814371257485, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1631, "step": 4237 }, { "epoch": 6.344311377245509, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1523, "step": 4238 }, { "epoch": 6.345808383233533, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1514, "step": 4239 }, { "epoch": 6.347305389221557, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1608, "step": 4240 }, { "epoch": 6.348802395209581, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1496, "step": 4241 }, { "epoch": 6.350299401197605, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1597, "step": 4242 }, { "epoch": 6.3517964071856285, "grad_norm": 0.162109375, "learning_rate": 0.0008, "loss": 1.1648, "step": 4243 }, { "epoch": 6.3532934131736525, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1621, "step": 4244 }, { "epoch": 6.354790419161676, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.16, "step": 4245 }, { "epoch": 6.3562874251497, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1579, "step": 4246 }, { "epoch": 6.357784431137724, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1555, "step": 4247 }, { "epoch": 6.359281437125748, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1642, "step": 4248 }, { "epoch": 6.360778443113772, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1527, "step": 4249 }, { "epoch": 6.362275449101796, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1577, "step": 4250 }, { "epoch": 6.36377245508982, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.164, "step": 4251 }, { "epoch": 6.365269461077844, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1638, "step": 4252 }, { "epoch": 6.366766467065868, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1522, "step": 4253 }, { "epoch": 6.368263473053892, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1581, "step": 4254 }, { "epoch": 6.369760479041916, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1551, "step": 4255 }, { "epoch": 6.37125748502994, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1614, "step": 4256 }, { "epoch": 6.3727544910179645, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1607, "step": 4257 }, { "epoch": 6.374251497005988, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1656, "step": 4258 }, { "epoch": 6.375748502994012, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1595, "step": 4259 }, { "epoch": 6.3772455089820355, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1634, "step": 4260 }, { "epoch": 6.37874251497006, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1592, "step": 4261 }, { "epoch": 6.380239520958084, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1552, "step": 4262 }, { "epoch": 6.381736526946108, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1594, "step": 4263 }, { "epoch": 6.383233532934132, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1537, "step": 4264 }, { "epoch": 6.384730538922156, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1587, "step": 4265 }, { "epoch": 6.38622754491018, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1563, "step": 4266 }, { "epoch": 6.387724550898204, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1584, "step": 4267 }, { "epoch": 6.389221556886228, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1535, "step": 4268 }, { "epoch": 6.390718562874252, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1542, "step": 4269 }, { "epoch": 6.392215568862276, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1551, "step": 4270 }, { "epoch": 6.3937125748503, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1549, "step": 4271 }, { "epoch": 6.395209580838324, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1618, "step": 4272 }, { "epoch": 6.3967065868263475, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1636, "step": 4273 }, { "epoch": 6.3982035928143715, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1565, "step": 4274 }, { "epoch": 6.399700598802395, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1576, "step": 4275 }, { "epoch": 6.401197604790419, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1573, "step": 4276 }, { "epoch": 6.402694610778443, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1584, "step": 4277 }, { "epoch": 6.404191616766467, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1567, "step": 4278 }, { "epoch": 6.405688622754491, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1526, "step": 4279 }, { "epoch": 6.407185628742515, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1588, "step": 4280 }, { "epoch": 6.408682634730539, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1481, "step": 4281 }, { "epoch": 6.410179640718563, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.158, "step": 4282 }, { "epoch": 6.411676646706587, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1634, "step": 4283 }, { "epoch": 6.413173652694611, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1627, "step": 4284 }, { "epoch": 6.414670658682635, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1573, "step": 4285 }, { "epoch": 6.416167664670659, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1544, "step": 4286 }, { "epoch": 6.417664670658683, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1604, "step": 4287 }, { "epoch": 6.419161676646707, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1597, "step": 4288 }, { "epoch": 6.4206586826347305, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1538, "step": 4289 }, { "epoch": 6.422155688622754, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1552, "step": 4290 }, { "epoch": 6.423652694610778, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1526, "step": 4291 }, { "epoch": 6.425149700598802, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1494, "step": 4292 }, { "epoch": 6.426646706586826, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1622, "step": 4293 }, { "epoch": 6.42814371257485, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1595, "step": 4294 }, { "epoch": 6.429640718562874, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.16, "step": 4295 }, { "epoch": 6.431137724550898, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1626, "step": 4296 }, { "epoch": 6.432634730538922, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1581, "step": 4297 }, { "epoch": 6.434131736526946, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1586, "step": 4298 }, { "epoch": 6.43562874251497, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1548, "step": 4299 }, { "epoch": 6.437125748502994, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1582, "step": 4300 }, { "epoch": 6.438622754491018, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1614, "step": 4301 }, { "epoch": 6.440119760479042, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1546, "step": 4302 }, { "epoch": 6.441616766467066, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1541, "step": 4303 }, { "epoch": 6.4431137724550895, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1555, "step": 4304 }, { "epoch": 6.4446107784431135, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1659, "step": 4305 }, { "epoch": 6.446107784431137, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1582, "step": 4306 }, { "epoch": 6.447604790419161, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1545, "step": 4307 }, { "epoch": 6.449101796407185, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1563, "step": 4308 }, { "epoch": 6.450598802395209, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.154, "step": 4309 }, { "epoch": 6.452095808383233, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1584, "step": 4310 }, { "epoch": 6.453592814371257, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1546, "step": 4311 }, { "epoch": 6.455089820359281, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1559, "step": 4312 }, { "epoch": 6.456586826347305, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1598, "step": 4313 }, { "epoch": 6.45808383233533, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.155, "step": 4314 }, { "epoch": 6.459580838323353, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1643, "step": 4315 }, { "epoch": 6.461077844311378, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1656, "step": 4316 }, { "epoch": 6.462574850299402, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.154, "step": 4317 }, { "epoch": 6.4640718562874255, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1538, "step": 4318 }, { "epoch": 6.4655688622754495, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1549, "step": 4319 }, { "epoch": 6.467065868263473, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1515, "step": 4320 }, { "epoch": 6.468562874251497, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1612, "step": 4321 }, { "epoch": 6.470059880239521, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1575, "step": 4322 }, { "epoch": 6.471556886227545, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1539, "step": 4323 }, { "epoch": 6.473053892215569, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1593, "step": 4324 }, { "epoch": 6.474550898203593, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1573, "step": 4325 }, { "epoch": 6.476047904191617, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1544, "step": 4326 }, { "epoch": 6.477544910179641, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1603, "step": 4327 }, { "epoch": 6.479041916167665, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.152, "step": 4328 }, { "epoch": 6.480538922155689, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1593, "step": 4329 }, { "epoch": 6.482035928143713, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1549, "step": 4330 }, { "epoch": 6.483532934131737, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1569, "step": 4331 }, { "epoch": 6.485029940119761, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1608, "step": 4332 }, { "epoch": 6.486526946107785, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1566, "step": 4333 }, { "epoch": 6.4880239520958085, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1527, "step": 4334 }, { "epoch": 6.4895209580838324, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.155, "step": 4335 }, { "epoch": 6.491017964071856, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1571, "step": 4336 }, { "epoch": 6.49251497005988, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.164, "step": 4337 }, { "epoch": 6.494011976047904, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1586, "step": 4338 }, { "epoch": 6.495508982035928, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1599, "step": 4339 }, { "epoch": 6.497005988023952, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1585, "step": 4340 }, { "epoch": 6.498502994011976, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1538, "step": 4341 }, { "epoch": 6.5, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1678, "step": 4342 }, { "epoch": 6.501497005988024, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1581, "step": 4343 }, { "epoch": 6.502994011976048, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1525, "step": 4344 }, { "epoch": 6.504491017964072, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1574, "step": 4345 }, { "epoch": 6.505988023952096, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1606, "step": 4346 }, { "epoch": 6.50748502994012, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1603, "step": 4347 }, { "epoch": 6.508982035928144, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1531, "step": 4348 }, { "epoch": 6.5104790419161676, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1566, "step": 4349 }, { "epoch": 6.5119760479041915, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.167, "step": 4350 }, { "epoch": 6.513473053892215, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1611, "step": 4351 }, { "epoch": 6.514970059880239, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1631, "step": 4352 }, { "epoch": 6.516467065868263, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1601, "step": 4353 }, { "epoch": 6.517964071856287, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1552, "step": 4354 }, { "epoch": 6.519461077844311, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1639, "step": 4355 }, { "epoch": 6.520958083832335, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1594, "step": 4356 }, { "epoch": 6.522455089820359, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1558, "step": 4357 }, { "epoch": 6.523952095808383, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1518, "step": 4358 }, { "epoch": 6.525449101796407, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.161, "step": 4359 }, { "epoch": 6.526946107784431, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1512, "step": 4360 }, { "epoch": 6.528443113772455, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1539, "step": 4361 }, { "epoch": 6.529940119760479, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1594, "step": 4362 }, { "epoch": 6.531437125748503, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1534, "step": 4363 }, { "epoch": 6.532934131736527, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1594, "step": 4364 }, { "epoch": 6.5344311377245505, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1609, "step": 4365 }, { "epoch": 6.5359281437125745, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1527, "step": 4366 }, { "epoch": 6.537425149700599, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1621, "step": 4367 }, { "epoch": 6.538922155688622, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1558, "step": 4368 }, { "epoch": 6.540419161676647, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1523, "step": 4369 }, { "epoch": 6.54191616766467, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.16, "step": 4370 }, { "epoch": 6.543413173652695, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1581, "step": 4371 }, { "epoch": 6.544910179640718, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1609, "step": 4372 }, { "epoch": 6.546407185628743, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1549, "step": 4373 }, { "epoch": 6.547904191616767, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1553, "step": 4374 }, { "epoch": 6.549401197604791, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1587, "step": 4375 }, { "epoch": 6.550898203592815, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1594, "step": 4376 }, { "epoch": 6.552395209580839, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1549, "step": 4377 }, { "epoch": 6.553892215568863, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1624, "step": 4378 }, { "epoch": 6.5553892215568865, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1482, "step": 4379 }, { "epoch": 6.5568862275449105, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1549, "step": 4380 }, { "epoch": 6.558383233532934, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1581, "step": 4381 }, { "epoch": 6.559880239520958, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1574, "step": 4382 }, { "epoch": 6.561377245508982, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1543, "step": 4383 }, { "epoch": 6.562874251497006, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1551, "step": 4384 }, { "epoch": 6.56437125748503, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1602, "step": 4385 }, { "epoch": 6.565868263473054, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1543, "step": 4386 }, { "epoch": 6.567365269461078, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1471, "step": 4387 }, { "epoch": 6.568862275449102, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1591, "step": 4388 }, { "epoch": 6.570359281437126, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1577, "step": 4389 }, { "epoch": 6.57185628742515, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1595, "step": 4390 }, { "epoch": 6.573353293413174, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1568, "step": 4391 }, { "epoch": 6.574850299401198, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1532, "step": 4392 }, { "epoch": 6.576347305389222, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1553, "step": 4393 }, { "epoch": 6.577844311377246, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1607, "step": 4394 }, { "epoch": 6.5793413173652695, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1612, "step": 4395 }, { "epoch": 6.580838323353293, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1565, "step": 4396 }, { "epoch": 6.582335329341317, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1573, "step": 4397 }, { "epoch": 6.583832335329341, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1559, "step": 4398 }, { "epoch": 6.585329341317365, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1592, "step": 4399 }, { "epoch": 6.586826347305389, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1546, "step": 4400 }, { "epoch": 6.588323353293413, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1595, "step": 4401 }, { "epoch": 6.589820359281437, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1601, "step": 4402 }, { "epoch": 6.591317365269461, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1617, "step": 4403 }, { "epoch": 6.592814371257485, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1613, "step": 4404 }, { "epoch": 6.594311377245509, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1602, "step": 4405 }, { "epoch": 6.595808383233533, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1538, "step": 4406 }, { "epoch": 6.597305389221557, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1544, "step": 4407 }, { "epoch": 6.598802395209581, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.168, "step": 4408 }, { "epoch": 6.600299401197605, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1551, "step": 4409 }, { "epoch": 6.6017964071856285, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1592, "step": 4410 }, { "epoch": 6.6032934131736525, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1539, "step": 4411 }, { "epoch": 6.604790419161676, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1595, "step": 4412 }, { "epoch": 6.6062874251497, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1621, "step": 4413 }, { "epoch": 6.607784431137724, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1509, "step": 4414 }, { "epoch": 6.609281437125748, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1551, "step": 4415 }, { "epoch": 6.610778443113772, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1618, "step": 4416 }, { "epoch": 6.612275449101796, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1591, "step": 4417 }, { "epoch": 6.61377245508982, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1645, "step": 4418 }, { "epoch": 6.615269461077844, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1576, "step": 4419 }, { "epoch": 6.616766467065868, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1515, "step": 4420 }, { "epoch": 6.618263473053892, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1526, "step": 4421 }, { "epoch": 6.619760479041916, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1571, "step": 4422 }, { "epoch": 6.62125748502994, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1571, "step": 4423 }, { "epoch": 6.6227544910179645, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1573, "step": 4424 }, { "epoch": 6.624251497005988, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.159, "step": 4425 }, { "epoch": 6.625748502994012, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1497, "step": 4426 }, { "epoch": 6.6272455089820355, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1553, "step": 4427 }, { "epoch": 6.62874251497006, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1602, "step": 4428 }, { "epoch": 6.630239520958084, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1577, "step": 4429 }, { "epoch": 6.631736526946108, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1624, "step": 4430 }, { "epoch": 6.633233532934132, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1586, "step": 4431 }, { "epoch": 6.634730538922156, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1607, "step": 4432 }, { "epoch": 6.63622754491018, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1593, "step": 4433 }, { "epoch": 6.637724550898204, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.159, "step": 4434 }, { "epoch": 6.639221556886228, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.161, "step": 4435 }, { "epoch": 6.640718562874252, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1548, "step": 4436 }, { "epoch": 6.642215568862276, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.16, "step": 4437 }, { "epoch": 6.6437125748503, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1565, "step": 4438 }, { "epoch": 6.645209580838324, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1585, "step": 4439 }, { "epoch": 6.6467065868263475, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1621, "step": 4440 }, { "epoch": 6.6482035928143715, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1557, "step": 4441 }, { "epoch": 6.649700598802395, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1554, "step": 4442 }, { "epoch": 6.651197604790419, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1594, "step": 4443 }, { "epoch": 6.652694610778443, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1563, "step": 4444 }, { "epoch": 6.654191616766467, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1597, "step": 4445 }, { "epoch": 6.655688622754491, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1533, "step": 4446 }, { "epoch": 6.657185628742515, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1591, "step": 4447 }, { "epoch": 6.658682634730539, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1534, "step": 4448 }, { "epoch": 6.660179640718563, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1552, "step": 4449 }, { "epoch": 6.661676646706587, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1556, "step": 4450 }, { "epoch": 6.663173652694611, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.156, "step": 4451 }, { "epoch": 6.664670658682635, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1491, "step": 4452 }, { "epoch": 6.666167664670659, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1492, "step": 4453 }, { "epoch": 6.667664670658683, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1548, "step": 4454 }, { "epoch": 6.669161676646707, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1529, "step": 4455 }, { "epoch": 6.6706586826347305, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1581, "step": 4456 }, { "epoch": 6.672155688622754, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1643, "step": 4457 }, { "epoch": 6.673652694610778, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1506, "step": 4458 }, { "epoch": 6.675149700598802, "grad_norm": 0.171875, "learning_rate": 0.0008, "loss": 1.1503, "step": 4459 }, { "epoch": 6.676646706586826, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.1637, "step": 4460 }, { "epoch": 6.67814371257485, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1612, "step": 4461 }, { "epoch": 6.679640718562874, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1584, "step": 4462 }, { "epoch": 6.681137724550898, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.154, "step": 4463 }, { "epoch": 6.682634730538922, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1519, "step": 4464 }, { "epoch": 6.684131736526946, "grad_norm": 0.15234375, "learning_rate": 0.0008, "loss": 1.161, "step": 4465 }, { "epoch": 6.68562874251497, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.1594, "step": 4466 }, { "epoch": 6.687125748502994, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1658, "step": 4467 }, { "epoch": 6.688622754491018, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1567, "step": 4468 }, { "epoch": 6.690119760479042, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1535, "step": 4469 }, { "epoch": 6.691616766467066, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1498, "step": 4470 }, { "epoch": 6.6931137724550895, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.153, "step": 4471 }, { "epoch": 6.6946107784431135, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.159, "step": 4472 }, { "epoch": 6.696107784431137, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1575, "step": 4473 }, { "epoch": 6.697604790419161, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1593, "step": 4474 }, { "epoch": 6.699101796407185, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1571, "step": 4475 }, { "epoch": 6.700598802395209, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1571, "step": 4476 }, { "epoch": 6.702095808383233, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1659, "step": 4477 }, { "epoch": 6.703592814371257, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.155, "step": 4478 }, { "epoch": 6.705089820359282, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1515, "step": 4479 }, { "epoch": 6.706586826347305, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1579, "step": 4480 }, { "epoch": 6.70808383233533, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1581, "step": 4481 }, { "epoch": 6.709580838323353, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1598, "step": 4482 }, { "epoch": 6.711077844311378, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1664, "step": 4483 }, { "epoch": 6.712574850299401, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1508, "step": 4484 }, { "epoch": 6.7140718562874255, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1535, "step": 4485 }, { "epoch": 6.7155688622754495, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1566, "step": 4486 }, { "epoch": 6.717065868263473, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1643, "step": 4487 }, { "epoch": 6.718562874251497, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1602, "step": 4488 }, { "epoch": 6.720059880239521, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1543, "step": 4489 }, { "epoch": 6.721556886227545, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1533, "step": 4490 }, { "epoch": 6.723053892215569, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1545, "step": 4491 }, { "epoch": 6.724550898203593, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1579, "step": 4492 }, { "epoch": 6.726047904191617, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1598, "step": 4493 }, { "epoch": 6.727544910179641, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1546, "step": 4494 }, { "epoch": 6.729041916167665, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1438, "step": 4495 }, { "epoch": 6.730538922155689, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1524, "step": 4496 }, { "epoch": 6.732035928143713, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1575, "step": 4497 }, { "epoch": 6.733532934131737, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1545, "step": 4498 }, { "epoch": 6.735029940119761, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1511, "step": 4499 }, { "epoch": 6.736526946107785, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1617, "step": 4500 }, { "epoch": 6.7380239520958085, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.16, "step": 4501 }, { "epoch": 6.7395209580838324, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1516, "step": 4502 }, { "epoch": 6.741017964071856, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1585, "step": 4503 }, { "epoch": 6.74251497005988, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1543, "step": 4504 }, { "epoch": 6.744011976047904, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.162, "step": 4505 }, { "epoch": 6.745508982035928, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1656, "step": 4506 }, { "epoch": 6.747005988023952, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1579, "step": 4507 }, { "epoch": 6.748502994011976, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1574, "step": 4508 }, { "epoch": 6.75, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1523, "step": 4509 }, { "epoch": 6.751497005988024, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1478, "step": 4510 }, { "epoch": 6.752994011976048, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1555, "step": 4511 }, { "epoch": 6.754491017964072, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.158, "step": 4512 }, { "epoch": 6.755988023952096, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1506, "step": 4513 }, { "epoch": 6.75748502994012, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1531, "step": 4514 }, { "epoch": 6.758982035928144, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1534, "step": 4515 }, { "epoch": 6.7604790419161676, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1555, "step": 4516 }, { "epoch": 6.7619760479041915, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1594, "step": 4517 }, { "epoch": 6.763473053892215, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1586, "step": 4518 }, { "epoch": 6.764970059880239, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1582, "step": 4519 }, { "epoch": 6.766467065868263, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.151, "step": 4520 }, { "epoch": 6.767964071856287, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1528, "step": 4521 }, { "epoch": 6.769461077844311, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1633, "step": 4522 }, { "epoch": 6.770958083832335, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1611, "step": 4523 }, { "epoch": 6.772455089820359, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1554, "step": 4524 }, { "epoch": 6.773952095808383, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1598, "step": 4525 }, { "epoch": 6.775449101796407, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1608, "step": 4526 }, { "epoch": 6.776946107784431, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.161, "step": 4527 }, { "epoch": 6.778443113772455, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1546, "step": 4528 }, { "epoch": 6.779940119760479, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1599, "step": 4529 }, { "epoch": 6.781437125748503, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1596, "step": 4530 }, { "epoch": 6.782934131736527, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1571, "step": 4531 }, { "epoch": 6.7844311377245505, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1494, "step": 4532 }, { "epoch": 6.7859281437125745, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1545, "step": 4533 }, { "epoch": 6.787425149700599, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1542, "step": 4534 }, { "epoch": 6.788922155688622, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1573, "step": 4535 }, { "epoch": 6.790419161676647, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1541, "step": 4536 }, { "epoch": 6.79191616766467, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1551, "step": 4537 }, { "epoch": 6.793413173652695, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1603, "step": 4538 }, { "epoch": 6.794910179640718, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1561, "step": 4539 }, { "epoch": 6.796407185628743, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.154, "step": 4540 }, { "epoch": 6.797904191616767, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1503, "step": 4541 }, { "epoch": 6.799401197604791, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1623, "step": 4542 }, { "epoch": 6.800898203592815, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1581, "step": 4543 }, { "epoch": 6.802395209580839, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1547, "step": 4544 }, { "epoch": 6.803892215568863, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1584, "step": 4545 }, { "epoch": 6.8053892215568865, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1583, "step": 4546 }, { "epoch": 6.8068862275449105, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1564, "step": 4547 }, { "epoch": 6.808383233532934, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1581, "step": 4548 }, { "epoch": 6.809880239520958, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1533, "step": 4549 }, { "epoch": 6.811377245508982, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1602, "step": 4550 }, { "epoch": 6.812874251497006, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1553, "step": 4551 }, { "epoch": 6.81437125748503, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1543, "step": 4552 }, { "epoch": 6.815868263473054, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1487, "step": 4553 }, { "epoch": 6.817365269461078, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1563, "step": 4554 }, { "epoch": 6.818862275449102, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1582, "step": 4555 }, { "epoch": 6.820359281437126, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1554, "step": 4556 }, { "epoch": 6.82185628742515, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1592, "step": 4557 }, { "epoch": 6.823353293413174, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1589, "step": 4558 }, { "epoch": 6.824850299401198, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1565, "step": 4559 }, { "epoch": 6.826347305389222, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1527, "step": 4560 }, { "epoch": 6.827844311377246, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1562, "step": 4561 }, { "epoch": 6.8293413173652695, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.158, "step": 4562 }, { "epoch": 6.830838323353293, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1606, "step": 4563 }, { "epoch": 6.832335329341317, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1528, "step": 4564 }, { "epoch": 6.833832335329341, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.156, "step": 4565 }, { "epoch": 6.835329341317365, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1537, "step": 4566 }, { "epoch": 6.836826347305389, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1559, "step": 4567 }, { "epoch": 6.838323353293413, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1572, "step": 4568 }, { "epoch": 6.839820359281437, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1549, "step": 4569 }, { "epoch": 6.841317365269461, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1569, "step": 4570 }, { "epoch": 6.842814371257485, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1552, "step": 4571 }, { "epoch": 6.844311377245509, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1471, "step": 4572 }, { "epoch": 6.845808383233533, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1608, "step": 4573 }, { "epoch": 6.847305389221557, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1536, "step": 4574 }, { "epoch": 6.848802395209581, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1531, "step": 4575 }, { "epoch": 6.850299401197605, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1484, "step": 4576 }, { "epoch": 6.8517964071856285, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1551, "step": 4577 }, { "epoch": 6.8532934131736525, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.153, "step": 4578 }, { "epoch": 6.854790419161676, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.156, "step": 4579 }, { "epoch": 6.8562874251497, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1617, "step": 4580 }, { "epoch": 6.857784431137724, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1581, "step": 4581 }, { "epoch": 6.859281437125748, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1596, "step": 4582 }, { "epoch": 6.860778443113772, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.16, "step": 4583 }, { "epoch": 6.862275449101796, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1532, "step": 4584 }, { "epoch": 6.86377245508982, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1509, "step": 4585 }, { "epoch": 6.865269461077844, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1598, "step": 4586 }, { "epoch": 6.866766467065868, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1492, "step": 4587 }, { "epoch": 6.868263473053892, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1619, "step": 4588 }, { "epoch": 6.869760479041916, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1594, "step": 4589 }, { "epoch": 6.87125748502994, "grad_norm": 0.15625, "learning_rate": 0.0008, "loss": 1.1551, "step": 4590 }, { "epoch": 6.8727544910179645, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.161, "step": 4591 }, { "epoch": 6.874251497005988, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.153, "step": 4592 }, { "epoch": 6.875748502994012, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1583, "step": 4593 }, { "epoch": 6.8772455089820355, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1583, "step": 4594 }, { "epoch": 6.87874251497006, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1581, "step": 4595 }, { "epoch": 6.880239520958084, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1581, "step": 4596 }, { "epoch": 6.881736526946108, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1548, "step": 4597 }, { "epoch": 6.883233532934132, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1519, "step": 4598 }, { "epoch": 6.884730538922156, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1505, "step": 4599 }, { "epoch": 6.88622754491018, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1521, "step": 4600 }, { "epoch": 6.887724550898204, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1516, "step": 4601 }, { "epoch": 6.889221556886228, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1588, "step": 4602 }, { "epoch": 6.890718562874252, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1535, "step": 4603 }, { "epoch": 6.892215568862276, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1571, "step": 4604 }, { "epoch": 6.8937125748503, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1528, "step": 4605 }, { "epoch": 6.895209580838324, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1529, "step": 4606 }, { "epoch": 6.8967065868263475, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1538, "step": 4607 }, { "epoch": 6.8982035928143715, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1546, "step": 4608 }, { "epoch": 6.899700598802395, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1516, "step": 4609 }, { "epoch": 6.901197604790419, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.15, "step": 4610 }, { "epoch": 6.902694610778443, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1537, "step": 4611 }, { "epoch": 6.904191616766467, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1603, "step": 4612 }, { "epoch": 6.905688622754491, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.158, "step": 4613 }, { "epoch": 6.907185628742515, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1535, "step": 4614 }, { "epoch": 6.908682634730539, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1538, "step": 4615 }, { "epoch": 6.910179640718563, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1578, "step": 4616 }, { "epoch": 6.911676646706587, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1522, "step": 4617 }, { "epoch": 6.913173652694611, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1618, "step": 4618 }, { "epoch": 6.914670658682635, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1523, "step": 4619 }, { "epoch": 6.916167664670659, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.157, "step": 4620 }, { "epoch": 6.917664670658683, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1597, "step": 4621 }, { "epoch": 6.919161676646707, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1594, "step": 4622 }, { "epoch": 6.9206586826347305, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.161, "step": 4623 }, { "epoch": 6.922155688622754, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1541, "step": 4624 }, { "epoch": 6.923652694610778, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1573, "step": 4625 }, { "epoch": 6.925149700598802, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1644, "step": 4626 }, { "epoch": 6.926646706586826, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1543, "step": 4627 }, { "epoch": 6.92814371257485, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1557, "step": 4628 }, { "epoch": 6.929640718562874, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1597, "step": 4629 }, { "epoch": 6.931137724550898, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.151, "step": 4630 }, { "epoch": 6.932634730538922, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1566, "step": 4631 }, { "epoch": 6.934131736526946, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1605, "step": 4632 }, { "epoch": 6.93562874251497, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.159, "step": 4633 }, { "epoch": 6.937125748502994, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.153, "step": 4634 }, { "epoch": 6.938622754491018, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1532, "step": 4635 }, { "epoch": 6.940119760479042, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1555, "step": 4636 }, { "epoch": 6.941616766467066, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1646, "step": 4637 }, { "epoch": 6.9431137724550895, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1598, "step": 4638 }, { "epoch": 6.9446107784431135, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1531, "step": 4639 }, { "epoch": 6.946107784431137, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1601, "step": 4640 }, { "epoch": 6.947604790419161, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1565, "step": 4641 }, { "epoch": 6.949101796407185, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.153, "step": 4642 }, { "epoch": 6.950598802395209, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1601, "step": 4643 }, { "epoch": 6.952095808383233, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1555, "step": 4644 }, { "epoch": 6.953592814371257, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.16, "step": 4645 }, { "epoch": 6.955089820359282, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1599, "step": 4646 }, { "epoch": 6.956586826347305, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1533, "step": 4647 }, { "epoch": 6.95808383233533, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1597, "step": 4648 }, { "epoch": 6.959580838323353, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1539, "step": 4649 }, { "epoch": 6.961077844311378, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1583, "step": 4650 }, { "epoch": 6.962574850299401, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1547, "step": 4651 }, { "epoch": 6.9640718562874255, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.16, "step": 4652 }, { "epoch": 6.9655688622754495, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1578, "step": 4653 }, { "epoch": 6.967065868263473, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1522, "step": 4654 }, { "epoch": 6.968562874251497, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1494, "step": 4655 }, { "epoch": 6.970059880239521, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1634, "step": 4656 }, { "epoch": 6.971556886227545, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1506, "step": 4657 }, { "epoch": 6.973053892215569, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1566, "step": 4658 }, { "epoch": 6.974550898203593, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1505, "step": 4659 }, { "epoch": 6.976047904191617, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1625, "step": 4660 }, { "epoch": 6.977544910179641, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1603, "step": 4661 }, { "epoch": 6.979041916167665, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.157, "step": 4662 }, { "epoch": 6.980538922155689, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1574, "step": 4663 }, { "epoch": 6.982035928143713, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1509, "step": 4664 }, { "epoch": 6.983532934131737, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1619, "step": 4665 }, { "epoch": 6.985029940119761, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1596, "step": 4666 }, { "epoch": 6.986526946107785, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1537, "step": 4667 }, { "epoch": 6.9880239520958085, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.157, "step": 4668 }, { "epoch": 6.9895209580838324, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1564, "step": 4669 }, { "epoch": 6.991017964071856, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1537, "step": 4670 }, { "epoch": 6.99251497005988, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1597, "step": 4671 }, { "epoch": 6.994011976047904, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1584, "step": 4672 }, { "epoch": 6.995508982035928, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1563, "step": 4673 }, { "epoch": 6.997005988023952, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1561, "step": 4674 }, { "epoch": 6.998502994011976, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1582, "step": 4675 }, { "epoch": 7.0, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1549, "step": 4676 }, { "epoch": 7.001497005988024, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1562, "step": 4677 }, { "epoch": 7.002994011976048, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1631, "step": 4678 }, { "epoch": 7.004491017964072, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1555, "step": 4679 }, { "epoch": 7.005988023952096, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1591, "step": 4680 }, { "epoch": 7.00748502994012, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1508, "step": 4681 }, { "epoch": 7.008982035928144, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1619, "step": 4682 }, { "epoch": 7.0104790419161676, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1511, "step": 4683 }, { "epoch": 7.0119760479041915, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1594, "step": 4684 }, { "epoch": 7.013473053892215, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1527, "step": 4685 }, { "epoch": 7.014970059880239, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.148, "step": 4686 }, { "epoch": 7.016467065868263, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1663, "step": 4687 }, { "epoch": 7.017964071856287, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1596, "step": 4688 }, { "epoch": 7.019461077844311, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1545, "step": 4689 }, { "epoch": 7.020958083832335, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1554, "step": 4690 }, { "epoch": 7.022455089820359, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1544, "step": 4691 }, { "epoch": 7.023952095808383, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1513, "step": 4692 }, { "epoch": 7.025449101796407, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1531, "step": 4693 }, { "epoch": 7.026946107784431, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.155, "step": 4694 }, { "epoch": 7.028443113772455, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1591, "step": 4695 }, { "epoch": 7.029940119760479, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1539, "step": 4696 }, { "epoch": 7.031437125748503, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1614, "step": 4697 }, { "epoch": 7.032934131736527, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1565, "step": 4698 }, { "epoch": 7.0344311377245505, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1586, "step": 4699 }, { "epoch": 7.0359281437125745, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1545, "step": 4700 }, { "epoch": 7.037425149700598, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1591, "step": 4701 }, { "epoch": 7.038922155688622, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1565, "step": 4702 }, { "epoch": 7.040419161676646, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1458, "step": 4703 }, { "epoch": 7.04191616766467, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1548, "step": 4704 }, { "epoch": 7.043413173652695, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1565, "step": 4705 }, { "epoch": 7.044910179640719, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1562, "step": 4706 }, { "epoch": 7.046407185628743, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1436, "step": 4707 }, { "epoch": 7.047904191616767, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1601, "step": 4708 }, { "epoch": 7.049401197604791, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1557, "step": 4709 }, { "epoch": 7.050898203592815, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1563, "step": 4710 }, { "epoch": 7.052395209580839, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1543, "step": 4711 }, { "epoch": 7.053892215568863, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.1478, "step": 4712 }, { "epoch": 7.0553892215568865, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1536, "step": 4713 }, { "epoch": 7.0568862275449105, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1559, "step": 4714 }, { "epoch": 7.058383233532934, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1525, "step": 4715 }, { "epoch": 7.059880239520958, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1574, "step": 4716 }, { "epoch": 7.061377245508982, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1513, "step": 4717 }, { "epoch": 7.062874251497006, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1463, "step": 4718 }, { "epoch": 7.06437125748503, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1553, "step": 4719 }, { "epoch": 7.065868263473054, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1536, "step": 4720 }, { "epoch": 7.067365269461078, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1573, "step": 4721 }, { "epoch": 7.068862275449102, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1471, "step": 4722 }, { "epoch": 7.070359281437126, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1593, "step": 4723 }, { "epoch": 7.07185628742515, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1575, "step": 4724 }, { "epoch": 7.073353293413174, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1511, "step": 4725 }, { "epoch": 7.074850299401198, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1533, "step": 4726 }, { "epoch": 7.076347305389222, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1526, "step": 4727 }, { "epoch": 7.077844311377246, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1583, "step": 4728 }, { "epoch": 7.0793413173652695, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1575, "step": 4729 }, { "epoch": 7.080838323353293, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1613, "step": 4730 }, { "epoch": 7.082335329341317, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1535, "step": 4731 }, { "epoch": 7.083832335329341, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1531, "step": 4732 }, { "epoch": 7.085329341317365, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1531, "step": 4733 }, { "epoch": 7.086826347305389, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1531, "step": 4734 }, { "epoch": 7.088323353293413, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1573, "step": 4735 }, { "epoch": 7.089820359281437, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1488, "step": 4736 }, { "epoch": 7.091317365269461, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1579, "step": 4737 }, { "epoch": 7.092814371257485, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1467, "step": 4738 }, { "epoch": 7.094311377245509, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1503, "step": 4739 }, { "epoch": 7.095808383233533, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1583, "step": 4740 }, { "epoch": 7.097305389221557, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.155, "step": 4741 }, { "epoch": 7.098802395209581, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.161, "step": 4742 }, { "epoch": 7.100299401197605, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1586, "step": 4743 }, { "epoch": 7.1017964071856285, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1537, "step": 4744 }, { "epoch": 7.1032934131736525, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1561, "step": 4745 }, { "epoch": 7.104790419161676, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1508, "step": 4746 }, { "epoch": 7.1062874251497, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1526, "step": 4747 }, { "epoch": 7.107784431137724, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1542, "step": 4748 }, { "epoch": 7.109281437125748, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1536, "step": 4749 }, { "epoch": 7.110778443113772, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1584, "step": 4750 }, { "epoch": 7.112275449101796, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1592, "step": 4751 }, { "epoch": 7.11377245508982, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.154, "step": 4752 }, { "epoch": 7.115269461077844, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1497, "step": 4753 }, { "epoch": 7.116766467065868, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1592, "step": 4754 }, { "epoch": 7.118263473053892, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1583, "step": 4755 }, { "epoch": 7.119760479041916, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1606, "step": 4756 }, { "epoch": 7.12125748502994, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1548, "step": 4757 }, { "epoch": 7.122754491017964, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.162, "step": 4758 }, { "epoch": 7.124251497005988, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1553, "step": 4759 }, { "epoch": 7.125748502994012, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1515, "step": 4760 }, { "epoch": 7.127245508982036, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1557, "step": 4761 }, { "epoch": 7.12874251497006, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1598, "step": 4762 }, { "epoch": 7.130239520958084, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1617, "step": 4763 }, { "epoch": 7.131736526946108, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1573, "step": 4764 }, { "epoch": 7.133233532934132, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1538, "step": 4765 }, { "epoch": 7.134730538922156, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1546, "step": 4766 }, { "epoch": 7.13622754491018, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1484, "step": 4767 }, { "epoch": 7.137724550898204, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1601, "step": 4768 }, { "epoch": 7.139221556886228, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1497, "step": 4769 }, { "epoch": 7.140718562874252, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1504, "step": 4770 }, { "epoch": 7.142215568862276, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1535, "step": 4771 }, { "epoch": 7.1437125748503, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1549, "step": 4772 }, { "epoch": 7.145209580838324, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1534, "step": 4773 }, { "epoch": 7.1467065868263475, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1618, "step": 4774 }, { "epoch": 7.1482035928143715, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1539, "step": 4775 }, { "epoch": 7.149700598802395, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1516, "step": 4776 }, { "epoch": 7.151197604790419, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1566, "step": 4777 }, { "epoch": 7.152694610778443, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1668, "step": 4778 }, { "epoch": 7.154191616766467, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1537, "step": 4779 }, { "epoch": 7.155688622754491, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.157, "step": 4780 }, { "epoch": 7.157185628742515, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1592, "step": 4781 }, { "epoch": 7.158682634730539, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1589, "step": 4782 }, { "epoch": 7.160179640718563, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1576, "step": 4783 }, { "epoch": 7.161676646706587, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1563, "step": 4784 }, { "epoch": 7.163173652694611, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1587, "step": 4785 }, { "epoch": 7.164670658682635, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1552, "step": 4786 }, { "epoch": 7.166167664670659, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1552, "step": 4787 }, { "epoch": 7.167664670658683, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1511, "step": 4788 }, { "epoch": 7.169161676646707, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1534, "step": 4789 }, { "epoch": 7.1706586826347305, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1609, "step": 4790 }, { "epoch": 7.172155688622754, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1517, "step": 4791 }, { "epoch": 7.173652694610778, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1526, "step": 4792 }, { "epoch": 7.175149700598802, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1555, "step": 4793 }, { "epoch": 7.176646706586826, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1546, "step": 4794 }, { "epoch": 7.17814371257485, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1558, "step": 4795 }, { "epoch": 7.179640718562874, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1475, "step": 4796 }, { "epoch": 7.181137724550898, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1588, "step": 4797 }, { "epoch": 7.182634730538922, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1516, "step": 4798 }, { "epoch": 7.184131736526946, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1535, "step": 4799 }, { "epoch": 7.18562874251497, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1521, "step": 4800 }, { "epoch": 7.187125748502994, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1533, "step": 4801 }, { "epoch": 7.188622754491018, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.159, "step": 4802 }, { "epoch": 7.190119760479042, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1566, "step": 4803 }, { "epoch": 7.191616766467066, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1496, "step": 4804 }, { "epoch": 7.1931137724550895, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1581, "step": 4805 }, { "epoch": 7.1946107784431135, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1511, "step": 4806 }, { "epoch": 7.196107784431137, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1503, "step": 4807 }, { "epoch": 7.197604790419161, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1568, "step": 4808 }, { "epoch": 7.199101796407185, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1549, "step": 4809 }, { "epoch": 7.200598802395209, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1572, "step": 4810 }, { "epoch": 7.202095808383233, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1534, "step": 4811 }, { "epoch": 7.203592814371257, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1575, "step": 4812 }, { "epoch": 7.205089820359281, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1495, "step": 4813 }, { "epoch": 7.206586826347305, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1621, "step": 4814 }, { "epoch": 7.20808383233533, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1536, "step": 4815 }, { "epoch": 7.209580838323353, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.157, "step": 4816 }, { "epoch": 7.211077844311378, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1567, "step": 4817 }, { "epoch": 7.212574850299402, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1612, "step": 4818 }, { "epoch": 7.2140718562874255, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1609, "step": 4819 }, { "epoch": 7.2155688622754495, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1537, "step": 4820 }, { "epoch": 7.217065868263473, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1504, "step": 4821 }, { "epoch": 7.218562874251497, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1577, "step": 4822 }, { "epoch": 7.220059880239521, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1554, "step": 4823 }, { "epoch": 7.221556886227545, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1477, "step": 4824 }, { "epoch": 7.223053892215569, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1529, "step": 4825 }, { "epoch": 7.224550898203593, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1577, "step": 4826 }, { "epoch": 7.226047904191617, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1526, "step": 4827 }, { "epoch": 7.227544910179641, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1559, "step": 4828 }, { "epoch": 7.229041916167665, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1482, "step": 4829 }, { "epoch": 7.230538922155689, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1467, "step": 4830 }, { "epoch": 7.232035928143713, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.1556, "step": 4831 }, { "epoch": 7.233532934131737, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1531, "step": 4832 }, { "epoch": 7.235029940119761, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1585, "step": 4833 }, { "epoch": 7.236526946107785, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1571, "step": 4834 }, { "epoch": 7.2380239520958085, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1588, "step": 4835 }, { "epoch": 7.2395209580838324, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1494, "step": 4836 }, { "epoch": 7.241017964071856, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1565, "step": 4837 }, { "epoch": 7.24251497005988, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1589, "step": 4838 }, { "epoch": 7.244011976047904, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1554, "step": 4839 }, { "epoch": 7.245508982035928, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1621, "step": 4840 }, { "epoch": 7.247005988023952, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1549, "step": 4841 }, { "epoch": 7.248502994011976, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1494, "step": 4842 }, { "epoch": 7.25, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.158, "step": 4843 }, { "epoch": 7.251497005988024, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1578, "step": 4844 }, { "epoch": 7.252994011976048, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1535, "step": 4845 }, { "epoch": 7.254491017964072, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1522, "step": 4846 }, { "epoch": 7.255988023952096, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1529, "step": 4847 }, { "epoch": 7.25748502994012, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1543, "step": 4848 }, { "epoch": 7.258982035928144, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1499, "step": 4849 }, { "epoch": 7.2604790419161676, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1507, "step": 4850 }, { "epoch": 7.2619760479041915, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1522, "step": 4851 }, { "epoch": 7.263473053892215, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1579, "step": 4852 }, { "epoch": 7.264970059880239, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1562, "step": 4853 }, { "epoch": 7.266467065868263, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1587, "step": 4854 }, { "epoch": 7.267964071856287, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1567, "step": 4855 }, { "epoch": 7.269461077844311, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1528, "step": 4856 }, { "epoch": 7.270958083832335, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1474, "step": 4857 }, { "epoch": 7.272455089820359, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1577, "step": 4858 }, { "epoch": 7.273952095808383, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1507, "step": 4859 }, { "epoch": 7.275449101796407, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1529, "step": 4860 }, { "epoch": 7.276946107784431, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1519, "step": 4861 }, { "epoch": 7.278443113772455, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1512, "step": 4862 }, { "epoch": 7.279940119760479, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.1605, "step": 4863 }, { "epoch": 7.281437125748503, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1574, "step": 4864 }, { "epoch": 7.282934131736527, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1557, "step": 4865 }, { "epoch": 7.2844311377245505, "grad_norm": 0.050048828125, "learning_rate": 0.0008, "loss": 1.1523, "step": 4866 }, { "epoch": 7.2859281437125745, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1522, "step": 4867 }, { "epoch": 7.287425149700598, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1508, "step": 4868 }, { "epoch": 7.288922155688622, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1553, "step": 4869 }, { "epoch": 7.290419161676647, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1511, "step": 4870 }, { "epoch": 7.29191616766467, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1581, "step": 4871 }, { "epoch": 7.293413173652695, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1597, "step": 4872 }, { "epoch": 7.294910179640719, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1465, "step": 4873 }, { "epoch": 7.296407185628743, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1543, "step": 4874 }, { "epoch": 7.297904191616767, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1465, "step": 4875 }, { "epoch": 7.299401197604791, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.152, "step": 4876 }, { "epoch": 7.300898203592815, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1587, "step": 4877 }, { "epoch": 7.302395209580839, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1497, "step": 4878 }, { "epoch": 7.303892215568863, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1579, "step": 4879 }, { "epoch": 7.3053892215568865, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1602, "step": 4880 }, { "epoch": 7.3068862275449105, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1535, "step": 4881 }, { "epoch": 7.308383233532934, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1496, "step": 4882 }, { "epoch": 7.309880239520958, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.155, "step": 4883 }, { "epoch": 7.311377245508982, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1537, "step": 4884 }, { "epoch": 7.312874251497006, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1519, "step": 4885 }, { "epoch": 7.31437125748503, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1578, "step": 4886 }, { "epoch": 7.315868263473054, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.1556, "step": 4887 }, { "epoch": 7.317365269461078, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1555, "step": 4888 }, { "epoch": 7.318862275449102, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1554, "step": 4889 }, { "epoch": 7.320359281437126, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1546, "step": 4890 }, { "epoch": 7.32185628742515, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1576, "step": 4891 }, { "epoch": 7.323353293413174, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1531, "step": 4892 }, { "epoch": 7.324850299401198, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1521, "step": 4893 }, { "epoch": 7.326347305389222, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1536, "step": 4894 }, { "epoch": 7.327844311377246, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.15, "step": 4895 }, { "epoch": 7.3293413173652695, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1538, "step": 4896 }, { "epoch": 7.330838323353293, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1549, "step": 4897 }, { "epoch": 7.332335329341317, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.159, "step": 4898 }, { "epoch": 7.333832335329341, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1529, "step": 4899 }, { "epoch": 7.335329341317365, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1522, "step": 4900 }, { "epoch": 7.336826347305389, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1528, "step": 4901 }, { "epoch": 7.338323353293413, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.158, "step": 4902 }, { "epoch": 7.339820359281437, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1536, "step": 4903 }, { "epoch": 7.341317365269461, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1512, "step": 4904 }, { "epoch": 7.342814371257485, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1505, "step": 4905 }, { "epoch": 7.344311377245509, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1591, "step": 4906 }, { "epoch": 7.345808383233533, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1577, "step": 4907 }, { "epoch": 7.347305389221557, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1586, "step": 4908 }, { "epoch": 7.348802395209581, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1479, "step": 4909 }, { "epoch": 7.350299401197605, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1548, "step": 4910 }, { "epoch": 7.3517964071856285, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1536, "step": 4911 }, { "epoch": 7.3532934131736525, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1571, "step": 4912 }, { "epoch": 7.354790419161676, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1525, "step": 4913 }, { "epoch": 7.3562874251497, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1569, "step": 4914 }, { "epoch": 7.357784431137724, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1532, "step": 4915 }, { "epoch": 7.359281437125748, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1605, "step": 4916 }, { "epoch": 7.360778443113772, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.155, "step": 4917 }, { "epoch": 7.362275449101796, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1516, "step": 4918 }, { "epoch": 7.36377245508982, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1624, "step": 4919 }, { "epoch": 7.365269461077844, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1643, "step": 4920 }, { "epoch": 7.366766467065868, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.154, "step": 4921 }, { "epoch": 7.368263473053892, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1529, "step": 4922 }, { "epoch": 7.369760479041916, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1547, "step": 4923 }, { "epoch": 7.37125748502994, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1528, "step": 4924 }, { "epoch": 7.3727544910179645, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1562, "step": 4925 }, { "epoch": 7.374251497005988, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1527, "step": 4926 }, { "epoch": 7.375748502994012, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1542, "step": 4927 }, { "epoch": 7.3772455089820355, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1605, "step": 4928 }, { "epoch": 7.37874251497006, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.157, "step": 4929 }, { "epoch": 7.380239520958084, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1589, "step": 4930 }, { "epoch": 7.381736526946108, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1571, "step": 4931 }, { "epoch": 7.383233532934132, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1519, "step": 4932 }, { "epoch": 7.384730538922156, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1595, "step": 4933 }, { "epoch": 7.38622754491018, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1609, "step": 4934 }, { "epoch": 7.387724550898204, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1568, "step": 4935 }, { "epoch": 7.389221556886228, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1547, "step": 4936 }, { "epoch": 7.390718562874252, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1574, "step": 4937 }, { "epoch": 7.392215568862276, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1567, "step": 4938 }, { "epoch": 7.3937125748503, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1606, "step": 4939 }, { "epoch": 7.395209580838324, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1541, "step": 4940 }, { "epoch": 7.3967065868263475, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1536, "step": 4941 }, { "epoch": 7.3982035928143715, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.156, "step": 4942 }, { "epoch": 7.399700598802395, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1561, "step": 4943 }, { "epoch": 7.401197604790419, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1527, "step": 4944 }, { "epoch": 7.402694610778443, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1549, "step": 4945 }, { "epoch": 7.404191616766467, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1525, "step": 4946 }, { "epoch": 7.405688622754491, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1526, "step": 4947 }, { "epoch": 7.407185628742515, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1601, "step": 4948 }, { "epoch": 7.408682634730539, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1565, "step": 4949 }, { "epoch": 7.410179640718563, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1547, "step": 4950 }, { "epoch": 7.411676646706587, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1625, "step": 4951 }, { "epoch": 7.413173652694611, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1565, "step": 4952 }, { "epoch": 7.414670658682635, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1525, "step": 4953 }, { "epoch": 7.416167664670659, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1553, "step": 4954 }, { "epoch": 7.417664670658683, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1555, "step": 4955 }, { "epoch": 7.419161676646707, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1535, "step": 4956 }, { "epoch": 7.4206586826347305, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1544, "step": 4957 }, { "epoch": 7.422155688622754, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1524, "step": 4958 }, { "epoch": 7.423652694610778, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1517, "step": 4959 }, { "epoch": 7.425149700598802, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1557, "step": 4960 }, { "epoch": 7.426646706586826, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1564, "step": 4961 }, { "epoch": 7.42814371257485, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1551, "step": 4962 }, { "epoch": 7.429640718562874, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1564, "step": 4963 }, { "epoch": 7.431137724550898, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1555, "step": 4964 }, { "epoch": 7.432634730538922, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1546, "step": 4965 }, { "epoch": 7.434131736526946, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1525, "step": 4966 }, { "epoch": 7.43562874251497, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1527, "step": 4967 }, { "epoch": 7.437125748502994, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1528, "step": 4968 }, { "epoch": 7.438622754491018, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1581, "step": 4969 }, { "epoch": 7.440119760479042, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1575, "step": 4970 }, { "epoch": 7.441616766467066, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1509, "step": 4971 }, { "epoch": 7.4431137724550895, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1593, "step": 4972 }, { "epoch": 7.4446107784431135, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1572, "step": 4973 }, { "epoch": 7.446107784431137, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1537, "step": 4974 }, { "epoch": 7.447604790419161, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1556, "step": 4975 }, { "epoch": 7.449101796407185, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1523, "step": 4976 }, { "epoch": 7.450598802395209, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1577, "step": 4977 }, { "epoch": 7.452095808383233, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1539, "step": 4978 }, { "epoch": 7.453592814371257, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1581, "step": 4979 }, { "epoch": 7.455089820359281, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1529, "step": 4980 }, { "epoch": 7.456586826347305, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.154, "step": 4981 }, { "epoch": 7.45808383233533, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1578, "step": 4982 }, { "epoch": 7.459580838323353, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1541, "step": 4983 }, { "epoch": 7.461077844311378, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1547, "step": 4984 }, { "epoch": 7.462574850299402, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1493, "step": 4985 }, { "epoch": 7.4640718562874255, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1526, "step": 4986 }, { "epoch": 7.4655688622754495, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1585, "step": 4987 }, { "epoch": 7.467065868263473, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.151, "step": 4988 }, { "epoch": 7.468562874251497, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.148, "step": 4989 }, { "epoch": 7.470059880239521, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1526, "step": 4990 }, { "epoch": 7.471556886227545, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1556, "step": 4991 }, { "epoch": 7.473053892215569, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1493, "step": 4992 }, { "epoch": 7.474550898203593, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1503, "step": 4993 }, { "epoch": 7.476047904191617, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1577, "step": 4994 }, { "epoch": 7.477544910179641, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1559, "step": 4995 }, { "epoch": 7.479041916167665, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.161, "step": 4996 }, { "epoch": 7.480538922155689, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1527, "step": 4997 }, { "epoch": 7.482035928143713, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1498, "step": 4998 }, { "epoch": 7.483532934131737, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1608, "step": 4999 }, { "epoch": 7.485029940119761, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1498, "step": 5000 }, { "epoch": 7.486526946107785, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.1603, "step": 5001 }, { "epoch": 7.4880239520958085, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1504, "step": 5002 }, { "epoch": 7.4895209580838324, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1586, "step": 5003 }, { "epoch": 7.491017964071856, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1525, "step": 5004 }, { "epoch": 7.49251497005988, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1533, "step": 5005 }, { "epoch": 7.494011976047904, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1571, "step": 5006 }, { "epoch": 7.495508982035928, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1496, "step": 5007 }, { "epoch": 7.497005988023952, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1509, "step": 5008 }, { "epoch": 7.498502994011976, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1565, "step": 5009 }, { "epoch": 7.5, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1549, "step": 5010 }, { "epoch": 7.501497005988024, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1522, "step": 5011 }, { "epoch": 7.502994011976048, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1506, "step": 5012 }, { "epoch": 7.504491017964072, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1601, "step": 5013 }, { "epoch": 7.505988023952096, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1518, "step": 5014 }, { "epoch": 7.50748502994012, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1482, "step": 5015 }, { "epoch": 7.508982035928144, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1543, "step": 5016 }, { "epoch": 7.5104790419161676, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1515, "step": 5017 }, { "epoch": 7.5119760479041915, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1535, "step": 5018 }, { "epoch": 7.513473053892215, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1588, "step": 5019 }, { "epoch": 7.514970059880239, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1551, "step": 5020 }, { "epoch": 7.516467065868263, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1561, "step": 5021 }, { "epoch": 7.517964071856287, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1597, "step": 5022 }, { "epoch": 7.519461077844311, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1508, "step": 5023 }, { "epoch": 7.520958083832335, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1607, "step": 5024 }, { "epoch": 7.522455089820359, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1558, "step": 5025 }, { "epoch": 7.523952095808383, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1571, "step": 5026 }, { "epoch": 7.525449101796407, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1478, "step": 5027 }, { "epoch": 7.526946107784431, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1558, "step": 5028 }, { "epoch": 7.528443113772455, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1474, "step": 5029 }, { "epoch": 7.529940119760479, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1594, "step": 5030 }, { "epoch": 7.531437125748503, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1532, "step": 5031 }, { "epoch": 7.532934131736527, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1543, "step": 5032 }, { "epoch": 7.5344311377245505, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1494, "step": 5033 }, { "epoch": 7.5359281437125745, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.151, "step": 5034 }, { "epoch": 7.537425149700599, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1571, "step": 5035 }, { "epoch": 7.538922155688622, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1534, "step": 5036 }, { "epoch": 7.540419161676647, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.154, "step": 5037 }, { "epoch": 7.54191616766467, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.153, "step": 5038 }, { "epoch": 7.543413173652695, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1552, "step": 5039 }, { "epoch": 7.544910179640718, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1517, "step": 5040 }, { "epoch": 7.546407185628743, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1564, "step": 5041 }, { "epoch": 7.547904191616767, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1511, "step": 5042 }, { "epoch": 7.549401197604791, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1621, "step": 5043 }, { "epoch": 7.550898203592815, "grad_norm": 0.150390625, "learning_rate": 0.0008, "loss": 1.1553, "step": 5044 }, { "epoch": 7.552395209580839, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1554, "step": 5045 }, { "epoch": 7.553892215568863, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1514, "step": 5046 }, { "epoch": 7.5553892215568865, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1526, "step": 5047 }, { "epoch": 7.5568862275449105, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1536, "step": 5048 }, { "epoch": 7.558383233532934, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1518, "step": 5049 }, { "epoch": 7.559880239520958, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1516, "step": 5050 }, { "epoch": 7.561377245508982, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1502, "step": 5051 }, { "epoch": 7.562874251497006, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1632, "step": 5052 }, { "epoch": 7.56437125748503, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1573, "step": 5053 }, { "epoch": 7.565868263473054, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1562, "step": 5054 }, { "epoch": 7.567365269461078, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1554, "step": 5055 }, { "epoch": 7.568862275449102, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1486, "step": 5056 }, { "epoch": 7.570359281437126, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1515, "step": 5057 }, { "epoch": 7.57185628742515, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1573, "step": 5058 }, { "epoch": 7.573353293413174, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1492, "step": 5059 }, { "epoch": 7.574850299401198, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1541, "step": 5060 }, { "epoch": 7.576347305389222, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1565, "step": 5061 }, { "epoch": 7.577844311377246, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1528, "step": 5062 }, { "epoch": 7.5793413173652695, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1573, "step": 5063 }, { "epoch": 7.580838323353293, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1485, "step": 5064 }, { "epoch": 7.582335329341317, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.158, "step": 5065 }, { "epoch": 7.583832335329341, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1546, "step": 5066 }, { "epoch": 7.585329341317365, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1594, "step": 5067 }, { "epoch": 7.586826347305389, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1464, "step": 5068 }, { "epoch": 7.588323353293413, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1609, "step": 5069 }, { "epoch": 7.589820359281437, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1507, "step": 5070 }, { "epoch": 7.591317365269461, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1529, "step": 5071 }, { "epoch": 7.592814371257485, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1569, "step": 5072 }, { "epoch": 7.594311377245509, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1539, "step": 5073 }, { "epoch": 7.595808383233533, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1567, "step": 5074 }, { "epoch": 7.597305389221557, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1541, "step": 5075 }, { "epoch": 7.598802395209581, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1552, "step": 5076 }, { "epoch": 7.600299401197605, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1561, "step": 5077 }, { "epoch": 7.6017964071856285, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1497, "step": 5078 }, { "epoch": 7.6032934131736525, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1526, "step": 5079 }, { "epoch": 7.604790419161676, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1509, "step": 5080 }, { "epoch": 7.6062874251497, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1587, "step": 5081 }, { "epoch": 7.607784431137724, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1575, "step": 5082 }, { "epoch": 7.609281437125748, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1558, "step": 5083 }, { "epoch": 7.610778443113772, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1531, "step": 5084 }, { "epoch": 7.612275449101796, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1543, "step": 5085 }, { "epoch": 7.61377245508982, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1513, "step": 5086 }, { "epoch": 7.615269461077844, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1526, "step": 5087 }, { "epoch": 7.616766467065868, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1542, "step": 5088 }, { "epoch": 7.618263473053892, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1566, "step": 5089 }, { "epoch": 7.619760479041916, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1576, "step": 5090 }, { "epoch": 7.62125748502994, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1569, "step": 5091 }, { "epoch": 7.6227544910179645, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1499, "step": 5092 }, { "epoch": 7.624251497005988, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1519, "step": 5093 }, { "epoch": 7.625748502994012, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1571, "step": 5094 }, { "epoch": 7.6272455089820355, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1593, "step": 5095 }, { "epoch": 7.62874251497006, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1534, "step": 5096 }, { "epoch": 7.630239520958084, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1558, "step": 5097 }, { "epoch": 7.631736526946108, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1581, "step": 5098 }, { "epoch": 7.633233532934132, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1498, "step": 5099 }, { "epoch": 7.634730538922156, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1544, "step": 5100 }, { "epoch": 7.63622754491018, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1543, "step": 5101 }, { "epoch": 7.637724550898204, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1543, "step": 5102 }, { "epoch": 7.639221556886228, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1584, "step": 5103 }, { "epoch": 7.640718562874252, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1593, "step": 5104 }, { "epoch": 7.642215568862276, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1502, "step": 5105 }, { "epoch": 7.6437125748503, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1489, "step": 5106 }, { "epoch": 7.645209580838324, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1513, "step": 5107 }, { "epoch": 7.6467065868263475, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1491, "step": 5108 }, { "epoch": 7.6482035928143715, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1459, "step": 5109 }, { "epoch": 7.649700598802395, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.153, "step": 5110 }, { "epoch": 7.651197604790419, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1512, "step": 5111 }, { "epoch": 7.652694610778443, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1509, "step": 5112 }, { "epoch": 7.654191616766467, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1611, "step": 5113 }, { "epoch": 7.655688622754491, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1487, "step": 5114 }, { "epoch": 7.657185628742515, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1462, "step": 5115 }, { "epoch": 7.658682634730539, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.149, "step": 5116 }, { "epoch": 7.660179640718563, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1611, "step": 5117 }, { "epoch": 7.661676646706587, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1514, "step": 5118 }, { "epoch": 7.663173652694611, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1589, "step": 5119 }, { "epoch": 7.664670658682635, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1586, "step": 5120 }, { "epoch": 7.666167664670659, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1599, "step": 5121 }, { "epoch": 7.667664670658683, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1518, "step": 5122 }, { "epoch": 7.669161676646707, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1478, "step": 5123 }, { "epoch": 7.6706586826347305, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1525, "step": 5124 }, { "epoch": 7.672155688622754, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.141, "step": 5125 }, { "epoch": 7.673652694610778, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1471, "step": 5126 }, { "epoch": 7.675149700598802, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1518, "step": 5127 }, { "epoch": 7.676646706586826, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1426, "step": 5128 }, { "epoch": 7.67814371257485, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1522, "step": 5129 }, { "epoch": 7.679640718562874, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1531, "step": 5130 }, { "epoch": 7.681137724550898, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1545, "step": 5131 }, { "epoch": 7.682634730538922, "grad_norm": 0.169921875, "learning_rate": 0.0008, "loss": 1.1594, "step": 5132 }, { "epoch": 7.684131736526946, "grad_norm": 0.20703125, "learning_rate": 0.0008, "loss": 1.1525, "step": 5133 }, { "epoch": 7.68562874251497, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.1554, "step": 5134 }, { "epoch": 7.687125748502994, "grad_norm": 0.1533203125, "learning_rate": 0.0008, "loss": 1.1445, "step": 5135 }, { "epoch": 7.688622754491018, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1553, "step": 5136 }, { "epoch": 7.690119760479042, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.1538, "step": 5137 }, { "epoch": 7.691616766467066, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1503, "step": 5138 }, { "epoch": 7.6931137724550895, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1522, "step": 5139 }, { "epoch": 7.6946107784431135, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1517, "step": 5140 }, { "epoch": 7.696107784431137, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1503, "step": 5141 }, { "epoch": 7.697604790419161, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1572, "step": 5142 }, { "epoch": 7.699101796407185, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1553, "step": 5143 }, { "epoch": 7.700598802395209, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1573, "step": 5144 }, { "epoch": 7.702095808383233, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1563, "step": 5145 }, { "epoch": 7.703592814371257, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1511, "step": 5146 }, { "epoch": 7.705089820359282, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1617, "step": 5147 }, { "epoch": 7.706586826347305, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1532, "step": 5148 }, { "epoch": 7.70808383233533, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1493, "step": 5149 }, { "epoch": 7.709580838323353, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1498, "step": 5150 }, { "epoch": 7.711077844311378, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1504, "step": 5151 }, { "epoch": 7.712574850299401, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1604, "step": 5152 }, { "epoch": 7.7140718562874255, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1522, "step": 5153 }, { "epoch": 7.7155688622754495, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1548, "step": 5154 }, { "epoch": 7.717065868263473, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1512, "step": 5155 }, { "epoch": 7.718562874251497, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1542, "step": 5156 }, { "epoch": 7.720059880239521, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1612, "step": 5157 }, { "epoch": 7.721556886227545, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1566, "step": 5158 }, { "epoch": 7.723053892215569, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1565, "step": 5159 }, { "epoch": 7.724550898203593, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1512, "step": 5160 }, { "epoch": 7.726047904191617, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1535, "step": 5161 }, { "epoch": 7.727544910179641, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1553, "step": 5162 }, { "epoch": 7.729041916167665, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1561, "step": 5163 }, { "epoch": 7.730538922155689, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1491, "step": 5164 }, { "epoch": 7.732035928143713, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1479, "step": 5165 }, { "epoch": 7.733532934131737, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1564, "step": 5166 }, { "epoch": 7.735029940119761, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1539, "step": 5167 }, { "epoch": 7.736526946107785, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1571, "step": 5168 }, { "epoch": 7.7380239520958085, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1563, "step": 5169 }, { "epoch": 7.7395209580838324, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1483, "step": 5170 }, { "epoch": 7.741017964071856, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1503, "step": 5171 }, { "epoch": 7.74251497005988, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1554, "step": 5172 }, { "epoch": 7.744011976047904, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1518, "step": 5173 }, { "epoch": 7.745508982035928, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1496, "step": 5174 }, { "epoch": 7.747005988023952, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1557, "step": 5175 }, { "epoch": 7.748502994011976, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1551, "step": 5176 }, { "epoch": 7.75, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1593, "step": 5177 }, { "epoch": 7.751497005988024, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1452, "step": 5178 }, { "epoch": 7.752994011976048, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1509, "step": 5179 }, { "epoch": 7.754491017964072, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1499, "step": 5180 }, { "epoch": 7.755988023952096, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.146, "step": 5181 }, { "epoch": 7.75748502994012, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1512, "step": 5182 }, { "epoch": 7.758982035928144, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1612, "step": 5183 }, { "epoch": 7.7604790419161676, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1576, "step": 5184 }, { "epoch": 7.7619760479041915, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1499, "step": 5185 }, { "epoch": 7.763473053892215, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1542, "step": 5186 }, { "epoch": 7.764970059880239, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1563, "step": 5187 }, { "epoch": 7.766467065868263, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1486, "step": 5188 }, { "epoch": 7.767964071856287, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1566, "step": 5189 }, { "epoch": 7.769461077844311, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1623, "step": 5190 }, { "epoch": 7.770958083832335, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1573, "step": 5191 }, { "epoch": 7.772455089820359, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1501, "step": 5192 }, { "epoch": 7.773952095808383, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1585, "step": 5193 }, { "epoch": 7.775449101796407, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1511, "step": 5194 }, { "epoch": 7.776946107784431, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1552, "step": 5195 }, { "epoch": 7.778443113772455, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1528, "step": 5196 }, { "epoch": 7.779940119760479, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1542, "step": 5197 }, { "epoch": 7.781437125748503, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1513, "step": 5198 }, { "epoch": 7.782934131736527, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1529, "step": 5199 }, { "epoch": 7.7844311377245505, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1542, "step": 5200 }, { "epoch": 7.7859281437125745, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.155, "step": 5201 }, { "epoch": 7.787425149700599, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1542, "step": 5202 }, { "epoch": 7.788922155688622, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1554, "step": 5203 }, { "epoch": 7.790419161676647, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1474, "step": 5204 }, { "epoch": 7.79191616766467, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1557, "step": 5205 }, { "epoch": 7.793413173652695, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1521, "step": 5206 }, { "epoch": 7.794910179640718, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1564, "step": 5207 }, { "epoch": 7.796407185628743, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.146, "step": 5208 }, { "epoch": 7.797904191616767, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1574, "step": 5209 }, { "epoch": 7.799401197604791, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1561, "step": 5210 }, { "epoch": 7.800898203592815, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1508, "step": 5211 }, { "epoch": 7.802395209580839, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1517, "step": 5212 }, { "epoch": 7.803892215568863, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1525, "step": 5213 }, { "epoch": 7.8053892215568865, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1514, "step": 5214 }, { "epoch": 7.8068862275449105, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1517, "step": 5215 }, { "epoch": 7.808383233532934, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1433, "step": 5216 }, { "epoch": 7.809880239520958, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1537, "step": 5217 }, { "epoch": 7.811377245508982, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1505, "step": 5218 }, { "epoch": 7.812874251497006, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1565, "step": 5219 }, { "epoch": 7.81437125748503, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1511, "step": 5220 }, { "epoch": 7.815868263473054, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1471, "step": 5221 }, { "epoch": 7.817365269461078, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1552, "step": 5222 }, { "epoch": 7.818862275449102, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1455, "step": 5223 }, { "epoch": 7.820359281437126, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1499, "step": 5224 }, { "epoch": 7.82185628742515, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1558, "step": 5225 }, { "epoch": 7.823353293413174, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1515, "step": 5226 }, { "epoch": 7.824850299401198, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1523, "step": 5227 }, { "epoch": 7.826347305389222, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1542, "step": 5228 }, { "epoch": 7.827844311377246, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1532, "step": 5229 }, { "epoch": 7.8293413173652695, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.152, "step": 5230 }, { "epoch": 7.830838323353293, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1518, "step": 5231 }, { "epoch": 7.832335329341317, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.151, "step": 5232 }, { "epoch": 7.833832335329341, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1586, "step": 5233 }, { "epoch": 7.835329341317365, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1484, "step": 5234 }, { "epoch": 7.836826347305389, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1487, "step": 5235 }, { "epoch": 7.838323353293413, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1498, "step": 5236 }, { "epoch": 7.839820359281437, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1504, "step": 5237 }, { "epoch": 7.841317365269461, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1531, "step": 5238 }, { "epoch": 7.842814371257485, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1551, "step": 5239 }, { "epoch": 7.844311377245509, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1478, "step": 5240 }, { "epoch": 7.845808383233533, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1506, "step": 5241 }, { "epoch": 7.847305389221557, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1521, "step": 5242 }, { "epoch": 7.848802395209581, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1473, "step": 5243 }, { "epoch": 7.850299401197605, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1546, "step": 5244 }, { "epoch": 7.8517964071856285, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1568, "step": 5245 }, { "epoch": 7.8532934131736525, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1507, "step": 5246 }, { "epoch": 7.854790419161676, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1453, "step": 5247 }, { "epoch": 7.8562874251497, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1583, "step": 5248 }, { "epoch": 7.857784431137724, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1497, "step": 5249 }, { "epoch": 7.859281437125748, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1478, "step": 5250 }, { "epoch": 7.860778443113772, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1493, "step": 5251 }, { "epoch": 7.862275449101796, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1561, "step": 5252 }, { "epoch": 7.86377245508982, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1557, "step": 5253 }, { "epoch": 7.865269461077844, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1601, "step": 5254 }, { "epoch": 7.866766467065868, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1528, "step": 5255 }, { "epoch": 7.868263473053892, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1501, "step": 5256 }, { "epoch": 7.869760479041916, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1586, "step": 5257 }, { "epoch": 7.87125748502994, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1545, "step": 5258 }, { "epoch": 7.8727544910179645, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1576, "step": 5259 }, { "epoch": 7.874251497005988, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1532, "step": 5260 }, { "epoch": 7.875748502994012, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1481, "step": 5261 }, { "epoch": 7.8772455089820355, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1514, "step": 5262 }, { "epoch": 7.87874251497006, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1514, "step": 5263 }, { "epoch": 7.880239520958084, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1575, "step": 5264 }, { "epoch": 7.881736526946108, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1581, "step": 5265 }, { "epoch": 7.883233532934132, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1552, "step": 5266 }, { "epoch": 7.884730538922156, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1505, "step": 5267 }, { "epoch": 7.88622754491018, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1509, "step": 5268 }, { "epoch": 7.887724550898204, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.152, "step": 5269 }, { "epoch": 7.889221556886228, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1553, "step": 5270 }, { "epoch": 7.890718562874252, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1582, "step": 5271 }, { "epoch": 7.892215568862276, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.156, "step": 5272 }, { "epoch": 7.8937125748503, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1514, "step": 5273 }, { "epoch": 7.895209580838324, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1585, "step": 5274 }, { "epoch": 7.8967065868263475, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1558, "step": 5275 }, { "epoch": 7.8982035928143715, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1551, "step": 5276 }, { "epoch": 7.899700598802395, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1466, "step": 5277 }, { "epoch": 7.901197604790419, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1516, "step": 5278 }, { "epoch": 7.902694610778443, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1582, "step": 5279 }, { "epoch": 7.904191616766467, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.153, "step": 5280 }, { "epoch": 7.905688622754491, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1525, "step": 5281 }, { "epoch": 7.907185628742515, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1518, "step": 5282 }, { "epoch": 7.908682634730539, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1496, "step": 5283 }, { "epoch": 7.910179640718563, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1518, "step": 5284 }, { "epoch": 7.911676646706587, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1541, "step": 5285 }, { "epoch": 7.913173652694611, "grad_norm": 0.049072265625, "learning_rate": 0.0008, "loss": 1.1513, "step": 5286 }, { "epoch": 7.914670658682635, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1522, "step": 5287 }, { "epoch": 7.916167664670659, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1534, "step": 5288 }, { "epoch": 7.917664670658683, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1559, "step": 5289 }, { "epoch": 7.919161676646707, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1511, "step": 5290 }, { "epoch": 7.9206586826347305, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1527, "step": 5291 }, { "epoch": 7.922155688622754, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1553, "step": 5292 }, { "epoch": 7.923652694610778, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1525, "step": 5293 }, { "epoch": 7.925149700598802, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1488, "step": 5294 }, { "epoch": 7.926646706586826, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1505, "step": 5295 }, { "epoch": 7.92814371257485, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.152, "step": 5296 }, { "epoch": 7.929640718562874, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1554, "step": 5297 }, { "epoch": 7.931137724550898, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1462, "step": 5298 }, { "epoch": 7.932634730538922, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1425, "step": 5299 }, { "epoch": 7.934131736526946, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1528, "step": 5300 }, { "epoch": 7.93562874251497, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1516, "step": 5301 }, { "epoch": 7.937125748502994, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1424, "step": 5302 }, { "epoch": 7.938622754491018, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1437, "step": 5303 }, { "epoch": 7.940119760479042, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1446, "step": 5304 }, { "epoch": 7.941616766467066, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1502, "step": 5305 }, { "epoch": 7.9431137724550895, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1485, "step": 5306 }, { "epoch": 7.9446107784431135, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1468, "step": 5307 }, { "epoch": 7.946107784431137, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1496, "step": 5308 }, { "epoch": 7.947604790419161, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1501, "step": 5309 }, { "epoch": 7.949101796407185, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1516, "step": 5310 }, { "epoch": 7.950598802395209, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.152, "step": 5311 }, { "epoch": 7.952095808383233, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1606, "step": 5312 }, { "epoch": 7.953592814371257, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1541, "step": 5313 }, { "epoch": 7.955089820359282, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1521, "step": 5314 }, { "epoch": 7.956586826347305, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1478, "step": 5315 }, { "epoch": 7.95808383233533, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1562, "step": 5316 }, { "epoch": 7.959580838323353, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.154, "step": 5317 }, { "epoch": 7.961077844311378, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1518, "step": 5318 }, { "epoch": 7.962574850299401, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1548, "step": 5319 }, { "epoch": 7.9640718562874255, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1507, "step": 5320 }, { "epoch": 7.9655688622754495, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1539, "step": 5321 }, { "epoch": 7.967065868263473, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1498, "step": 5322 }, { "epoch": 7.968562874251497, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1556, "step": 5323 }, { "epoch": 7.970059880239521, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1482, "step": 5324 }, { "epoch": 7.971556886227545, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1545, "step": 5325 }, { "epoch": 7.973053892215569, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1581, "step": 5326 }, { "epoch": 7.974550898203593, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1533, "step": 5327 }, { "epoch": 7.976047904191617, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1531, "step": 5328 }, { "epoch": 7.977544910179641, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1522, "step": 5329 }, { "epoch": 7.979041916167665, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1603, "step": 5330 }, { "epoch": 7.980538922155689, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1488, "step": 5331 }, { "epoch": 7.982035928143713, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.153, "step": 5332 }, { "epoch": 7.983532934131737, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1584, "step": 5333 }, { "epoch": 7.985029940119761, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.143, "step": 5334 }, { "epoch": 7.986526946107785, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1529, "step": 5335 }, { "epoch": 7.9880239520958085, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1563, "step": 5336 }, { "epoch": 7.9895209580838324, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1513, "step": 5337 }, { "epoch": 7.991017964071856, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1549, "step": 5338 }, { "epoch": 7.99251497005988, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1519, "step": 5339 }, { "epoch": 7.994011976047904, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1513, "step": 5340 }, { "epoch": 7.995508982035928, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1549, "step": 5341 }, { "epoch": 7.997005988023952, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1427, "step": 5342 }, { "epoch": 7.998502994011976, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1478, "step": 5343 }, { "epoch": 8.0, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1479, "step": 5344 }, { "epoch": 8.001497005988025, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1474, "step": 5345 }, { "epoch": 8.002994011976048, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1516, "step": 5346 }, { "epoch": 8.004491017964073, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1493, "step": 5347 }, { "epoch": 8.005988023952096, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1444, "step": 5348 }, { "epoch": 8.00748502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1553, "step": 5349 }, { "epoch": 8.008982035928144, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1521, "step": 5350 }, { "epoch": 8.010479041916168, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1549, "step": 5351 }, { "epoch": 8.011976047904191, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1572, "step": 5352 }, { "epoch": 8.013473053892216, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1505, "step": 5353 }, { "epoch": 8.01497005988024, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1518, "step": 5354 }, { "epoch": 8.016467065868264, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1525, "step": 5355 }, { "epoch": 8.017964071856287, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1541, "step": 5356 }, { "epoch": 8.019461077844312, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1514, "step": 5357 }, { "epoch": 8.020958083832335, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1521, "step": 5358 }, { "epoch": 8.02245508982036, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1489, "step": 5359 }, { "epoch": 8.023952095808383, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1494, "step": 5360 }, { "epoch": 8.025449101796408, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1511, "step": 5361 }, { "epoch": 8.02694610778443, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1567, "step": 5362 }, { "epoch": 8.028443113772456, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1485, "step": 5363 }, { "epoch": 8.029940119760479, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1529, "step": 5364 }, { "epoch": 8.031437125748504, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1472, "step": 5365 }, { "epoch": 8.032934131736527, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.156, "step": 5366 }, { "epoch": 8.034431137724551, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1605, "step": 5367 }, { "epoch": 8.035928143712574, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1501, "step": 5368 }, { "epoch": 8.0374251497006, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.15, "step": 5369 }, { "epoch": 8.038922155688622, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1559, "step": 5370 }, { "epoch": 8.040419161676647, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1546, "step": 5371 }, { "epoch": 8.04191616766467, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1456, "step": 5372 }, { "epoch": 8.043413173652695, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1459, "step": 5373 }, { "epoch": 8.044910179640718, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1536, "step": 5374 }, { "epoch": 8.046407185628743, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1474, "step": 5375 }, { "epoch": 8.047904191616766, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1511, "step": 5376 }, { "epoch": 8.04940119760479, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1578, "step": 5377 }, { "epoch": 8.050898203592814, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1604, "step": 5378 }, { "epoch": 8.052395209580839, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1497, "step": 5379 }, { "epoch": 8.053892215568862, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.144, "step": 5380 }, { "epoch": 8.055389221556887, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1554, "step": 5381 }, { "epoch": 8.05688622754491, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1545, "step": 5382 }, { "epoch": 8.058383233532934, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.158, "step": 5383 }, { "epoch": 8.059880239520957, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1507, "step": 5384 }, { "epoch": 8.061377245508982, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1529, "step": 5385 }, { "epoch": 8.062874251497005, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1544, "step": 5386 }, { "epoch": 8.06437125748503, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1453, "step": 5387 }, { "epoch": 8.065868263473053, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1517, "step": 5388 }, { "epoch": 8.067365269461078, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1453, "step": 5389 }, { "epoch": 8.068862275449101, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1427, "step": 5390 }, { "epoch": 8.070359281437126, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1447, "step": 5391 }, { "epoch": 8.071856287425149, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1503, "step": 5392 }, { "epoch": 8.073353293413174, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1553, "step": 5393 }, { "epoch": 8.074850299401197, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1554, "step": 5394 }, { "epoch": 8.076347305389222, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1499, "step": 5395 }, { "epoch": 8.077844311377245, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1489, "step": 5396 }, { "epoch": 8.07934131736527, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.149, "step": 5397 }, { "epoch": 8.080838323353293, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1492, "step": 5398 }, { "epoch": 8.082335329341317, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1553, "step": 5399 }, { "epoch": 8.08383233532934, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1495, "step": 5400 }, { "epoch": 8.085329341317365, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1517, "step": 5401 }, { "epoch": 8.08682634730539, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1506, "step": 5402 }, { "epoch": 8.088323353293413, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1427, "step": 5403 }, { "epoch": 8.089820359281438, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1527, "step": 5404 }, { "epoch": 8.091317365269461, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1444, "step": 5405 }, { "epoch": 8.092814371257486, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1481, "step": 5406 }, { "epoch": 8.094311377245509, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1546, "step": 5407 }, { "epoch": 8.095808383233534, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1541, "step": 5408 }, { "epoch": 8.097305389221557, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1474, "step": 5409 }, { "epoch": 8.098802395209582, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1569, "step": 5410 }, { "epoch": 8.100299401197605, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1499, "step": 5411 }, { "epoch": 8.10179640718563, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1511, "step": 5412 }, { "epoch": 8.103293413173652, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1558, "step": 5413 }, { "epoch": 8.104790419161677, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1492, "step": 5414 }, { "epoch": 8.1062874251497, "grad_norm": 0.15234375, "learning_rate": 0.0008, "loss": 1.1491, "step": 5415 }, { "epoch": 8.107784431137725, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1527, "step": 5416 }, { "epoch": 8.109281437125748, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1494, "step": 5417 }, { "epoch": 8.110778443113773, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1574, "step": 5418 }, { "epoch": 8.112275449101796, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1582, "step": 5419 }, { "epoch": 8.113772455089821, "grad_norm": 0.15234375, "learning_rate": 0.0008, "loss": 1.1508, "step": 5420 }, { "epoch": 8.115269461077844, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1577, "step": 5421 }, { "epoch": 8.116766467065869, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1436, "step": 5422 }, { "epoch": 8.118263473053892, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1602, "step": 5423 }, { "epoch": 8.119760479041917, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1535, "step": 5424 }, { "epoch": 8.12125748502994, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1466, "step": 5425 }, { "epoch": 8.122754491017965, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1475, "step": 5426 }, { "epoch": 8.124251497005988, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.152, "step": 5427 }, { "epoch": 8.125748502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.152, "step": 5428 }, { "epoch": 8.127245508982035, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1501, "step": 5429 }, { "epoch": 8.12874251497006, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1518, "step": 5430 }, { "epoch": 8.130239520958083, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1504, "step": 5431 }, { "epoch": 8.131736526946108, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1496, "step": 5432 }, { "epoch": 8.133233532934131, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1576, "step": 5433 }, { "epoch": 8.134730538922156, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1551, "step": 5434 }, { "epoch": 8.136227544910179, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1482, "step": 5435 }, { "epoch": 8.137724550898204, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.155, "step": 5436 }, { "epoch": 8.139221556886227, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1517, "step": 5437 }, { "epoch": 8.140718562874252, "grad_norm": 0.1796875, "learning_rate": 0.0008, "loss": 1.1565, "step": 5438 }, { "epoch": 8.142215568862275, "grad_norm": 0.1943359375, "learning_rate": 0.0008, "loss": 1.1551, "step": 5439 }, { "epoch": 8.1437125748503, "grad_norm": 0.220703125, "learning_rate": 0.0008, "loss": 1.1553, "step": 5440 }, { "epoch": 8.145209580838323, "grad_norm": 0.2412109375, "learning_rate": 0.0008, "loss": 1.1542, "step": 5441 }, { "epoch": 8.146706586826348, "grad_norm": 0.25, "learning_rate": 0.0008, "loss": 1.1539, "step": 5442 }, { "epoch": 8.14820359281437, "grad_norm": 0.232421875, "learning_rate": 0.0008, "loss": 1.157, "step": 5443 }, { "epoch": 8.149700598802395, "grad_norm": 0.1591796875, "learning_rate": 0.0008, "loss": 1.152, "step": 5444 }, { "epoch": 8.151197604790418, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1522, "step": 5445 }, { "epoch": 8.152694610778443, "grad_norm": 0.212890625, "learning_rate": 0.0008, "loss": 1.1497, "step": 5446 }, { "epoch": 8.154191616766466, "grad_norm": 0.28515625, "learning_rate": 0.0008, "loss": 1.158, "step": 5447 }, { "epoch": 8.155688622754491, "grad_norm": 0.408203125, "learning_rate": 0.0008, "loss": 1.1523, "step": 5448 }, { "epoch": 8.157185628742514, "grad_norm": 0.734375, "learning_rate": 0.0008, "loss": 1.1533, "step": 5449 }, { "epoch": 8.158682634730539, "grad_norm": 1.28125, "learning_rate": 0.0008, "loss": 1.1855, "step": 5450 }, { "epoch": 8.160179640718562, "grad_norm": 0.671875, "learning_rate": 0.0008, "loss": 1.1727, "step": 5451 }, { "epoch": 8.161676646706587, "grad_norm": 0.369140625, "learning_rate": 0.0008, "loss": 1.161, "step": 5452 }, { "epoch": 8.16317365269461, "grad_norm": 0.5625, "learning_rate": 0.0008, "loss": 1.1676, "step": 5453 }, { "epoch": 8.164670658682635, "grad_norm": 0.1748046875, "learning_rate": 0.0008, "loss": 1.1607, "step": 5454 }, { "epoch": 8.16616766467066, "grad_norm": 0.47265625, "learning_rate": 0.0008, "loss": 1.1718, "step": 5455 }, { "epoch": 8.167664670658683, "grad_norm": 0.376953125, "learning_rate": 0.0008, "loss": 1.1675, "step": 5456 }, { "epoch": 8.169161676646707, "grad_norm": 0.259765625, "learning_rate": 0.0008, "loss": 1.1579, "step": 5457 }, { "epoch": 8.17065868263473, "grad_norm": 0.53125, "learning_rate": 0.0008, "loss": 1.162, "step": 5458 }, { "epoch": 8.172155688622755, "grad_norm": 0.37109375, "learning_rate": 0.0008, "loss": 1.1661, "step": 5459 }, { "epoch": 8.173652694610778, "grad_norm": 0.18359375, "learning_rate": 0.0008, "loss": 1.1658, "step": 5460 }, { "epoch": 8.175149700598803, "grad_norm": 0.28125, "learning_rate": 0.0008, "loss": 1.1597, "step": 5461 }, { "epoch": 8.176646706586826, "grad_norm": 0.287109375, "learning_rate": 0.0008, "loss": 1.1678, "step": 5462 }, { "epoch": 8.178143712574851, "grad_norm": 0.19921875, "learning_rate": 0.0008, "loss": 1.1571, "step": 5463 }, { "epoch": 8.179640718562874, "grad_norm": 0.2421875, "learning_rate": 0.0008, "loss": 1.1588, "step": 5464 }, { "epoch": 8.181137724550899, "grad_norm": 0.30078125, "learning_rate": 0.0008, "loss": 1.1642, "step": 5465 }, { "epoch": 8.182634730538922, "grad_norm": 0.1748046875, "learning_rate": 0.0008, "loss": 1.1643, "step": 5466 }, { "epoch": 8.184131736526947, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.1508, "step": 5467 }, { "epoch": 8.18562874251497, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1541, "step": 5468 }, { "epoch": 8.187125748502995, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1534, "step": 5469 }, { "epoch": 8.188622754491018, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1591, "step": 5470 }, { "epoch": 8.190119760479043, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1576, "step": 5471 }, { "epoch": 8.191616766467066, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.15, "step": 5472 }, { "epoch": 8.19311377245509, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1573, "step": 5473 }, { "epoch": 8.194610778443113, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1593, "step": 5474 }, { "epoch": 8.196107784431138, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1545, "step": 5475 }, { "epoch": 8.197604790419161, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1544, "step": 5476 }, { "epoch": 8.199101796407186, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1593, "step": 5477 }, { "epoch": 8.20059880239521, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1527, "step": 5478 }, { "epoch": 8.202095808383234, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1515, "step": 5479 }, { "epoch": 8.203592814371257, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1555, "step": 5480 }, { "epoch": 8.205089820359282, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1517, "step": 5481 }, { "epoch": 8.206586826347305, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1513, "step": 5482 }, { "epoch": 8.20808383233533, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1593, "step": 5483 }, { "epoch": 8.209580838323353, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1579, "step": 5484 }, { "epoch": 8.211077844311378, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1579, "step": 5485 }, { "epoch": 8.2125748502994, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1486, "step": 5486 }, { "epoch": 8.214071856287426, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1512, "step": 5487 }, { "epoch": 8.215568862275449, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1549, "step": 5488 }, { "epoch": 8.217065868263473, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1556, "step": 5489 }, { "epoch": 8.218562874251496, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1563, "step": 5490 }, { "epoch": 8.220059880239521, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1567, "step": 5491 }, { "epoch": 8.221556886227544, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1577, "step": 5492 }, { "epoch": 8.22305389221557, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1519, "step": 5493 }, { "epoch": 8.224550898203592, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1469, "step": 5494 }, { "epoch": 8.226047904191617, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1571, "step": 5495 }, { "epoch": 8.22754491017964, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1567, "step": 5496 }, { "epoch": 8.229041916167665, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1536, "step": 5497 }, { "epoch": 8.230538922155688, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1527, "step": 5498 }, { "epoch": 8.232035928143713, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1541, "step": 5499 }, { "epoch": 8.233532934131736, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1611, "step": 5500 }, { "epoch": 8.23502994011976, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1486, "step": 5501 }, { "epoch": 8.236526946107784, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1554, "step": 5502 }, { "epoch": 8.238023952095809, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1534, "step": 5503 }, { "epoch": 8.239520958083832, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1541, "step": 5504 }, { "epoch": 8.241017964071856, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1515, "step": 5505 }, { "epoch": 8.24251497005988, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1506, "step": 5506 }, { "epoch": 8.244011976047904, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1526, "step": 5507 }, { "epoch": 8.245508982035927, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1509, "step": 5508 }, { "epoch": 8.247005988023952, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1493, "step": 5509 }, { "epoch": 8.248502994011975, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1502, "step": 5510 }, { "epoch": 8.25, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1541, "step": 5511 }, { "epoch": 8.251497005988025, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.156, "step": 5512 }, { "epoch": 8.252994011976048, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1565, "step": 5513 }, { "epoch": 8.254491017964073, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1494, "step": 5514 }, { "epoch": 8.255988023952096, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1593, "step": 5515 }, { "epoch": 8.25748502994012, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1456, "step": 5516 }, { "epoch": 8.258982035928144, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1524, "step": 5517 }, { "epoch": 8.260479041916168, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1567, "step": 5518 }, { "epoch": 8.261976047904191, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1553, "step": 5519 }, { "epoch": 8.263473053892216, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.148, "step": 5520 }, { "epoch": 8.26497005988024, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1478, "step": 5521 }, { "epoch": 8.266467065868264, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1544, "step": 5522 }, { "epoch": 8.267964071856287, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1565, "step": 5523 }, { "epoch": 8.269461077844312, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1589, "step": 5524 }, { "epoch": 8.270958083832335, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1472, "step": 5525 }, { "epoch": 8.27245508982036, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1517, "step": 5526 }, { "epoch": 8.273952095808383, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1571, "step": 5527 }, { "epoch": 8.275449101796408, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1532, "step": 5528 }, { "epoch": 8.27694610778443, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1547, "step": 5529 }, { "epoch": 8.278443113772456, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1555, "step": 5530 }, { "epoch": 8.279940119760479, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1502, "step": 5531 }, { "epoch": 8.281437125748504, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1458, "step": 5532 }, { "epoch": 8.282934131736527, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1547, "step": 5533 }, { "epoch": 8.284431137724551, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1564, "step": 5534 }, { "epoch": 8.285928143712574, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1552, "step": 5535 }, { "epoch": 8.2874251497006, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1491, "step": 5536 }, { "epoch": 8.288922155688622, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1544, "step": 5537 }, { "epoch": 8.290419161676647, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1574, "step": 5538 }, { "epoch": 8.29191616766467, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1564, "step": 5539 }, { "epoch": 8.293413173652695, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1497, "step": 5540 }, { "epoch": 8.294910179640718, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1504, "step": 5541 }, { "epoch": 8.296407185628743, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1528, "step": 5542 }, { "epoch": 8.297904191616766, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1551, "step": 5543 }, { "epoch": 8.29940119760479, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1453, "step": 5544 }, { "epoch": 8.300898203592814, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1516, "step": 5545 }, { "epoch": 8.302395209580839, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1491, "step": 5546 }, { "epoch": 8.303892215568862, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1527, "step": 5547 }, { "epoch": 8.305389221556887, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1456, "step": 5548 }, { "epoch": 8.30688622754491, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1493, "step": 5549 }, { "epoch": 8.308383233532934, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1559, "step": 5550 }, { "epoch": 8.309880239520957, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1513, "step": 5551 }, { "epoch": 8.311377245508982, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1523, "step": 5552 }, { "epoch": 8.312874251497005, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1485, "step": 5553 }, { "epoch": 8.31437125748503, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1471, "step": 5554 }, { "epoch": 8.315868263473053, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1499, "step": 5555 }, { "epoch": 8.317365269461078, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.153, "step": 5556 }, { "epoch": 8.318862275449101, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1568, "step": 5557 }, { "epoch": 8.320359281437126, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.152, "step": 5558 }, { "epoch": 8.321856287425149, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1586, "step": 5559 }, { "epoch": 8.323353293413174, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1481, "step": 5560 }, { "epoch": 8.324850299401197, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1534, "step": 5561 }, { "epoch": 8.326347305389222, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1518, "step": 5562 }, { "epoch": 8.327844311377245, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1518, "step": 5563 }, { "epoch": 8.32934131736527, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1461, "step": 5564 }, { "epoch": 8.330838323353294, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1486, "step": 5565 }, { "epoch": 8.332335329341317, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1489, "step": 5566 }, { "epoch": 8.33383233532934, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1555, "step": 5567 }, { "epoch": 8.335329341317365, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1509, "step": 5568 }, { "epoch": 8.33682634730539, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1534, "step": 5569 }, { "epoch": 8.338323353293413, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1462, "step": 5570 }, { "epoch": 8.339820359281438, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1486, "step": 5571 }, { "epoch": 8.341317365269461, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1485, "step": 5572 }, { "epoch": 8.342814371257486, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1505, "step": 5573 }, { "epoch": 8.344311377245509, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1522, "step": 5574 }, { "epoch": 8.345808383233534, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1529, "step": 5575 }, { "epoch": 8.347305389221557, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.15, "step": 5576 }, { "epoch": 8.348802395209582, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1466, "step": 5577 }, { "epoch": 8.350299401197605, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1472, "step": 5578 }, { "epoch": 8.35179640718563, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1547, "step": 5579 }, { "epoch": 8.353293413173652, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1489, "step": 5580 }, { "epoch": 8.354790419161677, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.153, "step": 5581 }, { "epoch": 8.3562874251497, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1581, "step": 5582 }, { "epoch": 8.357784431137725, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.157, "step": 5583 }, { "epoch": 8.359281437125748, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1478, "step": 5584 }, { "epoch": 8.360778443113773, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.157, "step": 5585 }, { "epoch": 8.362275449101796, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1498, "step": 5586 }, { "epoch": 8.363772455089821, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1492, "step": 5587 }, { "epoch": 8.365269461077844, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1453, "step": 5588 }, { "epoch": 8.366766467065869, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1513, "step": 5589 }, { "epoch": 8.368263473053892, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1554, "step": 5590 }, { "epoch": 8.369760479041917, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.154, "step": 5591 }, { "epoch": 8.37125748502994, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1544, "step": 5592 }, { "epoch": 8.372754491017965, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1461, "step": 5593 }, { "epoch": 8.374251497005988, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1433, "step": 5594 }, { "epoch": 8.375748502994012, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1476, "step": 5595 }, { "epoch": 8.377245508982035, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1573, "step": 5596 }, { "epoch": 8.37874251497006, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1473, "step": 5597 }, { "epoch": 8.380239520958083, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1523, "step": 5598 }, { "epoch": 8.381736526946108, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1473, "step": 5599 }, { "epoch": 8.383233532934131, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1486, "step": 5600 }, { "epoch": 8.384730538922156, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1549, "step": 5601 }, { "epoch": 8.386227544910179, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1504, "step": 5602 }, { "epoch": 8.387724550898204, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1488, "step": 5603 }, { "epoch": 8.389221556886227, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1531, "step": 5604 }, { "epoch": 8.390718562874252, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1478, "step": 5605 }, { "epoch": 8.392215568862275, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1507, "step": 5606 }, { "epoch": 8.3937125748503, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1525, "step": 5607 }, { "epoch": 8.395209580838323, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1492, "step": 5608 }, { "epoch": 8.396706586826348, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1487, "step": 5609 }, { "epoch": 8.39820359281437, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1452, "step": 5610 }, { "epoch": 8.399700598802395, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1528, "step": 5611 }, { "epoch": 8.401197604790418, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1514, "step": 5612 }, { "epoch": 8.402694610778443, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1494, "step": 5613 }, { "epoch": 8.404191616766466, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1539, "step": 5614 }, { "epoch": 8.405688622754491, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1501, "step": 5615 }, { "epoch": 8.407185628742514, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1563, "step": 5616 }, { "epoch": 8.408682634730539, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1519, "step": 5617 }, { "epoch": 8.410179640718562, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1488, "step": 5618 }, { "epoch": 8.411676646706587, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1523, "step": 5619 }, { "epoch": 8.41317365269461, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1416, "step": 5620 }, { "epoch": 8.414670658682635, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1507, "step": 5621 }, { "epoch": 8.41616766467066, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1473, "step": 5622 }, { "epoch": 8.417664670658683, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1524, "step": 5623 }, { "epoch": 8.419161676646706, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1498, "step": 5624 }, { "epoch": 8.42065868263473, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1524, "step": 5625 }, { "epoch": 8.422155688622755, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1534, "step": 5626 }, { "epoch": 8.423652694610778, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1474, "step": 5627 }, { "epoch": 8.425149700598803, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1517, "step": 5628 }, { "epoch": 8.426646706586826, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1534, "step": 5629 }, { "epoch": 8.428143712574851, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1504, "step": 5630 }, { "epoch": 8.429640718562874, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1539, "step": 5631 }, { "epoch": 8.431137724550899, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1491, "step": 5632 }, { "epoch": 8.432634730538922, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1558, "step": 5633 }, { "epoch": 8.434131736526947, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1562, "step": 5634 }, { "epoch": 8.43562874251497, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1487, "step": 5635 }, { "epoch": 8.437125748502995, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1565, "step": 5636 }, { "epoch": 8.438622754491018, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1532, "step": 5637 }, { "epoch": 8.440119760479043, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1482, "step": 5638 }, { "epoch": 8.441616766467066, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.149, "step": 5639 }, { "epoch": 8.44311377245509, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1571, "step": 5640 }, { "epoch": 8.444610778443113, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1571, "step": 5641 }, { "epoch": 8.446107784431138, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1557, "step": 5642 }, { "epoch": 8.447604790419161, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1536, "step": 5643 }, { "epoch": 8.449101796407186, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1581, "step": 5644 }, { "epoch": 8.45059880239521, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1584, "step": 5645 }, { "epoch": 8.452095808383234, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1471, "step": 5646 }, { "epoch": 8.453592814371257, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1532, "step": 5647 }, { "epoch": 8.455089820359282, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1581, "step": 5648 }, { "epoch": 8.456586826347305, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1542, "step": 5649 }, { "epoch": 8.45808383233533, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1441, "step": 5650 }, { "epoch": 8.459580838323353, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1511, "step": 5651 }, { "epoch": 8.461077844311378, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1556, "step": 5652 }, { "epoch": 8.4625748502994, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.153, "step": 5653 }, { "epoch": 8.464071856287426, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1537, "step": 5654 }, { "epoch": 8.465568862275449, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1505, "step": 5655 }, { "epoch": 8.467065868263473, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1428, "step": 5656 }, { "epoch": 8.468562874251496, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1452, "step": 5657 }, { "epoch": 8.470059880239521, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1479, "step": 5658 }, { "epoch": 8.471556886227544, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1569, "step": 5659 }, { "epoch": 8.47305389221557, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1533, "step": 5660 }, { "epoch": 8.474550898203592, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1509, "step": 5661 }, { "epoch": 8.476047904191617, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1473, "step": 5662 }, { "epoch": 8.47754491017964, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1553, "step": 5663 }, { "epoch": 8.479041916167665, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1464, "step": 5664 }, { "epoch": 8.480538922155688, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1563, "step": 5665 }, { "epoch": 8.482035928143713, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1521, "step": 5666 }, { "epoch": 8.483532934131736, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1492, "step": 5667 }, { "epoch": 8.48502994011976, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.151, "step": 5668 }, { "epoch": 8.486526946107784, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1552, "step": 5669 }, { "epoch": 8.488023952095809, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1397, "step": 5670 }, { "epoch": 8.489520958083832, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1578, "step": 5671 }, { "epoch": 8.491017964071856, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1505, "step": 5672 }, { "epoch": 8.49251497005988, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1484, "step": 5673 }, { "epoch": 8.494011976047904, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1488, "step": 5674 }, { "epoch": 8.495508982035927, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.151, "step": 5675 }, { "epoch": 8.497005988023952, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1497, "step": 5676 }, { "epoch": 8.498502994011975, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1541, "step": 5677 }, { "epoch": 8.5, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1582, "step": 5678 }, { "epoch": 8.501497005988025, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1533, "step": 5679 }, { "epoch": 8.502994011976048, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1493, "step": 5680 }, { "epoch": 8.504491017964071, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1544, "step": 5681 }, { "epoch": 8.505988023952096, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1498, "step": 5682 }, { "epoch": 8.50748502994012, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1507, "step": 5683 }, { "epoch": 8.508982035928144, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1444, "step": 5684 }, { "epoch": 8.510479041916168, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1445, "step": 5685 }, { "epoch": 8.511976047904191, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1501, "step": 5686 }, { "epoch": 8.513473053892216, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1503, "step": 5687 }, { "epoch": 8.51497005988024, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1521, "step": 5688 }, { "epoch": 8.516467065868264, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1551, "step": 5689 }, { "epoch": 8.517964071856287, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1501, "step": 5690 }, { "epoch": 8.519461077844312, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1469, "step": 5691 }, { "epoch": 8.520958083832335, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1495, "step": 5692 }, { "epoch": 8.52245508982036, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1549, "step": 5693 }, { "epoch": 8.523952095808383, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.154, "step": 5694 }, { "epoch": 8.525449101796408, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1568, "step": 5695 }, { "epoch": 8.52694610778443, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1582, "step": 5696 }, { "epoch": 8.528443113772456, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1446, "step": 5697 }, { "epoch": 8.529940119760479, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1585, "step": 5698 }, { "epoch": 8.531437125748504, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1491, "step": 5699 }, { "epoch": 8.532934131736527, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1443, "step": 5700 }, { "epoch": 8.534431137724551, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1557, "step": 5701 }, { "epoch": 8.535928143712574, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1503, "step": 5702 }, { "epoch": 8.5374251497006, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1495, "step": 5703 }, { "epoch": 8.538922155688622, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1521, "step": 5704 }, { "epoch": 8.540419161676647, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1544, "step": 5705 }, { "epoch": 8.54191616766467, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1512, "step": 5706 }, { "epoch": 8.543413173652695, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1512, "step": 5707 }, { "epoch": 8.544910179640718, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1498, "step": 5708 }, { "epoch": 8.546407185628743, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.147, "step": 5709 }, { "epoch": 8.547904191616766, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1456, "step": 5710 }, { "epoch": 8.54940119760479, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1486, "step": 5711 }, { "epoch": 8.550898203592814, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1469, "step": 5712 }, { "epoch": 8.552395209580839, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.155, "step": 5713 }, { "epoch": 8.553892215568862, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1618, "step": 5714 }, { "epoch": 8.555389221556887, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1589, "step": 5715 }, { "epoch": 8.55688622754491, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1527, "step": 5716 }, { "epoch": 8.558383233532934, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.148, "step": 5717 }, { "epoch": 8.559880239520957, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1568, "step": 5718 }, { "epoch": 8.561377245508982, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1543, "step": 5719 }, { "epoch": 8.562874251497005, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1536, "step": 5720 }, { "epoch": 8.56437125748503, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1504, "step": 5721 }, { "epoch": 8.565868263473053, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.15, "step": 5722 }, { "epoch": 8.567365269461078, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.154, "step": 5723 }, { "epoch": 8.568862275449101, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1512, "step": 5724 }, { "epoch": 8.570359281437126, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1513, "step": 5725 }, { "epoch": 8.571856287425149, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1517, "step": 5726 }, { "epoch": 8.573353293413174, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1509, "step": 5727 }, { "epoch": 8.574850299401197, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.153, "step": 5728 }, { "epoch": 8.576347305389222, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1501, "step": 5729 }, { "epoch": 8.577844311377245, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1547, "step": 5730 }, { "epoch": 8.57934131736527, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1552, "step": 5731 }, { "epoch": 8.580838323353294, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1497, "step": 5732 }, { "epoch": 8.582335329341317, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1513, "step": 5733 }, { "epoch": 8.58383233532934, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.151, "step": 5734 }, { "epoch": 8.585329341317365, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1466, "step": 5735 }, { "epoch": 8.58682634730539, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1453, "step": 5736 }, { "epoch": 8.588323353293413, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1446, "step": 5737 }, { "epoch": 8.589820359281438, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.149, "step": 5738 }, { "epoch": 8.591317365269461, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1565, "step": 5739 }, { "epoch": 8.592814371257486, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1449, "step": 5740 }, { "epoch": 8.594311377245509, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1419, "step": 5741 }, { "epoch": 8.595808383233534, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1559, "step": 5742 }, { "epoch": 8.597305389221557, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1506, "step": 5743 }, { "epoch": 8.598802395209582, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1542, "step": 5744 }, { "epoch": 8.600299401197605, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1483, "step": 5745 }, { "epoch": 8.60179640718563, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1474, "step": 5746 }, { "epoch": 8.603293413173652, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1457, "step": 5747 }, { "epoch": 8.604790419161677, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1513, "step": 5748 }, { "epoch": 8.6062874251497, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1472, "step": 5749 }, { "epoch": 8.607784431137725, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1496, "step": 5750 }, { "epoch": 8.609281437125748, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1503, "step": 5751 }, { "epoch": 8.610778443113773, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.142, "step": 5752 }, { "epoch": 8.612275449101796, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1484, "step": 5753 }, { "epoch": 8.613772455089821, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.156, "step": 5754 }, { "epoch": 8.615269461077844, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1499, "step": 5755 }, { "epoch": 8.616766467065869, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1475, "step": 5756 }, { "epoch": 8.618263473053892, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1487, "step": 5757 }, { "epoch": 8.619760479041917, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1507, "step": 5758 }, { "epoch": 8.62125748502994, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1488, "step": 5759 }, { "epoch": 8.622754491017965, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1506, "step": 5760 }, { "epoch": 8.624251497005988, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1508, "step": 5761 }, { "epoch": 8.625748502994012, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1503, "step": 5762 }, { "epoch": 8.627245508982035, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1467, "step": 5763 }, { "epoch": 8.62874251497006, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1567, "step": 5764 }, { "epoch": 8.630239520958083, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1519, "step": 5765 }, { "epoch": 8.631736526946108, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1489, "step": 5766 }, { "epoch": 8.633233532934131, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1604, "step": 5767 }, { "epoch": 8.634730538922156, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1516, "step": 5768 }, { "epoch": 8.636227544910179, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1492, "step": 5769 }, { "epoch": 8.637724550898204, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.148, "step": 5770 }, { "epoch": 8.639221556886227, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1454, "step": 5771 }, { "epoch": 8.640718562874252, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1461, "step": 5772 }, { "epoch": 8.642215568862275, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1507, "step": 5773 }, { "epoch": 8.6437125748503, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1552, "step": 5774 }, { "epoch": 8.645209580838323, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1455, "step": 5775 }, { "epoch": 8.646706586826348, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1466, "step": 5776 }, { "epoch": 8.64820359281437, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1539, "step": 5777 }, { "epoch": 8.649700598802395, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1553, "step": 5778 }, { "epoch": 8.651197604790418, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1439, "step": 5779 }, { "epoch": 8.652694610778443, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1518, "step": 5780 }, { "epoch": 8.654191616766466, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1473, "step": 5781 }, { "epoch": 8.655688622754491, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1477, "step": 5782 }, { "epoch": 8.657185628742514, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1498, "step": 5783 }, { "epoch": 8.658682634730539, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1515, "step": 5784 }, { "epoch": 8.660179640718562, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1442, "step": 5785 }, { "epoch": 8.661676646706587, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.15, "step": 5786 }, { "epoch": 8.66317365269461, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.158, "step": 5787 }, { "epoch": 8.664670658682635, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.15, "step": 5788 }, { "epoch": 8.66616766467066, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.15, "step": 5789 }, { "epoch": 8.667664670658683, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1457, "step": 5790 }, { "epoch": 8.669161676646706, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1472, "step": 5791 }, { "epoch": 8.67065868263473, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1513, "step": 5792 }, { "epoch": 8.672155688622755, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1501, "step": 5793 }, { "epoch": 8.673652694610778, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1484, "step": 5794 }, { "epoch": 8.675149700598803, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1516, "step": 5795 }, { "epoch": 8.676646706586826, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1515, "step": 5796 }, { "epoch": 8.678143712574851, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.155, "step": 5797 }, { "epoch": 8.679640718562874, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1472, "step": 5798 }, { "epoch": 8.681137724550899, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.155, "step": 5799 }, { "epoch": 8.682634730538922, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.145, "step": 5800 }, { "epoch": 8.684131736526947, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1572, "step": 5801 }, { "epoch": 8.68562874251497, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1469, "step": 5802 }, { "epoch": 8.687125748502995, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1439, "step": 5803 }, { "epoch": 8.688622754491018, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1436, "step": 5804 }, { "epoch": 8.690119760479043, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1466, "step": 5805 }, { "epoch": 8.691616766467066, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.153, "step": 5806 }, { "epoch": 8.69311377245509, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1537, "step": 5807 }, { "epoch": 8.694610778443113, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1456, "step": 5808 }, { "epoch": 8.696107784431138, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1486, "step": 5809 }, { "epoch": 8.697604790419161, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1407, "step": 5810 }, { "epoch": 8.699101796407186, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1515, "step": 5811 }, { "epoch": 8.70059880239521, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1511, "step": 5812 }, { "epoch": 8.702095808383234, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1521, "step": 5813 }, { "epoch": 8.703592814371257, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1502, "step": 5814 }, { "epoch": 8.705089820359282, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1508, "step": 5815 }, { "epoch": 8.706586826347305, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.146, "step": 5816 }, { "epoch": 8.70808383233533, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1452, "step": 5817 }, { "epoch": 8.709580838323353, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1532, "step": 5818 }, { "epoch": 8.711077844311378, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1482, "step": 5819 }, { "epoch": 8.7125748502994, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1485, "step": 5820 }, { "epoch": 8.714071856287426, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.157, "step": 5821 }, { "epoch": 8.715568862275449, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.155, "step": 5822 }, { "epoch": 8.717065868263473, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1456, "step": 5823 }, { "epoch": 8.718562874251496, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.15, "step": 5824 }, { "epoch": 8.720059880239521, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1495, "step": 5825 }, { "epoch": 8.721556886227544, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.15, "step": 5826 }, { "epoch": 8.72305389221557, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1487, "step": 5827 }, { "epoch": 8.724550898203592, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1496, "step": 5828 }, { "epoch": 8.726047904191617, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1447, "step": 5829 }, { "epoch": 8.72754491017964, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1494, "step": 5830 }, { "epoch": 8.729041916167665, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1492, "step": 5831 }, { "epoch": 8.730538922155688, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1538, "step": 5832 }, { "epoch": 8.732035928143713, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1534, "step": 5833 }, { "epoch": 8.733532934131736, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1464, "step": 5834 }, { "epoch": 8.73502994011976, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1423, "step": 5835 }, { "epoch": 8.736526946107784, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1531, "step": 5836 }, { "epoch": 8.738023952095809, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1541, "step": 5837 }, { "epoch": 8.739520958083832, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1505, "step": 5838 }, { "epoch": 8.741017964071856, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1404, "step": 5839 }, { "epoch": 8.74251497005988, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1575, "step": 5840 }, { "epoch": 8.744011976047904, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1534, "step": 5841 }, { "epoch": 8.745508982035929, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1514, "step": 5842 }, { "epoch": 8.747005988023952, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1451, "step": 5843 }, { "epoch": 8.748502994011975, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1469, "step": 5844 }, { "epoch": 8.75, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1486, "step": 5845 }, { "epoch": 8.751497005988025, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1485, "step": 5846 }, { "epoch": 8.752994011976048, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1507, "step": 5847 }, { "epoch": 8.754491017964071, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1491, "step": 5848 }, { "epoch": 8.755988023952096, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1504, "step": 5849 }, { "epoch": 8.75748502994012, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1456, "step": 5850 }, { "epoch": 8.758982035928144, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1499, "step": 5851 }, { "epoch": 8.760479041916168, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1556, "step": 5852 }, { "epoch": 8.761976047904191, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1484, "step": 5853 }, { "epoch": 8.763473053892216, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1536, "step": 5854 }, { "epoch": 8.76497005988024, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1485, "step": 5855 }, { "epoch": 8.766467065868264, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1493, "step": 5856 }, { "epoch": 8.767964071856287, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1511, "step": 5857 }, { "epoch": 8.769461077844312, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.1478, "step": 5858 }, { "epoch": 8.770958083832335, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1536, "step": 5859 }, { "epoch": 8.77245508982036, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1544, "step": 5860 }, { "epoch": 8.773952095808383, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1555, "step": 5861 }, { "epoch": 8.775449101796408, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1465, "step": 5862 }, { "epoch": 8.77694610778443, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.158, "step": 5863 }, { "epoch": 8.778443113772456, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1471, "step": 5864 }, { "epoch": 8.779940119760479, "grad_norm": 0.173828125, "learning_rate": 0.0008, "loss": 1.1502, "step": 5865 }, { "epoch": 8.781437125748504, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1498, "step": 5866 }, { "epoch": 8.782934131736527, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1505, "step": 5867 }, { "epoch": 8.784431137724551, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.153, "step": 5868 }, { "epoch": 8.785928143712574, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1461, "step": 5869 }, { "epoch": 8.7874251497006, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1426, "step": 5870 }, { "epoch": 8.788922155688622, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1562, "step": 5871 }, { "epoch": 8.790419161676647, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1555, "step": 5872 }, { "epoch": 8.79191616766467, "grad_norm": 0.166015625, "learning_rate": 0.0008, "loss": 1.1455, "step": 5873 }, { "epoch": 8.793413173652695, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1514, "step": 5874 }, { "epoch": 8.794910179640718, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1552, "step": 5875 }, { "epoch": 8.796407185628743, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1485, "step": 5876 }, { "epoch": 8.797904191616766, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1542, "step": 5877 }, { "epoch": 8.79940119760479, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1525, "step": 5878 }, { "epoch": 8.800898203592814, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1472, "step": 5879 }, { "epoch": 8.802395209580839, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1497, "step": 5880 }, { "epoch": 8.803892215568862, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1531, "step": 5881 }, { "epoch": 8.805389221556887, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1555, "step": 5882 }, { "epoch": 8.80688622754491, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1483, "step": 5883 }, { "epoch": 8.808383233532934, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1504, "step": 5884 }, { "epoch": 8.809880239520957, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1491, "step": 5885 }, { "epoch": 8.811377245508982, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.1511, "step": 5886 }, { "epoch": 8.812874251497005, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1457, "step": 5887 }, { "epoch": 8.81437125748503, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1436, "step": 5888 }, { "epoch": 8.815868263473053, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1566, "step": 5889 }, { "epoch": 8.817365269461078, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1491, "step": 5890 }, { "epoch": 8.818862275449101, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1493, "step": 5891 }, { "epoch": 8.820359281437126, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1512, "step": 5892 }, { "epoch": 8.821856287425149, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1588, "step": 5893 }, { "epoch": 8.823353293413174, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1448, "step": 5894 }, { "epoch": 8.824850299401197, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1486, "step": 5895 }, { "epoch": 8.826347305389222, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1494, "step": 5896 }, { "epoch": 8.827844311377245, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1469, "step": 5897 }, { "epoch": 8.82934131736527, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1518, "step": 5898 }, { "epoch": 8.830838323353294, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.15, "step": 5899 }, { "epoch": 8.832335329341317, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1508, "step": 5900 }, { "epoch": 8.83383233532934, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1423, "step": 5901 }, { "epoch": 8.835329341317365, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1518, "step": 5902 }, { "epoch": 8.83682634730539, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1496, "step": 5903 }, { "epoch": 8.838323353293413, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1489, "step": 5904 }, { "epoch": 8.839820359281438, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.153, "step": 5905 }, { "epoch": 8.841317365269461, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1475, "step": 5906 }, { "epoch": 8.842814371257486, "grad_norm": 0.1787109375, "learning_rate": 0.0008, "loss": 1.1523, "step": 5907 }, { "epoch": 8.844311377245509, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1556, "step": 5908 }, { "epoch": 8.845808383233534, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1521, "step": 5909 }, { "epoch": 8.847305389221557, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.153, "step": 5910 }, { "epoch": 8.848802395209582, "grad_norm": 0.1865234375, "learning_rate": 0.0008, "loss": 1.1468, "step": 5911 }, { "epoch": 8.850299401197605, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1502, "step": 5912 }, { "epoch": 8.85179640718563, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1572, "step": 5913 }, { "epoch": 8.853293413173652, "grad_norm": 0.1689453125, "learning_rate": 0.0008, "loss": 1.1499, "step": 5914 }, { "epoch": 8.854790419161677, "grad_norm": 0.1865234375, "learning_rate": 0.0008, "loss": 1.1468, "step": 5915 }, { "epoch": 8.8562874251497, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1477, "step": 5916 }, { "epoch": 8.857784431137725, "grad_norm": 0.1591796875, "learning_rate": 0.0008, "loss": 1.1491, "step": 5917 }, { "epoch": 8.859281437125748, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1539, "step": 5918 }, { "epoch": 8.860778443113773, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.157, "step": 5919 }, { "epoch": 8.862275449101796, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1505, "step": 5920 }, { "epoch": 8.863772455089821, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1522, "step": 5921 }, { "epoch": 8.865269461077844, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1479, "step": 5922 }, { "epoch": 8.866766467065869, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1513, "step": 5923 }, { "epoch": 8.868263473053892, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1435, "step": 5924 }, { "epoch": 8.869760479041917, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1486, "step": 5925 }, { "epoch": 8.87125748502994, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1487, "step": 5926 }, { "epoch": 8.872754491017965, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.145, "step": 5927 }, { "epoch": 8.874251497005988, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1503, "step": 5928 }, { "epoch": 8.875748502994012, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1482, "step": 5929 }, { "epoch": 8.877245508982035, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1515, "step": 5930 }, { "epoch": 8.87874251497006, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1562, "step": 5931 }, { "epoch": 8.880239520958083, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1515, "step": 5932 }, { "epoch": 8.881736526946108, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.147, "step": 5933 }, { "epoch": 8.883233532934131, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1477, "step": 5934 }, { "epoch": 8.884730538922156, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.15, "step": 5935 }, { "epoch": 8.886227544910179, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.154, "step": 5936 }, { "epoch": 8.887724550898204, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1493, "step": 5937 }, { "epoch": 8.889221556886227, "grad_norm": 0.051513671875, "learning_rate": 0.0008, "loss": 1.1478, "step": 5938 }, { "epoch": 8.890718562874252, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1489, "step": 5939 }, { "epoch": 8.892215568862275, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1535, "step": 5940 }, { "epoch": 8.8937125748503, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1527, "step": 5941 }, { "epoch": 8.895209580838323, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1529, "step": 5942 }, { "epoch": 8.896706586826348, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1513, "step": 5943 }, { "epoch": 8.89820359281437, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1531, "step": 5944 }, { "epoch": 8.899700598802395, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1522, "step": 5945 }, { "epoch": 8.901197604790418, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1452, "step": 5946 }, { "epoch": 8.902694610778443, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1465, "step": 5947 }, { "epoch": 8.904191616766466, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1519, "step": 5948 }, { "epoch": 8.905688622754491, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1397, "step": 5949 }, { "epoch": 8.907185628742514, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1467, "step": 5950 }, { "epoch": 8.908682634730539, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1563, "step": 5951 }, { "epoch": 8.910179640718562, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1566, "step": 5952 }, { "epoch": 8.911676646706587, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1518, "step": 5953 }, { "epoch": 8.91317365269461, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1534, "step": 5954 }, { "epoch": 8.914670658682635, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1558, "step": 5955 }, { "epoch": 8.91616766467066, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1513, "step": 5956 }, { "epoch": 8.917664670658683, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1492, "step": 5957 }, { "epoch": 8.919161676646706, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1505, "step": 5958 }, { "epoch": 8.92065868263473, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1516, "step": 5959 }, { "epoch": 8.922155688622755, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1505, "step": 5960 }, { "epoch": 8.923652694610778, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1479, "step": 5961 }, { "epoch": 8.925149700598803, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1429, "step": 5962 }, { "epoch": 8.926646706586826, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1523, "step": 5963 }, { "epoch": 8.928143712574851, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1463, "step": 5964 }, { "epoch": 8.929640718562874, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1444, "step": 5965 }, { "epoch": 8.931137724550899, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1476, "step": 5966 }, { "epoch": 8.932634730538922, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1451, "step": 5967 }, { "epoch": 8.934131736526947, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1482, "step": 5968 }, { "epoch": 8.93562874251497, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1501, "step": 5969 }, { "epoch": 8.937125748502995, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1475, "step": 5970 }, { "epoch": 8.938622754491018, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1513, "step": 5971 }, { "epoch": 8.940119760479043, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1521, "step": 5972 }, { "epoch": 8.941616766467066, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1555, "step": 5973 }, { "epoch": 8.94311377245509, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1535, "step": 5974 }, { "epoch": 8.944610778443113, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1466, "step": 5975 }, { "epoch": 8.946107784431138, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1508, "step": 5976 }, { "epoch": 8.947604790419161, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1509, "step": 5977 }, { "epoch": 8.949101796407186, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1481, "step": 5978 }, { "epoch": 8.95059880239521, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1424, "step": 5979 }, { "epoch": 8.952095808383234, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.144, "step": 5980 }, { "epoch": 8.953592814371257, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1466, "step": 5981 }, { "epoch": 8.955089820359282, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1504, "step": 5982 }, { "epoch": 8.956586826347305, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.16, "step": 5983 }, { "epoch": 8.95808383233533, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1553, "step": 5984 }, { "epoch": 8.959580838323353, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1506, "step": 5985 }, { "epoch": 8.961077844311378, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1524, "step": 5986 }, { "epoch": 8.9625748502994, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1522, "step": 5987 }, { "epoch": 8.964071856287426, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1478, "step": 5988 }, { "epoch": 8.965568862275449, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1485, "step": 5989 }, { "epoch": 8.967065868263473, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.155, "step": 5990 }, { "epoch": 8.968562874251496, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1452, "step": 5991 }, { "epoch": 8.970059880239521, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1563, "step": 5992 }, { "epoch": 8.971556886227544, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1424, "step": 5993 }, { "epoch": 8.97305389221557, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1537, "step": 5994 }, { "epoch": 8.974550898203592, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1447, "step": 5995 }, { "epoch": 8.976047904191617, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1496, "step": 5996 }, { "epoch": 8.97754491017964, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1554, "step": 5997 }, { "epoch": 8.979041916167665, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1557, "step": 5998 }, { "epoch": 8.980538922155688, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1521, "step": 5999 }, { "epoch": 8.982035928143713, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1447, "step": 6000 }, { "epoch": 8.983532934131736, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1517, "step": 6001 }, { "epoch": 8.98502994011976, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.153, "step": 6002 }, { "epoch": 8.986526946107784, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1506, "step": 6003 }, { "epoch": 8.988023952095809, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1492, "step": 6004 }, { "epoch": 8.989520958083832, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.145, "step": 6005 }, { "epoch": 8.991017964071856, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1445, "step": 6006 }, { "epoch": 8.99251497005988, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.149, "step": 6007 }, { "epoch": 8.994011976047904, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1478, "step": 6008 }, { "epoch": 8.995508982035929, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1389, "step": 6009 }, { "epoch": 8.997005988023952, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1459, "step": 6010 }, { "epoch": 8.998502994011975, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1472, "step": 6011 }, { "epoch": 9.0, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1453, "step": 6012 }, { "epoch": 9.001497005988025, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1485, "step": 6013 }, { "epoch": 9.002994011976048, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1527, "step": 6014 }, { "epoch": 9.004491017964073, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1485, "step": 6015 }, { "epoch": 9.005988023952096, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1451, "step": 6016 }, { "epoch": 9.00748502994012, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1446, "step": 6017 }, { "epoch": 9.008982035928144, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1446, "step": 6018 }, { "epoch": 9.010479041916168, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1507, "step": 6019 }, { "epoch": 9.011976047904191, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1513, "step": 6020 }, { "epoch": 9.013473053892216, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1475, "step": 6021 }, { "epoch": 9.01497005988024, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1399, "step": 6022 }, { "epoch": 9.016467065868264, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1495, "step": 6023 }, { "epoch": 9.017964071856287, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1563, "step": 6024 }, { "epoch": 9.019461077844312, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1521, "step": 6025 }, { "epoch": 9.020958083832335, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.1486, "step": 6026 }, { "epoch": 9.02245508982036, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1519, "step": 6027 }, { "epoch": 9.023952095808383, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1488, "step": 6028 }, { "epoch": 9.025449101796408, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1462, "step": 6029 }, { "epoch": 9.02694610778443, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1444, "step": 6030 }, { "epoch": 9.028443113772456, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1512, "step": 6031 }, { "epoch": 9.029940119760479, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1476, "step": 6032 }, { "epoch": 9.031437125748504, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1503, "step": 6033 }, { "epoch": 9.032934131736527, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1481, "step": 6034 }, { "epoch": 9.034431137724551, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1484, "step": 6035 }, { "epoch": 9.035928143712574, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1539, "step": 6036 }, { "epoch": 9.0374251497006, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1437, "step": 6037 }, { "epoch": 9.038922155688622, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1482, "step": 6038 }, { "epoch": 9.040419161676647, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1558, "step": 6039 }, { "epoch": 9.04191616766467, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1449, "step": 6040 }, { "epoch": 9.043413173652695, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1496, "step": 6041 }, { "epoch": 9.044910179640718, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.146, "step": 6042 }, { "epoch": 9.046407185628743, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1496, "step": 6043 }, { "epoch": 9.047904191616766, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1468, "step": 6044 }, { "epoch": 9.04940119760479, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1504, "step": 6045 }, { "epoch": 9.050898203592814, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1533, "step": 6046 }, { "epoch": 9.052395209580839, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1485, "step": 6047 }, { "epoch": 9.053892215568862, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1496, "step": 6048 }, { "epoch": 9.055389221556887, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1497, "step": 6049 }, { "epoch": 9.05688622754491, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1509, "step": 6050 }, { "epoch": 9.058383233532934, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1478, "step": 6051 }, { "epoch": 9.059880239520957, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1359, "step": 6052 }, { "epoch": 9.061377245508982, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1521, "step": 6053 }, { "epoch": 9.062874251497005, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1528, "step": 6054 }, { "epoch": 9.06437125748503, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1458, "step": 6055 }, { "epoch": 9.065868263473053, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1457, "step": 6056 }, { "epoch": 9.067365269461078, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1464, "step": 6057 }, { "epoch": 9.068862275449101, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1456, "step": 6058 }, { "epoch": 9.070359281437126, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1485, "step": 6059 }, { "epoch": 9.071856287425149, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1493, "step": 6060 }, { "epoch": 9.073353293413174, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1427, "step": 6061 }, { "epoch": 9.074850299401197, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1497, "step": 6062 }, { "epoch": 9.076347305389222, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1583, "step": 6063 }, { "epoch": 9.077844311377245, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.145, "step": 6064 }, { "epoch": 9.07934131736527, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1447, "step": 6065 }, { "epoch": 9.080838323353293, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1438, "step": 6066 }, { "epoch": 9.082335329341317, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1545, "step": 6067 }, { "epoch": 9.08383233532934, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1534, "step": 6068 }, { "epoch": 9.085329341317365, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1601, "step": 6069 }, { "epoch": 9.08682634730539, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1495, "step": 6070 }, { "epoch": 9.088323353293413, "grad_norm": 0.1533203125, "learning_rate": 0.0008, "loss": 1.1472, "step": 6071 }, { "epoch": 9.089820359281438, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1492, "step": 6072 }, { "epoch": 9.091317365269461, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1494, "step": 6073 }, { "epoch": 9.092814371257486, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1547, "step": 6074 }, { "epoch": 9.094311377245509, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1479, "step": 6075 }, { "epoch": 9.095808383233534, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1526, "step": 6076 }, { "epoch": 9.097305389221557, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1548, "step": 6077 }, { "epoch": 9.098802395209582, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1477, "step": 6078 }, { "epoch": 9.100299401197605, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1405, "step": 6079 }, { "epoch": 9.10179640718563, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1546, "step": 6080 }, { "epoch": 9.103293413173652, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1466, "step": 6081 }, { "epoch": 9.104790419161677, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1502, "step": 6082 }, { "epoch": 9.1062874251497, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1453, "step": 6083 }, { "epoch": 9.107784431137725, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1468, "step": 6084 }, { "epoch": 9.109281437125748, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1506, "step": 6085 }, { "epoch": 9.110778443113773, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1477, "step": 6086 }, { "epoch": 9.112275449101796, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1401, "step": 6087 }, { "epoch": 9.113772455089821, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1436, "step": 6088 }, { "epoch": 9.115269461077844, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1593, "step": 6089 }, { "epoch": 9.116766467065869, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1517, "step": 6090 }, { "epoch": 9.118263473053892, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.142, "step": 6091 }, { "epoch": 9.119760479041917, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1557, "step": 6092 }, { "epoch": 9.12125748502994, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1469, "step": 6093 }, { "epoch": 9.122754491017965, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1515, "step": 6094 }, { "epoch": 9.124251497005988, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1503, "step": 6095 }, { "epoch": 9.125748502994012, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1503, "step": 6096 }, { "epoch": 9.127245508982035, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1435, "step": 6097 }, { "epoch": 9.12874251497006, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1541, "step": 6098 }, { "epoch": 9.130239520958083, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1507, "step": 6099 }, { "epoch": 9.131736526946108, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1433, "step": 6100 }, { "epoch": 9.133233532934131, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.153, "step": 6101 }, { "epoch": 9.134730538922156, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1438, "step": 6102 }, { "epoch": 9.136227544910179, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.1488, "step": 6103 }, { "epoch": 9.137724550898204, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.1498, "step": 6104 }, { "epoch": 9.139221556886227, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1462, "step": 6105 }, { "epoch": 9.140718562874252, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1578, "step": 6106 }, { "epoch": 9.142215568862275, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1524, "step": 6107 }, { "epoch": 9.1437125748503, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1556, "step": 6108 }, { "epoch": 9.145209580838323, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1516, "step": 6109 }, { "epoch": 9.146706586826348, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1469, "step": 6110 }, { "epoch": 9.14820359281437, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.1515, "step": 6111 }, { "epoch": 9.149700598802395, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1535, "step": 6112 }, { "epoch": 9.151197604790418, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1524, "step": 6113 }, { "epoch": 9.152694610778443, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1414, "step": 6114 }, { "epoch": 9.154191616766466, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1481, "step": 6115 }, { "epoch": 9.155688622754491, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1488, "step": 6116 }, { "epoch": 9.157185628742514, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1464, "step": 6117 }, { "epoch": 9.158682634730539, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1508, "step": 6118 }, { "epoch": 9.160179640718562, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1531, "step": 6119 }, { "epoch": 9.161676646706587, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1477, "step": 6120 }, { "epoch": 9.16317365269461, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1398, "step": 6121 }, { "epoch": 9.164670658682635, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1506, "step": 6122 }, { "epoch": 9.16616766467066, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1487, "step": 6123 }, { "epoch": 9.167664670658683, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1502, "step": 6124 }, { "epoch": 9.169161676646707, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1426, "step": 6125 }, { "epoch": 9.17065868263473, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1558, "step": 6126 }, { "epoch": 9.172155688622755, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1531, "step": 6127 }, { "epoch": 9.173652694610778, "grad_norm": 0.1708984375, "learning_rate": 0.0008, "loss": 1.1492, "step": 6128 }, { "epoch": 9.175149700598803, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1495, "step": 6129 }, { "epoch": 9.176646706586826, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1531, "step": 6130 }, { "epoch": 9.178143712574851, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1484, "step": 6131 }, { "epoch": 9.179640718562874, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1423, "step": 6132 }, { "epoch": 9.181137724550899, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1551, "step": 6133 }, { "epoch": 9.182634730538922, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.151, "step": 6134 }, { "epoch": 9.184131736526947, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1463, "step": 6135 }, { "epoch": 9.18562874251497, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1511, "step": 6136 }, { "epoch": 9.187125748502995, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1496, "step": 6137 }, { "epoch": 9.188622754491018, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1531, "step": 6138 }, { "epoch": 9.190119760479043, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1485, "step": 6139 }, { "epoch": 9.191616766467066, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1455, "step": 6140 }, { "epoch": 9.19311377245509, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1535, "step": 6141 }, { "epoch": 9.194610778443113, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1547, "step": 6142 }, { "epoch": 9.196107784431138, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1555, "step": 6143 }, { "epoch": 9.197604790419161, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1548, "step": 6144 }, { "epoch": 9.199101796407186, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1417, "step": 6145 }, { "epoch": 9.20059880239521, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.142, "step": 6146 }, { "epoch": 9.202095808383234, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.151, "step": 6147 }, { "epoch": 9.203592814371257, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.145, "step": 6148 }, { "epoch": 9.205089820359282, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1504, "step": 6149 }, { "epoch": 9.206586826347305, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1462, "step": 6150 }, { "epoch": 9.20808383233533, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.146, "step": 6151 }, { "epoch": 9.209580838323353, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1485, "step": 6152 }, { "epoch": 9.211077844311378, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1422, "step": 6153 }, { "epoch": 9.2125748502994, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1551, "step": 6154 }, { "epoch": 9.214071856287426, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1391, "step": 6155 }, { "epoch": 9.215568862275449, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1522, "step": 6156 }, { "epoch": 9.217065868263473, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1485, "step": 6157 }, { "epoch": 9.218562874251496, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1522, "step": 6158 }, { "epoch": 9.220059880239521, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1518, "step": 6159 }, { "epoch": 9.221556886227544, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1508, "step": 6160 }, { "epoch": 9.22305389221557, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1537, "step": 6161 }, { "epoch": 9.224550898203592, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1455, "step": 6162 }, { "epoch": 9.226047904191617, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1412, "step": 6163 }, { "epoch": 9.22754491017964, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1472, "step": 6164 }, { "epoch": 9.229041916167665, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1537, "step": 6165 }, { "epoch": 9.230538922155688, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1488, "step": 6166 }, { "epoch": 9.232035928143713, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1476, "step": 6167 }, { "epoch": 9.233532934131736, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1476, "step": 6168 }, { "epoch": 9.23502994011976, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1464, "step": 6169 }, { "epoch": 9.236526946107784, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1496, "step": 6170 }, { "epoch": 9.238023952095809, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1445, "step": 6171 }, { "epoch": 9.239520958083832, "grad_norm": 0.2099609375, "learning_rate": 0.0008, "loss": 1.1537, "step": 6172 }, { "epoch": 9.241017964071856, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1481, "step": 6173 }, { "epoch": 9.24251497005988, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1442, "step": 6174 }, { "epoch": 9.244011976047904, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1534, "step": 6175 }, { "epoch": 9.245508982035927, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1485, "step": 6176 }, { "epoch": 9.247005988023952, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1522, "step": 6177 }, { "epoch": 9.248502994011975, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1431, "step": 6178 }, { "epoch": 9.25, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1462, "step": 6179 }, { "epoch": 9.251497005988025, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1536, "step": 6180 }, { "epoch": 9.252994011976048, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1485, "step": 6181 }, { "epoch": 9.254491017964073, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1553, "step": 6182 }, { "epoch": 9.255988023952096, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1465, "step": 6183 }, { "epoch": 9.25748502994012, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1427, "step": 6184 }, { "epoch": 9.258982035928144, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1493, "step": 6185 }, { "epoch": 9.260479041916168, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1422, "step": 6186 }, { "epoch": 9.261976047904191, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1435, "step": 6187 }, { "epoch": 9.263473053892216, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1511, "step": 6188 }, { "epoch": 9.26497005988024, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1544, "step": 6189 }, { "epoch": 9.266467065868264, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1511, "step": 6190 }, { "epoch": 9.267964071856287, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1449, "step": 6191 }, { "epoch": 9.269461077844312, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1484, "step": 6192 }, { "epoch": 9.270958083832335, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1492, "step": 6193 }, { "epoch": 9.27245508982036, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1502, "step": 6194 }, { "epoch": 9.273952095808383, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1513, "step": 6195 }, { "epoch": 9.275449101796408, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1488, "step": 6196 }, { "epoch": 9.27694610778443, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1411, "step": 6197 }, { "epoch": 9.278443113772456, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1511, "step": 6198 }, { "epoch": 9.279940119760479, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1477, "step": 6199 }, { "epoch": 9.281437125748504, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1549, "step": 6200 }, { "epoch": 9.282934131736527, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1515, "step": 6201 }, { "epoch": 9.284431137724551, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1481, "step": 6202 }, { "epoch": 9.285928143712574, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1454, "step": 6203 }, { "epoch": 9.2874251497006, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1426, "step": 6204 }, { "epoch": 9.288922155688622, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1602, "step": 6205 }, { "epoch": 9.290419161676647, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.149, "step": 6206 }, { "epoch": 9.29191616766467, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1453, "step": 6207 }, { "epoch": 9.293413173652695, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1485, "step": 6208 }, { "epoch": 9.294910179640718, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1493, "step": 6209 }, { "epoch": 9.296407185628743, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1463, "step": 6210 }, { "epoch": 9.297904191616766, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1479, "step": 6211 }, { "epoch": 9.29940119760479, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1488, "step": 6212 }, { "epoch": 9.300898203592814, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1476, "step": 6213 }, { "epoch": 9.302395209580839, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1465, "step": 6214 }, { "epoch": 9.303892215568862, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1468, "step": 6215 }, { "epoch": 9.305389221556887, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1491, "step": 6216 }, { "epoch": 9.30688622754491, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1428, "step": 6217 }, { "epoch": 9.308383233532934, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1479, "step": 6218 }, { "epoch": 9.309880239520957, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1478, "step": 6219 }, { "epoch": 9.311377245508982, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1536, "step": 6220 }, { "epoch": 9.312874251497005, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1415, "step": 6221 }, { "epoch": 9.31437125748503, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1467, "step": 6222 }, { "epoch": 9.315868263473053, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1457, "step": 6223 }, { "epoch": 9.317365269461078, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1522, "step": 6224 }, { "epoch": 9.318862275449101, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1539, "step": 6225 }, { "epoch": 9.320359281437126, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1517, "step": 6226 }, { "epoch": 9.321856287425149, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1487, "step": 6227 }, { "epoch": 9.323353293413174, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1491, "step": 6228 }, { "epoch": 9.324850299401197, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1478, "step": 6229 }, { "epoch": 9.326347305389222, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1505, "step": 6230 }, { "epoch": 9.327844311377245, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1514, "step": 6231 }, { "epoch": 9.32934131736527, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1453, "step": 6232 }, { "epoch": 9.330838323353294, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1507, "step": 6233 }, { "epoch": 9.332335329341317, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1529, "step": 6234 }, { "epoch": 9.33383233532934, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1397, "step": 6235 }, { "epoch": 9.335329341317365, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1535, "step": 6236 }, { "epoch": 9.33682634730539, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1495, "step": 6237 }, { "epoch": 9.338323353293413, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1478, "step": 6238 }, { "epoch": 9.339820359281438, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1497, "step": 6239 }, { "epoch": 9.341317365269461, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1484, "step": 6240 }, { "epoch": 9.342814371257486, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.147, "step": 6241 }, { "epoch": 9.344311377245509, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1451, "step": 6242 }, { "epoch": 9.345808383233534, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1477, "step": 6243 }, { "epoch": 9.347305389221557, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1499, "step": 6244 }, { "epoch": 9.348802395209582, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1466, "step": 6245 }, { "epoch": 9.350299401197605, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1492, "step": 6246 }, { "epoch": 9.35179640718563, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1438, "step": 6247 }, { "epoch": 9.353293413173652, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1467, "step": 6248 }, { "epoch": 9.354790419161677, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1483, "step": 6249 }, { "epoch": 9.3562874251497, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1483, "step": 6250 }, { "epoch": 9.357784431137725, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1441, "step": 6251 }, { "epoch": 9.359281437125748, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1523, "step": 6252 }, { "epoch": 9.360778443113773, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1521, "step": 6253 }, { "epoch": 9.362275449101796, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1538, "step": 6254 }, { "epoch": 9.363772455089821, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1507, "step": 6255 }, { "epoch": 9.365269461077844, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1498, "step": 6256 }, { "epoch": 9.366766467065869, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.149, "step": 6257 }, { "epoch": 9.368263473053892, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.148, "step": 6258 }, { "epoch": 9.369760479041917, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1528, "step": 6259 }, { "epoch": 9.37125748502994, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1461, "step": 6260 }, { "epoch": 9.372754491017965, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1476, "step": 6261 }, { "epoch": 9.374251497005988, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1452, "step": 6262 }, { "epoch": 9.375748502994012, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1547, "step": 6263 }, { "epoch": 9.377245508982035, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1467, "step": 6264 }, { "epoch": 9.37874251497006, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1509, "step": 6265 }, { "epoch": 9.380239520958083, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1447, "step": 6266 }, { "epoch": 9.381736526946108, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1507, "step": 6267 }, { "epoch": 9.383233532934131, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.152, "step": 6268 }, { "epoch": 9.384730538922156, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1499, "step": 6269 }, { "epoch": 9.386227544910179, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1463, "step": 6270 }, { "epoch": 9.387724550898204, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1497, "step": 6271 }, { "epoch": 9.389221556886227, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1492, "step": 6272 }, { "epoch": 9.390718562874252, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1525, "step": 6273 }, { "epoch": 9.392215568862275, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.149, "step": 6274 }, { "epoch": 9.3937125748503, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1516, "step": 6275 }, { "epoch": 9.395209580838323, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1434, "step": 6276 }, { "epoch": 9.396706586826348, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1528, "step": 6277 }, { "epoch": 9.39820359281437, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.153, "step": 6278 }, { "epoch": 9.399700598802395, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.144, "step": 6279 }, { "epoch": 9.401197604790418, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1465, "step": 6280 }, { "epoch": 9.402694610778443, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1483, "step": 6281 }, { "epoch": 9.404191616766466, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1474, "step": 6282 }, { "epoch": 9.405688622754491, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1426, "step": 6283 }, { "epoch": 9.407185628742514, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1397, "step": 6284 }, { "epoch": 9.408682634730539, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1473, "step": 6285 }, { "epoch": 9.410179640718562, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1447, "step": 6286 }, { "epoch": 9.411676646706587, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1469, "step": 6287 }, { "epoch": 9.41317365269461, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1533, "step": 6288 }, { "epoch": 9.414670658682635, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1447, "step": 6289 }, { "epoch": 9.41616766467066, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1432, "step": 6290 }, { "epoch": 9.417664670658683, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1464, "step": 6291 }, { "epoch": 9.419161676646706, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1503, "step": 6292 }, { "epoch": 9.42065868263473, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1536, "step": 6293 }, { "epoch": 9.422155688622755, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1446, "step": 6294 }, { "epoch": 9.423652694610778, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1566, "step": 6295 }, { "epoch": 9.425149700598803, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1492, "step": 6296 }, { "epoch": 9.426646706586826, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1499, "step": 6297 }, { "epoch": 9.428143712574851, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1528, "step": 6298 }, { "epoch": 9.429640718562874, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1472, "step": 6299 }, { "epoch": 9.431137724550899, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1465, "step": 6300 }, { "epoch": 9.432634730538922, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.143, "step": 6301 }, { "epoch": 9.434131736526947, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1432, "step": 6302 }, { "epoch": 9.43562874251497, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1521, "step": 6303 }, { "epoch": 9.437125748502995, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1491, "step": 6304 }, { "epoch": 9.438622754491018, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1506, "step": 6305 }, { "epoch": 9.440119760479043, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1449, "step": 6306 }, { "epoch": 9.441616766467066, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1541, "step": 6307 }, { "epoch": 9.44311377245509, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1449, "step": 6308 }, { "epoch": 9.444610778443113, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.147, "step": 6309 }, { "epoch": 9.446107784431138, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.15, "step": 6310 }, { "epoch": 9.447604790419161, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1509, "step": 6311 }, { "epoch": 9.449101796407186, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1563, "step": 6312 }, { "epoch": 9.45059880239521, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1548, "step": 6313 }, { "epoch": 9.452095808383234, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1504, "step": 6314 }, { "epoch": 9.453592814371257, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1546, "step": 6315 }, { "epoch": 9.455089820359282, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1477, "step": 6316 }, { "epoch": 9.456586826347305, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1509, "step": 6317 }, { "epoch": 9.45808383233533, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1488, "step": 6318 }, { "epoch": 9.459580838323353, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1504, "step": 6319 }, { "epoch": 9.461077844311378, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.15, "step": 6320 }, { "epoch": 9.4625748502994, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1444, "step": 6321 }, { "epoch": 9.464071856287426, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1508, "step": 6322 }, { "epoch": 9.465568862275449, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1513, "step": 6323 }, { "epoch": 9.467065868263473, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1476, "step": 6324 }, { "epoch": 9.468562874251496, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1465, "step": 6325 }, { "epoch": 9.470059880239521, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1523, "step": 6326 }, { "epoch": 9.471556886227544, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1506, "step": 6327 }, { "epoch": 9.47305389221557, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1462, "step": 6328 }, { "epoch": 9.474550898203592, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1485, "step": 6329 }, { "epoch": 9.476047904191617, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1462, "step": 6330 }, { "epoch": 9.47754491017964, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1484, "step": 6331 }, { "epoch": 9.479041916167665, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1451, "step": 6332 }, { "epoch": 9.480538922155688, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1432, "step": 6333 }, { "epoch": 9.482035928143713, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1495, "step": 6334 }, { "epoch": 9.483532934131736, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1561, "step": 6335 }, { "epoch": 9.48502994011976, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1462, "step": 6336 }, { "epoch": 9.486526946107784, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1494, "step": 6337 }, { "epoch": 9.488023952095809, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1448, "step": 6338 }, { "epoch": 9.489520958083832, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1425, "step": 6339 }, { "epoch": 9.491017964071856, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1476, "step": 6340 }, { "epoch": 9.49251497005988, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1444, "step": 6341 }, { "epoch": 9.494011976047904, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1421, "step": 6342 }, { "epoch": 9.495508982035927, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1516, "step": 6343 }, { "epoch": 9.497005988023952, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1441, "step": 6344 }, { "epoch": 9.498502994011975, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1554, "step": 6345 }, { "epoch": 9.5, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1529, "step": 6346 }, { "epoch": 9.501497005988025, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1433, "step": 6347 }, { "epoch": 9.502994011976048, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1505, "step": 6348 }, { "epoch": 9.504491017964071, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1457, "step": 6349 }, { "epoch": 9.505988023952096, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1464, "step": 6350 }, { "epoch": 9.50748502994012, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1466, "step": 6351 }, { "epoch": 9.508982035928144, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1536, "step": 6352 }, { "epoch": 9.510479041916168, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.153, "step": 6353 }, { "epoch": 9.511976047904191, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1536, "step": 6354 }, { "epoch": 9.513473053892216, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1449, "step": 6355 }, { "epoch": 9.51497005988024, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1444, "step": 6356 }, { "epoch": 9.516467065868264, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1502, "step": 6357 }, { "epoch": 9.517964071856287, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.152, "step": 6358 }, { "epoch": 9.519461077844312, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1474, "step": 6359 }, { "epoch": 9.520958083832335, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1471, "step": 6360 }, { "epoch": 9.52245508982036, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1476, "step": 6361 }, { "epoch": 9.523952095808383, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1496, "step": 6362 }, { "epoch": 9.525449101796408, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1407, "step": 6363 }, { "epoch": 9.52694610778443, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1426, "step": 6364 }, { "epoch": 9.528443113772456, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1491, "step": 6365 }, { "epoch": 9.529940119760479, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1514, "step": 6366 }, { "epoch": 9.531437125748504, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1476, "step": 6367 }, { "epoch": 9.532934131736527, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1595, "step": 6368 }, { "epoch": 9.534431137724551, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1415, "step": 6369 }, { "epoch": 9.535928143712574, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1529, "step": 6370 }, { "epoch": 9.5374251497006, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1457, "step": 6371 }, { "epoch": 9.538922155688622, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1502, "step": 6372 }, { "epoch": 9.540419161676647, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1424, "step": 6373 }, { "epoch": 9.54191616766467, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1481, "step": 6374 }, { "epoch": 9.543413173652695, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1487, "step": 6375 }, { "epoch": 9.544910179640718, "grad_norm": 0.19140625, "learning_rate": 0.0008, "loss": 1.1488, "step": 6376 }, { "epoch": 9.546407185628743, "grad_norm": 0.1484375, "learning_rate": 0.0008, "loss": 1.1521, "step": 6377 }, { "epoch": 9.547904191616766, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1425, "step": 6378 }, { "epoch": 9.54940119760479, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1528, "step": 6379 }, { "epoch": 9.550898203592814, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1486, "step": 6380 }, { "epoch": 9.552395209580839, "grad_norm": 0.150390625, "learning_rate": 0.0008, "loss": 1.1534, "step": 6381 }, { "epoch": 9.553892215568862, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1472, "step": 6382 }, { "epoch": 9.555389221556887, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1469, "step": 6383 }, { "epoch": 9.55688622754491, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.141, "step": 6384 }, { "epoch": 9.558383233532934, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1543, "step": 6385 }, { "epoch": 9.559880239520957, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1479, "step": 6386 }, { "epoch": 9.561377245508982, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1446, "step": 6387 }, { "epoch": 9.562874251497005, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.151, "step": 6388 }, { "epoch": 9.56437125748503, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.149, "step": 6389 }, { "epoch": 9.565868263473053, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1507, "step": 6390 }, { "epoch": 9.567365269461078, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1458, "step": 6391 }, { "epoch": 9.568862275449101, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.139, "step": 6392 }, { "epoch": 9.570359281437126, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1511, "step": 6393 }, { "epoch": 9.571856287425149, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1516, "step": 6394 }, { "epoch": 9.573353293413174, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1533, "step": 6395 }, { "epoch": 9.574850299401197, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1472, "step": 6396 }, { "epoch": 9.576347305389222, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1505, "step": 6397 }, { "epoch": 9.577844311377245, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1492, "step": 6398 }, { "epoch": 9.57934131736527, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.144, "step": 6399 }, { "epoch": 9.580838323353294, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1504, "step": 6400 }, { "epoch": 9.582335329341317, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1561, "step": 6401 }, { "epoch": 9.58383233532934, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1478, "step": 6402 }, { "epoch": 9.585329341317365, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.149, "step": 6403 }, { "epoch": 9.58682634730539, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1508, "step": 6404 }, { "epoch": 9.588323353293413, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1493, "step": 6405 }, { "epoch": 9.589820359281438, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1529, "step": 6406 }, { "epoch": 9.591317365269461, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1547, "step": 6407 }, { "epoch": 9.592814371257486, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1502, "step": 6408 }, { "epoch": 9.594311377245509, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1451, "step": 6409 }, { "epoch": 9.595808383233534, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1503, "step": 6410 }, { "epoch": 9.597305389221557, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1466, "step": 6411 }, { "epoch": 9.598802395209582, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1479, "step": 6412 }, { "epoch": 9.600299401197605, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1509, "step": 6413 }, { "epoch": 9.60179640718563, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1454, "step": 6414 }, { "epoch": 9.603293413173652, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1494, "step": 6415 }, { "epoch": 9.604790419161677, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.151, "step": 6416 }, { "epoch": 9.6062874251497, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1486, "step": 6417 }, { "epoch": 9.607784431137725, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1536, "step": 6418 }, { "epoch": 9.609281437125748, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1529, "step": 6419 }, { "epoch": 9.610778443113773, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1573, "step": 6420 }, { "epoch": 9.612275449101796, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1429, "step": 6421 }, { "epoch": 9.613772455089821, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1494, "step": 6422 }, { "epoch": 9.615269461077844, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.148, "step": 6423 }, { "epoch": 9.616766467065869, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.147, "step": 6424 }, { "epoch": 9.618263473053892, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1444, "step": 6425 }, { "epoch": 9.619760479041917, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1495, "step": 6426 }, { "epoch": 9.62125748502994, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1527, "step": 6427 }, { "epoch": 9.622754491017965, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.15, "step": 6428 }, { "epoch": 9.624251497005988, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1491, "step": 6429 }, { "epoch": 9.625748502994012, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.142, "step": 6430 }, { "epoch": 9.627245508982035, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1427, "step": 6431 }, { "epoch": 9.62874251497006, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1438, "step": 6432 }, { "epoch": 9.630239520958083, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1528, "step": 6433 }, { "epoch": 9.631736526946108, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.15, "step": 6434 }, { "epoch": 9.633233532934131, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1508, "step": 6435 }, { "epoch": 9.634730538922156, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1469, "step": 6436 }, { "epoch": 9.636227544910179, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1424, "step": 6437 }, { "epoch": 9.637724550898204, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1496, "step": 6438 }, { "epoch": 9.639221556886227, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1514, "step": 6439 }, { "epoch": 9.640718562874252, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1447, "step": 6440 }, { "epoch": 9.642215568862275, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1506, "step": 6441 }, { "epoch": 9.6437125748503, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1533, "step": 6442 }, { "epoch": 9.645209580838323, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1448, "step": 6443 }, { "epoch": 9.646706586826348, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1524, "step": 6444 }, { "epoch": 9.64820359281437, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1497, "step": 6445 }, { "epoch": 9.649700598802395, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1524, "step": 6446 }, { "epoch": 9.651197604790418, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1475, "step": 6447 }, { "epoch": 9.652694610778443, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1482, "step": 6448 }, { "epoch": 9.654191616766466, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1499, "step": 6449 }, { "epoch": 9.655688622754491, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1554, "step": 6450 }, { "epoch": 9.657185628742514, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1495, "step": 6451 }, { "epoch": 9.658682634730539, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1481, "step": 6452 }, { "epoch": 9.660179640718562, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1449, "step": 6453 }, { "epoch": 9.661676646706587, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.143, "step": 6454 }, { "epoch": 9.66317365269461, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1506, "step": 6455 }, { "epoch": 9.664670658682635, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1495, "step": 6456 }, { "epoch": 9.66616766467066, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1534, "step": 6457 }, { "epoch": 9.667664670658683, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1417, "step": 6458 }, { "epoch": 9.669161676646706, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.147, "step": 6459 }, { "epoch": 9.67065868263473, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1502, "step": 6460 }, { "epoch": 9.672155688622755, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1504, "step": 6461 }, { "epoch": 9.673652694610778, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.145, "step": 6462 }, { "epoch": 9.675149700598803, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.148, "step": 6463 }, { "epoch": 9.676646706586826, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.154, "step": 6464 }, { "epoch": 9.678143712574851, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.15, "step": 6465 }, { "epoch": 9.679640718562874, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1403, "step": 6466 }, { "epoch": 9.681137724550899, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1453, "step": 6467 }, { "epoch": 9.682634730538922, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1516, "step": 6468 }, { "epoch": 9.684131736526947, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1468, "step": 6469 }, { "epoch": 9.68562874251497, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1531, "step": 6470 }, { "epoch": 9.687125748502995, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1536, "step": 6471 }, { "epoch": 9.688622754491018, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1474, "step": 6472 }, { "epoch": 9.690119760479043, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1506, "step": 6473 }, { "epoch": 9.691616766467066, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1452, "step": 6474 }, { "epoch": 9.69311377245509, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1529, "step": 6475 }, { "epoch": 9.694610778443113, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1455, "step": 6476 }, { "epoch": 9.696107784431138, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1497, "step": 6477 }, { "epoch": 9.697604790419161, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1486, "step": 6478 }, { "epoch": 9.699101796407186, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1455, "step": 6479 }, { "epoch": 9.70059880239521, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1456, "step": 6480 }, { "epoch": 9.702095808383234, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1513, "step": 6481 }, { "epoch": 9.703592814371257, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1539, "step": 6482 }, { "epoch": 9.705089820359282, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1419, "step": 6483 }, { "epoch": 9.706586826347305, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1447, "step": 6484 }, { "epoch": 9.70808383233533, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1402, "step": 6485 }, { "epoch": 9.709580838323353, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1497, "step": 6486 }, { "epoch": 9.711077844311378, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1444, "step": 6487 }, { "epoch": 9.7125748502994, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.152, "step": 6488 }, { "epoch": 9.714071856287426, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1462, "step": 6489 }, { "epoch": 9.715568862275449, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1524, "step": 6490 }, { "epoch": 9.717065868263473, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1506, "step": 6491 }, { "epoch": 9.718562874251496, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1446, "step": 6492 }, { "epoch": 9.720059880239521, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.152, "step": 6493 }, { "epoch": 9.721556886227544, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1498, "step": 6494 }, { "epoch": 9.72305389221557, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1547, "step": 6495 }, { "epoch": 9.724550898203592, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1498, "step": 6496 }, { "epoch": 9.726047904191617, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.15, "step": 6497 }, { "epoch": 9.72754491017964, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1546, "step": 6498 }, { "epoch": 9.729041916167665, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1518, "step": 6499 }, { "epoch": 9.730538922155688, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1491, "step": 6500 }, { "epoch": 9.732035928143713, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.152, "step": 6501 }, { "epoch": 9.733532934131736, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1445, "step": 6502 }, { "epoch": 9.73502994011976, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1529, "step": 6503 }, { "epoch": 9.736526946107784, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1463, "step": 6504 }, { "epoch": 9.738023952095809, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1503, "step": 6505 }, { "epoch": 9.739520958083832, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1482, "step": 6506 }, { "epoch": 9.741017964071856, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1531, "step": 6507 }, { "epoch": 9.74251497005988, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1524, "step": 6508 }, { "epoch": 9.744011976047904, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1553, "step": 6509 }, { "epoch": 9.745508982035929, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1439, "step": 6510 }, { "epoch": 9.747005988023952, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1412, "step": 6511 }, { "epoch": 9.748502994011975, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1495, "step": 6512 }, { "epoch": 9.75, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1472, "step": 6513 }, { "epoch": 9.751497005988025, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1488, "step": 6514 }, { "epoch": 9.752994011976048, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1455, "step": 6515 }, { "epoch": 9.754491017964071, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1451, "step": 6516 }, { "epoch": 9.755988023952096, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1501, "step": 6517 }, { "epoch": 9.75748502994012, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1465, "step": 6518 }, { "epoch": 9.758982035928144, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1444, "step": 6519 }, { "epoch": 9.760479041916168, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1437, "step": 6520 }, { "epoch": 9.761976047904191, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1484, "step": 6521 }, { "epoch": 9.763473053892216, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.146, "step": 6522 }, { "epoch": 9.76497005988024, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1453, "step": 6523 }, { "epoch": 9.766467065868264, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1504, "step": 6524 }, { "epoch": 9.767964071856287, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1496, "step": 6525 }, { "epoch": 9.769461077844312, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1426, "step": 6526 }, { "epoch": 9.770958083832335, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1439, "step": 6527 }, { "epoch": 9.77245508982036, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1399, "step": 6528 }, { "epoch": 9.773952095808383, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1392, "step": 6529 }, { "epoch": 9.775449101796408, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1467, "step": 6530 }, { "epoch": 9.77694610778443, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1453, "step": 6531 }, { "epoch": 9.778443113772456, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1436, "step": 6532 }, { "epoch": 9.779940119760479, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1455, "step": 6533 }, { "epoch": 9.781437125748504, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1434, "step": 6534 }, { "epoch": 9.782934131736527, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1525, "step": 6535 }, { "epoch": 9.784431137724551, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1519, "step": 6536 }, { "epoch": 9.785928143712574, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1483, "step": 6537 }, { "epoch": 9.7874251497006, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1547, "step": 6538 }, { "epoch": 9.788922155688622, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1502, "step": 6539 }, { "epoch": 9.790419161676647, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1443, "step": 6540 }, { "epoch": 9.79191616766467, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1515, "step": 6541 }, { "epoch": 9.793413173652695, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1434, "step": 6542 }, { "epoch": 9.794910179640718, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1519, "step": 6543 }, { "epoch": 9.796407185628743, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1464, "step": 6544 }, { "epoch": 9.797904191616766, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1543, "step": 6545 }, { "epoch": 9.79940119760479, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1457, "step": 6546 }, { "epoch": 9.800898203592814, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1472, "step": 6547 }, { "epoch": 9.802395209580839, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1461, "step": 6548 }, { "epoch": 9.803892215568862, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1492, "step": 6549 }, { "epoch": 9.805389221556887, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1488, "step": 6550 }, { "epoch": 9.80688622754491, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1441, "step": 6551 }, { "epoch": 9.808383233532934, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1475, "step": 6552 }, { "epoch": 9.809880239520957, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1448, "step": 6553 }, { "epoch": 9.811377245508982, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1492, "step": 6554 }, { "epoch": 9.812874251497005, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1432, "step": 6555 }, { "epoch": 9.81437125748503, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1392, "step": 6556 }, { "epoch": 9.815868263473053, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1544, "step": 6557 }, { "epoch": 9.817365269461078, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.145, "step": 6558 }, { "epoch": 9.818862275449101, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1411, "step": 6559 }, { "epoch": 9.820359281437126, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1509, "step": 6560 }, { "epoch": 9.821856287425149, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1569, "step": 6561 }, { "epoch": 9.823353293413174, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1481, "step": 6562 }, { "epoch": 9.824850299401197, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.149, "step": 6563 }, { "epoch": 9.826347305389222, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1517, "step": 6564 }, { "epoch": 9.827844311377245, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1402, "step": 6565 }, { "epoch": 9.82934131736527, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1462, "step": 6566 }, { "epoch": 9.830838323353294, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1443, "step": 6567 }, { "epoch": 9.832335329341317, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1482, "step": 6568 }, { "epoch": 9.83383233532934, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1503, "step": 6569 }, { "epoch": 9.835329341317365, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1428, "step": 6570 }, { "epoch": 9.83682634730539, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1446, "step": 6571 }, { "epoch": 9.838323353293413, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1527, "step": 6572 }, { "epoch": 9.839820359281438, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1519, "step": 6573 }, { "epoch": 9.841317365269461, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1401, "step": 6574 }, { "epoch": 9.842814371257486, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.146, "step": 6575 }, { "epoch": 9.844311377245509, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1456, "step": 6576 }, { "epoch": 9.845808383233534, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1503, "step": 6577 }, { "epoch": 9.847305389221557, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1461, "step": 6578 }, { "epoch": 9.848802395209582, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.156, "step": 6579 }, { "epoch": 9.850299401197605, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1491, "step": 6580 }, { "epoch": 9.85179640718563, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.153, "step": 6581 }, { "epoch": 9.853293413173652, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1416, "step": 6582 }, { "epoch": 9.854790419161677, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.147, "step": 6583 }, { "epoch": 9.8562874251497, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1441, "step": 6584 }, { "epoch": 9.857784431137725, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1558, "step": 6585 }, { "epoch": 9.859281437125748, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1487, "step": 6586 }, { "epoch": 9.860778443113773, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1418, "step": 6587 }, { "epoch": 9.862275449101796, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1469, "step": 6588 }, { "epoch": 9.863772455089821, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1525, "step": 6589 }, { "epoch": 9.865269461077844, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1485, "step": 6590 }, { "epoch": 9.866766467065869, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1449, "step": 6591 }, { "epoch": 9.868263473053892, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.146, "step": 6592 }, { "epoch": 9.869760479041917, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1475, "step": 6593 }, { "epoch": 9.87125748502994, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1427, "step": 6594 }, { "epoch": 9.872754491017965, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1513, "step": 6595 }, { "epoch": 9.874251497005988, "grad_norm": 0.16796875, "learning_rate": 0.0008, "loss": 1.1521, "step": 6596 }, { "epoch": 9.875748502994012, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1469, "step": 6597 }, { "epoch": 9.877245508982035, "grad_norm": 0.1708984375, "learning_rate": 0.0008, "loss": 1.1529, "step": 6598 }, { "epoch": 9.87874251497006, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1485, "step": 6599 }, { "epoch": 9.880239520958083, "grad_norm": 0.1767578125, "learning_rate": 0.0008, "loss": 1.1482, "step": 6600 }, { "epoch": 9.881736526946108, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1455, "step": 6601 }, { "epoch": 9.883233532934131, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1468, "step": 6602 }, { "epoch": 9.884730538922156, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1513, "step": 6603 }, { "epoch": 9.886227544910179, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1471, "step": 6604 }, { "epoch": 9.887724550898204, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1467, "step": 6605 }, { "epoch": 9.889221556886227, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1416, "step": 6606 }, { "epoch": 9.890718562874252, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.1488, "step": 6607 }, { "epoch": 9.892215568862275, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1528, "step": 6608 }, { "epoch": 9.8937125748503, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1397, "step": 6609 }, { "epoch": 9.895209580838323, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1397, "step": 6610 }, { "epoch": 9.896706586826348, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1445, "step": 6611 }, { "epoch": 9.89820359281437, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1435, "step": 6612 }, { "epoch": 9.899700598802395, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1458, "step": 6613 }, { "epoch": 9.901197604790418, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1478, "step": 6614 }, { "epoch": 9.902694610778443, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1496, "step": 6615 }, { "epoch": 9.904191616766466, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1472, "step": 6616 }, { "epoch": 9.905688622754491, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1438, "step": 6617 }, { "epoch": 9.907185628742514, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1436, "step": 6618 }, { "epoch": 9.908682634730539, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1451, "step": 6619 }, { "epoch": 9.910179640718562, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1501, "step": 6620 }, { "epoch": 9.911676646706587, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.149, "step": 6621 }, { "epoch": 9.91317365269461, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1426, "step": 6622 }, { "epoch": 9.914670658682635, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1461, "step": 6623 }, { "epoch": 9.91616766467066, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1422, "step": 6624 }, { "epoch": 9.917664670658683, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1383, "step": 6625 }, { "epoch": 9.919161676646706, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1507, "step": 6626 }, { "epoch": 9.92065868263473, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.149, "step": 6627 }, { "epoch": 9.922155688622755, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1458, "step": 6628 }, { "epoch": 9.923652694610778, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1502, "step": 6629 }, { "epoch": 9.925149700598803, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1522, "step": 6630 }, { "epoch": 9.926646706586826, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1494, "step": 6631 }, { "epoch": 9.928143712574851, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1499, "step": 6632 }, { "epoch": 9.929640718562874, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1543, "step": 6633 }, { "epoch": 9.931137724550899, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1469, "step": 6634 }, { "epoch": 9.932634730538922, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1521, "step": 6635 }, { "epoch": 9.934131736526947, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1494, "step": 6636 }, { "epoch": 9.93562874251497, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1475, "step": 6637 }, { "epoch": 9.937125748502995, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1461, "step": 6638 }, { "epoch": 9.938622754491018, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.139, "step": 6639 }, { "epoch": 9.940119760479043, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1503, "step": 6640 }, { "epoch": 9.941616766467066, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1446, "step": 6641 }, { "epoch": 9.94311377245509, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1512, "step": 6642 }, { "epoch": 9.944610778443113, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1475, "step": 6643 }, { "epoch": 9.946107784431138, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1455, "step": 6644 }, { "epoch": 9.947604790419161, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.149, "step": 6645 }, { "epoch": 9.949101796407186, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.1491, "step": 6646 }, { "epoch": 9.95059880239521, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1484, "step": 6647 }, { "epoch": 9.952095808383234, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1515, "step": 6648 }, { "epoch": 9.953592814371257, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1437, "step": 6649 }, { "epoch": 9.955089820359282, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1392, "step": 6650 }, { "epoch": 9.956586826347305, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1523, "step": 6651 }, { "epoch": 9.95808383233533, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1506, "step": 6652 }, { "epoch": 9.959580838323353, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1437, "step": 6653 }, { "epoch": 9.961077844311378, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1458, "step": 6654 }, { "epoch": 9.9625748502994, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1505, "step": 6655 }, { "epoch": 9.964071856287426, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1478, "step": 6656 }, { "epoch": 9.965568862275449, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1472, "step": 6657 }, { "epoch": 9.967065868263473, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1443, "step": 6658 }, { "epoch": 9.968562874251496, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1525, "step": 6659 }, { "epoch": 9.970059880239521, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1514, "step": 6660 }, { "epoch": 9.971556886227544, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1445, "step": 6661 }, { "epoch": 9.97305389221557, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1457, "step": 6662 }, { "epoch": 9.974550898203592, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1431, "step": 6663 }, { "epoch": 9.976047904191617, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1464, "step": 6664 }, { "epoch": 9.97754491017964, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1487, "step": 6665 }, { "epoch": 9.979041916167665, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1377, "step": 6666 }, { "epoch": 9.980538922155688, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1529, "step": 6667 }, { "epoch": 9.982035928143713, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1482, "step": 6668 }, { "epoch": 9.983532934131736, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1427, "step": 6669 }, { "epoch": 9.98502994011976, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1426, "step": 6670 }, { "epoch": 9.986526946107784, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1406, "step": 6671 }, { "epoch": 9.988023952095809, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1491, "step": 6672 }, { "epoch": 9.989520958083832, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1507, "step": 6673 }, { "epoch": 9.991017964071856, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1489, "step": 6674 }, { "epoch": 9.99251497005988, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1475, "step": 6675 }, { "epoch": 9.994011976047904, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1514, "step": 6676 }, { "epoch": 9.995508982035929, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1473, "step": 6677 }, { "epoch": 9.997005988023952, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1484, "step": 6678 }, { "epoch": 9.998502994011975, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.145, "step": 6679 }, { "epoch": 10.0, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1508, "step": 6680 }, { "epoch": 10.001497005988025, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1464, "step": 6681 }, { "epoch": 10.002994011976048, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1435, "step": 6682 }, { "epoch": 10.004491017964073, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1439, "step": 6683 }, { "epoch": 10.005988023952096, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1412, "step": 6684 }, { "epoch": 10.00748502994012, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1464, "step": 6685 }, { "epoch": 10.008982035928144, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1431, "step": 6686 }, { "epoch": 10.010479041916168, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1457, "step": 6687 }, { "epoch": 10.011976047904191, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1471, "step": 6688 }, { "epoch": 10.013473053892216, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.147, "step": 6689 }, { "epoch": 10.01497005988024, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1474, "step": 6690 }, { "epoch": 10.016467065868264, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1436, "step": 6691 }, { "epoch": 10.017964071856287, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1505, "step": 6692 }, { "epoch": 10.019461077844312, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1372, "step": 6693 }, { "epoch": 10.020958083832335, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1507, "step": 6694 }, { "epoch": 10.02245508982036, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1481, "step": 6695 }, { "epoch": 10.023952095808383, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1422, "step": 6696 }, { "epoch": 10.025449101796408, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1417, "step": 6697 }, { "epoch": 10.02694610778443, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1416, "step": 6698 }, { "epoch": 10.028443113772456, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1417, "step": 6699 }, { "epoch": 10.029940119760479, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1425, "step": 6700 }, { "epoch": 10.031437125748504, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1527, "step": 6701 }, { "epoch": 10.032934131736527, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1449, "step": 6702 }, { "epoch": 10.034431137724551, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1462, "step": 6703 }, { "epoch": 10.035928143712574, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1518, "step": 6704 }, { "epoch": 10.0374251497006, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1432, "step": 6705 }, { "epoch": 10.038922155688622, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1461, "step": 6706 }, { "epoch": 10.040419161676647, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1462, "step": 6707 }, { "epoch": 10.04191616766467, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1503, "step": 6708 }, { "epoch": 10.043413173652695, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1481, "step": 6709 }, { "epoch": 10.044910179640718, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.141, "step": 6710 }, { "epoch": 10.046407185628743, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1438, "step": 6711 }, { "epoch": 10.047904191616766, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1575, "step": 6712 }, { "epoch": 10.04940119760479, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1482, "step": 6713 }, { "epoch": 10.050898203592814, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.142, "step": 6714 }, { "epoch": 10.052395209580839, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1472, "step": 6715 }, { "epoch": 10.053892215568862, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1563, "step": 6716 }, { "epoch": 10.055389221556887, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1493, "step": 6717 }, { "epoch": 10.05688622754491, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1474, "step": 6718 }, { "epoch": 10.058383233532934, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1445, "step": 6719 }, { "epoch": 10.059880239520957, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1456, "step": 6720 }, { "epoch": 10.061377245508982, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1479, "step": 6721 }, { "epoch": 10.062874251497005, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1466, "step": 6722 }, { "epoch": 10.06437125748503, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1412, "step": 6723 }, { "epoch": 10.065868263473053, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1419, "step": 6724 }, { "epoch": 10.067365269461078, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1439, "step": 6725 }, { "epoch": 10.068862275449101, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1401, "step": 6726 }, { "epoch": 10.070359281437126, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1452, "step": 6727 }, { "epoch": 10.071856287425149, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1511, "step": 6728 }, { "epoch": 10.073353293413174, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1511, "step": 6729 }, { "epoch": 10.074850299401197, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.149, "step": 6730 }, { "epoch": 10.076347305389222, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1447, "step": 6731 }, { "epoch": 10.077844311377245, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1584, "step": 6732 }, { "epoch": 10.07934131736527, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1383, "step": 6733 }, { "epoch": 10.080838323353293, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1512, "step": 6734 }, { "epoch": 10.082335329341317, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.146, "step": 6735 }, { "epoch": 10.08383233532934, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1554, "step": 6736 }, { "epoch": 10.085329341317365, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1446, "step": 6737 }, { "epoch": 10.08682634730539, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1475, "step": 6738 }, { "epoch": 10.088323353293413, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1484, "step": 6739 }, { "epoch": 10.089820359281438, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1487, "step": 6740 }, { "epoch": 10.091317365269461, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1488, "step": 6741 }, { "epoch": 10.092814371257486, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1368, "step": 6742 }, { "epoch": 10.094311377245509, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1445, "step": 6743 }, { "epoch": 10.095808383233534, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1476, "step": 6744 }, { "epoch": 10.097305389221557, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1478, "step": 6745 }, { "epoch": 10.098802395209582, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1381, "step": 6746 }, { "epoch": 10.100299401197605, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1421, "step": 6747 }, { "epoch": 10.10179640718563, "grad_norm": 0.049072265625, "learning_rate": 0.0008, "loss": 1.1513, "step": 6748 }, { "epoch": 10.103293413173652, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1393, "step": 6749 }, { "epoch": 10.104790419161677, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1485, "step": 6750 }, { "epoch": 10.1062874251497, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1517, "step": 6751 }, { "epoch": 10.107784431137725, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1514, "step": 6752 }, { "epoch": 10.109281437125748, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1469, "step": 6753 }, { "epoch": 10.110778443113773, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1511, "step": 6754 }, { "epoch": 10.112275449101796, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1447, "step": 6755 }, { "epoch": 10.113772455089821, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1442, "step": 6756 }, { "epoch": 10.115269461077844, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1439, "step": 6757 }, { "epoch": 10.116766467065869, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1422, "step": 6758 }, { "epoch": 10.118263473053892, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1417, "step": 6759 }, { "epoch": 10.119760479041917, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1448, "step": 6760 }, { "epoch": 10.12125748502994, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.149, "step": 6761 }, { "epoch": 10.122754491017965, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1496, "step": 6762 }, { "epoch": 10.124251497005988, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1459, "step": 6763 }, { "epoch": 10.125748502994012, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1473, "step": 6764 }, { "epoch": 10.127245508982035, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1452, "step": 6765 }, { "epoch": 10.12874251497006, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1424, "step": 6766 }, { "epoch": 10.130239520958083, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1408, "step": 6767 }, { "epoch": 10.131736526946108, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1473, "step": 6768 }, { "epoch": 10.133233532934131, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1455, "step": 6769 }, { "epoch": 10.134730538922156, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1453, "step": 6770 }, { "epoch": 10.136227544910179, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1524, "step": 6771 }, { "epoch": 10.137724550898204, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.145, "step": 6772 }, { "epoch": 10.139221556886227, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1444, "step": 6773 }, { "epoch": 10.140718562874252, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1531, "step": 6774 }, { "epoch": 10.142215568862275, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1484, "step": 6775 }, { "epoch": 10.1437125748503, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1494, "step": 6776 }, { "epoch": 10.145209580838323, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1478, "step": 6777 }, { "epoch": 10.146706586826348, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1453, "step": 6778 }, { "epoch": 10.14820359281437, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1401, "step": 6779 }, { "epoch": 10.149700598802395, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1524, "step": 6780 }, { "epoch": 10.151197604790418, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1487, "step": 6781 }, { "epoch": 10.152694610778443, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1432, "step": 6782 }, { "epoch": 10.154191616766466, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.153, "step": 6783 }, { "epoch": 10.155688622754491, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1483, "step": 6784 }, { "epoch": 10.157185628742514, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1409, "step": 6785 }, { "epoch": 10.158682634730539, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.1497, "step": 6786 }, { "epoch": 10.160179640718562, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.1495, "step": 6787 }, { "epoch": 10.161676646706587, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.147, "step": 6788 }, { "epoch": 10.16317365269461, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1488, "step": 6789 }, { "epoch": 10.164670658682635, "grad_norm": 0.1484375, "learning_rate": 0.0008, "loss": 1.1476, "step": 6790 }, { "epoch": 10.16616766467066, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1499, "step": 6791 }, { "epoch": 10.167664670658683, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.149, "step": 6792 }, { "epoch": 10.169161676646707, "grad_norm": 0.166015625, "learning_rate": 0.0008, "loss": 1.1456, "step": 6793 }, { "epoch": 10.17065868263473, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.142, "step": 6794 }, { "epoch": 10.172155688622755, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.146, "step": 6795 }, { "epoch": 10.173652694610778, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1441, "step": 6796 }, { "epoch": 10.175149700598803, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1451, "step": 6797 }, { "epoch": 10.176646706586826, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1466, "step": 6798 }, { "epoch": 10.178143712574851, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1486, "step": 6799 }, { "epoch": 10.179640718562874, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1467, "step": 6800 }, { "epoch": 10.181137724550899, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1472, "step": 6801 }, { "epoch": 10.182634730538922, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1467, "step": 6802 }, { "epoch": 10.184131736526947, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.154, "step": 6803 }, { "epoch": 10.18562874251497, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1449, "step": 6804 }, { "epoch": 10.187125748502995, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1418, "step": 6805 }, { "epoch": 10.188622754491018, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1448, "step": 6806 }, { "epoch": 10.190119760479043, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1416, "step": 6807 }, { "epoch": 10.191616766467066, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1439, "step": 6808 }, { "epoch": 10.19311377245509, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1534, "step": 6809 }, { "epoch": 10.194610778443113, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1484, "step": 6810 }, { "epoch": 10.196107784431138, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1499, "step": 6811 }, { "epoch": 10.197604790419161, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1454, "step": 6812 }, { "epoch": 10.199101796407186, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1462, "step": 6813 }, { "epoch": 10.20059880239521, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1482, "step": 6814 }, { "epoch": 10.202095808383234, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1458, "step": 6815 }, { "epoch": 10.203592814371257, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.151, "step": 6816 }, { "epoch": 10.205089820359282, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1435, "step": 6817 }, { "epoch": 10.206586826347305, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1454, "step": 6818 }, { "epoch": 10.20808383233533, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.147, "step": 6819 }, { "epoch": 10.209580838323353, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1428, "step": 6820 }, { "epoch": 10.211077844311378, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1447, "step": 6821 }, { "epoch": 10.2125748502994, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1473, "step": 6822 }, { "epoch": 10.214071856287426, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1444, "step": 6823 }, { "epoch": 10.215568862275449, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1444, "step": 6824 }, { "epoch": 10.217065868263473, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1459, "step": 6825 }, { "epoch": 10.218562874251496, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1499, "step": 6826 }, { "epoch": 10.220059880239521, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.151, "step": 6827 }, { "epoch": 10.221556886227544, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1467, "step": 6828 }, { "epoch": 10.22305389221557, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1528, "step": 6829 }, { "epoch": 10.224550898203592, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1364, "step": 6830 }, { "epoch": 10.226047904191617, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1446, "step": 6831 }, { "epoch": 10.22754491017964, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1431, "step": 6832 }, { "epoch": 10.229041916167665, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1454, "step": 6833 }, { "epoch": 10.230538922155688, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.145, "step": 6834 }, { "epoch": 10.232035928143713, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1486, "step": 6835 }, { "epoch": 10.233532934131736, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1437, "step": 6836 }, { "epoch": 10.23502994011976, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1452, "step": 6837 }, { "epoch": 10.236526946107784, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1448, "step": 6838 }, { "epoch": 10.238023952095809, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1494, "step": 6839 }, { "epoch": 10.239520958083832, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1535, "step": 6840 }, { "epoch": 10.241017964071856, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1449, "step": 6841 }, { "epoch": 10.24251497005988, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1437, "step": 6842 }, { "epoch": 10.244011976047904, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1467, "step": 6843 }, { "epoch": 10.245508982035927, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1483, "step": 6844 }, { "epoch": 10.247005988023952, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1405, "step": 6845 }, { "epoch": 10.248502994011975, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1482, "step": 6846 }, { "epoch": 10.25, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1472, "step": 6847 }, { "epoch": 10.251497005988025, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1412, "step": 6848 }, { "epoch": 10.252994011976048, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.15, "step": 6849 }, { "epoch": 10.254491017964073, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1385, "step": 6850 }, { "epoch": 10.255988023952096, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1424, "step": 6851 }, { "epoch": 10.25748502994012, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.152, "step": 6852 }, { "epoch": 10.258982035928144, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1482, "step": 6853 }, { "epoch": 10.260479041916168, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1439, "step": 6854 }, { "epoch": 10.261976047904191, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1445, "step": 6855 }, { "epoch": 10.263473053892216, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1438, "step": 6856 }, { "epoch": 10.26497005988024, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.142, "step": 6857 }, { "epoch": 10.266467065868264, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1515, "step": 6858 }, { "epoch": 10.267964071856287, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1478, "step": 6859 }, { "epoch": 10.269461077844312, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1546, "step": 6860 }, { "epoch": 10.270958083832335, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.145, "step": 6861 }, { "epoch": 10.27245508982036, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1438, "step": 6862 }, { "epoch": 10.273952095808383, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1389, "step": 6863 }, { "epoch": 10.275449101796408, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1472, "step": 6864 }, { "epoch": 10.27694610778443, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1502, "step": 6865 }, { "epoch": 10.278443113772456, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.152, "step": 6866 }, { "epoch": 10.279940119760479, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1431, "step": 6867 }, { "epoch": 10.281437125748504, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1454, "step": 6868 }, { "epoch": 10.282934131736527, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1401, "step": 6869 }, { "epoch": 10.284431137724551, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1491, "step": 6870 }, { "epoch": 10.285928143712574, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1513, "step": 6871 }, { "epoch": 10.2874251497006, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.145, "step": 6872 }, { "epoch": 10.288922155688622, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.14, "step": 6873 }, { "epoch": 10.290419161676647, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1411, "step": 6874 }, { "epoch": 10.29191616766467, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1505, "step": 6875 }, { "epoch": 10.293413173652695, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1518, "step": 6876 }, { "epoch": 10.294910179640718, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1509, "step": 6877 }, { "epoch": 10.296407185628743, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.143, "step": 6878 }, { "epoch": 10.297904191616766, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1418, "step": 6879 }, { "epoch": 10.29940119760479, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1519, "step": 6880 }, { "epoch": 10.300898203592814, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1508, "step": 6881 }, { "epoch": 10.302395209580839, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1531, "step": 6882 }, { "epoch": 10.303892215568862, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1499, "step": 6883 }, { "epoch": 10.305389221556887, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1417, "step": 6884 }, { "epoch": 10.30688622754491, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1386, "step": 6885 }, { "epoch": 10.308383233532934, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1396, "step": 6886 }, { "epoch": 10.309880239520957, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1515, "step": 6887 }, { "epoch": 10.311377245508982, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1448, "step": 6888 }, { "epoch": 10.312874251497005, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1396, "step": 6889 }, { "epoch": 10.31437125748503, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1448, "step": 6890 }, { "epoch": 10.315868263473053, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1459, "step": 6891 }, { "epoch": 10.317365269461078, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.145, "step": 6892 }, { "epoch": 10.318862275449101, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1465, "step": 6893 }, { "epoch": 10.320359281437126, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1536, "step": 6894 }, { "epoch": 10.321856287425149, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1463, "step": 6895 }, { "epoch": 10.323353293413174, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1442, "step": 6896 }, { "epoch": 10.324850299401197, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1423, "step": 6897 }, { "epoch": 10.326347305389222, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1455, "step": 6898 }, { "epoch": 10.327844311377245, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1455, "step": 6899 }, { "epoch": 10.32934131736527, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1488, "step": 6900 }, { "epoch": 10.330838323353294, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1513, "step": 6901 }, { "epoch": 10.332335329341317, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1456, "step": 6902 }, { "epoch": 10.33383233532934, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.152, "step": 6903 }, { "epoch": 10.335329341317365, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.142, "step": 6904 }, { "epoch": 10.33682634730539, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1491, "step": 6905 }, { "epoch": 10.338323353293413, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1463, "step": 6906 }, { "epoch": 10.339820359281438, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1457, "step": 6907 }, { "epoch": 10.341317365269461, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1424, "step": 6908 }, { "epoch": 10.342814371257486, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1475, "step": 6909 }, { "epoch": 10.344311377245509, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1545, "step": 6910 }, { "epoch": 10.345808383233534, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1414, "step": 6911 }, { "epoch": 10.347305389221557, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1449, "step": 6912 }, { "epoch": 10.348802395209582, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1457, "step": 6913 }, { "epoch": 10.350299401197605, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1425, "step": 6914 }, { "epoch": 10.35179640718563, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1493, "step": 6915 }, { "epoch": 10.353293413173652, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1494, "step": 6916 }, { "epoch": 10.354790419161677, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1454, "step": 6917 }, { "epoch": 10.3562874251497, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1396, "step": 6918 }, { "epoch": 10.357784431137725, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1491, "step": 6919 }, { "epoch": 10.359281437125748, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1505, "step": 6920 }, { "epoch": 10.360778443113773, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1493, "step": 6921 }, { "epoch": 10.362275449101796, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1391, "step": 6922 }, { "epoch": 10.363772455089821, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1476, "step": 6923 }, { "epoch": 10.365269461077844, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1419, "step": 6924 }, { "epoch": 10.366766467065869, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1419, "step": 6925 }, { "epoch": 10.368263473053892, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.14, "step": 6926 }, { "epoch": 10.369760479041917, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1411, "step": 6927 }, { "epoch": 10.37125748502994, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.141, "step": 6928 }, { "epoch": 10.372754491017965, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1442, "step": 6929 }, { "epoch": 10.374251497005988, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1389, "step": 6930 }, { "epoch": 10.375748502994012, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1356, "step": 6931 }, { "epoch": 10.377245508982035, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1518, "step": 6932 }, { "epoch": 10.37874251497006, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.151, "step": 6933 }, { "epoch": 10.380239520958083, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1443, "step": 6934 }, { "epoch": 10.381736526946108, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1515, "step": 6935 }, { "epoch": 10.383233532934131, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1513, "step": 6936 }, { "epoch": 10.384730538922156, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.147, "step": 6937 }, { "epoch": 10.386227544910179, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1477, "step": 6938 }, { "epoch": 10.387724550898204, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.151, "step": 6939 }, { "epoch": 10.389221556886227, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1454, "step": 6940 }, { "epoch": 10.390718562874252, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1413, "step": 6941 }, { "epoch": 10.392215568862275, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1432, "step": 6942 }, { "epoch": 10.3937125748503, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1455, "step": 6943 }, { "epoch": 10.395209580838323, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1551, "step": 6944 }, { "epoch": 10.396706586826348, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1546, "step": 6945 }, { "epoch": 10.39820359281437, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1487, "step": 6946 }, { "epoch": 10.399700598802395, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1475, "step": 6947 }, { "epoch": 10.401197604790418, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1408, "step": 6948 }, { "epoch": 10.402694610778443, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1484, "step": 6949 }, { "epoch": 10.404191616766466, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1473, "step": 6950 }, { "epoch": 10.405688622754491, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1428, "step": 6951 }, { "epoch": 10.407185628742514, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1508, "step": 6952 }, { "epoch": 10.408682634730539, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1507, "step": 6953 }, { "epoch": 10.410179640718562, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.155, "step": 6954 }, { "epoch": 10.411676646706587, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1447, "step": 6955 }, { "epoch": 10.41317365269461, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1424, "step": 6956 }, { "epoch": 10.414670658682635, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.139, "step": 6957 }, { "epoch": 10.41616766467066, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1452, "step": 6958 }, { "epoch": 10.417664670658683, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1435, "step": 6959 }, { "epoch": 10.419161676646706, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1442, "step": 6960 }, { "epoch": 10.42065868263473, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.151, "step": 6961 }, { "epoch": 10.422155688622755, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1423, "step": 6962 }, { "epoch": 10.423652694610778, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1451, "step": 6963 }, { "epoch": 10.425149700598803, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1393, "step": 6964 }, { "epoch": 10.426646706586826, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1423, "step": 6965 }, { "epoch": 10.428143712574851, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1515, "step": 6966 }, { "epoch": 10.429640718562874, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1444, "step": 6967 }, { "epoch": 10.431137724550899, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1477, "step": 6968 }, { "epoch": 10.432634730538922, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1451, "step": 6969 }, { "epoch": 10.434131736526947, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1445, "step": 6970 }, { "epoch": 10.43562874251497, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1387, "step": 6971 }, { "epoch": 10.437125748502995, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1439, "step": 6972 }, { "epoch": 10.438622754491018, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1436, "step": 6973 }, { "epoch": 10.440119760479043, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1455, "step": 6974 }, { "epoch": 10.441616766467066, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1526, "step": 6975 }, { "epoch": 10.44311377245509, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1461, "step": 6976 }, { "epoch": 10.444610778443113, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1466, "step": 6977 }, { "epoch": 10.446107784431138, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1391, "step": 6978 }, { "epoch": 10.447604790419161, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1432, "step": 6979 }, { "epoch": 10.449101796407186, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1432, "step": 6980 }, { "epoch": 10.45059880239521, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.145, "step": 6981 }, { "epoch": 10.452095808383234, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1453, "step": 6982 }, { "epoch": 10.453592814371257, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1494, "step": 6983 }, { "epoch": 10.455089820359282, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1485, "step": 6984 }, { "epoch": 10.456586826347305, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1433, "step": 6985 }, { "epoch": 10.45808383233533, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.138, "step": 6986 }, { "epoch": 10.459580838323353, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1438, "step": 6987 }, { "epoch": 10.461077844311378, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1428, "step": 6988 }, { "epoch": 10.4625748502994, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1473, "step": 6989 }, { "epoch": 10.464071856287426, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1501, "step": 6990 }, { "epoch": 10.465568862275449, "grad_norm": 0.15625, "learning_rate": 0.0008, "loss": 1.1496, "step": 6991 }, { "epoch": 10.467065868263473, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1479, "step": 6992 }, { "epoch": 10.468562874251496, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1461, "step": 6993 }, { "epoch": 10.470059880239521, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1388, "step": 6994 }, { "epoch": 10.471556886227544, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1572, "step": 6995 }, { "epoch": 10.47305389221557, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1441, "step": 6996 }, { "epoch": 10.474550898203592, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1408, "step": 6997 }, { "epoch": 10.476047904191617, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1436, "step": 6998 }, { "epoch": 10.47754491017964, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.138, "step": 6999 }, { "epoch": 10.479041916167665, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1375, "step": 7000 }, { "epoch": 10.480538922155688, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1498, "step": 7001 }, { "epoch": 10.482035928143713, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.14, "step": 7002 }, { "epoch": 10.483532934131736, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1432, "step": 7003 }, { "epoch": 10.48502994011976, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1471, "step": 7004 }, { "epoch": 10.486526946107784, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1468, "step": 7005 }, { "epoch": 10.488023952095809, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1512, "step": 7006 }, { "epoch": 10.489520958083832, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1449, "step": 7007 }, { "epoch": 10.491017964071856, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1484, "step": 7008 }, { "epoch": 10.49251497005988, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.149, "step": 7009 }, { "epoch": 10.494011976047904, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.14, "step": 7010 }, { "epoch": 10.495508982035927, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1456, "step": 7011 }, { "epoch": 10.497005988023952, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1427, "step": 7012 }, { "epoch": 10.498502994011975, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1424, "step": 7013 }, { "epoch": 10.5, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1442, "step": 7014 }, { "epoch": 10.501497005988025, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.139, "step": 7015 }, { "epoch": 10.502994011976048, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.148, "step": 7016 }, { "epoch": 10.504491017964071, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1432, "step": 7017 }, { "epoch": 10.505988023952096, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1409, "step": 7018 }, { "epoch": 10.50748502994012, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1478, "step": 7019 }, { "epoch": 10.508982035928144, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1442, "step": 7020 }, { "epoch": 10.510479041916168, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1444, "step": 7021 }, { "epoch": 10.511976047904191, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1504, "step": 7022 }, { "epoch": 10.513473053892216, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1451, "step": 7023 }, { "epoch": 10.51497005988024, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1506, "step": 7024 }, { "epoch": 10.516467065868264, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1518, "step": 7025 }, { "epoch": 10.517964071856287, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1452, "step": 7026 }, { "epoch": 10.519461077844312, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1413, "step": 7027 }, { "epoch": 10.520958083832335, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1526, "step": 7028 }, { "epoch": 10.52245508982036, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1438, "step": 7029 }, { "epoch": 10.523952095808383, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1495, "step": 7030 }, { "epoch": 10.525449101796408, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1452, "step": 7031 }, { "epoch": 10.52694610778443, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1477, "step": 7032 }, { "epoch": 10.528443113772456, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1506, "step": 7033 }, { "epoch": 10.529940119760479, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1457, "step": 7034 }, { "epoch": 10.531437125748504, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.153, "step": 7035 }, { "epoch": 10.532934131736527, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1432, "step": 7036 }, { "epoch": 10.534431137724551, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1438, "step": 7037 }, { "epoch": 10.535928143712574, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1429, "step": 7038 }, { "epoch": 10.5374251497006, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1401, "step": 7039 }, { "epoch": 10.538922155688622, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1499, "step": 7040 }, { "epoch": 10.540419161676647, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1463, "step": 7041 }, { "epoch": 10.54191616766467, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1445, "step": 7042 }, { "epoch": 10.543413173652695, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1499, "step": 7043 }, { "epoch": 10.544910179640718, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1432, "step": 7044 }, { "epoch": 10.546407185628743, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1441, "step": 7045 }, { "epoch": 10.547904191616766, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1432, "step": 7046 }, { "epoch": 10.54940119760479, "grad_norm": 0.1796875, "learning_rate": 0.0008, "loss": 1.1376, "step": 7047 }, { "epoch": 10.550898203592814, "grad_norm": 0.1689453125, "learning_rate": 0.0008, "loss": 1.1445, "step": 7048 }, { "epoch": 10.552395209580839, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1392, "step": 7049 }, { "epoch": 10.553892215568862, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.141, "step": 7050 }, { "epoch": 10.555389221556887, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.1396, "step": 7051 }, { "epoch": 10.55688622754491, "grad_norm": 0.26953125, "learning_rate": 0.0008, "loss": 1.1431, "step": 7052 }, { "epoch": 10.558383233532934, "grad_norm": 0.28125, "learning_rate": 0.0008, "loss": 1.1436, "step": 7053 }, { "epoch": 10.559880239520957, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1453, "step": 7054 }, { "epoch": 10.561377245508982, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1498, "step": 7055 }, { "epoch": 10.562874251497005, "grad_norm": 0.2109375, "learning_rate": 0.0008, "loss": 1.1433, "step": 7056 }, { "epoch": 10.56437125748503, "grad_norm": 0.240234375, "learning_rate": 0.0008, "loss": 1.1476, "step": 7057 }, { "epoch": 10.565868263473053, "grad_norm": 0.1845703125, "learning_rate": 0.0008, "loss": 1.1449, "step": 7058 }, { "epoch": 10.567365269461078, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1484, "step": 7059 }, { "epoch": 10.568862275449101, "grad_norm": 0.2490234375, "learning_rate": 0.0008, "loss": 1.15, "step": 7060 }, { "epoch": 10.570359281437126, "grad_norm": 0.22265625, "learning_rate": 0.0008, "loss": 1.1556, "step": 7061 }, { "epoch": 10.571856287425149, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1548, "step": 7062 }, { "epoch": 10.573353293413174, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1445, "step": 7063 }, { "epoch": 10.574850299401197, "grad_norm": 0.1875, "learning_rate": 0.0008, "loss": 1.1503, "step": 7064 }, { "epoch": 10.576347305389222, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1491, "step": 7065 }, { "epoch": 10.577844311377245, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1516, "step": 7066 }, { "epoch": 10.57934131736527, "grad_norm": 0.1962890625, "learning_rate": 0.0008, "loss": 1.1539, "step": 7067 }, { "epoch": 10.580838323353294, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.1531, "step": 7068 }, { "epoch": 10.582335329341317, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1469, "step": 7069 }, { "epoch": 10.58383233532934, "grad_norm": 0.166015625, "learning_rate": 0.0008, "loss": 1.1496, "step": 7070 }, { "epoch": 10.585329341317365, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1434, "step": 7071 }, { "epoch": 10.58682634730539, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1503, "step": 7072 }, { "epoch": 10.588323353293413, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1433, "step": 7073 }, { "epoch": 10.589820359281438, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1445, "step": 7074 }, { "epoch": 10.591317365269461, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1451, "step": 7075 }, { "epoch": 10.592814371257486, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.15, "step": 7076 }, { "epoch": 10.594311377245509, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1413, "step": 7077 }, { "epoch": 10.595808383233534, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.147, "step": 7078 }, { "epoch": 10.597305389221557, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1495, "step": 7079 }, { "epoch": 10.598802395209582, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1446, "step": 7080 }, { "epoch": 10.600299401197605, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1443, "step": 7081 }, { "epoch": 10.60179640718563, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.14, "step": 7082 }, { "epoch": 10.603293413173652, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.141, "step": 7083 }, { "epoch": 10.604790419161677, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1508, "step": 7084 }, { "epoch": 10.6062874251497, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1384, "step": 7085 }, { "epoch": 10.607784431137725, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1486, "step": 7086 }, { "epoch": 10.609281437125748, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1471, "step": 7087 }, { "epoch": 10.610778443113773, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1474, "step": 7088 }, { "epoch": 10.612275449101796, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1423, "step": 7089 }, { "epoch": 10.613772455089821, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1437, "step": 7090 }, { "epoch": 10.615269461077844, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1482, "step": 7091 }, { "epoch": 10.616766467065869, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1536, "step": 7092 }, { "epoch": 10.618263473053892, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1414, "step": 7093 }, { "epoch": 10.619760479041917, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1436, "step": 7094 }, { "epoch": 10.62125748502994, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.143, "step": 7095 }, { "epoch": 10.622754491017965, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1537, "step": 7096 }, { "epoch": 10.624251497005988, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1526, "step": 7097 }, { "epoch": 10.625748502994012, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1506, "step": 7098 }, { "epoch": 10.627245508982035, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1474, "step": 7099 }, { "epoch": 10.62874251497006, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1501, "step": 7100 }, { "epoch": 10.630239520958083, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1412, "step": 7101 }, { "epoch": 10.631736526946108, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1432, "step": 7102 }, { "epoch": 10.633233532934131, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1466, "step": 7103 }, { "epoch": 10.634730538922156, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1468, "step": 7104 }, { "epoch": 10.636227544910179, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1445, "step": 7105 }, { "epoch": 10.637724550898204, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1403, "step": 7106 }, { "epoch": 10.639221556886227, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1471, "step": 7107 }, { "epoch": 10.640718562874252, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1463, "step": 7108 }, { "epoch": 10.642215568862275, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1492, "step": 7109 }, { "epoch": 10.6437125748503, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1489, "step": 7110 }, { "epoch": 10.645209580838323, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1396, "step": 7111 }, { "epoch": 10.646706586826348, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.143, "step": 7112 }, { "epoch": 10.64820359281437, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1478, "step": 7113 }, { "epoch": 10.649700598802395, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1451, "step": 7114 }, { "epoch": 10.651197604790418, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1484, "step": 7115 }, { "epoch": 10.652694610778443, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1462, "step": 7116 }, { "epoch": 10.654191616766466, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1533, "step": 7117 }, { "epoch": 10.655688622754491, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1477, "step": 7118 }, { "epoch": 10.657185628742514, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1478, "step": 7119 }, { "epoch": 10.658682634730539, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1491, "step": 7120 }, { "epoch": 10.660179640718562, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1427, "step": 7121 }, { "epoch": 10.661676646706587, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1535, "step": 7122 }, { "epoch": 10.66317365269461, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.148, "step": 7123 }, { "epoch": 10.664670658682635, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1388, "step": 7124 }, { "epoch": 10.66616766467066, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1454, "step": 7125 }, { "epoch": 10.667664670658683, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1458, "step": 7126 }, { "epoch": 10.669161676646706, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1493, "step": 7127 }, { "epoch": 10.67065868263473, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1451, "step": 7128 }, { "epoch": 10.672155688622755, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1357, "step": 7129 }, { "epoch": 10.673652694610778, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1449, "step": 7130 }, { "epoch": 10.675149700598803, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1415, "step": 7131 }, { "epoch": 10.676646706586826, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1428, "step": 7132 }, { "epoch": 10.678143712574851, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.142, "step": 7133 }, { "epoch": 10.679640718562874, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1471, "step": 7134 }, { "epoch": 10.681137724550899, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1405, "step": 7135 }, { "epoch": 10.682634730538922, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1475, "step": 7136 }, { "epoch": 10.684131736526947, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1434, "step": 7137 }, { "epoch": 10.68562874251497, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1422, "step": 7138 }, { "epoch": 10.687125748502995, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1503, "step": 7139 }, { "epoch": 10.688622754491018, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1455, "step": 7140 }, { "epoch": 10.690119760479043, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1437, "step": 7141 }, { "epoch": 10.691616766467066, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1426, "step": 7142 }, { "epoch": 10.69311377245509, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.149, "step": 7143 }, { "epoch": 10.694610778443113, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1402, "step": 7144 }, { "epoch": 10.696107784431138, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1478, "step": 7145 }, { "epoch": 10.697604790419161, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1449, "step": 7146 }, { "epoch": 10.699101796407186, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1387, "step": 7147 }, { "epoch": 10.70059880239521, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1379, "step": 7148 }, { "epoch": 10.702095808383234, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1451, "step": 7149 }, { "epoch": 10.703592814371257, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1433, "step": 7150 }, { "epoch": 10.705089820359282, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1406, "step": 7151 }, { "epoch": 10.706586826347305, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.145, "step": 7152 }, { "epoch": 10.70808383233533, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1463, "step": 7153 }, { "epoch": 10.709580838323353, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1442, "step": 7154 }, { "epoch": 10.711077844311378, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1458, "step": 7155 }, { "epoch": 10.7125748502994, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1431, "step": 7156 }, { "epoch": 10.714071856287426, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1446, "step": 7157 }, { "epoch": 10.715568862275449, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1451, "step": 7158 }, { "epoch": 10.717065868263473, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1498, "step": 7159 }, { "epoch": 10.718562874251496, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1448, "step": 7160 }, { "epoch": 10.720059880239521, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1415, "step": 7161 }, { "epoch": 10.721556886227544, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1472, "step": 7162 }, { "epoch": 10.72305389221557, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1448, "step": 7163 }, { "epoch": 10.724550898203592, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1464, "step": 7164 }, { "epoch": 10.726047904191617, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1449, "step": 7165 }, { "epoch": 10.72754491017964, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1476, "step": 7166 }, { "epoch": 10.729041916167665, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1434, "step": 7167 }, { "epoch": 10.730538922155688, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1437, "step": 7168 }, { "epoch": 10.732035928143713, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1471, "step": 7169 }, { "epoch": 10.733532934131736, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1467, "step": 7170 }, { "epoch": 10.73502994011976, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1498, "step": 7171 }, { "epoch": 10.736526946107784, "grad_norm": 0.1904296875, "learning_rate": 0.0008, "loss": 1.1464, "step": 7172 }, { "epoch": 10.738023952095809, "grad_norm": 0.166015625, "learning_rate": 0.0008, "loss": 1.147, "step": 7173 }, { "epoch": 10.739520958083832, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1513, "step": 7174 }, { "epoch": 10.741017964071856, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1428, "step": 7175 }, { "epoch": 10.74251497005988, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1466, "step": 7176 }, { "epoch": 10.744011976047904, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1447, "step": 7177 }, { "epoch": 10.745508982035929, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1501, "step": 7178 }, { "epoch": 10.747005988023952, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.146, "step": 7179 }, { "epoch": 10.748502994011975, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.145, "step": 7180 }, { "epoch": 10.75, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.142, "step": 7181 }, { "epoch": 10.751497005988025, "grad_norm": 0.166015625, "learning_rate": 0.0008, "loss": 1.1461, "step": 7182 }, { "epoch": 10.752994011976048, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1452, "step": 7183 }, { "epoch": 10.754491017964071, "grad_norm": 0.1904296875, "learning_rate": 0.0008, "loss": 1.1468, "step": 7184 }, { "epoch": 10.755988023952096, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1471, "step": 7185 }, { "epoch": 10.75748502994012, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1472, "step": 7186 }, { "epoch": 10.758982035928144, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1452, "step": 7187 }, { "epoch": 10.760479041916168, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1444, "step": 7188 }, { "epoch": 10.761976047904191, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1504, "step": 7189 }, { "epoch": 10.763473053892216, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1413, "step": 7190 }, { "epoch": 10.76497005988024, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.146, "step": 7191 }, { "epoch": 10.766467065868264, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1426, "step": 7192 }, { "epoch": 10.767964071856287, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.151, "step": 7193 }, { "epoch": 10.769461077844312, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1447, "step": 7194 }, { "epoch": 10.770958083832335, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1495, "step": 7195 }, { "epoch": 10.77245508982036, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1507, "step": 7196 }, { "epoch": 10.773952095808383, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1462, "step": 7197 }, { "epoch": 10.775449101796408, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1417, "step": 7198 }, { "epoch": 10.77694610778443, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1485, "step": 7199 }, { "epoch": 10.778443113772456, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.143, "step": 7200 }, { "epoch": 10.779940119760479, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1451, "step": 7201 }, { "epoch": 10.781437125748504, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1478, "step": 7202 }, { "epoch": 10.782934131736527, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1457, "step": 7203 }, { "epoch": 10.784431137724551, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1466, "step": 7204 }, { "epoch": 10.785928143712574, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1401, "step": 7205 }, { "epoch": 10.7874251497006, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1473, "step": 7206 }, { "epoch": 10.788922155688622, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1401, "step": 7207 }, { "epoch": 10.790419161676647, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1475, "step": 7208 }, { "epoch": 10.79191616766467, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1512, "step": 7209 }, { "epoch": 10.793413173652695, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1432, "step": 7210 }, { "epoch": 10.794910179640718, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1426, "step": 7211 }, { "epoch": 10.796407185628743, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1373, "step": 7212 }, { "epoch": 10.797904191616766, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1454, "step": 7213 }, { "epoch": 10.79940119760479, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1442, "step": 7214 }, { "epoch": 10.800898203592814, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1408, "step": 7215 }, { "epoch": 10.802395209580839, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1462, "step": 7216 }, { "epoch": 10.803892215568862, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1468, "step": 7217 }, { "epoch": 10.805389221556887, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.144, "step": 7218 }, { "epoch": 10.80688622754491, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1487, "step": 7219 }, { "epoch": 10.808383233532934, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1416, "step": 7220 }, { "epoch": 10.809880239520957, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1472, "step": 7221 }, { "epoch": 10.811377245508982, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1464, "step": 7222 }, { "epoch": 10.812874251497005, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1451, "step": 7223 }, { "epoch": 10.81437125748503, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1536, "step": 7224 }, { "epoch": 10.815868263473053, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1377, "step": 7225 }, { "epoch": 10.817365269461078, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1475, "step": 7226 }, { "epoch": 10.818862275449101, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1486, "step": 7227 }, { "epoch": 10.820359281437126, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1464, "step": 7228 }, { "epoch": 10.821856287425149, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1416, "step": 7229 }, { "epoch": 10.823353293413174, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1468, "step": 7230 }, { "epoch": 10.824850299401197, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1483, "step": 7231 }, { "epoch": 10.826347305389222, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1466, "step": 7232 }, { "epoch": 10.827844311377245, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1457, "step": 7233 }, { "epoch": 10.82934131736527, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1467, "step": 7234 }, { "epoch": 10.830838323353294, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1442, "step": 7235 }, { "epoch": 10.832335329341317, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1488, "step": 7236 }, { "epoch": 10.83383233532934, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1459, "step": 7237 }, { "epoch": 10.835329341317365, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1437, "step": 7238 }, { "epoch": 10.83682634730539, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1471, "step": 7239 }, { "epoch": 10.838323353293413, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1518, "step": 7240 }, { "epoch": 10.839820359281438, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1419, "step": 7241 }, { "epoch": 10.841317365269461, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.152, "step": 7242 }, { "epoch": 10.842814371257486, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1436, "step": 7243 }, { "epoch": 10.844311377245509, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1462, "step": 7244 }, { "epoch": 10.845808383233534, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1421, "step": 7245 }, { "epoch": 10.847305389221557, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1456, "step": 7246 }, { "epoch": 10.848802395209582, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1403, "step": 7247 }, { "epoch": 10.850299401197605, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1393, "step": 7248 }, { "epoch": 10.85179640718563, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1455, "step": 7249 }, { "epoch": 10.853293413173652, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1451, "step": 7250 }, { "epoch": 10.854790419161677, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1445, "step": 7251 }, { "epoch": 10.8562874251497, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1451, "step": 7252 }, { "epoch": 10.857784431137725, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1415, "step": 7253 }, { "epoch": 10.859281437125748, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1396, "step": 7254 }, { "epoch": 10.860778443113773, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1447, "step": 7255 }, { "epoch": 10.862275449101796, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.138, "step": 7256 }, { "epoch": 10.863772455089821, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1492, "step": 7257 }, { "epoch": 10.865269461077844, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1456, "step": 7258 }, { "epoch": 10.866766467065869, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1336, "step": 7259 }, { "epoch": 10.868263473053892, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1498, "step": 7260 }, { "epoch": 10.869760479041917, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1424, "step": 7261 }, { "epoch": 10.87125748502994, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1471, "step": 7262 }, { "epoch": 10.872754491017965, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1517, "step": 7263 }, { "epoch": 10.874251497005988, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1442, "step": 7264 }, { "epoch": 10.875748502994012, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1453, "step": 7265 }, { "epoch": 10.877245508982035, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1462, "step": 7266 }, { "epoch": 10.87874251497006, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1414, "step": 7267 }, { "epoch": 10.880239520958083, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1448, "step": 7268 }, { "epoch": 10.881736526946108, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1424, "step": 7269 }, { "epoch": 10.883233532934131, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.148, "step": 7270 }, { "epoch": 10.884730538922156, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1405, "step": 7271 }, { "epoch": 10.886227544910179, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1424, "step": 7272 }, { "epoch": 10.887724550898204, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1484, "step": 7273 }, { "epoch": 10.889221556886227, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1451, "step": 7274 }, { "epoch": 10.890718562874252, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1469, "step": 7275 }, { "epoch": 10.892215568862275, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1415, "step": 7276 }, { "epoch": 10.8937125748503, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1467, "step": 7277 }, { "epoch": 10.895209580838323, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1455, "step": 7278 }, { "epoch": 10.896706586826348, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1466, "step": 7279 }, { "epoch": 10.89820359281437, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1509, "step": 7280 }, { "epoch": 10.899700598802395, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1453, "step": 7281 }, { "epoch": 10.901197604790418, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1495, "step": 7282 }, { "epoch": 10.902694610778443, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1442, "step": 7283 }, { "epoch": 10.904191616766466, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1433, "step": 7284 }, { "epoch": 10.905688622754491, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1469, "step": 7285 }, { "epoch": 10.907185628742514, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.144, "step": 7286 }, { "epoch": 10.908682634730539, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1445, "step": 7287 }, { "epoch": 10.910179640718562, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1441, "step": 7288 }, { "epoch": 10.911676646706587, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1513, "step": 7289 }, { "epoch": 10.91317365269461, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1395, "step": 7290 }, { "epoch": 10.914670658682635, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1465, "step": 7291 }, { "epoch": 10.91616766467066, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.145, "step": 7292 }, { "epoch": 10.917664670658683, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1415, "step": 7293 }, { "epoch": 10.919161676646706, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1459, "step": 7294 }, { "epoch": 10.92065868263473, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1402, "step": 7295 }, { "epoch": 10.922155688622755, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1419, "step": 7296 }, { "epoch": 10.923652694610778, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.15, "step": 7297 }, { "epoch": 10.925149700598803, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1369, "step": 7298 }, { "epoch": 10.926646706586826, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1424, "step": 7299 }, { "epoch": 10.928143712574851, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1536, "step": 7300 }, { "epoch": 10.929640718562874, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1485, "step": 7301 }, { "epoch": 10.931137724550899, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1436, "step": 7302 }, { "epoch": 10.932634730538922, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1474, "step": 7303 }, { "epoch": 10.934131736526947, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1453, "step": 7304 }, { "epoch": 10.93562874251497, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1461, "step": 7305 }, { "epoch": 10.937125748502995, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1429, "step": 7306 }, { "epoch": 10.938622754491018, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1468, "step": 7307 }, { "epoch": 10.940119760479043, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1454, "step": 7308 }, { "epoch": 10.941616766467066, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1422, "step": 7309 }, { "epoch": 10.94311377245509, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.142, "step": 7310 }, { "epoch": 10.944610778443113, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1442, "step": 7311 }, { "epoch": 10.946107784431138, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.139, "step": 7312 }, { "epoch": 10.947604790419161, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1474, "step": 7313 }, { "epoch": 10.949101796407186, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1478, "step": 7314 }, { "epoch": 10.95059880239521, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1513, "step": 7315 }, { "epoch": 10.952095808383234, "grad_norm": 0.04931640625, "learning_rate": 0.0008, "loss": 1.1411, "step": 7316 }, { "epoch": 10.953592814371257, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1426, "step": 7317 }, { "epoch": 10.955089820359282, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1396, "step": 7318 }, { "epoch": 10.956586826347305, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1419, "step": 7319 }, { "epoch": 10.95808383233533, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1459, "step": 7320 }, { "epoch": 10.959580838323353, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1437, "step": 7321 }, { "epoch": 10.961077844311378, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1476, "step": 7322 }, { "epoch": 10.9625748502994, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.145, "step": 7323 }, { "epoch": 10.964071856287426, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1483, "step": 7324 }, { "epoch": 10.965568862275449, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.143, "step": 7325 }, { "epoch": 10.967065868263473, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1456, "step": 7326 }, { "epoch": 10.968562874251496, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1455, "step": 7327 }, { "epoch": 10.970059880239521, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1376, "step": 7328 }, { "epoch": 10.971556886227544, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.143, "step": 7329 }, { "epoch": 10.97305389221557, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1394, "step": 7330 }, { "epoch": 10.974550898203592, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.141, "step": 7331 }, { "epoch": 10.976047904191617, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1466, "step": 7332 }, { "epoch": 10.97754491017964, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1469, "step": 7333 }, { "epoch": 10.979041916167665, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1514, "step": 7334 }, { "epoch": 10.980538922155688, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1397, "step": 7335 }, { "epoch": 10.982035928143713, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1378, "step": 7336 }, { "epoch": 10.983532934131736, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.142, "step": 7337 }, { "epoch": 10.98502994011976, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1447, "step": 7338 }, { "epoch": 10.986526946107784, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1484, "step": 7339 }, { "epoch": 10.988023952095809, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1473, "step": 7340 }, { "epoch": 10.989520958083832, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1475, "step": 7341 }, { "epoch": 10.991017964071856, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1426, "step": 7342 }, { "epoch": 10.99251497005988, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1407, "step": 7343 }, { "epoch": 10.994011976047904, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1458, "step": 7344 }, { "epoch": 10.995508982035929, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1433, "step": 7345 }, { "epoch": 10.997005988023952, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1411, "step": 7346 }, { "epoch": 10.998502994011975, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1384, "step": 7347 }, { "epoch": 11.0, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1422, "step": 7348 }, { "epoch": 11.001497005988025, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1442, "step": 7349 }, { "epoch": 11.002994011976048, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1447, "step": 7350 }, { "epoch": 11.004491017964073, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1396, "step": 7351 }, { "epoch": 11.005988023952096, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1493, "step": 7352 }, { "epoch": 11.00748502994012, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1477, "step": 7353 }, { "epoch": 11.008982035928144, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1402, "step": 7354 }, { "epoch": 11.010479041916168, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1495, "step": 7355 }, { "epoch": 11.011976047904191, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1398, "step": 7356 }, { "epoch": 11.013473053892216, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1416, "step": 7357 }, { "epoch": 11.01497005988024, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.143, "step": 7358 }, { "epoch": 11.016467065868264, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.141, "step": 7359 }, { "epoch": 11.017964071856287, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1423, "step": 7360 }, { "epoch": 11.019461077844312, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1466, "step": 7361 }, { "epoch": 11.020958083832335, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1423, "step": 7362 }, { "epoch": 11.02245508982036, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1492, "step": 7363 }, { "epoch": 11.023952095808383, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.141, "step": 7364 }, { "epoch": 11.025449101796408, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1469, "step": 7365 }, { "epoch": 11.02694610778443, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1447, "step": 7366 }, { "epoch": 11.028443113772456, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1405, "step": 7367 }, { "epoch": 11.029940119760479, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1435, "step": 7368 }, { "epoch": 11.031437125748504, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1386, "step": 7369 }, { "epoch": 11.032934131736527, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1501, "step": 7370 }, { "epoch": 11.034431137724551, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1431, "step": 7371 }, { "epoch": 11.035928143712574, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1369, "step": 7372 }, { "epoch": 11.0374251497006, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1437, "step": 7373 }, { "epoch": 11.038922155688622, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1417, "step": 7374 }, { "epoch": 11.040419161676647, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1365, "step": 7375 }, { "epoch": 11.04191616766467, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1436, "step": 7376 }, { "epoch": 11.043413173652695, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1462, "step": 7377 }, { "epoch": 11.044910179640718, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1412, "step": 7378 }, { "epoch": 11.046407185628743, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1425, "step": 7379 }, { "epoch": 11.047904191616766, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.146, "step": 7380 }, { "epoch": 11.04940119760479, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1526, "step": 7381 }, { "epoch": 11.050898203592814, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1386, "step": 7382 }, { "epoch": 11.052395209580839, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1376, "step": 7383 }, { "epoch": 11.053892215568862, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1526, "step": 7384 }, { "epoch": 11.055389221556887, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1439, "step": 7385 }, { "epoch": 11.05688622754491, "grad_norm": 0.049560546875, "learning_rate": 0.0008, "loss": 1.146, "step": 7386 }, { "epoch": 11.058383233532934, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1452, "step": 7387 }, { "epoch": 11.059880239520957, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.136, "step": 7388 }, { "epoch": 11.061377245508982, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1465, "step": 7389 }, { "epoch": 11.062874251497005, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1463, "step": 7390 }, { "epoch": 11.06437125748503, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.151, "step": 7391 }, { "epoch": 11.065868263473053, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1463, "step": 7392 }, { "epoch": 11.067365269461078, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1458, "step": 7393 }, { "epoch": 11.068862275449101, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.15, "step": 7394 }, { "epoch": 11.070359281437126, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1432, "step": 7395 }, { "epoch": 11.071856287425149, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1456, "step": 7396 }, { "epoch": 11.073353293413174, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.141, "step": 7397 }, { "epoch": 11.074850299401197, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1402, "step": 7398 }, { "epoch": 11.076347305389222, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1408, "step": 7399 }, { "epoch": 11.077844311377245, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1428, "step": 7400 }, { "epoch": 11.07934131736527, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1494, "step": 7401 }, { "epoch": 11.080838323353293, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1455, "step": 7402 }, { "epoch": 11.082335329341317, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1502, "step": 7403 }, { "epoch": 11.08383233532934, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1481, "step": 7404 }, { "epoch": 11.085329341317365, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1426, "step": 7405 }, { "epoch": 11.08682634730539, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1448, "step": 7406 }, { "epoch": 11.088323353293413, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.147, "step": 7407 }, { "epoch": 11.089820359281438, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1443, "step": 7408 }, { "epoch": 11.091317365269461, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1495, "step": 7409 }, { "epoch": 11.092814371257486, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1446, "step": 7410 }, { "epoch": 11.094311377245509, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1468, "step": 7411 }, { "epoch": 11.095808383233534, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1382, "step": 7412 }, { "epoch": 11.097305389221557, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1411, "step": 7413 }, { "epoch": 11.098802395209582, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1494, "step": 7414 }, { "epoch": 11.100299401197605, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1432, "step": 7415 }, { "epoch": 11.10179640718563, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1383, "step": 7416 }, { "epoch": 11.103293413173652, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1392, "step": 7417 }, { "epoch": 11.104790419161677, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1453, "step": 7418 }, { "epoch": 11.1062874251497, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1443, "step": 7419 }, { "epoch": 11.107784431137725, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1451, "step": 7420 }, { "epoch": 11.109281437125748, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1365, "step": 7421 }, { "epoch": 11.110778443113773, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1432, "step": 7422 }, { "epoch": 11.112275449101796, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1455, "step": 7423 }, { "epoch": 11.113772455089821, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1456, "step": 7424 }, { "epoch": 11.115269461077844, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1407, "step": 7425 }, { "epoch": 11.116766467065869, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1517, "step": 7426 }, { "epoch": 11.118263473053892, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1422, "step": 7427 }, { "epoch": 11.119760479041917, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1443, "step": 7428 }, { "epoch": 11.12125748502994, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1442, "step": 7429 }, { "epoch": 11.122754491017965, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1427, "step": 7430 }, { "epoch": 11.124251497005988, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1419, "step": 7431 }, { "epoch": 11.125748502994012, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1475, "step": 7432 }, { "epoch": 11.127245508982035, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1479, "step": 7433 }, { "epoch": 11.12874251497006, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1435, "step": 7434 }, { "epoch": 11.130239520958083, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1479, "step": 7435 }, { "epoch": 11.131736526946108, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1397, "step": 7436 }, { "epoch": 11.133233532934131, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.145, "step": 7437 }, { "epoch": 11.134730538922156, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1484, "step": 7438 }, { "epoch": 11.136227544910179, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1442, "step": 7439 }, { "epoch": 11.137724550898204, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1472, "step": 7440 }, { "epoch": 11.139221556886227, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1443, "step": 7441 }, { "epoch": 11.140718562874252, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1455, "step": 7442 }, { "epoch": 11.142215568862275, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1423, "step": 7443 }, { "epoch": 11.1437125748503, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1487, "step": 7444 }, { "epoch": 11.145209580838323, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1515, "step": 7445 }, { "epoch": 11.146706586826348, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1387, "step": 7446 }, { "epoch": 11.14820359281437, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1477, "step": 7447 }, { "epoch": 11.149700598802395, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1421, "step": 7448 }, { "epoch": 11.151197604790418, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1368, "step": 7449 }, { "epoch": 11.152694610778443, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1481, "step": 7450 }, { "epoch": 11.154191616766466, "grad_norm": 0.14453125, "learning_rate": 0.0008, "loss": 1.1481, "step": 7451 }, { "epoch": 11.155688622754491, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1433, "step": 7452 }, { "epoch": 11.157185628742514, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1414, "step": 7453 }, { "epoch": 11.158682634730539, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1433, "step": 7454 }, { "epoch": 11.160179640718562, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1533, "step": 7455 }, { "epoch": 11.161676646706587, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.147, "step": 7456 }, { "epoch": 11.16317365269461, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1459, "step": 7457 }, { "epoch": 11.164670658682635, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1419, "step": 7458 }, { "epoch": 11.16616766467066, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1556, "step": 7459 }, { "epoch": 11.167664670658683, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1391, "step": 7460 }, { "epoch": 11.169161676646707, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1461, "step": 7461 }, { "epoch": 11.17065868263473, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1474, "step": 7462 }, { "epoch": 11.172155688622755, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.144, "step": 7463 }, { "epoch": 11.173652694610778, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1422, "step": 7464 }, { "epoch": 11.175149700598803, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1464, "step": 7465 }, { "epoch": 11.176646706586826, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1424, "step": 7466 }, { "epoch": 11.178143712574851, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1493, "step": 7467 }, { "epoch": 11.179640718562874, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1441, "step": 7468 }, { "epoch": 11.181137724550899, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1437, "step": 7469 }, { "epoch": 11.182634730538922, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1388, "step": 7470 }, { "epoch": 11.184131736526947, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1506, "step": 7471 }, { "epoch": 11.18562874251497, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1477, "step": 7472 }, { "epoch": 11.187125748502995, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1452, "step": 7473 }, { "epoch": 11.188622754491018, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1419, "step": 7474 }, { "epoch": 11.190119760479043, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1414, "step": 7475 }, { "epoch": 11.191616766467066, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1515, "step": 7476 }, { "epoch": 11.19311377245509, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1436, "step": 7477 }, { "epoch": 11.194610778443113, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1475, "step": 7478 }, { "epoch": 11.196107784431138, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1406, "step": 7479 }, { "epoch": 11.197604790419161, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1451, "step": 7480 }, { "epoch": 11.199101796407186, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1486, "step": 7481 }, { "epoch": 11.20059880239521, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1442, "step": 7482 }, { "epoch": 11.202095808383234, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1428, "step": 7483 }, { "epoch": 11.203592814371257, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1441, "step": 7484 }, { "epoch": 11.205089820359282, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1391, "step": 7485 }, { "epoch": 11.206586826347305, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.15, "step": 7486 }, { "epoch": 11.20808383233533, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1425, "step": 7487 }, { "epoch": 11.209580838323353, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1404, "step": 7488 }, { "epoch": 11.211077844311378, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1447, "step": 7489 }, { "epoch": 11.2125748502994, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1471, "step": 7490 }, { "epoch": 11.214071856287426, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1443, "step": 7491 }, { "epoch": 11.215568862275449, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1417, "step": 7492 }, { "epoch": 11.217065868263473, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1419, "step": 7493 }, { "epoch": 11.218562874251496, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1367, "step": 7494 }, { "epoch": 11.220059880239521, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1523, "step": 7495 }, { "epoch": 11.221556886227544, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1427, "step": 7496 }, { "epoch": 11.22305389221557, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1453, "step": 7497 }, { "epoch": 11.224550898203592, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1469, "step": 7498 }, { "epoch": 11.226047904191617, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1459, "step": 7499 }, { "epoch": 11.22754491017964, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.146, "step": 7500 }, { "epoch": 11.229041916167665, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1409, "step": 7501 }, { "epoch": 11.230538922155688, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1408, "step": 7502 }, { "epoch": 11.232035928143713, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.141, "step": 7503 }, { "epoch": 11.233532934131736, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1366, "step": 7504 }, { "epoch": 11.23502994011976, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.145, "step": 7505 }, { "epoch": 11.236526946107784, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1445, "step": 7506 }, { "epoch": 11.238023952095809, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.145, "step": 7507 }, { "epoch": 11.239520958083832, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1541, "step": 7508 }, { "epoch": 11.241017964071856, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1386, "step": 7509 }, { "epoch": 11.24251497005988, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1449, "step": 7510 }, { "epoch": 11.244011976047904, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1494, "step": 7511 }, { "epoch": 11.245508982035927, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1387, "step": 7512 }, { "epoch": 11.247005988023952, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1489, "step": 7513 }, { "epoch": 11.248502994011975, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1505, "step": 7514 }, { "epoch": 11.25, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1377, "step": 7515 }, { "epoch": 11.251497005988025, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1468, "step": 7516 }, { "epoch": 11.252994011976048, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1445, "step": 7517 }, { "epoch": 11.254491017964073, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1434, "step": 7518 }, { "epoch": 11.255988023952096, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1366, "step": 7519 }, { "epoch": 11.25748502994012, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1464, "step": 7520 }, { "epoch": 11.258982035928144, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1468, "step": 7521 }, { "epoch": 11.260479041916168, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1419, "step": 7522 }, { "epoch": 11.261976047904191, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1494, "step": 7523 }, { "epoch": 11.263473053892216, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1391, "step": 7524 }, { "epoch": 11.26497005988024, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1478, "step": 7525 }, { "epoch": 11.266467065868264, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.148, "step": 7526 }, { "epoch": 11.267964071856287, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1334, "step": 7527 }, { "epoch": 11.269461077844312, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1419, "step": 7528 }, { "epoch": 11.270958083832335, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1465, "step": 7529 }, { "epoch": 11.27245508982036, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.145, "step": 7530 }, { "epoch": 11.273952095808383, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.141, "step": 7531 }, { "epoch": 11.275449101796408, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1441, "step": 7532 }, { "epoch": 11.27694610778443, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1397, "step": 7533 }, { "epoch": 11.278443113772456, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1448, "step": 7534 }, { "epoch": 11.279940119760479, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1437, "step": 7535 }, { "epoch": 11.281437125748504, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 7536 }, { "epoch": 11.282934131736527, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1429, "step": 7537 }, { "epoch": 11.284431137724551, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.144, "step": 7538 }, { "epoch": 11.285928143712574, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1434, "step": 7539 }, { "epoch": 11.2874251497006, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1424, "step": 7540 }, { "epoch": 11.288922155688622, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1461, "step": 7541 }, { "epoch": 11.290419161676647, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1451, "step": 7542 }, { "epoch": 11.29191616766467, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1435, "step": 7543 }, { "epoch": 11.293413173652695, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1532, "step": 7544 }, { "epoch": 11.294910179640718, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1496, "step": 7545 }, { "epoch": 11.296407185628743, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1403, "step": 7546 }, { "epoch": 11.297904191616766, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1388, "step": 7547 }, { "epoch": 11.29940119760479, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1409, "step": 7548 }, { "epoch": 11.300898203592814, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1445, "step": 7549 }, { "epoch": 11.302395209580839, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1406, "step": 7550 }, { "epoch": 11.303892215568862, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1506, "step": 7551 }, { "epoch": 11.305389221556887, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1368, "step": 7552 }, { "epoch": 11.30688622754491, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1415, "step": 7553 }, { "epoch": 11.308383233532934, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1457, "step": 7554 }, { "epoch": 11.309880239520957, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1405, "step": 7555 }, { "epoch": 11.311377245508982, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1348, "step": 7556 }, { "epoch": 11.312874251497005, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1385, "step": 7557 }, { "epoch": 11.31437125748503, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.139, "step": 7558 }, { "epoch": 11.315868263473053, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1418, "step": 7559 }, { "epoch": 11.317365269461078, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1437, "step": 7560 }, { "epoch": 11.318862275449101, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1361, "step": 7561 }, { "epoch": 11.320359281437126, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1451, "step": 7562 }, { "epoch": 11.321856287425149, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1477, "step": 7563 }, { "epoch": 11.323353293413174, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1439, "step": 7564 }, { "epoch": 11.324850299401197, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1456, "step": 7565 }, { "epoch": 11.326347305389222, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1433, "step": 7566 }, { "epoch": 11.327844311377245, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1387, "step": 7567 }, { "epoch": 11.32934131736527, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1478, "step": 7568 }, { "epoch": 11.330838323353294, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1424, "step": 7569 }, { "epoch": 11.332335329341317, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1446, "step": 7570 }, { "epoch": 11.33383233532934, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1439, "step": 7571 }, { "epoch": 11.335329341317365, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1421, "step": 7572 }, { "epoch": 11.33682634730539, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1453, "step": 7573 }, { "epoch": 11.338323353293413, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1489, "step": 7574 }, { "epoch": 11.339820359281438, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1457, "step": 7575 }, { "epoch": 11.341317365269461, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.143, "step": 7576 }, { "epoch": 11.342814371257486, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1389, "step": 7577 }, { "epoch": 11.344311377245509, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1417, "step": 7578 }, { "epoch": 11.345808383233534, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1467, "step": 7579 }, { "epoch": 11.347305389221557, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1439, "step": 7580 }, { "epoch": 11.348802395209582, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1409, "step": 7581 }, { "epoch": 11.350299401197605, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1397, "step": 7582 }, { "epoch": 11.35179640718563, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.143, "step": 7583 }, { "epoch": 11.353293413173652, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1491, "step": 7584 }, { "epoch": 11.354790419161677, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1428, "step": 7585 }, { "epoch": 11.3562874251497, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1473, "step": 7586 }, { "epoch": 11.357784431137725, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1408, "step": 7587 }, { "epoch": 11.359281437125748, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1416, "step": 7588 }, { "epoch": 11.360778443113773, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.145, "step": 7589 }, { "epoch": 11.362275449101796, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1449, "step": 7590 }, { "epoch": 11.363772455089821, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1453, "step": 7591 }, { "epoch": 11.365269461077844, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1458, "step": 7592 }, { "epoch": 11.366766467065869, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1435, "step": 7593 }, { "epoch": 11.368263473053892, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1478, "step": 7594 }, { "epoch": 11.369760479041917, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1459, "step": 7595 }, { "epoch": 11.37125748502994, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1362, "step": 7596 }, { "epoch": 11.372754491017965, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1393, "step": 7597 }, { "epoch": 11.374251497005988, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1496, "step": 7598 }, { "epoch": 11.375748502994012, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1496, "step": 7599 }, { "epoch": 11.377245508982035, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1427, "step": 7600 }, { "epoch": 11.37874251497006, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1448, "step": 7601 }, { "epoch": 11.380239520958083, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1459, "step": 7602 }, { "epoch": 11.381736526946108, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1474, "step": 7603 }, { "epoch": 11.383233532934131, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1367, "step": 7604 }, { "epoch": 11.384730538922156, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1431, "step": 7605 }, { "epoch": 11.386227544910179, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1437, "step": 7606 }, { "epoch": 11.387724550898204, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1432, "step": 7607 }, { "epoch": 11.389221556886227, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1422, "step": 7608 }, { "epoch": 11.390718562874252, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1437, "step": 7609 }, { "epoch": 11.392215568862275, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1502, "step": 7610 }, { "epoch": 11.3937125748503, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1469, "step": 7611 }, { "epoch": 11.395209580838323, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1408, "step": 7612 }, { "epoch": 11.396706586826348, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1381, "step": 7613 }, { "epoch": 11.39820359281437, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1408, "step": 7614 }, { "epoch": 11.399700598802395, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1484, "step": 7615 }, { "epoch": 11.401197604790418, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1441, "step": 7616 }, { "epoch": 11.402694610778443, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1415, "step": 7617 }, { "epoch": 11.404191616766466, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1464, "step": 7618 }, { "epoch": 11.405688622754491, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1436, "step": 7619 }, { "epoch": 11.407185628742514, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1391, "step": 7620 }, { "epoch": 11.408682634730539, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1406, "step": 7621 }, { "epoch": 11.410179640718562, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1396, "step": 7622 }, { "epoch": 11.411676646706587, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1385, "step": 7623 }, { "epoch": 11.41317365269461, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1442, "step": 7624 }, { "epoch": 11.414670658682635, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1386, "step": 7625 }, { "epoch": 11.41616766467066, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1409, "step": 7626 }, { "epoch": 11.417664670658683, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1415, "step": 7627 }, { "epoch": 11.419161676646706, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1414, "step": 7628 }, { "epoch": 11.42065868263473, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1471, "step": 7629 }, { "epoch": 11.422155688622755, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1464, "step": 7630 }, { "epoch": 11.423652694610778, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1386, "step": 7631 }, { "epoch": 11.425149700598803, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1339, "step": 7632 }, { "epoch": 11.426646706586826, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1363, "step": 7633 }, { "epoch": 11.428143712574851, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1391, "step": 7634 }, { "epoch": 11.429640718562874, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1475, "step": 7635 }, { "epoch": 11.431137724550899, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1475, "step": 7636 }, { "epoch": 11.432634730538922, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1438, "step": 7637 }, { "epoch": 11.434131736526947, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1412, "step": 7638 }, { "epoch": 11.43562874251497, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1459, "step": 7639 }, { "epoch": 11.437125748502995, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1435, "step": 7640 }, { "epoch": 11.438622754491018, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1392, "step": 7641 }, { "epoch": 11.440119760479043, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1445, "step": 7642 }, { "epoch": 11.441616766467066, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1426, "step": 7643 }, { "epoch": 11.44311377245509, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1459, "step": 7644 }, { "epoch": 11.444610778443113, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1422, "step": 7645 }, { "epoch": 11.446107784431138, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1415, "step": 7646 }, { "epoch": 11.447604790419161, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.145, "step": 7647 }, { "epoch": 11.449101796407186, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1438, "step": 7648 }, { "epoch": 11.45059880239521, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1403, "step": 7649 }, { "epoch": 11.452095808383234, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1416, "step": 7650 }, { "epoch": 11.453592814371257, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1461, "step": 7651 }, { "epoch": 11.455089820359282, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1468, "step": 7652 }, { "epoch": 11.456586826347305, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1477, "step": 7653 }, { "epoch": 11.45808383233533, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1507, "step": 7654 }, { "epoch": 11.459580838323353, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1478, "step": 7655 }, { "epoch": 11.461077844311378, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1419, "step": 7656 }, { "epoch": 11.4625748502994, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1417, "step": 7657 }, { "epoch": 11.464071856287426, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1457, "step": 7658 }, { "epoch": 11.465568862275449, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1415, "step": 7659 }, { "epoch": 11.467065868263473, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1483, "step": 7660 }, { "epoch": 11.468562874251496, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1438, "step": 7661 }, { "epoch": 11.470059880239521, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.139, "step": 7662 }, { "epoch": 11.471556886227544, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1403, "step": 7663 }, { "epoch": 11.47305389221557, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1441, "step": 7664 }, { "epoch": 11.474550898203592, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1445, "step": 7665 }, { "epoch": 11.476047904191617, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1379, "step": 7666 }, { "epoch": 11.47754491017964, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1413, "step": 7667 }, { "epoch": 11.479041916167665, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1437, "step": 7668 }, { "epoch": 11.480538922155688, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1442, "step": 7669 }, { "epoch": 11.482035928143713, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1436, "step": 7670 }, { "epoch": 11.483532934131736, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1447, "step": 7671 }, { "epoch": 11.48502994011976, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1388, "step": 7672 }, { "epoch": 11.486526946107784, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1436, "step": 7673 }, { "epoch": 11.488023952095809, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1393, "step": 7674 }, { "epoch": 11.489520958083832, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.1462, "step": 7675 }, { "epoch": 11.491017964071856, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1421, "step": 7676 }, { "epoch": 11.49251497005988, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1406, "step": 7677 }, { "epoch": 11.494011976047904, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1389, "step": 7678 }, { "epoch": 11.495508982035927, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1493, "step": 7679 }, { "epoch": 11.497005988023952, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1456, "step": 7680 }, { "epoch": 11.498502994011975, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1491, "step": 7681 }, { "epoch": 11.5, "grad_norm": 0.051025390625, "learning_rate": 0.0008, "loss": 1.1402, "step": 7682 }, { "epoch": 11.501497005988025, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1405, "step": 7683 }, { "epoch": 11.502994011976048, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1403, "step": 7684 }, { "epoch": 11.504491017964071, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.141, "step": 7685 }, { "epoch": 11.505988023952096, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1388, "step": 7686 }, { "epoch": 11.50748502994012, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1427, "step": 7687 }, { "epoch": 11.508982035928144, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1435, "step": 7688 }, { "epoch": 11.510479041916168, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1387, "step": 7689 }, { "epoch": 11.511976047904191, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1471, "step": 7690 }, { "epoch": 11.513473053892216, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.146, "step": 7691 }, { "epoch": 11.51497005988024, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1477, "step": 7692 }, { "epoch": 11.516467065868264, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1435, "step": 7693 }, { "epoch": 11.517964071856287, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1417, "step": 7694 }, { "epoch": 11.519461077844312, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1428, "step": 7695 }, { "epoch": 11.520958083832335, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.138, "step": 7696 }, { "epoch": 11.52245508982036, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.135, "step": 7697 }, { "epoch": 11.523952095808383, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1437, "step": 7698 }, { "epoch": 11.525449101796408, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1498, "step": 7699 }, { "epoch": 11.52694610778443, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1491, "step": 7700 }, { "epoch": 11.528443113772456, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1364, "step": 7701 }, { "epoch": 11.529940119760479, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1454, "step": 7702 }, { "epoch": 11.531437125748504, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1389, "step": 7703 }, { "epoch": 11.532934131736527, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1485, "step": 7704 }, { "epoch": 11.534431137724551, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.143, "step": 7705 }, { "epoch": 11.535928143712574, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1384, "step": 7706 }, { "epoch": 11.5374251497006, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1403, "step": 7707 }, { "epoch": 11.538922155688622, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1481, "step": 7708 }, { "epoch": 11.540419161676647, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1391, "step": 7709 }, { "epoch": 11.54191616766467, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1467, "step": 7710 }, { "epoch": 11.543413173652695, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1412, "step": 7711 }, { "epoch": 11.544910179640718, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1452, "step": 7712 }, { "epoch": 11.546407185628743, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1438, "step": 7713 }, { "epoch": 11.547904191616766, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1455, "step": 7714 }, { "epoch": 11.54940119760479, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1397, "step": 7715 }, { "epoch": 11.550898203592814, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1436, "step": 7716 }, { "epoch": 11.552395209580839, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1456, "step": 7717 }, { "epoch": 11.553892215568862, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1409, "step": 7718 }, { "epoch": 11.555389221556887, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1402, "step": 7719 }, { "epoch": 11.55688622754491, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1458, "step": 7720 }, { "epoch": 11.558383233532934, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1427, "step": 7721 }, { "epoch": 11.559880239520957, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1382, "step": 7722 }, { "epoch": 11.561377245508982, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1451, "step": 7723 }, { "epoch": 11.562874251497005, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.147, "step": 7724 }, { "epoch": 11.56437125748503, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1411, "step": 7725 }, { "epoch": 11.565868263473053, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1519, "step": 7726 }, { "epoch": 11.567365269461078, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1456, "step": 7727 }, { "epoch": 11.568862275449101, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1418, "step": 7728 }, { "epoch": 11.570359281437126, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1397, "step": 7729 }, { "epoch": 11.571856287425149, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1389, "step": 7730 }, { "epoch": 11.573353293413174, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1419, "step": 7731 }, { "epoch": 11.574850299401197, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1466, "step": 7732 }, { "epoch": 11.576347305389222, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1418, "step": 7733 }, { "epoch": 11.577844311377245, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1487, "step": 7734 }, { "epoch": 11.57934131736527, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.142, "step": 7735 }, { "epoch": 11.580838323353294, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1363, "step": 7736 }, { "epoch": 11.582335329341317, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1384, "step": 7737 }, { "epoch": 11.58383233532934, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1452, "step": 7738 }, { "epoch": 11.585329341317365, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1361, "step": 7739 }, { "epoch": 11.58682634730539, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1329, "step": 7740 }, { "epoch": 11.588323353293413, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1477, "step": 7741 }, { "epoch": 11.589820359281438, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1378, "step": 7742 }, { "epoch": 11.591317365269461, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1458, "step": 7743 }, { "epoch": 11.592814371257486, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1477, "step": 7744 }, { "epoch": 11.594311377245509, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1374, "step": 7745 }, { "epoch": 11.595808383233534, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1431, "step": 7746 }, { "epoch": 11.597305389221557, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1451, "step": 7747 }, { "epoch": 11.598802395209582, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1448, "step": 7748 }, { "epoch": 11.600299401197605, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1379, "step": 7749 }, { "epoch": 11.60179640718563, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1464, "step": 7750 }, { "epoch": 11.603293413173652, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1382, "step": 7751 }, { "epoch": 11.604790419161677, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1424, "step": 7752 }, { "epoch": 11.6062874251497, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1471, "step": 7753 }, { "epoch": 11.607784431137725, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1368, "step": 7754 }, { "epoch": 11.609281437125748, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1442, "step": 7755 }, { "epoch": 11.610778443113773, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1403, "step": 7756 }, { "epoch": 11.612275449101796, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1433, "step": 7757 }, { "epoch": 11.613772455089821, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.138, "step": 7758 }, { "epoch": 11.615269461077844, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1395, "step": 7759 }, { "epoch": 11.616766467065869, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1439, "step": 7760 }, { "epoch": 11.618263473053892, "grad_norm": 0.05224609375, "learning_rate": 0.0008, "loss": 1.1427, "step": 7761 }, { "epoch": 11.619760479041917, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.142, "step": 7762 }, { "epoch": 11.62125748502994, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1446, "step": 7763 }, { "epoch": 11.622754491017965, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.144, "step": 7764 }, { "epoch": 11.624251497005988, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1461, "step": 7765 }, { "epoch": 11.625748502994012, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1524, "step": 7766 }, { "epoch": 11.627245508982035, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1465, "step": 7767 }, { "epoch": 11.62874251497006, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1477, "step": 7768 }, { "epoch": 11.630239520958083, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1453, "step": 7769 }, { "epoch": 11.631736526946108, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1473, "step": 7770 }, { "epoch": 11.633233532934131, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1377, "step": 7771 }, { "epoch": 11.634730538922156, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1406, "step": 7772 }, { "epoch": 11.636227544910179, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1454, "step": 7773 }, { "epoch": 11.637724550898204, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1435, "step": 7774 }, { "epoch": 11.639221556886227, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1459, "step": 7775 }, { "epoch": 11.640718562874252, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1453, "step": 7776 }, { "epoch": 11.642215568862275, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1421, "step": 7777 }, { "epoch": 11.6437125748503, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1412, "step": 7778 }, { "epoch": 11.645209580838323, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1361, "step": 7779 }, { "epoch": 11.646706586826348, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1408, "step": 7780 }, { "epoch": 11.64820359281437, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1402, "step": 7781 }, { "epoch": 11.649700598802395, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1456, "step": 7782 }, { "epoch": 11.651197604790418, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1497, "step": 7783 }, { "epoch": 11.652694610778443, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.1496, "step": 7784 }, { "epoch": 11.654191616766466, "grad_norm": 0.051513671875, "learning_rate": 0.0008, "loss": 1.1371, "step": 7785 }, { "epoch": 11.655688622754491, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1473, "step": 7786 }, { "epoch": 11.657185628742514, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1414, "step": 7787 }, { "epoch": 11.658682634730539, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1491, "step": 7788 }, { "epoch": 11.660179640718562, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1421, "step": 7789 }, { "epoch": 11.661676646706587, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1465, "step": 7790 }, { "epoch": 11.66317365269461, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1422, "step": 7791 }, { "epoch": 11.664670658682635, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1415, "step": 7792 }, { "epoch": 11.66616766467066, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1466, "step": 7793 }, { "epoch": 11.667664670658683, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1454, "step": 7794 }, { "epoch": 11.669161676646706, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1436, "step": 7795 }, { "epoch": 11.67065868263473, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1478, "step": 7796 }, { "epoch": 11.672155688622755, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1426, "step": 7797 }, { "epoch": 11.673652694610778, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1425, "step": 7798 }, { "epoch": 11.675149700598803, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1348, "step": 7799 }, { "epoch": 11.676646706586826, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1386, "step": 7800 }, { "epoch": 11.678143712574851, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1454, "step": 7801 }, { "epoch": 11.679640718562874, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1433, "step": 7802 }, { "epoch": 11.681137724550899, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1399, "step": 7803 }, { "epoch": 11.682634730538922, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1473, "step": 7804 }, { "epoch": 11.684131736526947, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1409, "step": 7805 }, { "epoch": 11.68562874251497, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1427, "step": 7806 }, { "epoch": 11.687125748502995, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1351, "step": 7807 }, { "epoch": 11.688622754491018, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.145, "step": 7808 }, { "epoch": 11.690119760479043, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1394, "step": 7809 }, { "epoch": 11.691616766467066, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1378, "step": 7810 }, { "epoch": 11.69311377245509, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1399, "step": 7811 }, { "epoch": 11.694610778443113, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1495, "step": 7812 }, { "epoch": 11.696107784431138, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.147, "step": 7813 }, { "epoch": 11.697604790419161, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1419, "step": 7814 }, { "epoch": 11.699101796407186, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1394, "step": 7815 }, { "epoch": 11.70059880239521, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1455, "step": 7816 }, { "epoch": 11.702095808383234, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1548, "step": 7817 }, { "epoch": 11.703592814371257, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1489, "step": 7818 }, { "epoch": 11.705089820359282, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1419, "step": 7819 }, { "epoch": 11.706586826347305, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1447, "step": 7820 }, { "epoch": 11.70808383233533, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1427, "step": 7821 }, { "epoch": 11.709580838323353, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1391, "step": 7822 }, { "epoch": 11.711077844311378, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1397, "step": 7823 }, { "epoch": 11.7125748502994, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1437, "step": 7824 }, { "epoch": 11.714071856287426, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1431, "step": 7825 }, { "epoch": 11.715568862275449, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1473, "step": 7826 }, { "epoch": 11.717065868263473, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1396, "step": 7827 }, { "epoch": 11.718562874251496, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1491, "step": 7828 }, { "epoch": 11.720059880239521, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1462, "step": 7829 }, { "epoch": 11.721556886227544, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1438, "step": 7830 }, { "epoch": 11.72305389221557, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.148, "step": 7831 }, { "epoch": 11.724550898203592, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1408, "step": 7832 }, { "epoch": 11.726047904191617, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1444, "step": 7833 }, { "epoch": 11.72754491017964, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1405, "step": 7834 }, { "epoch": 11.729041916167665, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1383, "step": 7835 }, { "epoch": 11.730538922155688, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1458, "step": 7836 }, { "epoch": 11.732035928143713, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.142, "step": 7837 }, { "epoch": 11.733532934131736, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1405, "step": 7838 }, { "epoch": 11.73502994011976, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1444, "step": 7839 }, { "epoch": 11.736526946107784, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1451, "step": 7840 }, { "epoch": 11.738023952095809, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1359, "step": 7841 }, { "epoch": 11.739520958083832, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1438, "step": 7842 }, { "epoch": 11.741017964071856, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1351, "step": 7843 }, { "epoch": 11.74251497005988, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.145, "step": 7844 }, { "epoch": 11.744011976047904, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1415, "step": 7845 }, { "epoch": 11.745508982035929, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1421, "step": 7846 }, { "epoch": 11.747005988023952, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1416, "step": 7847 }, { "epoch": 11.748502994011975, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1454, "step": 7848 }, { "epoch": 11.75, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1426, "step": 7849 }, { "epoch": 11.751497005988025, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.145, "step": 7850 }, { "epoch": 11.752994011976048, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.143, "step": 7851 }, { "epoch": 11.754491017964071, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1376, "step": 7852 }, { "epoch": 11.755988023952096, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1444, "step": 7853 }, { "epoch": 11.75748502994012, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.14, "step": 7854 }, { "epoch": 11.758982035928144, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.142, "step": 7855 }, { "epoch": 11.760479041916168, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1427, "step": 7856 }, { "epoch": 11.761976047904191, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1449, "step": 7857 }, { "epoch": 11.763473053892216, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1457, "step": 7858 }, { "epoch": 11.76497005988024, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1445, "step": 7859 }, { "epoch": 11.766467065868264, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1401, "step": 7860 }, { "epoch": 11.767964071856287, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1457, "step": 7861 }, { "epoch": 11.769461077844312, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1434, "step": 7862 }, { "epoch": 11.770958083832335, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1444, "step": 7863 }, { "epoch": 11.77245508982036, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1407, "step": 7864 }, { "epoch": 11.773952095808383, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1377, "step": 7865 }, { "epoch": 11.775449101796408, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1487, "step": 7866 }, { "epoch": 11.77694610778443, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1471, "step": 7867 }, { "epoch": 11.778443113772456, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1397, "step": 7868 }, { "epoch": 11.779940119760479, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.1483, "step": 7869 }, { "epoch": 11.781437125748504, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1394, "step": 7870 }, { "epoch": 11.782934131736527, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1378, "step": 7871 }, { "epoch": 11.784431137724551, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1486, "step": 7872 }, { "epoch": 11.785928143712574, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1469, "step": 7873 }, { "epoch": 11.7874251497006, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1385, "step": 7874 }, { "epoch": 11.788922155688622, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1456, "step": 7875 }, { "epoch": 11.790419161676647, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1471, "step": 7876 }, { "epoch": 11.79191616766467, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1407, "step": 7877 }, { "epoch": 11.793413173652695, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1419, "step": 7878 }, { "epoch": 11.794910179640718, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1421, "step": 7879 }, { "epoch": 11.796407185628743, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1454, "step": 7880 }, { "epoch": 11.797904191616766, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1467, "step": 7881 }, { "epoch": 11.79940119760479, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1409, "step": 7882 }, { "epoch": 11.800898203592814, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1465, "step": 7883 }, { "epoch": 11.802395209580839, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1518, "step": 7884 }, { "epoch": 11.803892215568862, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1376, "step": 7885 }, { "epoch": 11.805389221556887, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1458, "step": 7886 }, { "epoch": 11.80688622754491, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1376, "step": 7887 }, { "epoch": 11.808383233532934, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1508, "step": 7888 }, { "epoch": 11.809880239520957, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1447, "step": 7889 }, { "epoch": 11.811377245508982, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1427, "step": 7890 }, { "epoch": 11.812874251497005, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1437, "step": 7891 }, { "epoch": 11.81437125748503, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1421, "step": 7892 }, { "epoch": 11.815868263473053, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.147, "step": 7893 }, { "epoch": 11.817365269461078, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1366, "step": 7894 }, { "epoch": 11.818862275449101, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1433, "step": 7895 }, { "epoch": 11.820359281437126, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1401, "step": 7896 }, { "epoch": 11.821856287425149, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1448, "step": 7897 }, { "epoch": 11.823353293413174, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1462, "step": 7898 }, { "epoch": 11.824850299401197, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1465, "step": 7899 }, { "epoch": 11.826347305389222, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1396, "step": 7900 }, { "epoch": 11.827844311377245, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1393, "step": 7901 }, { "epoch": 11.82934131736527, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1384, "step": 7902 }, { "epoch": 11.830838323353294, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1494, "step": 7903 }, { "epoch": 11.832335329341317, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1453, "step": 7904 }, { "epoch": 11.83383233532934, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1418, "step": 7905 }, { "epoch": 11.835329341317365, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1385, "step": 7906 }, { "epoch": 11.83682634730539, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1434, "step": 7907 }, { "epoch": 11.838323353293413, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.143, "step": 7908 }, { "epoch": 11.839820359281438, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1404, "step": 7909 }, { "epoch": 11.841317365269461, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.137, "step": 7910 }, { "epoch": 11.842814371257486, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.1461, "step": 7911 }, { "epoch": 11.844311377245509, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1481, "step": 7912 }, { "epoch": 11.845808383233534, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.146, "step": 7913 }, { "epoch": 11.847305389221557, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1433, "step": 7914 }, { "epoch": 11.848802395209582, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1427, "step": 7915 }, { "epoch": 11.850299401197605, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1415, "step": 7916 }, { "epoch": 11.85179640718563, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1414, "step": 7917 }, { "epoch": 11.853293413173652, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1399, "step": 7918 }, { "epoch": 11.854790419161677, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1448, "step": 7919 }, { "epoch": 11.8562874251497, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1382, "step": 7920 }, { "epoch": 11.857784431137725, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.138, "step": 7921 }, { "epoch": 11.859281437125748, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1455, "step": 7922 }, { "epoch": 11.860778443113773, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1459, "step": 7923 }, { "epoch": 11.862275449101796, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1446, "step": 7924 }, { "epoch": 11.863772455089821, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1503, "step": 7925 }, { "epoch": 11.865269461077844, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1455, "step": 7926 }, { "epoch": 11.866766467065869, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1505, "step": 7927 }, { "epoch": 11.868263473053892, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1418, "step": 7928 }, { "epoch": 11.869760479041917, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1405, "step": 7929 }, { "epoch": 11.87125748502994, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1465, "step": 7930 }, { "epoch": 11.872754491017965, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1379, "step": 7931 }, { "epoch": 11.874251497005988, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1452, "step": 7932 }, { "epoch": 11.875748502994012, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1434, "step": 7933 }, { "epoch": 11.877245508982035, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1441, "step": 7934 }, { "epoch": 11.87874251497006, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1401, "step": 7935 }, { "epoch": 11.880239520958083, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1434, "step": 7936 }, { "epoch": 11.881736526946108, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1427, "step": 7937 }, { "epoch": 11.883233532934131, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1422, "step": 7938 }, { "epoch": 11.884730538922156, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1427, "step": 7939 }, { "epoch": 11.886227544910179, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1443, "step": 7940 }, { "epoch": 11.887724550898204, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1426, "step": 7941 }, { "epoch": 11.889221556886227, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1456, "step": 7942 }, { "epoch": 11.890718562874252, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1337, "step": 7943 }, { "epoch": 11.892215568862275, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1394, "step": 7944 }, { "epoch": 11.8937125748503, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1403, "step": 7945 }, { "epoch": 11.895209580838323, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1413, "step": 7946 }, { "epoch": 11.896706586826348, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1434, "step": 7947 }, { "epoch": 11.89820359281437, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1369, "step": 7948 }, { "epoch": 11.899700598802395, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.144, "step": 7949 }, { "epoch": 11.901197604790418, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1422, "step": 7950 }, { "epoch": 11.902694610778443, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1432, "step": 7951 }, { "epoch": 11.904191616766466, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1401, "step": 7952 }, { "epoch": 11.905688622754491, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1384, "step": 7953 }, { "epoch": 11.907185628742514, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1434, "step": 7954 }, { "epoch": 11.908682634730539, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1425, "step": 7955 }, { "epoch": 11.910179640718562, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1442, "step": 7956 }, { "epoch": 11.911676646706587, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1401, "step": 7957 }, { "epoch": 11.91317365269461, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1446, "step": 7958 }, { "epoch": 11.914670658682635, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1412, "step": 7959 }, { "epoch": 11.91616766467066, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1388, "step": 7960 }, { "epoch": 11.917664670658683, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1401, "step": 7961 }, { "epoch": 11.919161676646706, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1448, "step": 7962 }, { "epoch": 11.92065868263473, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.142, "step": 7963 }, { "epoch": 11.922155688622755, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1364, "step": 7964 }, { "epoch": 11.923652694610778, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1348, "step": 7965 }, { "epoch": 11.925149700598803, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1425, "step": 7966 }, { "epoch": 11.926646706586826, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1372, "step": 7967 }, { "epoch": 11.928143712574851, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1397, "step": 7968 }, { "epoch": 11.929640718562874, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1386, "step": 7969 }, { "epoch": 11.931137724550899, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1407, "step": 7970 }, { "epoch": 11.932634730538922, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1425, "step": 7971 }, { "epoch": 11.934131736526947, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1487, "step": 7972 }, { "epoch": 11.93562874251497, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1433, "step": 7973 }, { "epoch": 11.937125748502995, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1472, "step": 7974 }, { "epoch": 11.938622754491018, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1441, "step": 7975 }, { "epoch": 11.940119760479043, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1424, "step": 7976 }, { "epoch": 11.941616766467066, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1441, "step": 7977 }, { "epoch": 11.94311377245509, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1387, "step": 7978 }, { "epoch": 11.944610778443113, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1465, "step": 7979 }, { "epoch": 11.946107784431138, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.144, "step": 7980 }, { "epoch": 11.947604790419161, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1489, "step": 7981 }, { "epoch": 11.949101796407186, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1439, "step": 7982 }, { "epoch": 11.95059880239521, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1425, "step": 7983 }, { "epoch": 11.952095808383234, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1383, "step": 7984 }, { "epoch": 11.953592814371257, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1467, "step": 7985 }, { "epoch": 11.955089820359282, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1425, "step": 7986 }, { "epoch": 11.956586826347305, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1384, "step": 7987 }, { "epoch": 11.95808383233533, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1446, "step": 7988 }, { "epoch": 11.959580838323353, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1458, "step": 7989 }, { "epoch": 11.961077844311378, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1357, "step": 7990 }, { "epoch": 11.9625748502994, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1409, "step": 7991 }, { "epoch": 11.964071856287426, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1429, "step": 7992 }, { "epoch": 11.965568862275449, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1496, "step": 7993 }, { "epoch": 11.967065868263473, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1397, "step": 7994 }, { "epoch": 11.968562874251496, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1388, "step": 7995 }, { "epoch": 11.970059880239521, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1428, "step": 7996 }, { "epoch": 11.971556886227544, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1493, "step": 7997 }, { "epoch": 11.97305389221557, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1395, "step": 7998 }, { "epoch": 11.974550898203592, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1389, "step": 7999 }, { "epoch": 11.976047904191617, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1449, "step": 8000 }, { "epoch": 11.97754491017964, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1411, "step": 8001 }, { "epoch": 11.979041916167665, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1411, "step": 8002 }, { "epoch": 11.980538922155688, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1433, "step": 8003 }, { "epoch": 11.982035928143713, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1366, "step": 8004 }, { "epoch": 11.983532934131736, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1421, "step": 8005 }, { "epoch": 11.98502994011976, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1414, "step": 8006 }, { "epoch": 11.986526946107784, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1452, "step": 8007 }, { "epoch": 11.988023952095809, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1423, "step": 8008 }, { "epoch": 11.989520958083832, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1432, "step": 8009 }, { "epoch": 11.991017964071856, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1429, "step": 8010 }, { "epoch": 11.99251497005988, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1396, "step": 8011 }, { "epoch": 11.994011976047904, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1457, "step": 8012 }, { "epoch": 11.995508982035929, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1432, "step": 8013 }, { "epoch": 11.997005988023952, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1387, "step": 8014 }, { "epoch": 11.998502994011975, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1461, "step": 8015 }, { "epoch": 12.0, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1404, "step": 8016 }, { "epoch": 12.001497005988025, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1418, "step": 8017 }, { "epoch": 12.002994011976048, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1431, "step": 8018 }, { "epoch": 12.004491017964073, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1443, "step": 8019 }, { "epoch": 12.005988023952096, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1435, "step": 8020 }, { "epoch": 12.00748502994012, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.137, "step": 8021 }, { "epoch": 12.008982035928144, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1447, "step": 8022 }, { "epoch": 12.010479041916168, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1493, "step": 8023 }, { "epoch": 12.011976047904191, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1408, "step": 8024 }, { "epoch": 12.013473053892216, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1362, "step": 8025 }, { "epoch": 12.01497005988024, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1451, "step": 8026 }, { "epoch": 12.016467065868264, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1384, "step": 8027 }, { "epoch": 12.017964071856287, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1422, "step": 8028 }, { "epoch": 12.019461077844312, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1438, "step": 8029 }, { "epoch": 12.020958083832335, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.143, "step": 8030 }, { "epoch": 12.02245508982036, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.145, "step": 8031 }, { "epoch": 12.023952095808383, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1446, "step": 8032 }, { "epoch": 12.025449101796408, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.142, "step": 8033 }, { "epoch": 12.02694610778443, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1441, "step": 8034 }, { "epoch": 12.028443113772456, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1374, "step": 8035 }, { "epoch": 12.029940119760479, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1424, "step": 8036 }, { "epoch": 12.031437125748504, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1437, "step": 8037 }, { "epoch": 12.032934131736527, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1403, "step": 8038 }, { "epoch": 12.034431137724551, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1352, "step": 8039 }, { "epoch": 12.035928143712574, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1424, "step": 8040 }, { "epoch": 12.0374251497006, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1388, "step": 8041 }, { "epoch": 12.038922155688622, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1404, "step": 8042 }, { "epoch": 12.040419161676647, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1436, "step": 8043 }, { "epoch": 12.04191616766467, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1419, "step": 8044 }, { "epoch": 12.043413173652695, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1417, "step": 8045 }, { "epoch": 12.044910179640718, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.136, "step": 8046 }, { "epoch": 12.046407185628743, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1436, "step": 8047 }, { "epoch": 12.047904191616766, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1418, "step": 8048 }, { "epoch": 12.04940119760479, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1424, "step": 8049 }, { "epoch": 12.050898203592814, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1452, "step": 8050 }, { "epoch": 12.052395209580839, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1498, "step": 8051 }, { "epoch": 12.053892215568862, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1422, "step": 8052 }, { "epoch": 12.055389221556887, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1463, "step": 8053 }, { "epoch": 12.05688622754491, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1412, "step": 8054 }, { "epoch": 12.058383233532934, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1419, "step": 8055 }, { "epoch": 12.059880239520957, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1438, "step": 8056 }, { "epoch": 12.061377245508982, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1407, "step": 8057 }, { "epoch": 12.062874251497005, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1396, "step": 8058 }, { "epoch": 12.06437125748503, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1382, "step": 8059 }, { "epoch": 12.065868263473053, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1435, "step": 8060 }, { "epoch": 12.067365269461078, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1423, "step": 8061 }, { "epoch": 12.068862275449101, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1507, "step": 8062 }, { "epoch": 12.070359281437126, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1411, "step": 8063 }, { "epoch": 12.071856287425149, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1398, "step": 8064 }, { "epoch": 12.073353293413174, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1448, "step": 8065 }, { "epoch": 12.074850299401197, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1347, "step": 8066 }, { "epoch": 12.076347305389222, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1351, "step": 8067 }, { "epoch": 12.077844311377245, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1427, "step": 8068 }, { "epoch": 12.07934131736527, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1383, "step": 8069 }, { "epoch": 12.080838323353293, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1462, "step": 8070 }, { "epoch": 12.082335329341317, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1376, "step": 8071 }, { "epoch": 12.08383233532934, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.142, "step": 8072 }, { "epoch": 12.085329341317365, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1444, "step": 8073 }, { "epoch": 12.08682634730539, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1405, "step": 8074 }, { "epoch": 12.088323353293413, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1438, "step": 8075 }, { "epoch": 12.089820359281438, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1387, "step": 8076 }, { "epoch": 12.091317365269461, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1384, "step": 8077 }, { "epoch": 12.092814371257486, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1414, "step": 8078 }, { "epoch": 12.094311377245509, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1387, "step": 8079 }, { "epoch": 12.095808383233534, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1456, "step": 8080 }, { "epoch": 12.097305389221557, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1428, "step": 8081 }, { "epoch": 12.098802395209582, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.138, "step": 8082 }, { "epoch": 12.100299401197605, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1368, "step": 8083 }, { "epoch": 12.10179640718563, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1466, "step": 8084 }, { "epoch": 12.103293413173652, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1426, "step": 8085 }, { "epoch": 12.104790419161677, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1483, "step": 8086 }, { "epoch": 12.1062874251497, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1458, "step": 8087 }, { "epoch": 12.107784431137725, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.141, "step": 8088 }, { "epoch": 12.109281437125748, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1429, "step": 8089 }, { "epoch": 12.110778443113773, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1417, "step": 8090 }, { "epoch": 12.112275449101796, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1407, "step": 8091 }, { "epoch": 12.113772455089821, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1425, "step": 8092 }, { "epoch": 12.115269461077844, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1392, "step": 8093 }, { "epoch": 12.116766467065869, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1403, "step": 8094 }, { "epoch": 12.118263473053892, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1391, "step": 8095 }, { "epoch": 12.119760479041917, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1407, "step": 8096 }, { "epoch": 12.12125748502994, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1395, "step": 8097 }, { "epoch": 12.122754491017965, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1418, "step": 8098 }, { "epoch": 12.124251497005988, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1437, "step": 8099 }, { "epoch": 12.125748502994012, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1386, "step": 8100 }, { "epoch": 12.127245508982035, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1443, "step": 8101 }, { "epoch": 12.12874251497006, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1497, "step": 8102 }, { "epoch": 12.130239520958083, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1449, "step": 8103 }, { "epoch": 12.131736526946108, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1475, "step": 8104 }, { "epoch": 12.133233532934131, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1443, "step": 8105 }, { "epoch": 12.134730538922156, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1465, "step": 8106 }, { "epoch": 12.136227544910179, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.146, "step": 8107 }, { "epoch": 12.137724550898204, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1372, "step": 8108 }, { "epoch": 12.139221556886227, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1381, "step": 8109 }, { "epoch": 12.140718562874252, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1415, "step": 8110 }, { "epoch": 12.142215568862275, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1439, "step": 8111 }, { "epoch": 12.1437125748503, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1427, "step": 8112 }, { "epoch": 12.145209580838323, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1424, "step": 8113 }, { "epoch": 12.146706586826348, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1437, "step": 8114 }, { "epoch": 12.14820359281437, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1397, "step": 8115 }, { "epoch": 12.149700598802395, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1361, "step": 8116 }, { "epoch": 12.151197604790418, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1431, "step": 8117 }, { "epoch": 12.152694610778443, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1408, "step": 8118 }, { "epoch": 12.154191616766466, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.144, "step": 8119 }, { "epoch": 12.155688622754491, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1462, "step": 8120 }, { "epoch": 12.157185628742514, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.145, "step": 8121 }, { "epoch": 12.158682634730539, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1404, "step": 8122 }, { "epoch": 12.160179640718562, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1378, "step": 8123 }, { "epoch": 12.161676646706587, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1426, "step": 8124 }, { "epoch": 12.16317365269461, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1405, "step": 8125 }, { "epoch": 12.164670658682635, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1449, "step": 8126 }, { "epoch": 12.16616766467066, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1395, "step": 8127 }, { "epoch": 12.167664670658683, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1486, "step": 8128 }, { "epoch": 12.169161676646707, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1418, "step": 8129 }, { "epoch": 12.17065868263473, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1426, "step": 8130 }, { "epoch": 12.172155688622755, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1384, "step": 8131 }, { "epoch": 12.173652694610778, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1481, "step": 8132 }, { "epoch": 12.175149700598803, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1451, "step": 8133 }, { "epoch": 12.176646706586826, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1474, "step": 8134 }, { "epoch": 12.178143712574851, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1461, "step": 8135 }, { "epoch": 12.179640718562874, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1418, "step": 8136 }, { "epoch": 12.181137724550899, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1414, "step": 8137 }, { "epoch": 12.182634730538922, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1408, "step": 8138 }, { "epoch": 12.184131736526947, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1371, "step": 8139 }, { "epoch": 12.18562874251497, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1429, "step": 8140 }, { "epoch": 12.187125748502995, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1426, "step": 8141 }, { "epoch": 12.188622754491018, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.134, "step": 8142 }, { "epoch": 12.190119760479043, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1459, "step": 8143 }, { "epoch": 12.191616766467066, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1396, "step": 8144 }, { "epoch": 12.19311377245509, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1468, "step": 8145 }, { "epoch": 12.194610778443113, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1395, "step": 8146 }, { "epoch": 12.196107784431138, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1384, "step": 8147 }, { "epoch": 12.197604790419161, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.139, "step": 8148 }, { "epoch": 12.199101796407186, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1364, "step": 8149 }, { "epoch": 12.20059880239521, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1388, "step": 8150 }, { "epoch": 12.202095808383234, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1461, "step": 8151 }, { "epoch": 12.203592814371257, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1398, "step": 8152 }, { "epoch": 12.205089820359282, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.137, "step": 8153 }, { "epoch": 12.206586826347305, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1477, "step": 8154 }, { "epoch": 12.20808383233533, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1371, "step": 8155 }, { "epoch": 12.209580838323353, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.144, "step": 8156 }, { "epoch": 12.211077844311378, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1385, "step": 8157 }, { "epoch": 12.2125748502994, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1391, "step": 8158 }, { "epoch": 12.214071856287426, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.138, "step": 8159 }, { "epoch": 12.215568862275449, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1427, "step": 8160 }, { "epoch": 12.217065868263473, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1416, "step": 8161 }, { "epoch": 12.218562874251496, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1323, "step": 8162 }, { "epoch": 12.220059880239521, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.136, "step": 8163 }, { "epoch": 12.221556886227544, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1437, "step": 8164 }, { "epoch": 12.22305389221557, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1412, "step": 8165 }, { "epoch": 12.224550898203592, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1379, "step": 8166 }, { "epoch": 12.226047904191617, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1417, "step": 8167 }, { "epoch": 12.22754491017964, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1351, "step": 8168 }, { "epoch": 12.229041916167665, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.144, "step": 8169 }, { "epoch": 12.230538922155688, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1408, "step": 8170 }, { "epoch": 12.232035928143713, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1418, "step": 8171 }, { "epoch": 12.233532934131736, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1439, "step": 8172 }, { "epoch": 12.23502994011976, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1433, "step": 8173 }, { "epoch": 12.236526946107784, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1459, "step": 8174 }, { "epoch": 12.238023952095809, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.143, "step": 8175 }, { "epoch": 12.239520958083832, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1383, "step": 8176 }, { "epoch": 12.241017964071856, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1396, "step": 8177 }, { "epoch": 12.24251497005988, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1417, "step": 8178 }, { "epoch": 12.244011976047904, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1385, "step": 8179 }, { "epoch": 12.245508982035927, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.14, "step": 8180 }, { "epoch": 12.247005988023952, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1423, "step": 8181 }, { "epoch": 12.248502994011975, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1419, "step": 8182 }, { "epoch": 12.25, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1402, "step": 8183 }, { "epoch": 12.251497005988025, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.142, "step": 8184 }, { "epoch": 12.252994011976048, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.14, "step": 8185 }, { "epoch": 12.254491017964073, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1448, "step": 8186 }, { "epoch": 12.255988023952096, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.141, "step": 8187 }, { "epoch": 12.25748502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1413, "step": 8188 }, { "epoch": 12.258982035928144, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1465, "step": 8189 }, { "epoch": 12.260479041916168, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1454, "step": 8190 }, { "epoch": 12.261976047904191, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1396, "step": 8191 }, { "epoch": 12.263473053892216, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1339, "step": 8192 }, { "epoch": 12.26497005988024, "grad_norm": 0.051513671875, "learning_rate": 0.0008, "loss": 1.1465, "step": 8193 }, { "epoch": 12.266467065868264, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1371, "step": 8194 }, { "epoch": 12.267964071856287, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1386, "step": 8195 }, { "epoch": 12.269461077844312, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1458, "step": 8196 }, { "epoch": 12.270958083832335, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.134, "step": 8197 }, { "epoch": 12.27245508982036, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.147, "step": 8198 }, { "epoch": 12.273952095808383, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1417, "step": 8199 }, { "epoch": 12.275449101796408, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.141, "step": 8200 }, { "epoch": 12.27694610778443, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1423, "step": 8201 }, { "epoch": 12.278443113772456, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1389, "step": 8202 }, { "epoch": 12.279940119760479, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1375, "step": 8203 }, { "epoch": 12.281437125748504, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1371, "step": 8204 }, { "epoch": 12.282934131736527, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1453, "step": 8205 }, { "epoch": 12.284431137724551, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1361, "step": 8206 }, { "epoch": 12.285928143712574, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1409, "step": 8207 }, { "epoch": 12.2874251497006, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1392, "step": 8208 }, { "epoch": 12.288922155688622, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1435, "step": 8209 }, { "epoch": 12.290419161676647, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.14, "step": 8210 }, { "epoch": 12.29191616766467, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1391, "step": 8211 }, { "epoch": 12.293413173652695, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1397, "step": 8212 }, { "epoch": 12.294910179640718, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1404, "step": 8213 }, { "epoch": 12.296407185628743, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1441, "step": 8214 }, { "epoch": 12.297904191616766, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1407, "step": 8215 }, { "epoch": 12.29940119760479, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1471, "step": 8216 }, { "epoch": 12.300898203592814, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1474, "step": 8217 }, { "epoch": 12.302395209580839, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1425, "step": 8218 }, { "epoch": 12.303892215568862, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1498, "step": 8219 }, { "epoch": 12.305389221556887, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1372, "step": 8220 }, { "epoch": 12.30688622754491, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1341, "step": 8221 }, { "epoch": 12.308383233532934, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1323, "step": 8222 }, { "epoch": 12.309880239520957, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1369, "step": 8223 }, { "epoch": 12.311377245508982, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.141, "step": 8224 }, { "epoch": 12.312874251497005, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1432, "step": 8225 }, { "epoch": 12.31437125748503, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1409, "step": 8226 }, { "epoch": 12.315868263473053, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1408, "step": 8227 }, { "epoch": 12.317365269461078, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1437, "step": 8228 }, { "epoch": 12.318862275449101, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1381, "step": 8229 }, { "epoch": 12.320359281437126, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1364, "step": 8230 }, { "epoch": 12.321856287425149, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1458, "step": 8231 }, { "epoch": 12.323353293413174, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1365, "step": 8232 }, { "epoch": 12.324850299401197, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1436, "step": 8233 }, { "epoch": 12.326347305389222, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1397, "step": 8234 }, { "epoch": 12.327844311377245, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1379, "step": 8235 }, { "epoch": 12.32934131736527, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1392, "step": 8236 }, { "epoch": 12.330838323353294, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1486, "step": 8237 }, { "epoch": 12.332335329341317, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1474, "step": 8238 }, { "epoch": 12.33383233532934, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1375, "step": 8239 }, { "epoch": 12.335329341317365, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1441, "step": 8240 }, { "epoch": 12.33682634730539, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1388, "step": 8241 }, { "epoch": 12.338323353293413, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1414, "step": 8242 }, { "epoch": 12.339820359281438, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1423, "step": 8243 }, { "epoch": 12.341317365269461, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1449, "step": 8244 }, { "epoch": 12.342814371257486, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1374, "step": 8245 }, { "epoch": 12.344311377245509, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1383, "step": 8246 }, { "epoch": 12.345808383233534, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1396, "step": 8247 }, { "epoch": 12.347305389221557, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1459, "step": 8248 }, { "epoch": 12.348802395209582, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1414, "step": 8249 }, { "epoch": 12.350299401197605, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1379, "step": 8250 }, { "epoch": 12.35179640718563, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1407, "step": 8251 }, { "epoch": 12.353293413173652, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1387, "step": 8252 }, { "epoch": 12.354790419161677, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1389, "step": 8253 }, { "epoch": 12.3562874251497, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1407, "step": 8254 }, { "epoch": 12.357784431137725, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1453, "step": 8255 }, { "epoch": 12.359281437125748, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1398, "step": 8256 }, { "epoch": 12.360778443113773, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1445, "step": 8257 }, { "epoch": 12.362275449101796, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1496, "step": 8258 }, { "epoch": 12.363772455089821, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1383, "step": 8259 }, { "epoch": 12.365269461077844, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1399, "step": 8260 }, { "epoch": 12.366766467065869, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1447, "step": 8261 }, { "epoch": 12.368263473053892, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1375, "step": 8262 }, { "epoch": 12.369760479041917, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.136, "step": 8263 }, { "epoch": 12.37125748502994, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1403, "step": 8264 }, { "epoch": 12.372754491017965, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1358, "step": 8265 }, { "epoch": 12.374251497005988, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1351, "step": 8266 }, { "epoch": 12.375748502994012, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1411, "step": 8267 }, { "epoch": 12.377245508982035, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1409, "step": 8268 }, { "epoch": 12.37874251497006, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1432, "step": 8269 }, { "epoch": 12.380239520958083, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.143, "step": 8270 }, { "epoch": 12.381736526946108, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1461, "step": 8271 }, { "epoch": 12.383233532934131, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1439, "step": 8272 }, { "epoch": 12.384730538922156, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1459, "step": 8273 }, { "epoch": 12.386227544910179, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1396, "step": 8274 }, { "epoch": 12.387724550898204, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.144, "step": 8275 }, { "epoch": 12.389221556886227, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1441, "step": 8276 }, { "epoch": 12.390718562874252, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1406, "step": 8277 }, { "epoch": 12.392215568862275, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1466, "step": 8278 }, { "epoch": 12.3937125748503, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1457, "step": 8279 }, { "epoch": 12.395209580838323, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1391, "step": 8280 }, { "epoch": 12.396706586826348, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1491, "step": 8281 }, { "epoch": 12.39820359281437, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1406, "step": 8282 }, { "epoch": 12.399700598802395, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1492, "step": 8283 }, { "epoch": 12.401197604790418, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1417, "step": 8284 }, { "epoch": 12.402694610778443, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1387, "step": 8285 }, { "epoch": 12.404191616766466, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1444, "step": 8286 }, { "epoch": 12.405688622754491, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1372, "step": 8287 }, { "epoch": 12.407185628742514, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1447, "step": 8288 }, { "epoch": 12.408682634730539, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1449, "step": 8289 }, { "epoch": 12.410179640718562, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1455, "step": 8290 }, { "epoch": 12.411676646706587, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1439, "step": 8291 }, { "epoch": 12.41317365269461, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1369, "step": 8292 }, { "epoch": 12.414670658682635, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1372, "step": 8293 }, { "epoch": 12.41616766467066, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1466, "step": 8294 }, { "epoch": 12.417664670658683, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1427, "step": 8295 }, { "epoch": 12.419161676646706, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1499, "step": 8296 }, { "epoch": 12.42065868263473, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1431, "step": 8297 }, { "epoch": 12.422155688622755, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1434, "step": 8298 }, { "epoch": 12.423652694610778, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1402, "step": 8299 }, { "epoch": 12.425149700598803, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1438, "step": 8300 }, { "epoch": 12.426646706586826, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.138, "step": 8301 }, { "epoch": 12.428143712574851, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1423, "step": 8302 }, { "epoch": 12.429640718562874, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1403, "step": 8303 }, { "epoch": 12.431137724550899, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1427, "step": 8304 }, { "epoch": 12.432634730538922, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1422, "step": 8305 }, { "epoch": 12.434131736526947, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.142, "step": 8306 }, { "epoch": 12.43562874251497, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.139, "step": 8307 }, { "epoch": 12.437125748502995, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1449, "step": 8308 }, { "epoch": 12.438622754491018, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1351, "step": 8309 }, { "epoch": 12.440119760479043, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1507, "step": 8310 }, { "epoch": 12.441616766467066, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1382, "step": 8311 }, { "epoch": 12.44311377245509, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1448, "step": 8312 }, { "epoch": 12.444610778443113, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1435, "step": 8313 }, { "epoch": 12.446107784431138, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.139, "step": 8314 }, { "epoch": 12.447604790419161, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1368, "step": 8315 }, { "epoch": 12.449101796407186, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1424, "step": 8316 }, { "epoch": 12.45059880239521, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1432, "step": 8317 }, { "epoch": 12.452095808383234, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1417, "step": 8318 }, { "epoch": 12.453592814371257, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1372, "step": 8319 }, { "epoch": 12.455089820359282, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1402, "step": 8320 }, { "epoch": 12.456586826347305, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1431, "step": 8321 }, { "epoch": 12.45808383233533, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1352, "step": 8322 }, { "epoch": 12.459580838323353, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1392, "step": 8323 }, { "epoch": 12.461077844311378, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1405, "step": 8324 }, { "epoch": 12.4625748502994, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1445, "step": 8325 }, { "epoch": 12.464071856287426, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1441, "step": 8326 }, { "epoch": 12.465568862275449, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1371, "step": 8327 }, { "epoch": 12.467065868263473, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1448, "step": 8328 }, { "epoch": 12.468562874251496, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1421, "step": 8329 }, { "epoch": 12.470059880239521, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1467, "step": 8330 }, { "epoch": 12.471556886227544, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1411, "step": 8331 }, { "epoch": 12.47305389221557, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1396, "step": 8332 }, { "epoch": 12.474550898203592, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1396, "step": 8333 }, { "epoch": 12.476047904191617, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1409, "step": 8334 }, { "epoch": 12.47754491017964, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1379, "step": 8335 }, { "epoch": 12.479041916167665, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1405, "step": 8336 }, { "epoch": 12.480538922155688, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1408, "step": 8337 }, { "epoch": 12.482035928143713, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1423, "step": 8338 }, { "epoch": 12.483532934131736, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1385, "step": 8339 }, { "epoch": 12.48502994011976, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1435, "step": 8340 }, { "epoch": 12.486526946107784, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1441, "step": 8341 }, { "epoch": 12.488023952095809, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1377, "step": 8342 }, { "epoch": 12.489520958083832, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1337, "step": 8343 }, { "epoch": 12.491017964071856, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1413, "step": 8344 }, { "epoch": 12.49251497005988, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1463, "step": 8345 }, { "epoch": 12.494011976047904, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1472, "step": 8346 }, { "epoch": 12.495508982035927, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1418, "step": 8347 }, { "epoch": 12.497005988023952, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1385, "step": 8348 }, { "epoch": 12.498502994011975, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1434, "step": 8349 }, { "epoch": 12.5, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1396, "step": 8350 }, { "epoch": 12.501497005988025, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.143, "step": 8351 }, { "epoch": 12.502994011976048, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1392, "step": 8352 }, { "epoch": 12.504491017964071, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1404, "step": 8353 }, { "epoch": 12.505988023952096, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1475, "step": 8354 }, { "epoch": 12.50748502994012, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1453, "step": 8355 }, { "epoch": 12.508982035928144, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1454, "step": 8356 }, { "epoch": 12.510479041916168, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.141, "step": 8357 }, { "epoch": 12.511976047904191, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1352, "step": 8358 }, { "epoch": 12.513473053892216, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.141, "step": 8359 }, { "epoch": 12.51497005988024, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.141, "step": 8360 }, { "epoch": 12.516467065868264, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1448, "step": 8361 }, { "epoch": 12.517964071856287, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1411, "step": 8362 }, { "epoch": 12.519461077844312, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1382, "step": 8363 }, { "epoch": 12.520958083832335, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1437, "step": 8364 }, { "epoch": 12.52245508982036, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.138, "step": 8365 }, { "epoch": 12.523952095808383, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1527, "step": 8366 }, { "epoch": 12.525449101796408, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1456, "step": 8367 }, { "epoch": 12.52694610778443, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1341, "step": 8368 }, { "epoch": 12.528443113772456, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1481, "step": 8369 }, { "epoch": 12.529940119760479, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1436, "step": 8370 }, { "epoch": 12.531437125748504, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1417, "step": 8371 }, { "epoch": 12.532934131736527, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1391, "step": 8372 }, { "epoch": 12.534431137724551, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1435, "step": 8373 }, { "epoch": 12.535928143712574, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1403, "step": 8374 }, { "epoch": 12.5374251497006, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1396, "step": 8375 }, { "epoch": 12.538922155688622, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1395, "step": 8376 }, { "epoch": 12.540419161676647, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1407, "step": 8377 }, { "epoch": 12.54191616766467, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1412, "step": 8378 }, { "epoch": 12.543413173652695, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1444, "step": 8379 }, { "epoch": 12.544910179640718, "grad_norm": 0.1591796875, "learning_rate": 0.0008, "loss": 1.144, "step": 8380 }, { "epoch": 12.546407185628743, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1431, "step": 8381 }, { "epoch": 12.547904191616766, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.14, "step": 8382 }, { "epoch": 12.54940119760479, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1418, "step": 8383 }, { "epoch": 12.550898203592814, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1361, "step": 8384 }, { "epoch": 12.552395209580839, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.135, "step": 8385 }, { "epoch": 12.553892215568862, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1402, "step": 8386 }, { "epoch": 12.555389221556887, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1476, "step": 8387 }, { "epoch": 12.55688622754491, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1391, "step": 8388 }, { "epoch": 12.558383233532934, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1294, "step": 8389 }, { "epoch": 12.559880239520957, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.145, "step": 8390 }, { "epoch": 12.561377245508982, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1455, "step": 8391 }, { "epoch": 12.562874251497005, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1433, "step": 8392 }, { "epoch": 12.56437125748503, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.143, "step": 8393 }, { "epoch": 12.565868263473053, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1375, "step": 8394 }, { "epoch": 12.567365269461078, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1389, "step": 8395 }, { "epoch": 12.568862275449101, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1468, "step": 8396 }, { "epoch": 12.570359281437126, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1454, "step": 8397 }, { "epoch": 12.571856287425149, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1437, "step": 8398 }, { "epoch": 12.573353293413174, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1445, "step": 8399 }, { "epoch": 12.574850299401197, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1412, "step": 8400 }, { "epoch": 12.576347305389222, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1451, "step": 8401 }, { "epoch": 12.577844311377245, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.141, "step": 8402 }, { "epoch": 12.57934131736527, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.146, "step": 8403 }, { "epoch": 12.580838323353294, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1471, "step": 8404 }, { "epoch": 12.582335329341317, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1366, "step": 8405 }, { "epoch": 12.58383233532934, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1405, "step": 8406 }, { "epoch": 12.585329341317365, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1442, "step": 8407 }, { "epoch": 12.58682634730539, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1509, "step": 8408 }, { "epoch": 12.588323353293413, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1415, "step": 8409 }, { "epoch": 12.589820359281438, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1466, "step": 8410 }, { "epoch": 12.591317365269461, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1434, "step": 8411 }, { "epoch": 12.592814371257486, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.138, "step": 8412 }, { "epoch": 12.594311377245509, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1438, "step": 8413 }, { "epoch": 12.595808383233534, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1414, "step": 8414 }, { "epoch": 12.597305389221557, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1445, "step": 8415 }, { "epoch": 12.598802395209582, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.138, "step": 8416 }, { "epoch": 12.600299401197605, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1428, "step": 8417 }, { "epoch": 12.60179640718563, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1388, "step": 8418 }, { "epoch": 12.603293413173652, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1445, "step": 8419 }, { "epoch": 12.604790419161677, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1425, "step": 8420 }, { "epoch": 12.6062874251497, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1418, "step": 8421 }, { "epoch": 12.607784431137725, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1454, "step": 8422 }, { "epoch": 12.609281437125748, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1411, "step": 8423 }, { "epoch": 12.610778443113773, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1463, "step": 8424 }, { "epoch": 12.612275449101796, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1443, "step": 8425 }, { "epoch": 12.613772455089821, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1424, "step": 8426 }, { "epoch": 12.615269461077844, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1458, "step": 8427 }, { "epoch": 12.616766467065869, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1453, "step": 8428 }, { "epoch": 12.618263473053892, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1413, "step": 8429 }, { "epoch": 12.619760479041917, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1446, "step": 8430 }, { "epoch": 12.62125748502994, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1389, "step": 8431 }, { "epoch": 12.622754491017965, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.144, "step": 8432 }, { "epoch": 12.624251497005988, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1363, "step": 8433 }, { "epoch": 12.625748502994012, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1362, "step": 8434 }, { "epoch": 12.627245508982035, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1474, "step": 8435 }, { "epoch": 12.62874251497006, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1376, "step": 8436 }, { "epoch": 12.630239520958083, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1344, "step": 8437 }, { "epoch": 12.631736526946108, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1365, "step": 8438 }, { "epoch": 12.633233532934131, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1435, "step": 8439 }, { "epoch": 12.634730538922156, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1356, "step": 8440 }, { "epoch": 12.636227544910179, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1417, "step": 8441 }, { "epoch": 12.637724550898204, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.142, "step": 8442 }, { "epoch": 12.639221556886227, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1395, "step": 8443 }, { "epoch": 12.640718562874252, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.134, "step": 8444 }, { "epoch": 12.642215568862275, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1432, "step": 8445 }, { "epoch": 12.6437125748503, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1367, "step": 8446 }, { "epoch": 12.645209580838323, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1418, "step": 8447 }, { "epoch": 12.646706586826348, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1467, "step": 8448 }, { "epoch": 12.64820359281437, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.134, "step": 8449 }, { "epoch": 12.649700598802395, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1382, "step": 8450 }, { "epoch": 12.651197604790418, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1374, "step": 8451 }, { "epoch": 12.652694610778443, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1363, "step": 8452 }, { "epoch": 12.654191616766466, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1405, "step": 8453 }, { "epoch": 12.655688622754491, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1414, "step": 8454 }, { "epoch": 12.657185628742514, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1386, "step": 8455 }, { "epoch": 12.658682634730539, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1396, "step": 8456 }, { "epoch": 12.660179640718562, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1448, "step": 8457 }, { "epoch": 12.661676646706587, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1399, "step": 8458 }, { "epoch": 12.66317365269461, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1412, "step": 8459 }, { "epoch": 12.664670658682635, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1446, "step": 8460 }, { "epoch": 12.66616766467066, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1436, "step": 8461 }, { "epoch": 12.667664670658683, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1411, "step": 8462 }, { "epoch": 12.669161676646706, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.138, "step": 8463 }, { "epoch": 12.67065868263473, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1301, "step": 8464 }, { "epoch": 12.672155688622755, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1461, "step": 8465 }, { "epoch": 12.673652694610778, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1428, "step": 8466 }, { "epoch": 12.675149700598803, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1351, "step": 8467 }, { "epoch": 12.676646706586826, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1416, "step": 8468 }, { "epoch": 12.678143712574851, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1381, "step": 8469 }, { "epoch": 12.679640718562874, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1475, "step": 8470 }, { "epoch": 12.681137724550899, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1373, "step": 8471 }, { "epoch": 12.682634730538922, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1395, "step": 8472 }, { "epoch": 12.684131736526947, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1393, "step": 8473 }, { "epoch": 12.68562874251497, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1418, "step": 8474 }, { "epoch": 12.687125748502995, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1401, "step": 8475 }, { "epoch": 12.688622754491018, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1411, "step": 8476 }, { "epoch": 12.690119760479043, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1452, "step": 8477 }, { "epoch": 12.691616766467066, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.14, "step": 8478 }, { "epoch": 12.69311377245509, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1327, "step": 8479 }, { "epoch": 12.694610778443113, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1388, "step": 8480 }, { "epoch": 12.696107784431138, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1448, "step": 8481 }, { "epoch": 12.697604790419161, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1366, "step": 8482 }, { "epoch": 12.699101796407186, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.139, "step": 8483 }, { "epoch": 12.70059880239521, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1427, "step": 8484 }, { "epoch": 12.702095808383234, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1476, "step": 8485 }, { "epoch": 12.703592814371257, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1465, "step": 8486 }, { "epoch": 12.705089820359282, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1419, "step": 8487 }, { "epoch": 12.706586826347305, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.141, "step": 8488 }, { "epoch": 12.70808383233533, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1466, "step": 8489 }, { "epoch": 12.709580838323353, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.145, "step": 8490 }, { "epoch": 12.711077844311378, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1448, "step": 8491 }, { "epoch": 12.7125748502994, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1408, "step": 8492 }, { "epoch": 12.714071856287426, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1419, "step": 8493 }, { "epoch": 12.715568862275449, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1346, "step": 8494 }, { "epoch": 12.717065868263473, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1464, "step": 8495 }, { "epoch": 12.718562874251496, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1357, "step": 8496 }, { "epoch": 12.720059880239521, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1502, "step": 8497 }, { "epoch": 12.721556886227544, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.142, "step": 8498 }, { "epoch": 12.72305389221557, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1436, "step": 8499 }, { "epoch": 12.724550898203592, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1461, "step": 8500 }, { "epoch": 12.726047904191617, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1442, "step": 8501 }, { "epoch": 12.72754491017964, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.143, "step": 8502 }, { "epoch": 12.729041916167665, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1386, "step": 8503 }, { "epoch": 12.730538922155688, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1308, "step": 8504 }, { "epoch": 12.732035928143713, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1455, "step": 8505 }, { "epoch": 12.733532934131736, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1411, "step": 8506 }, { "epoch": 12.73502994011976, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1439, "step": 8507 }, { "epoch": 12.736526946107784, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1347, "step": 8508 }, { "epoch": 12.738023952095809, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1437, "step": 8509 }, { "epoch": 12.739520958083832, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1415, "step": 8510 }, { "epoch": 12.741017964071856, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1406, "step": 8511 }, { "epoch": 12.74251497005988, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1391, "step": 8512 }, { "epoch": 12.744011976047904, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1401, "step": 8513 }, { "epoch": 12.745508982035929, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.138, "step": 8514 }, { "epoch": 12.747005988023952, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1447, "step": 8515 }, { "epoch": 12.748502994011975, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1455, "step": 8516 }, { "epoch": 12.75, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1425, "step": 8517 }, { "epoch": 12.751497005988025, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1437, "step": 8518 }, { "epoch": 12.752994011976048, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1361, "step": 8519 }, { "epoch": 12.754491017964071, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1411, "step": 8520 }, { "epoch": 12.755988023952096, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1402, "step": 8521 }, { "epoch": 12.75748502994012, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1385, "step": 8522 }, { "epoch": 12.758982035928144, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1383, "step": 8523 }, { "epoch": 12.760479041916168, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1355, "step": 8524 }, { "epoch": 12.761976047904191, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1316, "step": 8525 }, { "epoch": 12.763473053892216, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1385, "step": 8526 }, { "epoch": 12.76497005988024, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1481, "step": 8527 }, { "epoch": 12.766467065868264, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1398, "step": 8528 }, { "epoch": 12.767964071856287, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1421, "step": 8529 }, { "epoch": 12.769461077844312, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1417, "step": 8530 }, { "epoch": 12.770958083832335, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1402, "step": 8531 }, { "epoch": 12.77245508982036, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1391, "step": 8532 }, { "epoch": 12.773952095808383, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1433, "step": 8533 }, { "epoch": 12.775449101796408, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1412, "step": 8534 }, { "epoch": 12.77694610778443, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1463, "step": 8535 }, { "epoch": 12.778443113772456, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1458, "step": 8536 }, { "epoch": 12.779940119760479, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1471, "step": 8537 }, { "epoch": 12.781437125748504, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1428, "step": 8538 }, { "epoch": 12.782934131736527, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1376, "step": 8539 }, { "epoch": 12.784431137724551, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.145, "step": 8540 }, { "epoch": 12.785928143712574, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1417, "step": 8541 }, { "epoch": 12.7874251497006, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1463, "step": 8542 }, { "epoch": 12.788922155688622, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1441, "step": 8543 }, { "epoch": 12.790419161676647, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1438, "step": 8544 }, { "epoch": 12.79191616766467, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1455, "step": 8545 }, { "epoch": 12.793413173652695, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1404, "step": 8546 }, { "epoch": 12.794910179640718, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1438, "step": 8547 }, { "epoch": 12.796407185628743, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1413, "step": 8548 }, { "epoch": 12.797904191616766, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1483, "step": 8549 }, { "epoch": 12.79940119760479, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1442, "step": 8550 }, { "epoch": 12.800898203592814, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1438, "step": 8551 }, { "epoch": 12.802395209580839, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1399, "step": 8552 }, { "epoch": 12.803892215568862, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1382, "step": 8553 }, { "epoch": 12.805389221556887, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1365, "step": 8554 }, { "epoch": 12.80688622754491, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.137, "step": 8555 }, { "epoch": 12.808383233532934, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1467, "step": 8556 }, { "epoch": 12.809880239520957, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.144, "step": 8557 }, { "epoch": 12.811377245508982, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1455, "step": 8558 }, { "epoch": 12.812874251497005, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1424, "step": 8559 }, { "epoch": 12.81437125748503, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1362, "step": 8560 }, { "epoch": 12.815868263473053, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1438, "step": 8561 }, { "epoch": 12.817365269461078, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1424, "step": 8562 }, { "epoch": 12.818862275449101, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1433, "step": 8563 }, { "epoch": 12.820359281437126, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1336, "step": 8564 }, { "epoch": 12.821856287425149, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1424, "step": 8565 }, { "epoch": 12.823353293413174, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1461, "step": 8566 }, { "epoch": 12.824850299401197, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1418, "step": 8567 }, { "epoch": 12.826347305389222, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1427, "step": 8568 }, { "epoch": 12.827844311377245, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1476, "step": 8569 }, { "epoch": 12.82934131736527, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1414, "step": 8570 }, { "epoch": 12.830838323353294, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1363, "step": 8571 }, { "epoch": 12.832335329341317, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1389, "step": 8572 }, { "epoch": 12.83383233532934, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1297, "step": 8573 }, { "epoch": 12.835329341317365, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1474, "step": 8574 }, { "epoch": 12.83682634730539, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1406, "step": 8575 }, { "epoch": 12.838323353293413, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1374, "step": 8576 }, { "epoch": 12.839820359281438, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1414, "step": 8577 }, { "epoch": 12.841317365269461, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1423, "step": 8578 }, { "epoch": 12.842814371257486, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1406, "step": 8579 }, { "epoch": 12.844311377245509, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1402, "step": 8580 }, { "epoch": 12.845808383233534, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1475, "step": 8581 }, { "epoch": 12.847305389221557, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1451, "step": 8582 }, { "epoch": 12.848802395209582, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.1442, "step": 8583 }, { "epoch": 12.850299401197605, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1408, "step": 8584 }, { "epoch": 12.85179640718563, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1384, "step": 8585 }, { "epoch": 12.853293413173652, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1411, "step": 8586 }, { "epoch": 12.854790419161677, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1399, "step": 8587 }, { "epoch": 12.8562874251497, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1486, "step": 8588 }, { "epoch": 12.857784431137725, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1376, "step": 8589 }, { "epoch": 12.859281437125748, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1443, "step": 8590 }, { "epoch": 12.860778443113773, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1394, "step": 8591 }, { "epoch": 12.862275449101796, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1389, "step": 8592 }, { "epoch": 12.863772455089821, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1445, "step": 8593 }, { "epoch": 12.865269461077844, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1446, "step": 8594 }, { "epoch": 12.866766467065869, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1384, "step": 8595 }, { "epoch": 12.868263473053892, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1416, "step": 8596 }, { "epoch": 12.869760479041917, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.14, "step": 8597 }, { "epoch": 12.87125748502994, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1387, "step": 8598 }, { "epoch": 12.872754491017965, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1439, "step": 8599 }, { "epoch": 12.874251497005988, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1384, "step": 8600 }, { "epoch": 12.875748502994012, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1417, "step": 8601 }, { "epoch": 12.877245508982035, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1374, "step": 8602 }, { "epoch": 12.87874251497006, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1381, "step": 8603 }, { "epoch": 12.880239520958083, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1453, "step": 8604 }, { "epoch": 12.881736526946108, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1418, "step": 8605 }, { "epoch": 12.883233532934131, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1428, "step": 8606 }, { "epoch": 12.884730538922156, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1455, "step": 8607 }, { "epoch": 12.886227544910179, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1406, "step": 8608 }, { "epoch": 12.887724550898204, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1401, "step": 8609 }, { "epoch": 12.889221556886227, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1298, "step": 8610 }, { "epoch": 12.890718562874252, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1399, "step": 8611 }, { "epoch": 12.892215568862275, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.142, "step": 8612 }, { "epoch": 12.8937125748503, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1334, "step": 8613 }, { "epoch": 12.895209580838323, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1456, "step": 8614 }, { "epoch": 12.896706586826348, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1378, "step": 8615 }, { "epoch": 12.89820359281437, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1411, "step": 8616 }, { "epoch": 12.899700598802395, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1372, "step": 8617 }, { "epoch": 12.901197604790418, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1396, "step": 8618 }, { "epoch": 12.902694610778443, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1336, "step": 8619 }, { "epoch": 12.904191616766466, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1366, "step": 8620 }, { "epoch": 12.905688622754491, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1431, "step": 8621 }, { "epoch": 12.907185628742514, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1394, "step": 8622 }, { "epoch": 12.908682634730539, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1425, "step": 8623 }, { "epoch": 12.910179640718562, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.143, "step": 8624 }, { "epoch": 12.911676646706587, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1388, "step": 8625 }, { "epoch": 12.91317365269461, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1402, "step": 8626 }, { "epoch": 12.914670658682635, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1302, "step": 8627 }, { "epoch": 12.91616766467066, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1358, "step": 8628 }, { "epoch": 12.917664670658683, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1443, "step": 8629 }, { "epoch": 12.919161676646706, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1429, "step": 8630 }, { "epoch": 12.92065868263473, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.145, "step": 8631 }, { "epoch": 12.922155688622755, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1471, "step": 8632 }, { "epoch": 12.923652694610778, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1492, "step": 8633 }, { "epoch": 12.925149700598803, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1409, "step": 8634 }, { "epoch": 12.926646706586826, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1439, "step": 8635 }, { "epoch": 12.928143712574851, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1356, "step": 8636 }, { "epoch": 12.929640718562874, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1394, "step": 8637 }, { "epoch": 12.931137724550899, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1372, "step": 8638 }, { "epoch": 12.932634730538922, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.143, "step": 8639 }, { "epoch": 12.934131736526947, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1369, "step": 8640 }, { "epoch": 12.93562874251497, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1408, "step": 8641 }, { "epoch": 12.937125748502995, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1389, "step": 8642 }, { "epoch": 12.938622754491018, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1373, "step": 8643 }, { "epoch": 12.940119760479043, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1354, "step": 8644 }, { "epoch": 12.941616766467066, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1349, "step": 8645 }, { "epoch": 12.94311377245509, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1423, "step": 8646 }, { "epoch": 12.944610778443113, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1431, "step": 8647 }, { "epoch": 12.946107784431138, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1411, "step": 8648 }, { "epoch": 12.947604790419161, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.134, "step": 8649 }, { "epoch": 12.949101796407186, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1391, "step": 8650 }, { "epoch": 12.95059880239521, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1443, "step": 8651 }, { "epoch": 12.952095808383234, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1376, "step": 8652 }, { "epoch": 12.953592814371257, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1392, "step": 8653 }, { "epoch": 12.955089820359282, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1451, "step": 8654 }, { "epoch": 12.956586826347305, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1428, "step": 8655 }, { "epoch": 12.95808383233533, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1423, "step": 8656 }, { "epoch": 12.959580838323353, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1435, "step": 8657 }, { "epoch": 12.961077844311378, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1397, "step": 8658 }, { "epoch": 12.9625748502994, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.136, "step": 8659 }, { "epoch": 12.964071856287426, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1386, "step": 8660 }, { "epoch": 12.965568862275449, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1405, "step": 8661 }, { "epoch": 12.967065868263473, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1443, "step": 8662 }, { "epoch": 12.968562874251496, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1389, "step": 8663 }, { "epoch": 12.970059880239521, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1429, "step": 8664 }, { "epoch": 12.971556886227544, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1407, "step": 8665 }, { "epoch": 12.97305389221557, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1421, "step": 8666 }, { "epoch": 12.974550898203592, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1374, "step": 8667 }, { "epoch": 12.976047904191617, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1372, "step": 8668 }, { "epoch": 12.97754491017964, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 8669 }, { "epoch": 12.979041916167665, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1406, "step": 8670 }, { "epoch": 12.980538922155688, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1436, "step": 8671 }, { "epoch": 12.982035928143713, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1382, "step": 8672 }, { "epoch": 12.983532934131736, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1442, "step": 8673 }, { "epoch": 12.98502994011976, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1443, "step": 8674 }, { "epoch": 12.986526946107784, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1416, "step": 8675 }, { "epoch": 12.988023952095809, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.143, "step": 8676 }, { "epoch": 12.989520958083832, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1317, "step": 8677 }, { "epoch": 12.991017964071856, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1388, "step": 8678 }, { "epoch": 12.99251497005988, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1426, "step": 8679 }, { "epoch": 12.994011976047904, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.146, "step": 8680 }, { "epoch": 12.995508982035929, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1354, "step": 8681 }, { "epoch": 12.997005988023952, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1413, "step": 8682 }, { "epoch": 12.998502994011975, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1402, "step": 8683 }, { "epoch": 13.0, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1391, "step": 8684 }, { "epoch": 13.001497005988025, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1427, "step": 8685 }, { "epoch": 13.002994011976048, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1372, "step": 8686 }, { "epoch": 13.004491017964073, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1372, "step": 8687 }, { "epoch": 13.005988023952096, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1409, "step": 8688 }, { "epoch": 13.00748502994012, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1393, "step": 8689 }, { "epoch": 13.008982035928144, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1348, "step": 8690 }, { "epoch": 13.010479041916168, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1418, "step": 8691 }, { "epoch": 13.011976047904191, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1423, "step": 8692 }, { "epoch": 13.013473053892216, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1388, "step": 8693 }, { "epoch": 13.01497005988024, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1395, "step": 8694 }, { "epoch": 13.016467065868264, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.144, "step": 8695 }, { "epoch": 13.017964071856287, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1438, "step": 8696 }, { "epoch": 13.019461077844312, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.137, "step": 8697 }, { "epoch": 13.020958083832335, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1331, "step": 8698 }, { "epoch": 13.02245508982036, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1335, "step": 8699 }, { "epoch": 13.023952095808383, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1447, "step": 8700 }, { "epoch": 13.025449101796408, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1335, "step": 8701 }, { "epoch": 13.02694610778443, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.139, "step": 8702 }, { "epoch": 13.028443113772456, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1396, "step": 8703 }, { "epoch": 13.029940119760479, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1361, "step": 8704 }, { "epoch": 13.031437125748504, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1388, "step": 8705 }, { "epoch": 13.032934131736527, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1428, "step": 8706 }, { "epoch": 13.034431137724551, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1436, "step": 8707 }, { "epoch": 13.035928143712574, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1409, "step": 8708 }, { "epoch": 13.0374251497006, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1415, "step": 8709 }, { "epoch": 13.038922155688622, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1344, "step": 8710 }, { "epoch": 13.040419161676647, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1349, "step": 8711 }, { "epoch": 13.04191616766467, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1392, "step": 8712 }, { "epoch": 13.043413173652695, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1412, "step": 8713 }, { "epoch": 13.044910179640718, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1436, "step": 8714 }, { "epoch": 13.046407185628743, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1358, "step": 8715 }, { "epoch": 13.047904191616766, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1394, "step": 8716 }, { "epoch": 13.04940119760479, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1382, "step": 8717 }, { "epoch": 13.050898203592814, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1408, "step": 8718 }, { "epoch": 13.052395209580839, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1407, "step": 8719 }, { "epoch": 13.053892215568862, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1442, "step": 8720 }, { "epoch": 13.055389221556887, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1418, "step": 8721 }, { "epoch": 13.05688622754491, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1354, "step": 8722 }, { "epoch": 13.058383233532934, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1411, "step": 8723 }, { "epoch": 13.059880239520957, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.14, "step": 8724 }, { "epoch": 13.061377245508982, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.145, "step": 8725 }, { "epoch": 13.062874251497005, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1421, "step": 8726 }, { "epoch": 13.06437125748503, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1463, "step": 8727 }, { "epoch": 13.065868263473053, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1319, "step": 8728 }, { "epoch": 13.067365269461078, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1407, "step": 8729 }, { "epoch": 13.068862275449101, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.133, "step": 8730 }, { "epoch": 13.070359281437126, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1369, "step": 8731 }, { "epoch": 13.071856287425149, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1447, "step": 8732 }, { "epoch": 13.073353293413174, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1388, "step": 8733 }, { "epoch": 13.074850299401197, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1469, "step": 8734 }, { "epoch": 13.076347305389222, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1368, "step": 8735 }, { "epoch": 13.077844311377245, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1459, "step": 8736 }, { "epoch": 13.07934131736527, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.142, "step": 8737 }, { "epoch": 13.080838323353293, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1417, "step": 8738 }, { "epoch": 13.082335329341317, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1403, "step": 8739 }, { "epoch": 13.08383233532934, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1359, "step": 8740 }, { "epoch": 13.085329341317365, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1422, "step": 8741 }, { "epoch": 13.08682634730539, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1406, "step": 8742 }, { "epoch": 13.088323353293413, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1382, "step": 8743 }, { "epoch": 13.089820359281438, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1471, "step": 8744 }, { "epoch": 13.091317365269461, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1358, "step": 8745 }, { "epoch": 13.092814371257486, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1419, "step": 8746 }, { "epoch": 13.094311377245509, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1361, "step": 8747 }, { "epoch": 13.095808383233534, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1357, "step": 8748 }, { "epoch": 13.097305389221557, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1378, "step": 8749 }, { "epoch": 13.098802395209582, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1374, "step": 8750 }, { "epoch": 13.100299401197605, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1357, "step": 8751 }, { "epoch": 13.10179640718563, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1464, "step": 8752 }, { "epoch": 13.103293413173652, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1409, "step": 8753 }, { "epoch": 13.104790419161677, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1351, "step": 8754 }, { "epoch": 13.1062874251497, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1355, "step": 8755 }, { "epoch": 13.107784431137725, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1343, "step": 8756 }, { "epoch": 13.109281437125748, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1341, "step": 8757 }, { "epoch": 13.110778443113773, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1374, "step": 8758 }, { "epoch": 13.112275449101796, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1476, "step": 8759 }, { "epoch": 13.113772455089821, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1398, "step": 8760 }, { "epoch": 13.115269461077844, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1417, "step": 8761 }, { "epoch": 13.116766467065869, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1419, "step": 8762 }, { "epoch": 13.118263473053892, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1484, "step": 8763 }, { "epoch": 13.119760479041917, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1444, "step": 8764 }, { "epoch": 13.12125748502994, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1443, "step": 8765 }, { "epoch": 13.122754491017965, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1401, "step": 8766 }, { "epoch": 13.124251497005988, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1385, "step": 8767 }, { "epoch": 13.125748502994012, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1418, "step": 8768 }, { "epoch": 13.127245508982035, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1378, "step": 8769 }, { "epoch": 13.12874251497006, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1327, "step": 8770 }, { "epoch": 13.130239520958083, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1411, "step": 8771 }, { "epoch": 13.131736526946108, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.135, "step": 8772 }, { "epoch": 13.133233532934131, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1408, "step": 8773 }, { "epoch": 13.134730538922156, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1457, "step": 8774 }, { "epoch": 13.136227544910179, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1448, "step": 8775 }, { "epoch": 13.137724550898204, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1364, "step": 8776 }, { "epoch": 13.139221556886227, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1342, "step": 8777 }, { "epoch": 13.140718562874252, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1362, "step": 8778 }, { "epoch": 13.142215568862275, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1402, "step": 8779 }, { "epoch": 13.1437125748503, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1357, "step": 8780 }, { "epoch": 13.145209580838323, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1392, "step": 8781 }, { "epoch": 13.146706586826348, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1464, "step": 8782 }, { "epoch": 13.14820359281437, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1457, "step": 8783 }, { "epoch": 13.149700598802395, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.136, "step": 8784 }, { "epoch": 13.151197604790418, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1413, "step": 8785 }, { "epoch": 13.152694610778443, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1405, "step": 8786 }, { "epoch": 13.154191616766466, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1366, "step": 8787 }, { "epoch": 13.155688622754491, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.141, "step": 8788 }, { "epoch": 13.157185628742514, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1407, "step": 8789 }, { "epoch": 13.158682634730539, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.144, "step": 8790 }, { "epoch": 13.160179640718562, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.135, "step": 8791 }, { "epoch": 13.161676646706587, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1398, "step": 8792 }, { "epoch": 13.16317365269461, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1454, "step": 8793 }, { "epoch": 13.164670658682635, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1347, "step": 8794 }, { "epoch": 13.16616766467066, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1439, "step": 8795 }, { "epoch": 13.167664670658683, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1399, "step": 8796 }, { "epoch": 13.169161676646707, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1397, "step": 8797 }, { "epoch": 13.17065868263473, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1344, "step": 8798 }, { "epoch": 13.172155688622755, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1415, "step": 8799 }, { "epoch": 13.173652694610778, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1421, "step": 8800 }, { "epoch": 13.175149700598803, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1411, "step": 8801 }, { "epoch": 13.176646706586826, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1387, "step": 8802 }, { "epoch": 13.178143712574851, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1412, "step": 8803 }, { "epoch": 13.179640718562874, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1409, "step": 8804 }, { "epoch": 13.181137724550899, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1348, "step": 8805 }, { "epoch": 13.182634730538922, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1387, "step": 8806 }, { "epoch": 13.184131736526947, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1452, "step": 8807 }, { "epoch": 13.18562874251497, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.139, "step": 8808 }, { "epoch": 13.187125748502995, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1376, "step": 8809 }, { "epoch": 13.188622754491018, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1389, "step": 8810 }, { "epoch": 13.190119760479043, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1389, "step": 8811 }, { "epoch": 13.191616766467066, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1312, "step": 8812 }, { "epoch": 13.19311377245509, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1359, "step": 8813 }, { "epoch": 13.194610778443113, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 8814 }, { "epoch": 13.196107784431138, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1413, "step": 8815 }, { "epoch": 13.197604790419161, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1405, "step": 8816 }, { "epoch": 13.199101796407186, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.141, "step": 8817 }, { "epoch": 13.20059880239521, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1407, "step": 8818 }, { "epoch": 13.202095808383234, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1378, "step": 8819 }, { "epoch": 13.203592814371257, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1385, "step": 8820 }, { "epoch": 13.205089820359282, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.14, "step": 8821 }, { "epoch": 13.206586826347305, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1367, "step": 8822 }, { "epoch": 13.20808383233533, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1454, "step": 8823 }, { "epoch": 13.209580838323353, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1434, "step": 8824 }, { "epoch": 13.211077844311378, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1381, "step": 8825 }, { "epoch": 13.2125748502994, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.142, "step": 8826 }, { "epoch": 13.214071856287426, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1437, "step": 8827 }, { "epoch": 13.215568862275449, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1345, "step": 8828 }, { "epoch": 13.217065868263473, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1483, "step": 8829 }, { "epoch": 13.218562874251496, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1409, "step": 8830 }, { "epoch": 13.220059880239521, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1336, "step": 8831 }, { "epoch": 13.221556886227544, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.143, "step": 8832 }, { "epoch": 13.22305389221557, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1414, "step": 8833 }, { "epoch": 13.224550898203592, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1399, "step": 8834 }, { "epoch": 13.226047904191617, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1387, "step": 8835 }, { "epoch": 13.22754491017964, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1398, "step": 8836 }, { "epoch": 13.229041916167665, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1478, "step": 8837 }, { "epoch": 13.230538922155688, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.14, "step": 8838 }, { "epoch": 13.232035928143713, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.145, "step": 8839 }, { "epoch": 13.233532934131736, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1383, "step": 8840 }, { "epoch": 13.23502994011976, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1404, "step": 8841 }, { "epoch": 13.236526946107784, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1438, "step": 8842 }, { "epoch": 13.238023952095809, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.139, "step": 8843 }, { "epoch": 13.239520958083832, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1393, "step": 8844 }, { "epoch": 13.241017964071856, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1376, "step": 8845 }, { "epoch": 13.24251497005988, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1386, "step": 8846 }, { "epoch": 13.244011976047904, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1391, "step": 8847 }, { "epoch": 13.245508982035927, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1481, "step": 8848 }, { "epoch": 13.247005988023952, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1423, "step": 8849 }, { "epoch": 13.248502994011975, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1423, "step": 8850 }, { "epoch": 13.25, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1403, "step": 8851 }, { "epoch": 13.251497005988025, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1409, "step": 8852 }, { "epoch": 13.252994011976048, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1323, "step": 8853 }, { "epoch": 13.254491017964073, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1398, "step": 8854 }, { "epoch": 13.255988023952096, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1371, "step": 8855 }, { "epoch": 13.25748502994012, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1433, "step": 8856 }, { "epoch": 13.258982035928144, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1388, "step": 8857 }, { "epoch": 13.260479041916168, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1437, "step": 8858 }, { "epoch": 13.261976047904191, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1389, "step": 8859 }, { "epoch": 13.263473053892216, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1389, "step": 8860 }, { "epoch": 13.26497005988024, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1354, "step": 8861 }, { "epoch": 13.266467065868264, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1382, "step": 8862 }, { "epoch": 13.267964071856287, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1429, "step": 8863 }, { "epoch": 13.269461077844312, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.135, "step": 8864 }, { "epoch": 13.270958083832335, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1388, "step": 8865 }, { "epoch": 13.27245508982036, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1474, "step": 8866 }, { "epoch": 13.273952095808383, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.143, "step": 8867 }, { "epoch": 13.275449101796408, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1478, "step": 8868 }, { "epoch": 13.27694610778443, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.134, "step": 8869 }, { "epoch": 13.278443113772456, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1431, "step": 8870 }, { "epoch": 13.279940119760479, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.141, "step": 8871 }, { "epoch": 13.281437125748504, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1327, "step": 8872 }, { "epoch": 13.282934131736527, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.132, "step": 8873 }, { "epoch": 13.284431137724551, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1346, "step": 8874 }, { "epoch": 13.285928143712574, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.136, "step": 8875 }, { "epoch": 13.2874251497006, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1342, "step": 8876 }, { "epoch": 13.288922155688622, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.134, "step": 8877 }, { "epoch": 13.290419161676647, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1415, "step": 8878 }, { "epoch": 13.29191616766467, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1398, "step": 8879 }, { "epoch": 13.293413173652695, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.141, "step": 8880 }, { "epoch": 13.294910179640718, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1385, "step": 8881 }, { "epoch": 13.296407185628743, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1432, "step": 8882 }, { "epoch": 13.297904191616766, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1372, "step": 8883 }, { "epoch": 13.29940119760479, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1414, "step": 8884 }, { "epoch": 13.300898203592814, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1335, "step": 8885 }, { "epoch": 13.302395209580839, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1425, "step": 8886 }, { "epoch": 13.303892215568862, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.146, "step": 8887 }, { "epoch": 13.305389221556887, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1407, "step": 8888 }, { "epoch": 13.30688622754491, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1381, "step": 8889 }, { "epoch": 13.308383233532934, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1421, "step": 8890 }, { "epoch": 13.309880239520957, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1487, "step": 8891 }, { "epoch": 13.311377245508982, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1403, "step": 8892 }, { "epoch": 13.312874251497005, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1414, "step": 8893 }, { "epoch": 13.31437125748503, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1319, "step": 8894 }, { "epoch": 13.315868263473053, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1294, "step": 8895 }, { "epoch": 13.317365269461078, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1409, "step": 8896 }, { "epoch": 13.318862275449101, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1371, "step": 8897 }, { "epoch": 13.320359281437126, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1472, "step": 8898 }, { "epoch": 13.321856287425149, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1407, "step": 8899 }, { "epoch": 13.323353293413174, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.135, "step": 8900 }, { "epoch": 13.324850299401197, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1419, "step": 8901 }, { "epoch": 13.326347305389222, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1421, "step": 8902 }, { "epoch": 13.327844311377245, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1417, "step": 8903 }, { "epoch": 13.32934131736527, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1367, "step": 8904 }, { "epoch": 13.330838323353294, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1405, "step": 8905 }, { "epoch": 13.332335329341317, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.138, "step": 8906 }, { "epoch": 13.33383233532934, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1362, "step": 8907 }, { "epoch": 13.335329341317365, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1487, "step": 8908 }, { "epoch": 13.33682634730539, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.139, "step": 8909 }, { "epoch": 13.338323353293413, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1422, "step": 8910 }, { "epoch": 13.339820359281438, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1438, "step": 8911 }, { "epoch": 13.341317365269461, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.143, "step": 8912 }, { "epoch": 13.342814371257486, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1378, "step": 8913 }, { "epoch": 13.344311377245509, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1399, "step": 8914 }, { "epoch": 13.345808383233534, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1362, "step": 8915 }, { "epoch": 13.347305389221557, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1395, "step": 8916 }, { "epoch": 13.348802395209582, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1458, "step": 8917 }, { "epoch": 13.350299401197605, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1423, "step": 8918 }, { "epoch": 13.35179640718563, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1411, "step": 8919 }, { "epoch": 13.353293413173652, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1411, "step": 8920 }, { "epoch": 13.354790419161677, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1462, "step": 8921 }, { "epoch": 13.3562874251497, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.141, "step": 8922 }, { "epoch": 13.357784431137725, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1359, "step": 8923 }, { "epoch": 13.359281437125748, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1364, "step": 8924 }, { "epoch": 13.360778443113773, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1318, "step": 8925 }, { "epoch": 13.362275449101796, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1377, "step": 8926 }, { "epoch": 13.363772455089821, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1357, "step": 8927 }, { "epoch": 13.365269461077844, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1399, "step": 8928 }, { "epoch": 13.366766467065869, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1458, "step": 8929 }, { "epoch": 13.368263473053892, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1484, "step": 8930 }, { "epoch": 13.369760479041917, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1444, "step": 8931 }, { "epoch": 13.37125748502994, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1393, "step": 8932 }, { "epoch": 13.372754491017965, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1404, "step": 8933 }, { "epoch": 13.374251497005988, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1364, "step": 8934 }, { "epoch": 13.375748502994012, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1399, "step": 8935 }, { "epoch": 13.377245508982035, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1353, "step": 8936 }, { "epoch": 13.37874251497006, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1434, "step": 8937 }, { "epoch": 13.380239520958083, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1325, "step": 8938 }, { "epoch": 13.381736526946108, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1466, "step": 8939 }, { "epoch": 13.383233532934131, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1382, "step": 8940 }, { "epoch": 13.384730538922156, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1383, "step": 8941 }, { "epoch": 13.386227544910179, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1404, "step": 8942 }, { "epoch": 13.387724550898204, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1375, "step": 8943 }, { "epoch": 13.389221556886227, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1373, "step": 8944 }, { "epoch": 13.390718562874252, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1419, "step": 8945 }, { "epoch": 13.392215568862275, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1397, "step": 8946 }, { "epoch": 13.3937125748503, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1387, "step": 8947 }, { "epoch": 13.395209580838323, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1376, "step": 8948 }, { "epoch": 13.396706586826348, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1434, "step": 8949 }, { "epoch": 13.39820359281437, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1439, "step": 8950 }, { "epoch": 13.399700598802395, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1467, "step": 8951 }, { "epoch": 13.401197604790418, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1418, "step": 8952 }, { "epoch": 13.402694610778443, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.147, "step": 8953 }, { "epoch": 13.404191616766466, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1434, "step": 8954 }, { "epoch": 13.405688622754491, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.139, "step": 8955 }, { "epoch": 13.407185628742514, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1394, "step": 8956 }, { "epoch": 13.408682634730539, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1311, "step": 8957 }, { "epoch": 13.410179640718562, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1445, "step": 8958 }, { "epoch": 13.411676646706587, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1366, "step": 8959 }, { "epoch": 13.41317365269461, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1447, "step": 8960 }, { "epoch": 13.414670658682635, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1388, "step": 8961 }, { "epoch": 13.41616766467066, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1424, "step": 8962 }, { "epoch": 13.417664670658683, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1366, "step": 8963 }, { "epoch": 13.419161676646706, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.131, "step": 8964 }, { "epoch": 13.42065868263473, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.138, "step": 8965 }, { "epoch": 13.422155688622755, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1377, "step": 8966 }, { "epoch": 13.423652694610778, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1436, "step": 8967 }, { "epoch": 13.425149700598803, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1352, "step": 8968 }, { "epoch": 13.426646706586826, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1355, "step": 8969 }, { "epoch": 13.428143712574851, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1357, "step": 8970 }, { "epoch": 13.429640718562874, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.135, "step": 8971 }, { "epoch": 13.431137724550899, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1388, "step": 8972 }, { "epoch": 13.432634730538922, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1444, "step": 8973 }, { "epoch": 13.434131736526947, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1372, "step": 8974 }, { "epoch": 13.43562874251497, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1467, "step": 8975 }, { "epoch": 13.437125748502995, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1367, "step": 8976 }, { "epoch": 13.438622754491018, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1357, "step": 8977 }, { "epoch": 13.440119760479043, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1374, "step": 8978 }, { "epoch": 13.441616766467066, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1427, "step": 8979 }, { "epoch": 13.44311377245509, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1401, "step": 8980 }, { "epoch": 13.444610778443113, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1377, "step": 8981 }, { "epoch": 13.446107784431138, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1372, "step": 8982 }, { "epoch": 13.447604790419161, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1404, "step": 8983 }, { "epoch": 13.449101796407186, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1363, "step": 8984 }, { "epoch": 13.45059880239521, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1459, "step": 8985 }, { "epoch": 13.452095808383234, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1397, "step": 8986 }, { "epoch": 13.453592814371257, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1467, "step": 8987 }, { "epoch": 13.455089820359282, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1404, "step": 8988 }, { "epoch": 13.456586826347305, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1438, "step": 8989 }, { "epoch": 13.45808383233533, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.144, "step": 8990 }, { "epoch": 13.459580838323353, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1395, "step": 8991 }, { "epoch": 13.461077844311378, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1374, "step": 8992 }, { "epoch": 13.4625748502994, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1432, "step": 8993 }, { "epoch": 13.464071856287426, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1448, "step": 8994 }, { "epoch": 13.465568862275449, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1426, "step": 8995 }, { "epoch": 13.467065868263473, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1377, "step": 8996 }, { "epoch": 13.468562874251496, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1382, "step": 8997 }, { "epoch": 13.470059880239521, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1382, "step": 8998 }, { "epoch": 13.471556886227544, "grad_norm": 0.048583984375, "learning_rate": 0.0008, "loss": 1.1429, "step": 8999 }, { "epoch": 13.47305389221557, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1345, "step": 9000 }, { "epoch": 13.474550898203592, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.142, "step": 9001 }, { "epoch": 13.476047904191617, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1375, "step": 9002 }, { "epoch": 13.47754491017964, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1432, "step": 9003 }, { "epoch": 13.479041916167665, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1427, "step": 9004 }, { "epoch": 13.480538922155688, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1382, "step": 9005 }, { "epoch": 13.482035928143713, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1405, "step": 9006 }, { "epoch": 13.483532934131736, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1338, "step": 9007 }, { "epoch": 13.48502994011976, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1382, "step": 9008 }, { "epoch": 13.486526946107784, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1339, "step": 9009 }, { "epoch": 13.488023952095809, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1412, "step": 9010 }, { "epoch": 13.489520958083832, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1413, "step": 9011 }, { "epoch": 13.491017964071856, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1409, "step": 9012 }, { "epoch": 13.49251497005988, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1424, "step": 9013 }, { "epoch": 13.494011976047904, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1367, "step": 9014 }, { "epoch": 13.495508982035927, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1448, "step": 9015 }, { "epoch": 13.497005988023952, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1413, "step": 9016 }, { "epoch": 13.498502994011975, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1353, "step": 9017 }, { "epoch": 13.5, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1412, "step": 9018 }, { "epoch": 13.501497005988025, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1346, "step": 9019 }, { "epoch": 13.502994011976048, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1366, "step": 9020 }, { "epoch": 13.504491017964071, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1426, "step": 9021 }, { "epoch": 13.505988023952096, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1357, "step": 9022 }, { "epoch": 13.50748502994012, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1359, "step": 9023 }, { "epoch": 13.508982035928144, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1344, "step": 9024 }, { "epoch": 13.510479041916168, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1389, "step": 9025 }, { "epoch": 13.511976047904191, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1425, "step": 9026 }, { "epoch": 13.513473053892216, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1412, "step": 9027 }, { "epoch": 13.51497005988024, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.145, "step": 9028 }, { "epoch": 13.516467065868264, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1372, "step": 9029 }, { "epoch": 13.517964071856287, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1379, "step": 9030 }, { "epoch": 13.519461077844312, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1431, "step": 9031 }, { "epoch": 13.520958083832335, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1395, "step": 9032 }, { "epoch": 13.52245508982036, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1416, "step": 9033 }, { "epoch": 13.523952095808383, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1431, "step": 9034 }, { "epoch": 13.525449101796408, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1433, "step": 9035 }, { "epoch": 13.52694610778443, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1373, "step": 9036 }, { "epoch": 13.528443113772456, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 9037 }, { "epoch": 13.529940119760479, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1388, "step": 9038 }, { "epoch": 13.531437125748504, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1408, "step": 9039 }, { "epoch": 13.532934131736527, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1405, "step": 9040 }, { "epoch": 13.534431137724551, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1342, "step": 9041 }, { "epoch": 13.535928143712574, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1409, "step": 9042 }, { "epoch": 13.5374251497006, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.147, "step": 9043 }, { "epoch": 13.538922155688622, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1399, "step": 9044 }, { "epoch": 13.540419161676647, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1449, "step": 9045 }, { "epoch": 13.54191616766467, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1365, "step": 9046 }, { "epoch": 13.543413173652695, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1342, "step": 9047 }, { "epoch": 13.544910179640718, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1391, "step": 9048 }, { "epoch": 13.546407185628743, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1349, "step": 9049 }, { "epoch": 13.547904191616766, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1386, "step": 9050 }, { "epoch": 13.54940119760479, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1356, "step": 9051 }, { "epoch": 13.550898203592814, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1397, "step": 9052 }, { "epoch": 13.552395209580839, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1293, "step": 9053 }, { "epoch": 13.553892215568862, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1358, "step": 9054 }, { "epoch": 13.555389221556887, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1461, "step": 9055 }, { "epoch": 13.55688622754491, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1394, "step": 9056 }, { "epoch": 13.558383233532934, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.141, "step": 9057 }, { "epoch": 13.559880239520957, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1443, "step": 9058 }, { "epoch": 13.561377245508982, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.1374, "step": 9059 }, { "epoch": 13.562874251497005, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1369, "step": 9060 }, { "epoch": 13.56437125748503, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1366, "step": 9061 }, { "epoch": 13.565868263473053, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1419, "step": 9062 }, { "epoch": 13.567365269461078, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.1416, "step": 9063 }, { "epoch": 13.568862275449101, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1385, "step": 9064 }, { "epoch": 13.570359281437126, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1344, "step": 9065 }, { "epoch": 13.571856287425149, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1365, "step": 9066 }, { "epoch": 13.573353293413174, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1385, "step": 9067 }, { "epoch": 13.574850299401197, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1364, "step": 9068 }, { "epoch": 13.576347305389222, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1373, "step": 9069 }, { "epoch": 13.577844311377245, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1343, "step": 9070 }, { "epoch": 13.57934131736527, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1374, "step": 9071 }, { "epoch": 13.580838323353294, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1323, "step": 9072 }, { "epoch": 13.582335329341317, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1439, "step": 9073 }, { "epoch": 13.58383233532934, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1356, "step": 9074 }, { "epoch": 13.585329341317365, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1368, "step": 9075 }, { "epoch": 13.58682634730539, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1382, "step": 9076 }, { "epoch": 13.588323353293413, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1335, "step": 9077 }, { "epoch": 13.589820359281438, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1403, "step": 9078 }, { "epoch": 13.591317365269461, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1423, "step": 9079 }, { "epoch": 13.592814371257486, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1428, "step": 9080 }, { "epoch": 13.594311377245509, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1397, "step": 9081 }, { "epoch": 13.595808383233534, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1374, "step": 9082 }, { "epoch": 13.597305389221557, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1375, "step": 9083 }, { "epoch": 13.598802395209582, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1389, "step": 9084 }, { "epoch": 13.600299401197605, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1442, "step": 9085 }, { "epoch": 13.60179640718563, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1447, "step": 9086 }, { "epoch": 13.603293413173652, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.13, "step": 9087 }, { "epoch": 13.604790419161677, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1404, "step": 9088 }, { "epoch": 13.6062874251497, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1377, "step": 9089 }, { "epoch": 13.607784431137725, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1353, "step": 9090 }, { "epoch": 13.609281437125748, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1343, "step": 9091 }, { "epoch": 13.610778443113773, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1405, "step": 9092 }, { "epoch": 13.612275449101796, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1389, "step": 9093 }, { "epoch": 13.613772455089821, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1392, "step": 9094 }, { "epoch": 13.615269461077844, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1427, "step": 9095 }, { "epoch": 13.616766467065869, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1312, "step": 9096 }, { "epoch": 13.618263473053892, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1372, "step": 9097 }, { "epoch": 13.619760479041917, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1424, "step": 9098 }, { "epoch": 13.62125748502994, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1381, "step": 9099 }, { "epoch": 13.622754491017965, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1408, "step": 9100 }, { "epoch": 13.624251497005988, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1409, "step": 9101 }, { "epoch": 13.625748502994012, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1415, "step": 9102 }, { "epoch": 13.627245508982035, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1378, "step": 9103 }, { "epoch": 13.62874251497006, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1338, "step": 9104 }, { "epoch": 13.630239520958083, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1357, "step": 9105 }, { "epoch": 13.631736526946108, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1424, "step": 9106 }, { "epoch": 13.633233532934131, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1394, "step": 9107 }, { "epoch": 13.634730538922156, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1405, "step": 9108 }, { "epoch": 13.636227544910179, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1417, "step": 9109 }, { "epoch": 13.637724550898204, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1362, "step": 9110 }, { "epoch": 13.639221556886227, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1328, "step": 9111 }, { "epoch": 13.640718562874252, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1367, "step": 9112 }, { "epoch": 13.642215568862275, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.137, "step": 9113 }, { "epoch": 13.6437125748503, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1384, "step": 9114 }, { "epoch": 13.645209580838323, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1402, "step": 9115 }, { "epoch": 13.646706586826348, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1428, "step": 9116 }, { "epoch": 13.64820359281437, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1361, "step": 9117 }, { "epoch": 13.649700598802395, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1413, "step": 9118 }, { "epoch": 13.651197604790418, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.137, "step": 9119 }, { "epoch": 13.652694610778443, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1358, "step": 9120 }, { "epoch": 13.654191616766466, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.143, "step": 9121 }, { "epoch": 13.655688622754491, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1365, "step": 9122 }, { "epoch": 13.657185628742514, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1383, "step": 9123 }, { "epoch": 13.658682634730539, "grad_norm": 0.1923828125, "learning_rate": 0.0008, "loss": 1.1396, "step": 9124 }, { "epoch": 13.660179640718562, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.1354, "step": 9125 }, { "epoch": 13.661676646706587, "grad_norm": 0.150390625, "learning_rate": 0.0008, "loss": 1.138, "step": 9126 }, { "epoch": 13.66317365269461, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1403, "step": 9127 }, { "epoch": 13.664670658682635, "grad_norm": 0.2177734375, "learning_rate": 0.0008, "loss": 1.1355, "step": 9128 }, { "epoch": 13.66616766467066, "grad_norm": 0.310546875, "learning_rate": 0.0008, "loss": 1.1381, "step": 9129 }, { "epoch": 13.667664670658683, "grad_norm": 0.451171875, "learning_rate": 0.0008, "loss": 1.1491, "step": 9130 }, { "epoch": 13.669161676646706, "grad_norm": 0.73828125, "learning_rate": 0.0008, "loss": 1.152, "step": 9131 }, { "epoch": 13.67065868263473, "grad_norm": 1.3984375, "learning_rate": 0.0008, "loss": 1.1754, "step": 9132 }, { "epoch": 13.672155688622755, "grad_norm": 0.56640625, "learning_rate": 0.0008, "loss": 1.155, "step": 9133 }, { "epoch": 13.673652694610778, "grad_norm": 0.37109375, "learning_rate": 0.0008, "loss": 1.1507, "step": 9134 }, { "epoch": 13.675149700598803, "grad_norm": 0.5859375, "learning_rate": 0.0008, "loss": 1.1575, "step": 9135 }, { "epoch": 13.676646706586826, "grad_norm": 0.478515625, "learning_rate": 0.0008, "loss": 1.1395, "step": 9136 }, { "epoch": 13.678143712574851, "grad_norm": 0.447265625, "learning_rate": 0.0008, "loss": 1.1566, "step": 9137 }, { "epoch": 13.679640718562874, "grad_norm": 0.1787109375, "learning_rate": 0.0008, "loss": 1.1481, "step": 9138 }, { "epoch": 13.681137724550899, "grad_norm": 0.2890625, "learning_rate": 0.0008, "loss": 1.1555, "step": 9139 }, { "epoch": 13.682634730538922, "grad_norm": 0.25, "learning_rate": 0.0008, "loss": 1.155, "step": 9140 }, { "epoch": 13.684131736526947, "grad_norm": 0.1650390625, "learning_rate": 0.0008, "loss": 1.1502, "step": 9141 }, { "epoch": 13.68562874251497, "grad_norm": 0.169921875, "learning_rate": 0.0008, "loss": 1.1448, "step": 9142 }, { "epoch": 13.687125748502995, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1403, "step": 9143 }, { "epoch": 13.688622754491018, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1551, "step": 9144 }, { "epoch": 13.690119760479043, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1438, "step": 9145 }, { "epoch": 13.691616766467066, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.1416, "step": 9146 }, { "epoch": 13.69311377245509, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1453, "step": 9147 }, { "epoch": 13.694610778443113, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.148, "step": 9148 }, { "epoch": 13.696107784431138, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1403, "step": 9149 }, { "epoch": 13.697604790419161, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1429, "step": 9150 }, { "epoch": 13.699101796407186, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1487, "step": 9151 }, { "epoch": 13.70059880239521, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1416, "step": 9152 }, { "epoch": 13.702095808383234, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1483, "step": 9153 }, { "epoch": 13.703592814371257, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1328, "step": 9154 }, { "epoch": 13.705089820359282, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1369, "step": 9155 }, { "epoch": 13.706586826347305, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1447, "step": 9156 }, { "epoch": 13.70808383233533, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1471, "step": 9157 }, { "epoch": 13.709580838323353, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1443, "step": 9158 }, { "epoch": 13.711077844311378, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1417, "step": 9159 }, { "epoch": 13.7125748502994, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1388, "step": 9160 }, { "epoch": 13.714071856287426, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1352, "step": 9161 }, { "epoch": 13.715568862275449, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1391, "step": 9162 }, { "epoch": 13.717065868263473, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1406, "step": 9163 }, { "epoch": 13.718562874251496, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1399, "step": 9164 }, { "epoch": 13.720059880239521, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1519, "step": 9165 }, { "epoch": 13.721556886227544, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.143, "step": 9166 }, { "epoch": 13.72305389221557, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1452, "step": 9167 }, { "epoch": 13.724550898203592, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.14, "step": 9168 }, { "epoch": 13.726047904191617, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1366, "step": 9169 }, { "epoch": 13.72754491017964, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1405, "step": 9170 }, { "epoch": 13.729041916167665, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.145, "step": 9171 }, { "epoch": 13.730538922155688, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1394, "step": 9172 }, { "epoch": 13.732035928143713, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.138, "step": 9173 }, { "epoch": 13.733532934131736, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1455, "step": 9174 }, { "epoch": 13.73502994011976, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1431, "step": 9175 }, { "epoch": 13.736526946107784, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1455, "step": 9176 }, { "epoch": 13.738023952095809, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1393, "step": 9177 }, { "epoch": 13.739520958083832, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1437, "step": 9178 }, { "epoch": 13.741017964071856, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1413, "step": 9179 }, { "epoch": 13.74251497005988, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1384, "step": 9180 }, { "epoch": 13.744011976047904, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1458, "step": 9181 }, { "epoch": 13.745508982035929, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.138, "step": 9182 }, { "epoch": 13.747005988023952, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1343, "step": 9183 }, { "epoch": 13.748502994011975, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1455, "step": 9184 }, { "epoch": 13.75, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.134, "step": 9185 }, { "epoch": 13.751497005988025, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1397, "step": 9186 }, { "epoch": 13.752994011976048, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.139, "step": 9187 }, { "epoch": 13.754491017964071, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1442, "step": 9188 }, { "epoch": 13.755988023952096, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1453, "step": 9189 }, { "epoch": 13.75748502994012, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1342, "step": 9190 }, { "epoch": 13.758982035928144, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1403, "step": 9191 }, { "epoch": 13.760479041916168, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1397, "step": 9192 }, { "epoch": 13.761976047904191, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1321, "step": 9193 }, { "epoch": 13.763473053892216, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1395, "step": 9194 }, { "epoch": 13.76497005988024, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1366, "step": 9195 }, { "epoch": 13.766467065868264, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1384, "step": 9196 }, { "epoch": 13.767964071856287, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1414, "step": 9197 }, { "epoch": 13.769461077844312, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1388, "step": 9198 }, { "epoch": 13.770958083832335, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1394, "step": 9199 }, { "epoch": 13.77245508982036, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1433, "step": 9200 }, { "epoch": 13.773952095808383, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1395, "step": 9201 }, { "epoch": 13.775449101796408, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1447, "step": 9202 }, { "epoch": 13.77694610778443, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1398, "step": 9203 }, { "epoch": 13.778443113772456, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1436, "step": 9204 }, { "epoch": 13.779940119760479, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.141, "step": 9205 }, { "epoch": 13.781437125748504, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1476, "step": 9206 }, { "epoch": 13.782934131736527, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1382, "step": 9207 }, { "epoch": 13.784431137724551, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1356, "step": 9208 }, { "epoch": 13.785928143712574, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1387, "step": 9209 }, { "epoch": 13.7874251497006, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1426, "step": 9210 }, { "epoch": 13.788922155688622, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1394, "step": 9211 }, { "epoch": 13.790419161676647, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1362, "step": 9212 }, { "epoch": 13.79191616766467, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.141, "step": 9213 }, { "epoch": 13.793413173652695, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1373, "step": 9214 }, { "epoch": 13.794910179640718, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1326, "step": 9215 }, { "epoch": 13.796407185628743, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1374, "step": 9216 }, { "epoch": 13.797904191616766, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.138, "step": 9217 }, { "epoch": 13.79940119760479, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1411, "step": 9218 }, { "epoch": 13.800898203592814, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1354, "step": 9219 }, { "epoch": 13.802395209580839, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1359, "step": 9220 }, { "epoch": 13.803892215568862, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1429, "step": 9221 }, { "epoch": 13.805389221556887, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1326, "step": 9222 }, { "epoch": 13.80688622754491, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1406, "step": 9223 }, { "epoch": 13.808383233532934, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1407, "step": 9224 }, { "epoch": 13.809880239520957, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.141, "step": 9225 }, { "epoch": 13.811377245508982, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1444, "step": 9226 }, { "epoch": 13.812874251497005, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1319, "step": 9227 }, { "epoch": 13.81437125748503, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1411, "step": 9228 }, { "epoch": 13.815868263473053, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1374, "step": 9229 }, { "epoch": 13.817365269461078, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1447, "step": 9230 }, { "epoch": 13.818862275449101, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1404, "step": 9231 }, { "epoch": 13.820359281437126, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1344, "step": 9232 }, { "epoch": 13.821856287425149, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1397, "step": 9233 }, { "epoch": 13.823353293413174, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1416, "step": 9234 }, { "epoch": 13.824850299401197, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.146, "step": 9235 }, { "epoch": 13.826347305389222, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1356, "step": 9236 }, { "epoch": 13.827844311377245, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1454, "step": 9237 }, { "epoch": 13.82934131736527, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.142, "step": 9238 }, { "epoch": 13.830838323353294, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1419, "step": 9239 }, { "epoch": 13.832335329341317, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1395, "step": 9240 }, { "epoch": 13.83383233532934, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1455, "step": 9241 }, { "epoch": 13.835329341317365, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1389, "step": 9242 }, { "epoch": 13.83682634730539, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1385, "step": 9243 }, { "epoch": 13.838323353293413, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1466, "step": 9244 }, { "epoch": 13.839820359281438, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1441, "step": 9245 }, { "epoch": 13.841317365269461, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1396, "step": 9246 }, { "epoch": 13.842814371257486, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1418, "step": 9247 }, { "epoch": 13.844311377245509, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.143, "step": 9248 }, { "epoch": 13.845808383233534, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1346, "step": 9249 }, { "epoch": 13.847305389221557, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.142, "step": 9250 }, { "epoch": 13.848802395209582, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1402, "step": 9251 }, { "epoch": 13.850299401197605, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1362, "step": 9252 }, { "epoch": 13.85179640718563, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1411, "step": 9253 }, { "epoch": 13.853293413173652, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1324, "step": 9254 }, { "epoch": 13.854790419161677, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1349, "step": 9255 }, { "epoch": 13.8562874251497, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1375, "step": 9256 }, { "epoch": 13.857784431137725, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1396, "step": 9257 }, { "epoch": 13.859281437125748, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1379, "step": 9258 }, { "epoch": 13.860778443113773, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1471, "step": 9259 }, { "epoch": 13.862275449101796, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1367, "step": 9260 }, { "epoch": 13.863772455089821, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1434, "step": 9261 }, { "epoch": 13.865269461077844, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.138, "step": 9262 }, { "epoch": 13.866766467065869, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.137, "step": 9263 }, { "epoch": 13.868263473053892, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1411, "step": 9264 }, { "epoch": 13.869760479041917, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1379, "step": 9265 }, { "epoch": 13.87125748502994, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1408, "step": 9266 }, { "epoch": 13.872754491017965, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1368, "step": 9267 }, { "epoch": 13.874251497005988, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1338, "step": 9268 }, { "epoch": 13.875748502994012, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1381, "step": 9269 }, { "epoch": 13.877245508982035, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1426, "step": 9270 }, { "epoch": 13.87874251497006, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1376, "step": 9271 }, { "epoch": 13.880239520958083, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.137, "step": 9272 }, { "epoch": 13.881736526946108, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1384, "step": 9273 }, { "epoch": 13.883233532934131, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1363, "step": 9274 }, { "epoch": 13.884730538922156, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1368, "step": 9275 }, { "epoch": 13.886227544910179, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1437, "step": 9276 }, { "epoch": 13.887724550898204, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1352, "step": 9277 }, { "epoch": 13.889221556886227, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1403, "step": 9278 }, { "epoch": 13.890718562874252, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1379, "step": 9279 }, { "epoch": 13.892215568862275, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1351, "step": 9280 }, { "epoch": 13.8937125748503, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1431, "step": 9281 }, { "epoch": 13.895209580838323, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1361, "step": 9282 }, { "epoch": 13.896706586826348, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1371, "step": 9283 }, { "epoch": 13.89820359281437, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1412, "step": 9284 }, { "epoch": 13.899700598802395, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1407, "step": 9285 }, { "epoch": 13.901197604790418, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1354, "step": 9286 }, { "epoch": 13.902694610778443, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 9287 }, { "epoch": 13.904191616766466, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.143, "step": 9288 }, { "epoch": 13.905688622754491, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1396, "step": 9289 }, { "epoch": 13.907185628742514, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1354, "step": 9290 }, { "epoch": 13.908682634730539, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1381, "step": 9291 }, { "epoch": 13.910179640718562, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1398, "step": 9292 }, { "epoch": 13.911676646706587, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1412, "step": 9293 }, { "epoch": 13.91317365269461, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1409, "step": 9294 }, { "epoch": 13.914670658682635, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1405, "step": 9295 }, { "epoch": 13.91616766467066, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1404, "step": 9296 }, { "epoch": 13.917664670658683, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1437, "step": 9297 }, { "epoch": 13.919161676646706, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1421, "step": 9298 }, { "epoch": 13.92065868263473, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1439, "step": 9299 }, { "epoch": 13.922155688622755, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1393, "step": 9300 }, { "epoch": 13.923652694610778, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1402, "step": 9301 }, { "epoch": 13.925149700598803, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1355, "step": 9302 }, { "epoch": 13.926646706586826, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1317, "step": 9303 }, { "epoch": 13.928143712574851, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1378, "step": 9304 }, { "epoch": 13.929640718562874, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1366, "step": 9305 }, { "epoch": 13.931137724550899, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1416, "step": 9306 }, { "epoch": 13.932634730538922, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1417, "step": 9307 }, { "epoch": 13.934131736526947, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1453, "step": 9308 }, { "epoch": 13.93562874251497, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1476, "step": 9309 }, { "epoch": 13.937125748502995, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1403, "step": 9310 }, { "epoch": 13.938622754491018, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1452, "step": 9311 }, { "epoch": 13.940119760479043, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1422, "step": 9312 }, { "epoch": 13.941616766467066, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1334, "step": 9313 }, { "epoch": 13.94311377245509, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1416, "step": 9314 }, { "epoch": 13.944610778443113, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1411, "step": 9315 }, { "epoch": 13.946107784431138, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1352, "step": 9316 }, { "epoch": 13.947604790419161, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1432, "step": 9317 }, { "epoch": 13.949101796407186, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.143, "step": 9318 }, { "epoch": 13.95059880239521, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1363, "step": 9319 }, { "epoch": 13.952095808383234, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1398, "step": 9320 }, { "epoch": 13.953592814371257, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1354, "step": 9321 }, { "epoch": 13.955089820359282, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1367, "step": 9322 }, { "epoch": 13.956586826347305, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1374, "step": 9323 }, { "epoch": 13.95808383233533, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1404, "step": 9324 }, { "epoch": 13.959580838323353, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1421, "step": 9325 }, { "epoch": 13.961077844311378, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1381, "step": 9326 }, { "epoch": 13.9625748502994, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1385, "step": 9327 }, { "epoch": 13.964071856287426, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1374, "step": 9328 }, { "epoch": 13.965568862275449, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1303, "step": 9329 }, { "epoch": 13.967065868263473, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.131, "step": 9330 }, { "epoch": 13.968562874251496, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1366, "step": 9331 }, { "epoch": 13.970059880239521, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1426, "step": 9332 }, { "epoch": 13.971556886227544, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1364, "step": 9333 }, { "epoch": 13.97305389221557, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1458, "step": 9334 }, { "epoch": 13.974550898203592, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.139, "step": 9335 }, { "epoch": 13.976047904191617, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1389, "step": 9336 }, { "epoch": 13.97754491017964, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1401, "step": 9337 }, { "epoch": 13.979041916167665, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1342, "step": 9338 }, { "epoch": 13.980538922155688, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1392, "step": 9339 }, { "epoch": 13.982035928143713, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1346, "step": 9340 }, { "epoch": 13.983532934131736, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1364, "step": 9341 }, { "epoch": 13.98502994011976, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.139, "step": 9342 }, { "epoch": 13.986526946107784, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.143, "step": 9343 }, { "epoch": 13.988023952095809, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.142, "step": 9344 }, { "epoch": 13.989520958083832, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1386, "step": 9345 }, { "epoch": 13.991017964071856, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1374, "step": 9346 }, { "epoch": 13.99251497005988, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1476, "step": 9347 }, { "epoch": 13.994011976047904, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1424, "step": 9348 }, { "epoch": 13.995508982035929, "grad_norm": 0.05078125, "learning_rate": 0.0008, "loss": 1.1418, "step": 9349 }, { "epoch": 13.997005988023952, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1399, "step": 9350 }, { "epoch": 13.998502994011975, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1425, "step": 9351 }, { "epoch": 14.0, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1443, "step": 9352 }, { "epoch": 14.001497005988025, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1446, "step": 9353 }, { "epoch": 14.002994011976048, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1345, "step": 9354 }, { "epoch": 14.004491017964073, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1434, "step": 9355 }, { "epoch": 14.005988023952096, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1307, "step": 9356 }, { "epoch": 14.00748502994012, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1419, "step": 9357 }, { "epoch": 14.008982035928144, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1299, "step": 9358 }, { "epoch": 14.010479041916168, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1316, "step": 9359 }, { "epoch": 14.011976047904191, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1351, "step": 9360 }, { "epoch": 14.013473053892216, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.146, "step": 9361 }, { "epoch": 14.01497005988024, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1422, "step": 9362 }, { "epoch": 14.016467065868264, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1433, "step": 9363 }, { "epoch": 14.017964071856287, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1401, "step": 9364 }, { "epoch": 14.019461077844312, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1408, "step": 9365 }, { "epoch": 14.020958083832335, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1395, "step": 9366 }, { "epoch": 14.02245508982036, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1459, "step": 9367 }, { "epoch": 14.023952095808383, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1393, "step": 9368 }, { "epoch": 14.025449101796408, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1397, "step": 9369 }, { "epoch": 14.02694610778443, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1353, "step": 9370 }, { "epoch": 14.028443113772456, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1415, "step": 9371 }, { "epoch": 14.029940119760479, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.136, "step": 9372 }, { "epoch": 14.031437125748504, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1351, "step": 9373 }, { "epoch": 14.032934131736527, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1334, "step": 9374 }, { "epoch": 14.034431137724551, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1345, "step": 9375 }, { "epoch": 14.035928143712574, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1366, "step": 9376 }, { "epoch": 14.0374251497006, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1414, "step": 9377 }, { "epoch": 14.038922155688622, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1481, "step": 9378 }, { "epoch": 14.040419161676647, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1352, "step": 9379 }, { "epoch": 14.04191616766467, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.137, "step": 9380 }, { "epoch": 14.043413173652695, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1364, "step": 9381 }, { "epoch": 14.044910179640718, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1347, "step": 9382 }, { "epoch": 14.046407185628743, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1403, "step": 9383 }, { "epoch": 14.047904191616766, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1443, "step": 9384 }, { "epoch": 14.04940119760479, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1371, "step": 9385 }, { "epoch": 14.050898203592814, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1385, "step": 9386 }, { "epoch": 14.052395209580839, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1404, "step": 9387 }, { "epoch": 14.053892215568862, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1349, "step": 9388 }, { "epoch": 14.055389221556887, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1346, "step": 9389 }, { "epoch": 14.05688622754491, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1409, "step": 9390 }, { "epoch": 14.058383233532934, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1417, "step": 9391 }, { "epoch": 14.059880239520957, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1346, "step": 9392 }, { "epoch": 14.061377245508982, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1408, "step": 9393 }, { "epoch": 14.062874251497005, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1393, "step": 9394 }, { "epoch": 14.06437125748503, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.132, "step": 9395 }, { "epoch": 14.065868263473053, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1382, "step": 9396 }, { "epoch": 14.067365269461078, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1372, "step": 9397 }, { "epoch": 14.068862275449101, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1393, "step": 9398 }, { "epoch": 14.070359281437126, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1343, "step": 9399 }, { "epoch": 14.071856287425149, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1429, "step": 9400 }, { "epoch": 14.073353293413174, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1356, "step": 9401 }, { "epoch": 14.074850299401197, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1418, "step": 9402 }, { "epoch": 14.076347305389222, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1427, "step": 9403 }, { "epoch": 14.077844311377245, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1388, "step": 9404 }, { "epoch": 14.07934131736527, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1461, "step": 9405 }, { "epoch": 14.080838323353293, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1329, "step": 9406 }, { "epoch": 14.082335329341317, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1444, "step": 9407 }, { "epoch": 14.08383233532934, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1353, "step": 9408 }, { "epoch": 14.085329341317365, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1465, "step": 9409 }, { "epoch": 14.08682634730539, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1302, "step": 9410 }, { "epoch": 14.088323353293413, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1397, "step": 9411 }, { "epoch": 14.089820359281438, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.138, "step": 9412 }, { "epoch": 14.091317365269461, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1413, "step": 9413 }, { "epoch": 14.092814371257486, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1374, "step": 9414 }, { "epoch": 14.094311377245509, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1391, "step": 9415 }, { "epoch": 14.095808383233534, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1394, "step": 9416 }, { "epoch": 14.097305389221557, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1397, "step": 9417 }, { "epoch": 14.098802395209582, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1444, "step": 9418 }, { "epoch": 14.100299401197605, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1459, "step": 9419 }, { "epoch": 14.10179640718563, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1376, "step": 9420 }, { "epoch": 14.103293413173652, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1393, "step": 9421 }, { "epoch": 14.104790419161677, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1404, "step": 9422 }, { "epoch": 14.1062874251497, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1363, "step": 9423 }, { "epoch": 14.107784431137725, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.139, "step": 9424 }, { "epoch": 14.109281437125748, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1444, "step": 9425 }, { "epoch": 14.110778443113773, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1368, "step": 9426 }, { "epoch": 14.112275449101796, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1358, "step": 9427 }, { "epoch": 14.113772455089821, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1424, "step": 9428 }, { "epoch": 14.115269461077844, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1377, "step": 9429 }, { "epoch": 14.116766467065869, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1391, "step": 9430 }, { "epoch": 14.118263473053892, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1316, "step": 9431 }, { "epoch": 14.119760479041917, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1423, "step": 9432 }, { "epoch": 14.12125748502994, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1423, "step": 9433 }, { "epoch": 14.122754491017965, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.134, "step": 9434 }, { "epoch": 14.124251497005988, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1397, "step": 9435 }, { "epoch": 14.125748502994012, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1376, "step": 9436 }, { "epoch": 14.127245508982035, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.139, "step": 9437 }, { "epoch": 14.12874251497006, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1388, "step": 9438 }, { "epoch": 14.130239520958083, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1434, "step": 9439 }, { "epoch": 14.131736526946108, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1336, "step": 9440 }, { "epoch": 14.133233532934131, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1409, "step": 9441 }, { "epoch": 14.134730538922156, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1394, "step": 9442 }, { "epoch": 14.136227544910179, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1426, "step": 9443 }, { "epoch": 14.137724550898204, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1355, "step": 9444 }, { "epoch": 14.139221556886227, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1422, "step": 9445 }, { "epoch": 14.140718562874252, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.134, "step": 9446 }, { "epoch": 14.142215568862275, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1389, "step": 9447 }, { "epoch": 14.1437125748503, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1432, "step": 9448 }, { "epoch": 14.145209580838323, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1377, "step": 9449 }, { "epoch": 14.146706586826348, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1394, "step": 9450 }, { "epoch": 14.14820359281437, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1324, "step": 9451 }, { "epoch": 14.149700598802395, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1476, "step": 9452 }, { "epoch": 14.151197604790418, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 9453 }, { "epoch": 14.152694610778443, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1375, "step": 9454 }, { "epoch": 14.154191616766466, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1378, "step": 9455 }, { "epoch": 14.155688622754491, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1398, "step": 9456 }, { "epoch": 14.157185628742514, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1343, "step": 9457 }, { "epoch": 14.158682634730539, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1433, "step": 9458 }, { "epoch": 14.160179640718562, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1384, "step": 9459 }, { "epoch": 14.161676646706587, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1363, "step": 9460 }, { "epoch": 14.16317365269461, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.135, "step": 9461 }, { "epoch": 14.164670658682635, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1397, "step": 9462 }, { "epoch": 14.16616766467066, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1367, "step": 9463 }, { "epoch": 14.167664670658683, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1416, "step": 9464 }, { "epoch": 14.169161676646707, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1359, "step": 9465 }, { "epoch": 14.17065868263473, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.136, "step": 9466 }, { "epoch": 14.172155688622755, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1359, "step": 9467 }, { "epoch": 14.173652694610778, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1371, "step": 9468 }, { "epoch": 14.175149700598803, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1404, "step": 9469 }, { "epoch": 14.176646706586826, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1369, "step": 9470 }, { "epoch": 14.178143712574851, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1362, "step": 9471 }, { "epoch": 14.179640718562874, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1377, "step": 9472 }, { "epoch": 14.181137724550899, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1425, "step": 9473 }, { "epoch": 14.182634730538922, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1458, "step": 9474 }, { "epoch": 14.184131736526947, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1402, "step": 9475 }, { "epoch": 14.18562874251497, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1438, "step": 9476 }, { "epoch": 14.187125748502995, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1366, "step": 9477 }, { "epoch": 14.188622754491018, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1338, "step": 9478 }, { "epoch": 14.190119760479043, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1355, "step": 9479 }, { "epoch": 14.191616766467066, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1355, "step": 9480 }, { "epoch": 14.19311377245509, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1355, "step": 9481 }, { "epoch": 14.194610778443113, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1369, "step": 9482 }, { "epoch": 14.196107784431138, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1369, "step": 9483 }, { "epoch": 14.197604790419161, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1334, "step": 9484 }, { "epoch": 14.199101796407186, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1369, "step": 9485 }, { "epoch": 14.20059880239521, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1366, "step": 9486 }, { "epoch": 14.202095808383234, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1398, "step": 9487 }, { "epoch": 14.203592814371257, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1392, "step": 9488 }, { "epoch": 14.205089820359282, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1368, "step": 9489 }, { "epoch": 14.206586826347305, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1445, "step": 9490 }, { "epoch": 14.20808383233533, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1434, "step": 9491 }, { "epoch": 14.209580838323353, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1362, "step": 9492 }, { "epoch": 14.211077844311378, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1405, "step": 9493 }, { "epoch": 14.2125748502994, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1352, "step": 9494 }, { "epoch": 14.214071856287426, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.136, "step": 9495 }, { "epoch": 14.215568862275449, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1442, "step": 9496 }, { "epoch": 14.217065868263473, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1337, "step": 9497 }, { "epoch": 14.218562874251496, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1379, "step": 9498 }, { "epoch": 14.220059880239521, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1423, "step": 9499 }, { "epoch": 14.221556886227544, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1276, "step": 9500 }, { "epoch": 14.22305389221557, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1372, "step": 9501 }, { "epoch": 14.224550898203592, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1443, "step": 9502 }, { "epoch": 14.226047904191617, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.137, "step": 9503 }, { "epoch": 14.22754491017964, "grad_norm": 0.15625, "learning_rate": 0.0008, "loss": 1.1429, "step": 9504 }, { "epoch": 14.229041916167665, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1399, "step": 9505 }, { "epoch": 14.230538922155688, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1388, "step": 9506 }, { "epoch": 14.232035928143713, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1337, "step": 9507 }, { "epoch": 14.233532934131736, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1423, "step": 9508 }, { "epoch": 14.23502994011976, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1341, "step": 9509 }, { "epoch": 14.236526946107784, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1339, "step": 9510 }, { "epoch": 14.238023952095809, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1353, "step": 9511 }, { "epoch": 14.239520958083832, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1404, "step": 9512 }, { "epoch": 14.241017964071856, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1336, "step": 9513 }, { "epoch": 14.24251497005988, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1348, "step": 9514 }, { "epoch": 14.244011976047904, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1326, "step": 9515 }, { "epoch": 14.245508982035927, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1401, "step": 9516 }, { "epoch": 14.247005988023952, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1406, "step": 9517 }, { "epoch": 14.248502994011975, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1319, "step": 9518 }, { "epoch": 14.25, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1396, "step": 9519 }, { "epoch": 14.251497005988025, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1352, "step": 9520 }, { "epoch": 14.252994011976048, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1383, "step": 9521 }, { "epoch": 14.254491017964073, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1357, "step": 9522 }, { "epoch": 14.255988023952096, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.139, "step": 9523 }, { "epoch": 14.25748502994012, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1404, "step": 9524 }, { "epoch": 14.258982035928144, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1417, "step": 9525 }, { "epoch": 14.260479041916168, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1341, "step": 9526 }, { "epoch": 14.261976047904191, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1399, "step": 9527 }, { "epoch": 14.263473053892216, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1392, "step": 9528 }, { "epoch": 14.26497005988024, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1426, "step": 9529 }, { "epoch": 14.266467065868264, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.138, "step": 9530 }, { "epoch": 14.267964071856287, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1355, "step": 9531 }, { "epoch": 14.269461077844312, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1383, "step": 9532 }, { "epoch": 14.270958083832335, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1409, "step": 9533 }, { "epoch": 14.27245508982036, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1347, "step": 9534 }, { "epoch": 14.273952095808383, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1364, "step": 9535 }, { "epoch": 14.275449101796408, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1376, "step": 9536 }, { "epoch": 14.27694610778443, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1407, "step": 9537 }, { "epoch": 14.278443113772456, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.137, "step": 9538 }, { "epoch": 14.279940119760479, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1341, "step": 9539 }, { "epoch": 14.281437125748504, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1419, "step": 9540 }, { "epoch": 14.282934131736527, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1381, "step": 9541 }, { "epoch": 14.284431137724551, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1417, "step": 9542 }, { "epoch": 14.285928143712574, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1296, "step": 9543 }, { "epoch": 14.2874251497006, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1411, "step": 9544 }, { "epoch": 14.288922155688622, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.143, "step": 9545 }, { "epoch": 14.290419161676647, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1343, "step": 9546 }, { "epoch": 14.29191616766467, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1385, "step": 9547 }, { "epoch": 14.293413173652695, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1362, "step": 9548 }, { "epoch": 14.294910179640718, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1367, "step": 9549 }, { "epoch": 14.296407185628743, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1399, "step": 9550 }, { "epoch": 14.297904191616766, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1333, "step": 9551 }, { "epoch": 14.29940119760479, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1347, "step": 9552 }, { "epoch": 14.300898203592814, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1379, "step": 9553 }, { "epoch": 14.302395209580839, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.138, "step": 9554 }, { "epoch": 14.303892215568862, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1353, "step": 9555 }, { "epoch": 14.305389221556887, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1425, "step": 9556 }, { "epoch": 14.30688622754491, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.14, "step": 9557 }, { "epoch": 14.308383233532934, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1351, "step": 9558 }, { "epoch": 14.309880239520957, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1394, "step": 9559 }, { "epoch": 14.311377245508982, "grad_norm": 0.193359375, "learning_rate": 0.0008, "loss": 1.1417, "step": 9560 }, { "epoch": 14.312874251497005, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1423, "step": 9561 }, { "epoch": 14.31437125748503, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1486, "step": 9562 }, { "epoch": 14.315868263473053, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1305, "step": 9563 }, { "epoch": 14.317365269461078, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1371, "step": 9564 }, { "epoch": 14.318862275449101, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1348, "step": 9565 }, { "epoch": 14.320359281437126, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1356, "step": 9566 }, { "epoch": 14.321856287425149, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1429, "step": 9567 }, { "epoch": 14.323353293413174, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1344, "step": 9568 }, { "epoch": 14.324850299401197, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1307, "step": 9569 }, { "epoch": 14.326347305389222, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.137, "step": 9570 }, { "epoch": 14.327844311377245, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1341, "step": 9571 }, { "epoch": 14.32934131736527, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1399, "step": 9572 }, { "epoch": 14.330838323353294, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1395, "step": 9573 }, { "epoch": 14.332335329341317, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1412, "step": 9574 }, { "epoch": 14.33383233532934, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1346, "step": 9575 }, { "epoch": 14.335329341317365, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1341, "step": 9576 }, { "epoch": 14.33682634730539, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1405, "step": 9577 }, { "epoch": 14.338323353293413, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1419, "step": 9578 }, { "epoch": 14.339820359281438, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1396, "step": 9579 }, { "epoch": 14.341317365269461, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.134, "step": 9580 }, { "epoch": 14.342814371257486, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.141, "step": 9581 }, { "epoch": 14.344311377245509, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1429, "step": 9582 }, { "epoch": 14.345808383233534, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1394, "step": 9583 }, { "epoch": 14.347305389221557, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1411, "step": 9584 }, { "epoch": 14.348802395209582, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1325, "step": 9585 }, { "epoch": 14.350299401197605, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1371, "step": 9586 }, { "epoch": 14.35179640718563, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1365, "step": 9587 }, { "epoch": 14.353293413173652, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1343, "step": 9588 }, { "epoch": 14.354790419161677, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1344, "step": 9589 }, { "epoch": 14.3562874251497, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1335, "step": 9590 }, { "epoch": 14.357784431137725, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1407, "step": 9591 }, { "epoch": 14.359281437125748, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1369, "step": 9592 }, { "epoch": 14.360778443113773, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1411, "step": 9593 }, { "epoch": 14.362275449101796, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1318, "step": 9594 }, { "epoch": 14.363772455089821, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1364, "step": 9595 }, { "epoch": 14.365269461077844, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1387, "step": 9596 }, { "epoch": 14.366766467065869, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1385, "step": 9597 }, { "epoch": 14.368263473053892, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1355, "step": 9598 }, { "epoch": 14.369760479041917, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1375, "step": 9599 }, { "epoch": 14.37125748502994, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1352, "step": 9600 }, { "epoch": 14.372754491017965, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1418, "step": 9601 }, { "epoch": 14.374251497005988, "grad_norm": 0.04931640625, "learning_rate": 0.0008, "loss": 1.1313, "step": 9602 }, { "epoch": 14.375748502994012, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1401, "step": 9603 }, { "epoch": 14.377245508982035, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.135, "step": 9604 }, { "epoch": 14.37874251497006, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1435, "step": 9605 }, { "epoch": 14.380239520958083, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1391, "step": 9606 }, { "epoch": 14.381736526946108, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1368, "step": 9607 }, { "epoch": 14.383233532934131, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.141, "step": 9608 }, { "epoch": 14.384730538922156, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1325, "step": 9609 }, { "epoch": 14.386227544910179, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1373, "step": 9610 }, { "epoch": 14.387724550898204, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1372, "step": 9611 }, { "epoch": 14.389221556886227, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1356, "step": 9612 }, { "epoch": 14.390718562874252, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1412, "step": 9613 }, { "epoch": 14.392215568862275, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1368, "step": 9614 }, { "epoch": 14.3937125748503, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1439, "step": 9615 }, { "epoch": 14.395209580838323, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1388, "step": 9616 }, { "epoch": 14.396706586826348, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.131, "step": 9617 }, { "epoch": 14.39820359281437, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1369, "step": 9618 }, { "epoch": 14.399700598802395, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.129, "step": 9619 }, { "epoch": 14.401197604790418, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1419, "step": 9620 }, { "epoch": 14.402694610778443, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1431, "step": 9621 }, { "epoch": 14.404191616766466, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1419, "step": 9622 }, { "epoch": 14.405688622754491, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1344, "step": 9623 }, { "epoch": 14.407185628742514, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1391, "step": 9624 }, { "epoch": 14.408682634730539, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1352, "step": 9625 }, { "epoch": 14.410179640718562, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1409, "step": 9626 }, { "epoch": 14.411676646706587, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1379, "step": 9627 }, { "epoch": 14.41317365269461, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1376, "step": 9628 }, { "epoch": 14.414670658682635, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1367, "step": 9629 }, { "epoch": 14.41616766467066, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1344, "step": 9630 }, { "epoch": 14.417664670658683, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1412, "step": 9631 }, { "epoch": 14.419161676646706, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1352, "step": 9632 }, { "epoch": 14.42065868263473, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1367, "step": 9633 }, { "epoch": 14.422155688622755, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1408, "step": 9634 }, { "epoch": 14.423652694610778, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1324, "step": 9635 }, { "epoch": 14.425149700598803, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1359, "step": 9636 }, { "epoch": 14.426646706586826, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1383, "step": 9637 }, { "epoch": 14.428143712574851, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1374, "step": 9638 }, { "epoch": 14.429640718562874, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1387, "step": 9639 }, { "epoch": 14.431137724550899, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1371, "step": 9640 }, { "epoch": 14.432634730538922, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1343, "step": 9641 }, { "epoch": 14.434131736526947, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1361, "step": 9642 }, { "epoch": 14.43562874251497, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1369, "step": 9643 }, { "epoch": 14.437125748502995, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1481, "step": 9644 }, { "epoch": 14.438622754491018, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1441, "step": 9645 }, { "epoch": 14.440119760479043, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1365, "step": 9646 }, { "epoch": 14.441616766467066, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1355, "step": 9647 }, { "epoch": 14.44311377245509, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1336, "step": 9648 }, { "epoch": 14.444610778443113, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1476, "step": 9649 }, { "epoch": 14.446107784431138, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1347, "step": 9650 }, { "epoch": 14.447604790419161, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1278, "step": 9651 }, { "epoch": 14.449101796407186, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1385, "step": 9652 }, { "epoch": 14.45059880239521, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1304, "step": 9653 }, { "epoch": 14.452095808383234, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1384, "step": 9654 }, { "epoch": 14.453592814371257, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1442, "step": 9655 }, { "epoch": 14.455089820359282, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1363, "step": 9656 }, { "epoch": 14.456586826347305, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1435, "step": 9657 }, { "epoch": 14.45808383233533, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1352, "step": 9658 }, { "epoch": 14.459580838323353, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1374, "step": 9659 }, { "epoch": 14.461077844311378, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1463, "step": 9660 }, { "epoch": 14.4625748502994, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1366, "step": 9661 }, { "epoch": 14.464071856287426, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1338, "step": 9662 }, { "epoch": 14.465568862275449, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1401, "step": 9663 }, { "epoch": 14.467065868263473, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1323, "step": 9664 }, { "epoch": 14.468562874251496, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1367, "step": 9665 }, { "epoch": 14.470059880239521, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1404, "step": 9666 }, { "epoch": 14.471556886227544, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1444, "step": 9667 }, { "epoch": 14.47305389221557, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1425, "step": 9668 }, { "epoch": 14.474550898203592, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.139, "step": 9669 }, { "epoch": 14.476047904191617, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1376, "step": 9670 }, { "epoch": 14.47754491017964, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1337, "step": 9671 }, { "epoch": 14.479041916167665, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1434, "step": 9672 }, { "epoch": 14.480538922155688, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1352, "step": 9673 }, { "epoch": 14.482035928143713, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1369, "step": 9674 }, { "epoch": 14.483532934131736, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1345, "step": 9675 }, { "epoch": 14.48502994011976, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.1413, "step": 9676 }, { "epoch": 14.486526946107784, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1346, "step": 9677 }, { "epoch": 14.488023952095809, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1328, "step": 9678 }, { "epoch": 14.489520958083832, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1409, "step": 9679 }, { "epoch": 14.491017964071856, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1432, "step": 9680 }, { "epoch": 14.49251497005988, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1356, "step": 9681 }, { "epoch": 14.494011976047904, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1398, "step": 9682 }, { "epoch": 14.495508982035927, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1382, "step": 9683 }, { "epoch": 14.497005988023952, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.134, "step": 9684 }, { "epoch": 14.498502994011975, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1372, "step": 9685 }, { "epoch": 14.5, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1395, "step": 9686 }, { "epoch": 14.501497005988025, "grad_norm": 0.1796875, "learning_rate": 0.0008, "loss": 1.1391, "step": 9687 }, { "epoch": 14.502994011976048, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1307, "step": 9688 }, { "epoch": 14.504491017964071, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1376, "step": 9689 }, { "epoch": 14.505988023952096, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1394, "step": 9690 }, { "epoch": 14.50748502994012, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1407, "step": 9691 }, { "epoch": 14.508982035928144, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1352, "step": 9692 }, { "epoch": 14.510479041916168, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1319, "step": 9693 }, { "epoch": 14.511976047904191, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1311, "step": 9694 }, { "epoch": 14.513473053892216, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1314, "step": 9695 }, { "epoch": 14.51497005988024, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1384, "step": 9696 }, { "epoch": 14.516467065868264, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1357, "step": 9697 }, { "epoch": 14.517964071856287, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1379, "step": 9698 }, { "epoch": 14.519461077844312, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1339, "step": 9699 }, { "epoch": 14.520958083832335, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1419, "step": 9700 }, { "epoch": 14.52245508982036, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1328, "step": 9701 }, { "epoch": 14.523952095808383, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1415, "step": 9702 }, { "epoch": 14.525449101796408, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1371, "step": 9703 }, { "epoch": 14.52694610778443, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1366, "step": 9704 }, { "epoch": 14.528443113772456, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.133, "step": 9705 }, { "epoch": 14.529940119760479, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.134, "step": 9706 }, { "epoch": 14.531437125748504, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1328, "step": 9707 }, { "epoch": 14.532934131736527, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1407, "step": 9708 }, { "epoch": 14.534431137724551, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1428, "step": 9709 }, { "epoch": 14.535928143712574, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1375, "step": 9710 }, { "epoch": 14.5374251497006, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1385, "step": 9711 }, { "epoch": 14.538922155688622, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.14, "step": 9712 }, { "epoch": 14.540419161676647, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.138, "step": 9713 }, { "epoch": 14.54191616766467, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1396, "step": 9714 }, { "epoch": 14.543413173652695, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.132, "step": 9715 }, { "epoch": 14.544910179640718, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1378, "step": 9716 }, { "epoch": 14.546407185628743, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1349, "step": 9717 }, { "epoch": 14.547904191616766, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1378, "step": 9718 }, { "epoch": 14.54940119760479, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.142, "step": 9719 }, { "epoch": 14.550898203592814, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1448, "step": 9720 }, { "epoch": 14.552395209580839, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1376, "step": 9721 }, { "epoch": 14.553892215568862, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1421, "step": 9722 }, { "epoch": 14.555389221556887, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1371, "step": 9723 }, { "epoch": 14.55688622754491, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1344, "step": 9724 }, { "epoch": 14.558383233532934, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1443, "step": 9725 }, { "epoch": 14.559880239520957, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1381, "step": 9726 }, { "epoch": 14.561377245508982, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1402, "step": 9727 }, { "epoch": 14.562874251497005, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1369, "step": 9728 }, { "epoch": 14.56437125748503, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1342, "step": 9729 }, { "epoch": 14.565868263473053, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1305, "step": 9730 }, { "epoch": 14.567365269461078, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.14, "step": 9731 }, { "epoch": 14.568862275449101, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1326, "step": 9732 }, { "epoch": 14.570359281437126, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1363, "step": 9733 }, { "epoch": 14.571856287425149, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1412, "step": 9734 }, { "epoch": 14.573353293413174, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1333, "step": 9735 }, { "epoch": 14.574850299401197, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1396, "step": 9736 }, { "epoch": 14.576347305389222, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1364, "step": 9737 }, { "epoch": 14.577844311377245, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.136, "step": 9738 }, { "epoch": 14.57934131736527, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1442, "step": 9739 }, { "epoch": 14.580838323353294, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1401, "step": 9740 }, { "epoch": 14.582335329341317, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1449, "step": 9741 }, { "epoch": 14.58383233532934, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1437, "step": 9742 }, { "epoch": 14.585329341317365, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.144, "step": 9743 }, { "epoch": 14.58682634730539, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1329, "step": 9744 }, { "epoch": 14.588323353293413, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1327, "step": 9745 }, { "epoch": 14.589820359281438, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1416, "step": 9746 }, { "epoch": 14.591317365269461, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1427, "step": 9747 }, { "epoch": 14.592814371257486, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1352, "step": 9748 }, { "epoch": 14.594311377245509, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1378, "step": 9749 }, { "epoch": 14.595808383233534, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1382, "step": 9750 }, { "epoch": 14.597305389221557, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.135, "step": 9751 }, { "epoch": 14.598802395209582, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1386, "step": 9752 }, { "epoch": 14.600299401197605, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1342, "step": 9753 }, { "epoch": 14.60179640718563, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1416, "step": 9754 }, { "epoch": 14.603293413173652, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1423, "step": 9755 }, { "epoch": 14.604790419161677, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1328, "step": 9756 }, { "epoch": 14.6062874251497, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1301, "step": 9757 }, { "epoch": 14.607784431137725, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1421, "step": 9758 }, { "epoch": 14.609281437125748, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1352, "step": 9759 }, { "epoch": 14.610778443113773, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1331, "step": 9760 }, { "epoch": 14.612275449101796, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1367, "step": 9761 }, { "epoch": 14.613772455089821, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1329, "step": 9762 }, { "epoch": 14.615269461077844, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1332, "step": 9763 }, { "epoch": 14.616766467065869, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.133, "step": 9764 }, { "epoch": 14.618263473053892, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1363, "step": 9765 }, { "epoch": 14.619760479041917, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1396, "step": 9766 }, { "epoch": 14.62125748502994, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1327, "step": 9767 }, { "epoch": 14.622754491017965, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1349, "step": 9768 }, { "epoch": 14.624251497005988, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1345, "step": 9769 }, { "epoch": 14.625748502994012, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1415, "step": 9770 }, { "epoch": 14.627245508982035, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1351, "step": 9771 }, { "epoch": 14.62874251497006, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1425, "step": 9772 }, { "epoch": 14.630239520958083, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1402, "step": 9773 }, { "epoch": 14.631736526946108, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1457, "step": 9774 }, { "epoch": 14.633233532934131, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1362, "step": 9775 }, { "epoch": 14.634730538922156, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1374, "step": 9776 }, { "epoch": 14.636227544910179, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1376, "step": 9777 }, { "epoch": 14.637724550898204, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1354, "step": 9778 }, { "epoch": 14.639221556886227, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1425, "step": 9779 }, { "epoch": 14.640718562874252, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1353, "step": 9780 }, { "epoch": 14.642215568862275, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1327, "step": 9781 }, { "epoch": 14.6437125748503, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1378, "step": 9782 }, { "epoch": 14.645209580838323, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 9783 }, { "epoch": 14.646706586826348, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1397, "step": 9784 }, { "epoch": 14.64820359281437, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1393, "step": 9785 }, { "epoch": 14.649700598802395, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1387, "step": 9786 }, { "epoch": 14.651197604790418, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1391, "step": 9787 }, { "epoch": 14.652694610778443, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1415, "step": 9788 }, { "epoch": 14.654191616766466, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.142, "step": 9789 }, { "epoch": 14.655688622754491, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.139, "step": 9790 }, { "epoch": 14.657185628742514, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1357, "step": 9791 }, { "epoch": 14.658682634730539, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1379, "step": 9792 }, { "epoch": 14.660179640718562, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1362, "step": 9793 }, { "epoch": 14.661676646706587, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1324, "step": 9794 }, { "epoch": 14.66317365269461, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1335, "step": 9795 }, { "epoch": 14.664670658682635, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1385, "step": 9796 }, { "epoch": 14.66616766467066, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1344, "step": 9797 }, { "epoch": 14.667664670658683, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1329, "step": 9798 }, { "epoch": 14.669161676646706, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1419, "step": 9799 }, { "epoch": 14.67065868263473, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1382, "step": 9800 }, { "epoch": 14.672155688622755, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.141, "step": 9801 }, { "epoch": 14.673652694610778, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.132, "step": 9802 }, { "epoch": 14.675149700598803, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.134, "step": 9803 }, { "epoch": 14.676646706586826, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1417, "step": 9804 }, { "epoch": 14.678143712574851, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.134, "step": 9805 }, { "epoch": 14.679640718562874, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1377, "step": 9806 }, { "epoch": 14.681137724550899, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1472, "step": 9807 }, { "epoch": 14.682634730538922, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1417, "step": 9808 }, { "epoch": 14.684131736526947, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1365, "step": 9809 }, { "epoch": 14.68562874251497, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1357, "step": 9810 }, { "epoch": 14.687125748502995, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1298, "step": 9811 }, { "epoch": 14.688622754491018, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1348, "step": 9812 }, { "epoch": 14.690119760479043, "grad_norm": 0.158203125, "learning_rate": 0.0008, "loss": 1.1369, "step": 9813 }, { "epoch": 14.691616766467066, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1315, "step": 9814 }, { "epoch": 14.69311377245509, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1408, "step": 9815 }, { "epoch": 14.694610778443113, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1383, "step": 9816 }, { "epoch": 14.696107784431138, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1444, "step": 9817 }, { "epoch": 14.697604790419161, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1381, "step": 9818 }, { "epoch": 14.699101796407186, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1377, "step": 9819 }, { "epoch": 14.70059880239521, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1384, "step": 9820 }, { "epoch": 14.702095808383234, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1301, "step": 9821 }, { "epoch": 14.703592814371257, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1443, "step": 9822 }, { "epoch": 14.705089820359282, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1397, "step": 9823 }, { "epoch": 14.706586826347305, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1364, "step": 9824 }, { "epoch": 14.70808383233533, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1434, "step": 9825 }, { "epoch": 14.709580838323353, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1411, "step": 9826 }, { "epoch": 14.711077844311378, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1385, "step": 9827 }, { "epoch": 14.7125748502994, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1339, "step": 9828 }, { "epoch": 14.714071856287426, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.138, "step": 9829 }, { "epoch": 14.715568862275449, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1369, "step": 9830 }, { "epoch": 14.717065868263473, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1412, "step": 9831 }, { "epoch": 14.718562874251496, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.146, "step": 9832 }, { "epoch": 14.720059880239521, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1425, "step": 9833 }, { "epoch": 14.721556886227544, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1374, "step": 9834 }, { "epoch": 14.72305389221557, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1437, "step": 9835 }, { "epoch": 14.724550898203592, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1417, "step": 9836 }, { "epoch": 14.726047904191617, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1368, "step": 9837 }, { "epoch": 14.72754491017964, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1375, "step": 9838 }, { "epoch": 14.729041916167665, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1313, "step": 9839 }, { "epoch": 14.730538922155688, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1403, "step": 9840 }, { "epoch": 14.732035928143713, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1425, "step": 9841 }, { "epoch": 14.733532934131736, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.131, "step": 9842 }, { "epoch": 14.73502994011976, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1366, "step": 9843 }, { "epoch": 14.736526946107784, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1409, "step": 9844 }, { "epoch": 14.738023952095809, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.142, "step": 9845 }, { "epoch": 14.739520958083832, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.135, "step": 9846 }, { "epoch": 14.741017964071856, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1405, "step": 9847 }, { "epoch": 14.74251497005988, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1279, "step": 9848 }, { "epoch": 14.744011976047904, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1379, "step": 9849 }, { "epoch": 14.745508982035929, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1362, "step": 9850 }, { "epoch": 14.747005988023952, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1342, "step": 9851 }, { "epoch": 14.748502994011975, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1441, "step": 9852 }, { "epoch": 14.75, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1375, "step": 9853 }, { "epoch": 14.751497005988025, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1305, "step": 9854 }, { "epoch": 14.752994011976048, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1414, "step": 9855 }, { "epoch": 14.754491017964071, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1388, "step": 9856 }, { "epoch": 14.755988023952096, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1398, "step": 9857 }, { "epoch": 14.75748502994012, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1392, "step": 9858 }, { "epoch": 14.758982035928144, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1314, "step": 9859 }, { "epoch": 14.760479041916168, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1369, "step": 9860 }, { "epoch": 14.761976047904191, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1375, "step": 9861 }, { "epoch": 14.763473053892216, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1394, "step": 9862 }, { "epoch": 14.76497005988024, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1404, "step": 9863 }, { "epoch": 14.766467065868264, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1385, "step": 9864 }, { "epoch": 14.767964071856287, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.143, "step": 9865 }, { "epoch": 14.769461077844312, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1394, "step": 9866 }, { "epoch": 14.770958083832335, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1415, "step": 9867 }, { "epoch": 14.77245508982036, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.138, "step": 9868 }, { "epoch": 14.773952095808383, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1321, "step": 9869 }, { "epoch": 14.775449101796408, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1379, "step": 9870 }, { "epoch": 14.77694610778443, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1401, "step": 9871 }, { "epoch": 14.778443113772456, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1429, "step": 9872 }, { "epoch": 14.779940119760479, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1346, "step": 9873 }, { "epoch": 14.781437125748504, "grad_norm": 0.0517578125, "learning_rate": 0.0008, "loss": 1.1366, "step": 9874 }, { "epoch": 14.782934131736527, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1408, "step": 9875 }, { "epoch": 14.784431137724551, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1383, "step": 9876 }, { "epoch": 14.785928143712574, "grad_norm": 0.16015625, "learning_rate": 0.0008, "loss": 1.1406, "step": 9877 }, { "epoch": 14.7874251497006, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1459, "step": 9878 }, { "epoch": 14.788922155688622, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1348, "step": 9879 }, { "epoch": 14.790419161676647, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1329, "step": 9880 }, { "epoch": 14.79191616766467, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1374, "step": 9881 }, { "epoch": 14.793413173652695, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1398, "step": 9882 }, { "epoch": 14.794910179640718, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1356, "step": 9883 }, { "epoch": 14.796407185628743, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1429, "step": 9884 }, { "epoch": 14.797904191616766, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1395, "step": 9885 }, { "epoch": 14.79940119760479, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1382, "step": 9886 }, { "epoch": 14.800898203592814, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1333, "step": 9887 }, { "epoch": 14.802395209580839, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1433, "step": 9888 }, { "epoch": 14.803892215568862, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1326, "step": 9889 }, { "epoch": 14.805389221556887, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1314, "step": 9890 }, { "epoch": 14.80688622754491, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1359, "step": 9891 }, { "epoch": 14.808383233532934, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1428, "step": 9892 }, { "epoch": 14.809880239520957, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.132, "step": 9893 }, { "epoch": 14.811377245508982, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1419, "step": 9894 }, { "epoch": 14.812874251497005, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1373, "step": 9895 }, { "epoch": 14.81437125748503, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1366, "step": 9896 }, { "epoch": 14.815868263473053, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1409, "step": 9897 }, { "epoch": 14.817365269461078, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1381, "step": 9898 }, { "epoch": 14.818862275449101, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1361, "step": 9899 }, { "epoch": 14.820359281437126, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1361, "step": 9900 }, { "epoch": 14.821856287425149, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1398, "step": 9901 }, { "epoch": 14.823353293413174, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.138, "step": 9902 }, { "epoch": 14.824850299401197, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1411, "step": 9903 }, { "epoch": 14.826347305389222, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1402, "step": 9904 }, { "epoch": 14.827844311377245, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1406, "step": 9905 }, { "epoch": 14.82934131736527, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1392, "step": 9906 }, { "epoch": 14.830838323353294, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1391, "step": 9907 }, { "epoch": 14.832335329341317, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1321, "step": 9908 }, { "epoch": 14.83383233532934, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1394, "step": 9909 }, { "epoch": 14.835329341317365, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1419, "step": 9910 }, { "epoch": 14.83682634730539, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1463, "step": 9911 }, { "epoch": 14.838323353293413, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1368, "step": 9912 }, { "epoch": 14.839820359281438, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1373, "step": 9913 }, { "epoch": 14.841317365269461, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1394, "step": 9914 }, { "epoch": 14.842814371257486, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1325, "step": 9915 }, { "epoch": 14.844311377245509, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.146, "step": 9916 }, { "epoch": 14.845808383233534, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1378, "step": 9917 }, { "epoch": 14.847305389221557, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1317, "step": 9918 }, { "epoch": 14.848802395209582, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1421, "step": 9919 }, { "epoch": 14.850299401197605, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1357, "step": 9920 }, { "epoch": 14.85179640718563, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1375, "step": 9921 }, { "epoch": 14.853293413173652, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1386, "step": 9922 }, { "epoch": 14.854790419161677, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.139, "step": 9923 }, { "epoch": 14.8562874251497, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1376, "step": 9924 }, { "epoch": 14.857784431137725, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1383, "step": 9925 }, { "epoch": 14.859281437125748, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1427, "step": 9926 }, { "epoch": 14.860778443113773, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.144, "step": 9927 }, { "epoch": 14.862275449101796, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1363, "step": 9928 }, { "epoch": 14.863772455089821, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1399, "step": 9929 }, { "epoch": 14.865269461077844, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1452, "step": 9930 }, { "epoch": 14.866766467065869, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1278, "step": 9931 }, { "epoch": 14.868263473053892, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1459, "step": 9932 }, { "epoch": 14.869760479041917, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1328, "step": 9933 }, { "epoch": 14.87125748502994, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1413, "step": 9934 }, { "epoch": 14.872754491017965, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1349, "step": 9935 }, { "epoch": 14.874251497005988, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1413, "step": 9936 }, { "epoch": 14.875748502994012, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1395, "step": 9937 }, { "epoch": 14.877245508982035, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1441, "step": 9938 }, { "epoch": 14.87874251497006, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1402, "step": 9939 }, { "epoch": 14.880239520958083, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1382, "step": 9940 }, { "epoch": 14.881736526946108, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1407, "step": 9941 }, { "epoch": 14.883233532934131, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1375, "step": 9942 }, { "epoch": 14.884730538922156, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1416, "step": 9943 }, { "epoch": 14.886227544910179, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1396, "step": 9944 }, { "epoch": 14.887724550898204, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1334, "step": 9945 }, { "epoch": 14.889221556886227, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1376, "step": 9946 }, { "epoch": 14.890718562874252, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.136, "step": 9947 }, { "epoch": 14.892215568862275, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1352, "step": 9948 }, { "epoch": 14.8937125748503, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1379, "step": 9949 }, { "epoch": 14.895209580838323, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1401, "step": 9950 }, { "epoch": 14.896706586826348, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1396, "step": 9951 }, { "epoch": 14.89820359281437, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1388, "step": 9952 }, { "epoch": 14.899700598802395, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1397, "step": 9953 }, { "epoch": 14.901197604790418, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1354, "step": 9954 }, { "epoch": 14.902694610778443, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1372, "step": 9955 }, { "epoch": 14.904191616766466, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1423, "step": 9956 }, { "epoch": 14.905688622754491, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1389, "step": 9957 }, { "epoch": 14.907185628742514, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1394, "step": 9958 }, { "epoch": 14.908682634730539, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1388, "step": 9959 }, { "epoch": 14.910179640718562, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1325, "step": 9960 }, { "epoch": 14.911676646706587, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1393, "step": 9961 }, { "epoch": 14.91317365269461, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1316, "step": 9962 }, { "epoch": 14.914670658682635, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1406, "step": 9963 }, { "epoch": 14.91616766467066, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1381, "step": 9964 }, { "epoch": 14.917664670658683, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1369, "step": 9965 }, { "epoch": 14.919161676646706, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1353, "step": 9966 }, { "epoch": 14.92065868263473, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1407, "step": 9967 }, { "epoch": 14.922155688622755, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1313, "step": 9968 }, { "epoch": 14.923652694610778, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.139, "step": 9969 }, { "epoch": 14.925149700598803, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1323, "step": 9970 }, { "epoch": 14.926646706586826, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.13, "step": 9971 }, { "epoch": 14.928143712574851, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1388, "step": 9972 }, { "epoch": 14.929640718562874, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1424, "step": 9973 }, { "epoch": 14.931137724550899, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1328, "step": 9974 }, { "epoch": 14.932634730538922, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1447, "step": 9975 }, { "epoch": 14.934131736526947, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1343, "step": 9976 }, { "epoch": 14.93562874251497, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.137, "step": 9977 }, { "epoch": 14.937125748502995, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1317, "step": 9978 }, { "epoch": 14.938622754491018, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1367, "step": 9979 }, { "epoch": 14.940119760479043, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1448, "step": 9980 }, { "epoch": 14.941616766467066, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.134, "step": 9981 }, { "epoch": 14.94311377245509, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1383, "step": 9982 }, { "epoch": 14.944610778443113, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1364, "step": 9983 }, { "epoch": 14.946107784431138, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1376, "step": 9984 }, { "epoch": 14.947604790419161, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1423, "step": 9985 }, { "epoch": 14.949101796407186, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1327, "step": 9986 }, { "epoch": 14.95059880239521, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1366, "step": 9987 }, { "epoch": 14.952095808383234, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1398, "step": 9988 }, { "epoch": 14.953592814371257, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1348, "step": 9989 }, { "epoch": 14.955089820359282, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1382, "step": 9990 }, { "epoch": 14.956586826347305, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1381, "step": 9991 }, { "epoch": 14.95808383233533, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1445, "step": 9992 }, { "epoch": 14.959580838323353, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1352, "step": 9993 }, { "epoch": 14.961077844311378, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.138, "step": 9994 }, { "epoch": 14.9625748502994, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1307, "step": 9995 }, { "epoch": 14.964071856287426, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.134, "step": 9996 }, { "epoch": 14.965568862275449, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1401, "step": 9997 }, { "epoch": 14.967065868263473, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1395, "step": 9998 }, { "epoch": 14.968562874251496, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1343, "step": 9999 }, { "epoch": 14.970059880239521, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1363, "step": 10000 }, { "epoch": 14.971556886227544, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.1435, "step": 10001 }, { "epoch": 14.97305389221557, "grad_norm": 0.7265625, "learning_rate": 0.0008, "loss": 1.1953, "step": 10002 }, { "epoch": 14.974550898203592, "grad_norm": 0.5703125, "learning_rate": 0.0008, "loss": 1.197, "step": 10003 }, { "epoch": 14.976047904191617, "grad_norm": 1.0859375, "learning_rate": 0.0008, "loss": 1.2462, "step": 10004 }, { "epoch": 14.97754491017964, "grad_norm": 0.8125, "learning_rate": 0.0008, "loss": 1.322, "step": 10005 }, { "epoch": 14.979041916167665, "grad_norm": 0.9453125, "learning_rate": 0.0008, "loss": 1.3121, "step": 10006 }, { "epoch": 14.980538922155688, "grad_norm": 1.90625, "learning_rate": 0.0008, "loss": 1.4386, "step": 10007 }, { "epoch": 14.982035928143713, "grad_norm": 0.91796875, "learning_rate": 0.0008, "loss": 1.415, "step": 10008 }, { "epoch": 14.983532934131736, "grad_norm": 1.421875, "learning_rate": 0.0008, "loss": 1.4384, "step": 10009 }, { "epoch": 14.98502994011976, "grad_norm": 1.09375, "learning_rate": 0.0008, "loss": 1.4024, "step": 10010 }, { "epoch": 14.986526946107784, "grad_norm": 1.4453125, "learning_rate": 0.0008, "loss": 1.4154, "step": 10011 }, { "epoch": 14.988023952095809, "grad_norm": 0.63671875, "learning_rate": 0.0008, "loss": 1.3349, "step": 10012 }, { "epoch": 14.989520958083832, "grad_norm": 0.71484375, "learning_rate": 0.0008, "loss": 1.3392, "step": 10013 }, { "epoch": 14.991017964071856, "grad_norm": 0.71875, "learning_rate": 0.0008, "loss": 1.3292, "step": 10014 }, { "epoch": 14.99251497005988, "grad_norm": 0.77734375, "learning_rate": 0.0008, "loss": 1.33, "step": 10015 }, { "epoch": 14.994011976047904, "grad_norm": 0.78125, "learning_rate": 0.0008, "loss": 1.3175, "step": 10016 }, { "epoch": 14.995508982035929, "grad_norm": 0.69921875, "learning_rate": 0.0008, "loss": 1.3063, "step": 10017 }, { "epoch": 14.997005988023952, "grad_norm": 0.73828125, "learning_rate": 0.0008, "loss": 1.2948, "step": 10018 }, { "epoch": 14.998502994011975, "grad_norm": 0.71484375, "learning_rate": 0.0008, "loss": 1.2935, "step": 10019 }, { "epoch": 15.0, "grad_norm": 0.6171875, "learning_rate": 0.0008, "loss": 1.3024, "step": 10020 }, { "epoch": 15.001497005988025, "grad_norm": 0.51953125, "learning_rate": 0.0008, "loss": 1.2768, "step": 10021 }, { "epoch": 15.002994011976048, "grad_norm": 0.5, "learning_rate": 0.0008, "loss": 1.2798, "step": 10022 }, { "epoch": 15.004491017964073, "grad_norm": 0.41796875, "learning_rate": 0.0008, "loss": 1.2622, "step": 10023 }, { "epoch": 15.005988023952096, "grad_norm": 0.4296875, "learning_rate": 0.0008, "loss": 1.2656, "step": 10024 }, { "epoch": 15.00748502994012, "grad_norm": 0.353515625, "learning_rate": 0.0008, "loss": 1.2513, "step": 10025 }, { "epoch": 15.008982035928144, "grad_norm": 0.330078125, "learning_rate": 0.0008, "loss": 1.2537, "step": 10026 }, { "epoch": 15.010479041916168, "grad_norm": 0.283203125, "learning_rate": 0.0008, "loss": 1.2473, "step": 10027 }, { "epoch": 15.011976047904191, "grad_norm": 0.279296875, "learning_rate": 0.0008, "loss": 1.2435, "step": 10028 }, { "epoch": 15.013473053892216, "grad_norm": 0.26171875, "learning_rate": 0.0008, "loss": 1.2405, "step": 10029 }, { "epoch": 15.01497005988024, "grad_norm": 0.2412109375, "learning_rate": 0.0008, "loss": 1.2356, "step": 10030 }, { "epoch": 15.016467065868264, "grad_norm": 0.2392578125, "learning_rate": 0.0008, "loss": 1.2365, "step": 10031 }, { "epoch": 15.017964071856287, "grad_norm": 0.2138671875, "learning_rate": 0.0008, "loss": 1.2272, "step": 10032 }, { "epoch": 15.019461077844312, "grad_norm": 0.20703125, "learning_rate": 0.0008, "loss": 1.2254, "step": 10033 }, { "epoch": 15.020958083832335, "grad_norm": 0.2177734375, "learning_rate": 0.0008, "loss": 1.2223, "step": 10034 }, { "epoch": 15.02245508982036, "grad_norm": 0.1962890625, "learning_rate": 0.0008, "loss": 1.2255, "step": 10035 }, { "epoch": 15.023952095808383, "grad_norm": 0.1806640625, "learning_rate": 0.0008, "loss": 1.2276, "step": 10036 }, { "epoch": 15.025449101796408, "grad_norm": 0.25390625, "learning_rate": 0.0008, "loss": 1.2096, "step": 10037 }, { "epoch": 15.02694610778443, "grad_norm": 0.220703125, "learning_rate": 0.0008, "loss": 1.2236, "step": 10038 }, { "epoch": 15.028443113772456, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.2168, "step": 10039 }, { "epoch": 15.029940119760479, "grad_norm": 0.1591796875, "learning_rate": 0.0008, "loss": 1.2155, "step": 10040 }, { "epoch": 15.031437125748504, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.2094, "step": 10041 }, { "epoch": 15.032934131736527, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.2182, "step": 10042 }, { "epoch": 15.034431137724551, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.2101, "step": 10043 }, { "epoch": 15.035928143712574, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.2025, "step": 10044 }, { "epoch": 15.0374251497006, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.2118, "step": 10045 }, { "epoch": 15.038922155688622, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.2121, "step": 10046 }, { "epoch": 15.040419161676647, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.2078, "step": 10047 }, { "epoch": 15.04191616766467, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1975, "step": 10048 }, { "epoch": 15.043413173652695, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.2134, "step": 10049 }, { "epoch": 15.044910179640718, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.2077, "step": 10050 }, { "epoch": 15.046407185628743, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.2121, "step": 10051 }, { "epoch": 15.047904191616766, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.2069, "step": 10052 }, { "epoch": 15.04940119760479, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.2011, "step": 10053 }, { "epoch": 15.050898203592814, "grad_norm": 0.1650390625, "learning_rate": 0.0008, "loss": 1.2021, "step": 10054 }, { "epoch": 15.052395209580839, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.2066, "step": 10055 }, { "epoch": 15.053892215568862, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.2001, "step": 10056 }, { "epoch": 15.055389221556887, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.2066, "step": 10057 }, { "epoch": 15.05688622754491, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1997, "step": 10058 }, { "epoch": 15.058383233532934, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.2026, "step": 10059 }, { "epoch": 15.059880239520957, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.1989, "step": 10060 }, { "epoch": 15.061377245508982, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.1982, "step": 10061 }, { "epoch": 15.062874251497005, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.195, "step": 10062 }, { "epoch": 15.06437125748503, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1939, "step": 10063 }, { "epoch": 15.065868263473053, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1967, "step": 10064 }, { "epoch": 15.067365269461078, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1991, "step": 10065 }, { "epoch": 15.068862275449101, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1929, "step": 10066 }, { "epoch": 15.070359281437126, "grad_norm": 0.162109375, "learning_rate": 0.0008, "loss": 1.1991, "step": 10067 }, { "epoch": 15.071856287425149, "grad_norm": 0.2080078125, "learning_rate": 0.0008, "loss": 1.1991, "step": 10068 }, { "epoch": 15.073353293413174, "grad_norm": 0.18359375, "learning_rate": 0.0008, "loss": 1.2034, "step": 10069 }, { "epoch": 15.074850299401197, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1983, "step": 10070 }, { "epoch": 15.076347305389222, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.2024, "step": 10071 }, { "epoch": 15.077844311377245, "grad_norm": 0.173828125, "learning_rate": 0.0008, "loss": 1.1969, "step": 10072 }, { "epoch": 15.07934131736527, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1962, "step": 10073 }, { "epoch": 15.080838323353293, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.2029, "step": 10074 }, { "epoch": 15.082335329341317, "grad_norm": 0.2421875, "learning_rate": 0.0008, "loss": 1.2002, "step": 10075 }, { "epoch": 15.08383233532934, "grad_norm": 0.271484375, "learning_rate": 0.0008, "loss": 1.2045, "step": 10076 }, { "epoch": 15.085329341317365, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1967, "step": 10077 }, { "epoch": 15.08682634730539, "grad_norm": 0.2080078125, "learning_rate": 0.0008, "loss": 1.1967, "step": 10078 }, { "epoch": 15.088323353293413, "grad_norm": 0.177734375, "learning_rate": 0.0008, "loss": 1.1964, "step": 10079 }, { "epoch": 15.089820359281438, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.1886, "step": 10080 }, { "epoch": 15.091317365269461, "grad_norm": 0.2080078125, "learning_rate": 0.0008, "loss": 1.2011, "step": 10081 }, { "epoch": 15.092814371257486, "grad_norm": 0.1875, "learning_rate": 0.0008, "loss": 1.1915, "step": 10082 }, { "epoch": 15.094311377245509, "grad_norm": 0.15625, "learning_rate": 0.0008, "loss": 1.1937, "step": 10083 }, { "epoch": 15.095808383233534, "grad_norm": 0.1943359375, "learning_rate": 0.0008, "loss": 1.1951, "step": 10084 }, { "epoch": 15.097305389221557, "grad_norm": 0.2001953125, "learning_rate": 0.0008, "loss": 1.1879, "step": 10085 }, { "epoch": 15.098802395209582, "grad_norm": 0.1728515625, "learning_rate": 0.0008, "loss": 1.1895, "step": 10086 }, { "epoch": 15.100299401197605, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1964, "step": 10087 }, { "epoch": 15.10179640718563, "grad_norm": 0.1875, "learning_rate": 0.0008, "loss": 1.1939, "step": 10088 }, { "epoch": 15.103293413173652, "grad_norm": 0.1796875, "learning_rate": 0.0008, "loss": 1.2019, "step": 10089 }, { "epoch": 15.104790419161677, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1935, "step": 10090 }, { "epoch": 15.1062874251497, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.1927, "step": 10091 }, { "epoch": 15.107784431137725, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.1872, "step": 10092 }, { "epoch": 15.109281437125748, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1908, "step": 10093 }, { "epoch": 15.110778443113773, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.185, "step": 10094 }, { "epoch": 15.112275449101796, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1925, "step": 10095 }, { "epoch": 15.113772455089821, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.187, "step": 10096 }, { "epoch": 15.115269461077844, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1898, "step": 10097 }, { "epoch": 15.116766467065869, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1908, "step": 10098 }, { "epoch": 15.118263473053892, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1927, "step": 10099 }, { "epoch": 15.119760479041917, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1933, "step": 10100 }, { "epoch": 15.12125748502994, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1867, "step": 10101 }, { "epoch": 15.122754491017965, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1823, "step": 10102 }, { "epoch": 15.124251497005988, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1814, "step": 10103 }, { "epoch": 15.125748502994012, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1911, "step": 10104 }, { "epoch": 15.127245508982035, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1813, "step": 10105 }, { "epoch": 15.12874251497006, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1893, "step": 10106 }, { "epoch": 15.130239520958083, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1838, "step": 10107 }, { "epoch": 15.131736526946108, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1872, "step": 10108 }, { "epoch": 15.133233532934131, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.1911, "step": 10109 }, { "epoch": 15.134730538922156, "grad_norm": 0.162109375, "learning_rate": 0.0008, "loss": 1.1896, "step": 10110 }, { "epoch": 15.136227544910179, "grad_norm": 0.2041015625, "learning_rate": 0.0008, "loss": 1.1874, "step": 10111 }, { "epoch": 15.137724550898204, "grad_norm": 0.185546875, "learning_rate": 0.0008, "loss": 1.1867, "step": 10112 }, { "epoch": 15.139221556886227, "grad_norm": 0.19921875, "learning_rate": 0.0008, "loss": 1.1852, "step": 10113 }, { "epoch": 15.140718562874252, "grad_norm": 0.232421875, "learning_rate": 0.0008, "loss": 1.1812, "step": 10114 }, { "epoch": 15.142215568862275, "grad_norm": 0.2451171875, "learning_rate": 0.0008, "loss": 1.1879, "step": 10115 }, { "epoch": 15.1437125748503, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.1872, "step": 10116 }, { "epoch": 15.145209580838323, "grad_norm": 0.177734375, "learning_rate": 0.0008, "loss": 1.1864, "step": 10117 }, { "epoch": 15.146706586826348, "grad_norm": 0.23828125, "learning_rate": 0.0008, "loss": 1.1896, "step": 10118 }, { "epoch": 15.14820359281437, "grad_norm": 0.357421875, "learning_rate": 0.0008, "loss": 1.1956, "step": 10119 }, { "epoch": 15.149700598802395, "grad_norm": 0.29296875, "learning_rate": 0.0008, "loss": 1.1972, "step": 10120 }, { "epoch": 15.151197604790418, "grad_norm": 0.171875, "learning_rate": 0.0008, "loss": 1.1886, "step": 10121 }, { "epoch": 15.152694610778443, "grad_norm": 0.2392578125, "learning_rate": 0.0008, "loss": 1.1964, "step": 10122 }, { "epoch": 15.154191616766466, "grad_norm": 0.275390625, "learning_rate": 0.0008, "loss": 1.1911, "step": 10123 }, { "epoch": 15.155688622754491, "grad_norm": 0.275390625, "learning_rate": 0.0008, "loss": 1.1974, "step": 10124 }, { "epoch": 15.157185628742514, "grad_norm": 0.263671875, "learning_rate": 0.0008, "loss": 1.1901, "step": 10125 }, { "epoch": 15.158682634730539, "grad_norm": 0.1630859375, "learning_rate": 0.0008, "loss": 1.1833, "step": 10126 }, { "epoch": 15.160179640718562, "grad_norm": 0.2314453125, "learning_rate": 0.0008, "loss": 1.184, "step": 10127 }, { "epoch": 15.161676646706587, "grad_norm": 0.23046875, "learning_rate": 0.0008, "loss": 1.1825, "step": 10128 }, { "epoch": 15.16317365269461, "grad_norm": 0.1884765625, "learning_rate": 0.0008, "loss": 1.1908, "step": 10129 }, { "epoch": 15.164670658682635, "grad_norm": 0.1865234375, "learning_rate": 0.0008, "loss": 1.186, "step": 10130 }, { "epoch": 15.16616766467066, "grad_norm": 0.25, "learning_rate": 0.0008, "loss": 1.1848, "step": 10131 }, { "epoch": 15.167664670658683, "grad_norm": 0.1806640625, "learning_rate": 0.0008, "loss": 1.1879, "step": 10132 }, { "epoch": 15.169161676646707, "grad_norm": 0.1494140625, "learning_rate": 0.0008, "loss": 1.1837, "step": 10133 }, { "epoch": 15.17065868263473, "grad_norm": 0.251953125, "learning_rate": 0.0008, "loss": 1.192, "step": 10134 }, { "epoch": 15.172155688622755, "grad_norm": 0.255859375, "learning_rate": 0.0008, "loss": 1.1921, "step": 10135 }, { "epoch": 15.173652694610778, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1854, "step": 10136 }, { "epoch": 15.175149700598803, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1867, "step": 10137 }, { "epoch": 15.176646706586826, "grad_norm": 0.173828125, "learning_rate": 0.0008, "loss": 1.187, "step": 10138 }, { "epoch": 15.178143712574851, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1886, "step": 10139 }, { "epoch": 15.179640718562874, "grad_norm": 0.169921875, "learning_rate": 0.0008, "loss": 1.1865, "step": 10140 }, { "epoch": 15.181137724550899, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1871, "step": 10141 }, { "epoch": 15.182634730538922, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1883, "step": 10142 }, { "epoch": 15.184131736526947, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1778, "step": 10143 }, { "epoch": 15.18562874251497, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.182, "step": 10144 }, { "epoch": 15.187125748502995, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1829, "step": 10145 }, { "epoch": 15.188622754491018, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1779, "step": 10146 }, { "epoch": 15.190119760479043, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1845, "step": 10147 }, { "epoch": 15.191616766467066, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1856, "step": 10148 }, { "epoch": 15.19311377245509, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1764, "step": 10149 }, { "epoch": 15.194610778443113, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1791, "step": 10150 }, { "epoch": 15.196107784431138, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1829, "step": 10151 }, { "epoch": 15.197604790419161, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1779, "step": 10152 }, { "epoch": 15.199101796407186, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1816, "step": 10153 }, { "epoch": 15.20059880239521, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1851, "step": 10154 }, { "epoch": 15.202095808383234, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1788, "step": 10155 }, { "epoch": 15.203592814371257, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1736, "step": 10156 }, { "epoch": 15.205089820359282, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1862, "step": 10157 }, { "epoch": 15.206586826347305, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.179, "step": 10158 }, { "epoch": 15.20808383233533, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1776, "step": 10159 }, { "epoch": 15.209580838323353, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.172, "step": 10160 }, { "epoch": 15.211077844311378, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.176, "step": 10161 }, { "epoch": 15.2125748502994, "grad_norm": 0.1591796875, "learning_rate": 0.0008, "loss": 1.1837, "step": 10162 }, { "epoch": 15.214071856287426, "grad_norm": 0.21484375, "learning_rate": 0.0008, "loss": 1.1855, "step": 10163 }, { "epoch": 15.215568862275449, "grad_norm": 0.259765625, "learning_rate": 0.0008, "loss": 1.1963, "step": 10164 }, { "epoch": 15.217065868263473, "grad_norm": 0.1904296875, "learning_rate": 0.0008, "loss": 1.1825, "step": 10165 }, { "epoch": 15.218562874251496, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.1807, "step": 10166 }, { "epoch": 15.220059880239521, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.1775, "step": 10167 }, { "epoch": 15.221556886227544, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1853, "step": 10168 }, { "epoch": 15.22305389221557, "grad_norm": 0.1796875, "learning_rate": 0.0008, "loss": 1.1803, "step": 10169 }, { "epoch": 15.224550898203592, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.188, "step": 10170 }, { "epoch": 15.226047904191617, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1799, "step": 10171 }, { "epoch": 15.22754491017964, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.175, "step": 10172 }, { "epoch": 15.229041916167665, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1806, "step": 10173 }, { "epoch": 15.230538922155688, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1758, "step": 10174 }, { "epoch": 15.232035928143713, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1775, "step": 10175 }, { "epoch": 15.233532934131736, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1863, "step": 10176 }, { "epoch": 15.23502994011976, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1771, "step": 10177 }, { "epoch": 15.236526946107784, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1763, "step": 10178 }, { "epoch": 15.238023952095809, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1832, "step": 10179 }, { "epoch": 15.239520958083832, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.183, "step": 10180 }, { "epoch": 15.241017964071856, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1787, "step": 10181 }, { "epoch": 15.24251497005988, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1767, "step": 10182 }, { "epoch": 15.244011976047904, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1749, "step": 10183 }, { "epoch": 15.245508982035927, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1788, "step": 10184 }, { "epoch": 15.247005988023952, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1799, "step": 10185 }, { "epoch": 15.248502994011975, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1837, "step": 10186 }, { "epoch": 15.25, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1717, "step": 10187 }, { "epoch": 15.251497005988025, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1763, "step": 10188 }, { "epoch": 15.252994011976048, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1751, "step": 10189 }, { "epoch": 15.254491017964073, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1748, "step": 10190 }, { "epoch": 15.255988023952096, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1754, "step": 10191 }, { "epoch": 15.25748502994012, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1802, "step": 10192 }, { "epoch": 15.258982035928144, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1813, "step": 10193 }, { "epoch": 15.260479041916168, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1771, "step": 10194 }, { "epoch": 15.261976047904191, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1715, "step": 10195 }, { "epoch": 15.263473053892216, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1703, "step": 10196 }, { "epoch": 15.26497005988024, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.166, "step": 10197 }, { "epoch": 15.266467065868264, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1747, "step": 10198 }, { "epoch": 15.267964071856287, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1742, "step": 10199 }, { "epoch": 15.269461077844312, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1746, "step": 10200 }, { "epoch": 15.270958083832335, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1698, "step": 10201 }, { "epoch": 15.27245508982036, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1697, "step": 10202 }, { "epoch": 15.273952095808383, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1774, "step": 10203 }, { "epoch": 15.275449101796408, "grad_norm": 0.212890625, "learning_rate": 0.0008, "loss": 1.1756, "step": 10204 }, { "epoch": 15.27694610778443, "grad_norm": 0.28515625, "learning_rate": 0.0008, "loss": 1.1756, "step": 10205 }, { "epoch": 15.278443113772456, "grad_norm": 0.4375, "learning_rate": 0.0008, "loss": 1.1811, "step": 10206 }, { "epoch": 15.279940119760479, "grad_norm": 0.66796875, "learning_rate": 0.0008, "loss": 1.1842, "step": 10207 }, { "epoch": 15.281437125748504, "grad_norm": 1.0625, "learning_rate": 0.0008, "loss": 1.2017, "step": 10208 }, { "epoch": 15.282934131736527, "grad_norm": 1.3203125, "learning_rate": 0.0008, "loss": 1.211, "step": 10209 }, { "epoch": 15.284431137724551, "grad_norm": 0.400390625, "learning_rate": 0.0008, "loss": 1.1962, "step": 10210 }, { "epoch": 15.285928143712574, "grad_norm": 0.7109375, "learning_rate": 0.0008, "loss": 1.2019, "step": 10211 }, { "epoch": 15.2874251497006, "grad_norm": 0.6171875, "learning_rate": 0.0008, "loss": 1.19, "step": 10212 }, { "epoch": 15.288922155688622, "grad_norm": 0.37890625, "learning_rate": 0.0008, "loss": 1.1935, "step": 10213 }, { "epoch": 15.290419161676647, "grad_norm": 0.58984375, "learning_rate": 0.0008, "loss": 1.1886, "step": 10214 }, { "epoch": 15.29191616766467, "grad_norm": 0.2421875, "learning_rate": 0.0008, "loss": 1.1925, "step": 10215 }, { "epoch": 15.293413173652695, "grad_norm": 0.361328125, "learning_rate": 0.0008, "loss": 1.1823, "step": 10216 }, { "epoch": 15.294910179640718, "grad_norm": 0.314453125, "learning_rate": 0.0008, "loss": 1.1803, "step": 10217 }, { "epoch": 15.296407185628743, "grad_norm": 0.2138671875, "learning_rate": 0.0008, "loss": 1.1814, "step": 10218 }, { "epoch": 15.297904191616766, "grad_norm": 0.216796875, "learning_rate": 0.0008, "loss": 1.1821, "step": 10219 }, { "epoch": 15.29940119760479, "grad_norm": 0.197265625, "learning_rate": 0.0008, "loss": 1.1714, "step": 10220 }, { "epoch": 15.300898203592814, "grad_norm": 0.1806640625, "learning_rate": 0.0008, "loss": 1.1753, "step": 10221 }, { "epoch": 15.302395209580839, "grad_norm": 0.1865234375, "learning_rate": 0.0008, "loss": 1.1812, "step": 10222 }, { "epoch": 15.303892215568862, "grad_norm": 0.1572265625, "learning_rate": 0.0008, "loss": 1.1746, "step": 10223 }, { "epoch": 15.305389221556887, "grad_norm": 0.1806640625, "learning_rate": 0.0008, "loss": 1.1741, "step": 10224 }, { "epoch": 15.30688622754491, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1798, "step": 10225 }, { "epoch": 15.308383233532934, "grad_norm": 0.162109375, "learning_rate": 0.0008, "loss": 1.1718, "step": 10226 }, { "epoch": 15.309880239520957, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1733, "step": 10227 }, { "epoch": 15.311377245508982, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.17, "step": 10228 }, { "epoch": 15.312874251497005, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1695, "step": 10229 }, { "epoch": 15.31437125748503, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1696, "step": 10230 }, { "epoch": 15.315868263473053, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1637, "step": 10231 }, { "epoch": 15.317365269461078, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1708, "step": 10232 }, { "epoch": 15.318862275449101, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.172, "step": 10233 }, { "epoch": 15.320359281437126, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1734, "step": 10234 }, { "epoch": 15.321856287425149, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1715, "step": 10235 }, { "epoch": 15.323353293413174, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1648, "step": 10236 }, { "epoch": 15.324850299401197, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1736, "step": 10237 }, { "epoch": 15.326347305389222, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1684, "step": 10238 }, { "epoch": 15.327844311377245, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1694, "step": 10239 }, { "epoch": 15.32934131736527, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1621, "step": 10240 }, { "epoch": 15.330838323353294, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1726, "step": 10241 }, { "epoch": 15.332335329341317, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1757, "step": 10242 }, { "epoch": 15.33383233532934, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1782, "step": 10243 }, { "epoch": 15.335329341317365, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1713, "step": 10244 }, { "epoch": 15.33682634730539, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1693, "step": 10245 }, { "epoch": 15.338323353293413, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1737, "step": 10246 }, { "epoch": 15.339820359281438, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1747, "step": 10247 }, { "epoch": 15.341317365269461, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1707, "step": 10248 }, { "epoch": 15.342814371257486, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.164, "step": 10249 }, { "epoch": 15.344311377245509, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1663, "step": 10250 }, { "epoch": 15.345808383233534, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1603, "step": 10251 }, { "epoch": 15.347305389221557, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1664, "step": 10252 }, { "epoch": 15.348802395209582, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1656, "step": 10253 }, { "epoch": 15.350299401197605, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1663, "step": 10254 }, { "epoch": 15.35179640718563, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1658, "step": 10255 }, { "epoch": 15.353293413173652, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1655, "step": 10256 }, { "epoch": 15.354790419161677, "grad_norm": 0.14453125, "learning_rate": 0.0008, "loss": 1.1631, "step": 10257 }, { "epoch": 15.3562874251497, "grad_norm": 0.181640625, "learning_rate": 0.0008, "loss": 1.1724, "step": 10258 }, { "epoch": 15.357784431137725, "grad_norm": 0.18359375, "learning_rate": 0.0008, "loss": 1.1674, "step": 10259 }, { "epoch": 15.359281437125748, "grad_norm": 0.1787109375, "learning_rate": 0.0008, "loss": 1.1656, "step": 10260 }, { "epoch": 15.360778443113773, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1653, "step": 10261 }, { "epoch": 15.362275449101796, "grad_norm": 0.208984375, "learning_rate": 0.0008, "loss": 1.1645, "step": 10262 }, { "epoch": 15.363772455089821, "grad_norm": 0.291015625, "learning_rate": 0.0008, "loss": 1.1746, "step": 10263 }, { "epoch": 15.365269461077844, "grad_norm": 0.29296875, "learning_rate": 0.0008, "loss": 1.1789, "step": 10264 }, { "epoch": 15.366766467065869, "grad_norm": 0.1748046875, "learning_rate": 0.0008, "loss": 1.1737, "step": 10265 }, { "epoch": 15.368263473053892, "grad_norm": 0.259765625, "learning_rate": 0.0008, "loss": 1.1677, "step": 10266 }, { "epoch": 15.369760479041917, "grad_norm": 0.2255859375, "learning_rate": 0.0008, "loss": 1.1637, "step": 10267 }, { "epoch": 15.37125748502994, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.1624, "step": 10268 }, { "epoch": 15.372754491017965, "grad_norm": 0.21875, "learning_rate": 0.0008, "loss": 1.1657, "step": 10269 }, { "epoch": 15.374251497005988, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.164, "step": 10270 }, { "epoch": 15.375748502994012, "grad_norm": 0.177734375, "learning_rate": 0.0008, "loss": 1.1686, "step": 10271 }, { "epoch": 15.377245508982035, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1666, "step": 10272 }, { "epoch": 15.37874251497006, "grad_norm": 0.15234375, "learning_rate": 0.0008, "loss": 1.1666, "step": 10273 }, { "epoch": 15.380239520958083, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.159, "step": 10274 }, { "epoch": 15.381736526946108, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.163, "step": 10275 }, { "epoch": 15.383233532934131, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1669, "step": 10276 }, { "epoch": 15.384730538922156, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1725, "step": 10277 }, { "epoch": 15.386227544910179, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.1594, "step": 10278 }, { "epoch": 15.387724550898204, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1586, "step": 10279 }, { "epoch": 15.389221556886227, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.1613, "step": 10280 }, { "epoch": 15.390718562874252, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1625, "step": 10281 }, { "epoch": 15.392215568862275, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1702, "step": 10282 }, { "epoch": 15.3937125748503, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1576, "step": 10283 }, { "epoch": 15.395209580838323, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1593, "step": 10284 }, { "epoch": 15.396706586826348, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1568, "step": 10285 }, { "epoch": 15.39820359281437, "grad_norm": 0.1875, "learning_rate": 0.0008, "loss": 1.1603, "step": 10286 }, { "epoch": 15.399700598802395, "grad_norm": 0.203125, "learning_rate": 0.0008, "loss": 1.1577, "step": 10287 }, { "epoch": 15.401197604790418, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1656, "step": 10288 }, { "epoch": 15.402694610778443, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1624, "step": 10289 }, { "epoch": 15.404191616766466, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1657, "step": 10290 }, { "epoch": 15.405688622754491, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1584, "step": 10291 }, { "epoch": 15.407185628742514, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1592, "step": 10292 }, { "epoch": 15.408682634730539, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1533, "step": 10293 }, { "epoch": 15.410179640718562, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1605, "step": 10294 }, { "epoch": 15.411676646706587, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1563, "step": 10295 }, { "epoch": 15.41317365269461, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1627, "step": 10296 }, { "epoch": 15.414670658682635, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1555, "step": 10297 }, { "epoch": 15.41616766467066, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.157, "step": 10298 }, { "epoch": 15.417664670658683, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1598, "step": 10299 }, { "epoch": 15.419161676646706, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1579, "step": 10300 }, { "epoch": 15.42065868263473, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1524, "step": 10301 }, { "epoch": 15.422155688622755, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1597, "step": 10302 }, { "epoch": 15.423652694610778, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1601, "step": 10303 }, { "epoch": 15.425149700598803, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1565, "step": 10304 }, { "epoch": 15.426646706586826, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1585, "step": 10305 }, { "epoch": 15.428143712574851, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1529, "step": 10306 }, { "epoch": 15.429640718562874, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1569, "step": 10307 }, { "epoch": 15.431137724550899, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1521, "step": 10308 }, { "epoch": 15.432634730538922, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1514, "step": 10309 }, { "epoch": 15.434131736526947, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1517, "step": 10310 }, { "epoch": 15.43562874251497, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1582, "step": 10311 }, { "epoch": 15.437125748502995, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1616, "step": 10312 }, { "epoch": 15.438622754491018, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1554, "step": 10313 }, { "epoch": 15.440119760479043, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.152, "step": 10314 }, { "epoch": 15.441616766467066, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1562, "step": 10315 }, { "epoch": 15.44311377245509, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1528, "step": 10316 }, { "epoch": 15.444610778443113, "grad_norm": 0.208984375, "learning_rate": 0.0008, "loss": 1.1558, "step": 10317 }, { "epoch": 15.446107784431138, "grad_norm": 0.24609375, "learning_rate": 0.0008, "loss": 1.1567, "step": 10318 }, { "epoch": 15.447604790419161, "grad_norm": 0.27734375, "learning_rate": 0.0008, "loss": 1.1573, "step": 10319 }, { "epoch": 15.449101796407186, "grad_norm": 0.291015625, "learning_rate": 0.0008, "loss": 1.1663, "step": 10320 }, { "epoch": 15.45059880239521, "grad_norm": 0.279296875, "learning_rate": 0.0008, "loss": 1.1592, "step": 10321 }, { "epoch": 15.452095808383234, "grad_norm": 0.16796875, "learning_rate": 0.0008, "loss": 1.1534, "step": 10322 }, { "epoch": 15.453592814371257, "grad_norm": 0.1884765625, "learning_rate": 0.0008, "loss": 1.1607, "step": 10323 }, { "epoch": 15.455089820359282, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1545, "step": 10324 }, { "epoch": 15.456586826347305, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1568, "step": 10325 }, { "epoch": 15.45808383233533, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1555, "step": 10326 }, { "epoch": 15.459580838323353, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1584, "step": 10327 }, { "epoch": 15.461077844311378, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.16, "step": 10328 }, { "epoch": 15.4625748502994, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1597, "step": 10329 }, { "epoch": 15.464071856287426, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1532, "step": 10330 }, { "epoch": 15.465568862275449, "grad_norm": 0.15625, "learning_rate": 0.0008, "loss": 1.1508, "step": 10331 }, { "epoch": 15.467065868263473, "grad_norm": 0.2197265625, "learning_rate": 0.0008, "loss": 1.1585, "step": 10332 }, { "epoch": 15.468562874251496, "grad_norm": 0.271484375, "learning_rate": 0.0008, "loss": 1.1617, "step": 10333 }, { "epoch": 15.470059880239521, "grad_norm": 0.1708984375, "learning_rate": 0.0008, "loss": 1.1609, "step": 10334 }, { "epoch": 15.471556886227544, "grad_norm": 0.30078125, "learning_rate": 0.0008, "loss": 1.1608, "step": 10335 }, { "epoch": 15.47305389221557, "grad_norm": 0.32421875, "learning_rate": 0.0008, "loss": 1.1592, "step": 10336 }, { "epoch": 15.474550898203592, "grad_norm": 0.34375, "learning_rate": 0.0008, "loss": 1.1598, "step": 10337 }, { "epoch": 15.476047904191617, "grad_norm": 0.2177734375, "learning_rate": 0.0008, "loss": 1.156, "step": 10338 }, { "epoch": 15.47754491017964, "grad_norm": 0.265625, "learning_rate": 0.0008, "loss": 1.16, "step": 10339 }, { "epoch": 15.479041916167665, "grad_norm": 0.2890625, "learning_rate": 0.0008, "loss": 1.161, "step": 10340 }, { "epoch": 15.480538922155688, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.1555, "step": 10341 }, { "epoch": 15.482035928143713, "grad_norm": 0.212890625, "learning_rate": 0.0008, "loss": 1.1576, "step": 10342 }, { "epoch": 15.483532934131736, "grad_norm": 0.2353515625, "learning_rate": 0.0008, "loss": 1.1636, "step": 10343 }, { "epoch": 15.48502994011976, "grad_norm": 0.1728515625, "learning_rate": 0.0008, "loss": 1.1559, "step": 10344 }, { "epoch": 15.486526946107784, "grad_norm": 0.25390625, "learning_rate": 0.0008, "loss": 1.1585, "step": 10345 }, { "epoch": 15.488023952095809, "grad_norm": 0.20703125, "learning_rate": 0.0008, "loss": 1.1547, "step": 10346 }, { "epoch": 15.489520958083832, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1529, "step": 10347 }, { "epoch": 15.491017964071856, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.1527, "step": 10348 }, { "epoch": 15.49251497005988, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1539, "step": 10349 }, { "epoch": 15.494011976047904, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1589, "step": 10350 }, { "epoch": 15.495508982035927, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.161, "step": 10351 }, { "epoch": 15.497005988023952, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.151, "step": 10352 }, { "epoch": 15.498502994011975, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1538, "step": 10353 }, { "epoch": 15.5, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1566, "step": 10354 }, { "epoch": 15.501497005988025, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.154, "step": 10355 }, { "epoch": 15.502994011976048, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.155, "step": 10356 }, { "epoch": 15.504491017964071, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1457, "step": 10357 }, { "epoch": 15.505988023952096, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1486, "step": 10358 }, { "epoch": 15.50748502994012, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1554, "step": 10359 }, { "epoch": 15.508982035928144, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1525, "step": 10360 }, { "epoch": 15.510479041916168, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1507, "step": 10361 }, { "epoch": 15.511976047904191, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1575, "step": 10362 }, { "epoch": 15.513473053892216, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.155, "step": 10363 }, { "epoch": 15.51497005988024, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1564, "step": 10364 }, { "epoch": 15.516467065868264, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.152, "step": 10365 }, { "epoch": 15.517964071856287, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1519, "step": 10366 }, { "epoch": 15.519461077844312, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1526, "step": 10367 }, { "epoch": 15.520958083832335, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1488, "step": 10368 }, { "epoch": 15.52245508982036, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1573, "step": 10369 }, { "epoch": 15.523952095808383, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1499, "step": 10370 }, { "epoch": 15.525449101796408, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1482, "step": 10371 }, { "epoch": 15.52694610778443, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1481, "step": 10372 }, { "epoch": 15.528443113772456, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1562, "step": 10373 }, { "epoch": 15.529940119760479, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1533, "step": 10374 }, { "epoch": 15.531437125748504, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.143, "step": 10375 }, { "epoch": 15.532934131736527, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1512, "step": 10376 }, { "epoch": 15.534431137724551, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1476, "step": 10377 }, { "epoch": 15.535928143712574, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1492, "step": 10378 }, { "epoch": 15.5374251497006, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1475, "step": 10379 }, { "epoch": 15.538922155688622, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1484, "step": 10380 }, { "epoch": 15.540419161676647, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1562, "step": 10381 }, { "epoch": 15.54191616766467, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1464, "step": 10382 }, { "epoch": 15.543413173652695, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1474, "step": 10383 }, { "epoch": 15.544910179640718, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1546, "step": 10384 }, { "epoch": 15.546407185628743, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1411, "step": 10385 }, { "epoch": 15.547904191616766, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1474, "step": 10386 }, { "epoch": 15.54940119760479, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1559, "step": 10387 }, { "epoch": 15.550898203592814, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1433, "step": 10388 }, { "epoch": 15.552395209580839, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1505, "step": 10389 }, { "epoch": 15.553892215568862, "grad_norm": 0.1630859375, "learning_rate": 0.0008, "loss": 1.1525, "step": 10390 }, { "epoch": 15.555389221556887, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1492, "step": 10391 }, { "epoch": 15.55688622754491, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1506, "step": 10392 }, { "epoch": 15.558383233532934, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1434, "step": 10393 }, { "epoch": 15.559880239520957, "grad_norm": 0.158203125, "learning_rate": 0.0008, "loss": 1.1532, "step": 10394 }, { "epoch": 15.561377245508982, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1472, "step": 10395 }, { "epoch": 15.562874251497005, "grad_norm": 0.1884765625, "learning_rate": 0.0008, "loss": 1.1519, "step": 10396 }, { "epoch": 15.56437125748503, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1481, "step": 10397 }, { "epoch": 15.565868263473053, "grad_norm": 0.16796875, "learning_rate": 0.0008, "loss": 1.153, "step": 10398 }, { "epoch": 15.567365269461078, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.1453, "step": 10399 }, { "epoch": 15.568862275449101, "grad_norm": 0.1533203125, "learning_rate": 0.0008, "loss": 1.1464, "step": 10400 }, { "epoch": 15.570359281437126, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1493, "step": 10401 }, { "epoch": 15.571856287425149, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.1427, "step": 10402 }, { "epoch": 15.573353293413174, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1519, "step": 10403 }, { "epoch": 15.574850299401197, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1465, "step": 10404 }, { "epoch": 15.576347305389222, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.148, "step": 10405 }, { "epoch": 15.577844311377245, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1472, "step": 10406 }, { "epoch": 15.57934131736527, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1467, "step": 10407 }, { "epoch": 15.580838323353294, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1441, "step": 10408 }, { "epoch": 15.582335329341317, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1511, "step": 10409 }, { "epoch": 15.58383233532934, "grad_norm": 0.166015625, "learning_rate": 0.0008, "loss": 1.1476, "step": 10410 }, { "epoch": 15.585329341317365, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1513, "step": 10411 }, { "epoch": 15.58682634730539, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1509, "step": 10412 }, { "epoch": 15.588323353293413, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1504, "step": 10413 }, { "epoch": 15.589820359281438, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1544, "step": 10414 }, { "epoch": 15.591317365269461, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.151, "step": 10415 }, { "epoch": 15.592814371257486, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1552, "step": 10416 }, { "epoch": 15.594311377245509, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1474, "step": 10417 }, { "epoch": 15.595808383233534, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1487, "step": 10418 }, { "epoch": 15.597305389221557, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1431, "step": 10419 }, { "epoch": 15.598802395209582, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1476, "step": 10420 }, { "epoch": 15.600299401197605, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1438, "step": 10421 }, { "epoch": 15.60179640718563, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1488, "step": 10422 }, { "epoch": 15.603293413173652, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.153, "step": 10423 }, { "epoch": 15.604790419161677, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1501, "step": 10424 }, { "epoch": 15.6062874251497, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1453, "step": 10425 }, { "epoch": 15.607784431137725, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1452, "step": 10426 }, { "epoch": 15.609281437125748, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1471, "step": 10427 }, { "epoch": 15.610778443113773, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1495, "step": 10428 }, { "epoch": 15.612275449101796, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1411, "step": 10429 }, { "epoch": 15.613772455089821, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.157, "step": 10430 }, { "epoch": 15.615269461077844, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1513, "step": 10431 }, { "epoch": 15.616766467065869, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.151, "step": 10432 }, { "epoch": 15.618263473053892, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1431, "step": 10433 }, { "epoch": 15.619760479041917, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1483, "step": 10434 }, { "epoch": 15.62125748502994, "grad_norm": 0.19921875, "learning_rate": 0.0008, "loss": 1.1463, "step": 10435 }, { "epoch": 15.622754491017965, "grad_norm": 0.16796875, "learning_rate": 0.0008, "loss": 1.1538, "step": 10436 }, { "epoch": 15.624251497005988, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1529, "step": 10437 }, { "epoch": 15.625748502994012, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1431, "step": 10438 }, { "epoch": 15.627245508982035, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.144, "step": 10439 }, { "epoch": 15.62874251497006, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1475, "step": 10440 }, { "epoch": 15.630239520958083, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1427, "step": 10441 }, { "epoch": 15.631736526946108, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.148, "step": 10442 }, { "epoch": 15.633233532934131, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1465, "step": 10443 }, { "epoch": 15.634730538922156, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.1437, "step": 10444 }, { "epoch": 15.636227544910179, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1385, "step": 10445 }, { "epoch": 15.637724550898204, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1467, "step": 10446 }, { "epoch": 15.639221556886227, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1459, "step": 10447 }, { "epoch": 15.640718562874252, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1487, "step": 10448 }, { "epoch": 15.642215568862275, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1483, "step": 10449 }, { "epoch": 15.6437125748503, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1457, "step": 10450 }, { "epoch": 15.645209580838323, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1418, "step": 10451 }, { "epoch": 15.646706586826348, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1379, "step": 10452 }, { "epoch": 15.64820359281437, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1488, "step": 10453 }, { "epoch": 15.649700598802395, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1457, "step": 10454 }, { "epoch": 15.651197604790418, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1458, "step": 10455 }, { "epoch": 15.652694610778443, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1448, "step": 10456 }, { "epoch": 15.654191616766466, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1507, "step": 10457 }, { "epoch": 15.655688622754491, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1427, "step": 10458 }, { "epoch": 15.657185628742514, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1534, "step": 10459 }, { "epoch": 15.658682634730539, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1493, "step": 10460 }, { "epoch": 15.660179640718562, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1466, "step": 10461 }, { "epoch": 15.661676646706587, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1468, "step": 10462 }, { "epoch": 15.66317365269461, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1452, "step": 10463 }, { "epoch": 15.664670658682635, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1491, "step": 10464 }, { "epoch": 15.66616766467066, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1479, "step": 10465 }, { "epoch": 15.667664670658683, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1441, "step": 10466 }, { "epoch": 15.669161676646706, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1496, "step": 10467 }, { "epoch": 15.67065868263473, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1452, "step": 10468 }, { "epoch": 15.672155688622755, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1437, "step": 10469 }, { "epoch": 15.673652694610778, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1394, "step": 10470 }, { "epoch": 15.675149700598803, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1458, "step": 10471 }, { "epoch": 15.676646706586826, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.151, "step": 10472 }, { "epoch": 15.678143712574851, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1455, "step": 10473 }, { "epoch": 15.679640718562874, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1443, "step": 10474 }, { "epoch": 15.681137724550899, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.1401, "step": 10475 }, { "epoch": 15.682634730538922, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1457, "step": 10476 }, { "epoch": 15.684131736526947, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1479, "step": 10477 }, { "epoch": 15.68562874251497, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1421, "step": 10478 }, { "epoch": 15.687125748502995, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1446, "step": 10479 }, { "epoch": 15.688622754491018, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1518, "step": 10480 }, { "epoch": 15.690119760479043, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1494, "step": 10481 }, { "epoch": 15.691616766467066, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1452, "step": 10482 }, { "epoch": 15.69311377245509, "grad_norm": 0.193359375, "learning_rate": 0.0008, "loss": 1.1471, "step": 10483 }, { "epoch": 15.694610778443113, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1422, "step": 10484 }, { "epoch": 15.696107784431138, "grad_norm": 0.158203125, "learning_rate": 0.0008, "loss": 1.1413, "step": 10485 }, { "epoch": 15.697604790419161, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1405, "step": 10486 }, { "epoch": 15.699101796407186, "grad_norm": 0.1572265625, "learning_rate": 0.0008, "loss": 1.1456, "step": 10487 }, { "epoch": 15.70059880239521, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1536, "step": 10488 }, { "epoch": 15.702095808383234, "grad_norm": 0.16015625, "learning_rate": 0.0008, "loss": 1.144, "step": 10489 }, { "epoch": 15.703592814371257, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.15, "step": 10490 }, { "epoch": 15.705089820359282, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1397, "step": 10491 }, { "epoch": 15.706586826347305, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1479, "step": 10492 }, { "epoch": 15.70808383233533, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1458, "step": 10493 }, { "epoch": 15.709580838323353, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1487, "step": 10494 }, { "epoch": 15.711077844311378, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.1413, "step": 10495 }, { "epoch": 15.7125748502994, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.14, "step": 10496 }, { "epoch": 15.714071856287426, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1449, "step": 10497 }, { "epoch": 15.715568862275449, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1458, "step": 10498 }, { "epoch": 15.717065868263473, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1502, "step": 10499 }, { "epoch": 15.718562874251496, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1477, "step": 10500 }, { "epoch": 15.720059880239521, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1439, "step": 10501 }, { "epoch": 15.721556886227544, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1439, "step": 10502 }, { "epoch": 15.72305389221557, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1473, "step": 10503 }, { "epoch": 15.724550898203592, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1467, "step": 10504 }, { "epoch": 15.726047904191617, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1422, "step": 10505 }, { "epoch": 15.72754491017964, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1403, "step": 10506 }, { "epoch": 15.729041916167665, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1448, "step": 10507 }, { "epoch": 15.730538922155688, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1419, "step": 10508 }, { "epoch": 15.732035928143713, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1428, "step": 10509 }, { "epoch": 15.733532934131736, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1477, "step": 10510 }, { "epoch": 15.73502994011976, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1471, "step": 10511 }, { "epoch": 15.736526946107784, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1534, "step": 10512 }, { "epoch": 15.738023952095809, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1514, "step": 10513 }, { "epoch": 15.739520958083832, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1449, "step": 10514 }, { "epoch": 15.741017964071856, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1422, "step": 10515 }, { "epoch": 15.74251497005988, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1485, "step": 10516 }, { "epoch": 15.744011976047904, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1467, "step": 10517 }, { "epoch": 15.745508982035929, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1424, "step": 10518 }, { "epoch": 15.747005988023952, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.138, "step": 10519 }, { "epoch": 15.748502994011975, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1437, "step": 10520 }, { "epoch": 15.75, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1447, "step": 10521 }, { "epoch": 15.751497005988025, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1414, "step": 10522 }, { "epoch": 15.752994011976048, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1486, "step": 10523 }, { "epoch": 15.754491017964071, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1414, "step": 10524 }, { "epoch": 15.755988023952096, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1365, "step": 10525 }, { "epoch": 15.75748502994012, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1507, "step": 10526 }, { "epoch": 15.758982035928144, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1415, "step": 10527 }, { "epoch": 15.760479041916168, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1464, "step": 10528 }, { "epoch": 15.761976047904191, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1475, "step": 10529 }, { "epoch": 15.763473053892216, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1485, "step": 10530 }, { "epoch": 15.76497005988024, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1409, "step": 10531 }, { "epoch": 15.766467065868264, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1464, "step": 10532 }, { "epoch": 15.767964071856287, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1394, "step": 10533 }, { "epoch": 15.769461077844312, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1496, "step": 10534 }, { "epoch": 15.770958083832335, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1402, "step": 10535 }, { "epoch": 15.77245508982036, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1477, "step": 10536 }, { "epoch": 15.773952095808383, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.144, "step": 10537 }, { "epoch": 15.775449101796408, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1485, "step": 10538 }, { "epoch": 15.77694610778443, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.142, "step": 10539 }, { "epoch": 15.778443113772456, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1416, "step": 10540 }, { "epoch": 15.779940119760479, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1424, "step": 10541 }, { "epoch": 15.781437125748504, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1412, "step": 10542 }, { "epoch": 15.782934131736527, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1454, "step": 10543 }, { "epoch": 15.784431137724551, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1494, "step": 10544 }, { "epoch": 15.785928143712574, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1421, "step": 10545 }, { "epoch": 15.7874251497006, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1412, "step": 10546 }, { "epoch": 15.788922155688622, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1499, "step": 10547 }, { "epoch": 15.790419161676647, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1399, "step": 10548 }, { "epoch": 15.79191616766467, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.146, "step": 10549 }, { "epoch": 15.793413173652695, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1474, "step": 10550 }, { "epoch": 15.794910179640718, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1454, "step": 10551 }, { "epoch": 15.796407185628743, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1452, "step": 10552 }, { "epoch": 15.797904191616766, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1505, "step": 10553 }, { "epoch": 15.79940119760479, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1438, "step": 10554 }, { "epoch": 15.800898203592814, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1446, "step": 10555 }, { "epoch": 15.802395209580839, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1507, "step": 10556 }, { "epoch": 15.803892215568862, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1467, "step": 10557 }, { "epoch": 15.805389221556887, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1439, "step": 10558 }, { "epoch": 15.80688622754491, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1466, "step": 10559 }, { "epoch": 15.808383233532934, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.148, "step": 10560 }, { "epoch": 15.809880239520957, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1465, "step": 10561 }, { "epoch": 15.811377245508982, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.141, "step": 10562 }, { "epoch": 15.812874251497005, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1447, "step": 10563 }, { "epoch": 15.81437125748503, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1439, "step": 10564 }, { "epoch": 15.815868263473053, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1533, "step": 10565 }, { "epoch": 15.817365269461078, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1393, "step": 10566 }, { "epoch": 15.818862275449101, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1441, "step": 10567 }, { "epoch": 15.820359281437126, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1385, "step": 10568 }, { "epoch": 15.821856287425149, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1419, "step": 10569 }, { "epoch": 15.823353293413174, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1478, "step": 10570 }, { "epoch": 15.824850299401197, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1389, "step": 10571 }, { "epoch": 15.826347305389222, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1359, "step": 10572 }, { "epoch": 15.827844311377245, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1461, "step": 10573 }, { "epoch": 15.82934131736527, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1375, "step": 10574 }, { "epoch": 15.830838323353294, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.144, "step": 10575 }, { "epoch": 15.832335329341317, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1383, "step": 10576 }, { "epoch": 15.83383233532934, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1435, "step": 10577 }, { "epoch": 15.835329341317365, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1417, "step": 10578 }, { "epoch": 15.83682634730539, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1489, "step": 10579 }, { "epoch": 15.838323353293413, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1457, "step": 10580 }, { "epoch": 15.839820359281438, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1444, "step": 10581 }, { "epoch": 15.841317365269461, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.14, "step": 10582 }, { "epoch": 15.842814371257486, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1501, "step": 10583 }, { "epoch": 15.844311377245509, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.145, "step": 10584 }, { "epoch": 15.845808383233534, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1432, "step": 10585 }, { "epoch": 15.847305389221557, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1445, "step": 10586 }, { "epoch": 15.848802395209582, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1474, "step": 10587 }, { "epoch": 15.850299401197605, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1506, "step": 10588 }, { "epoch": 15.85179640718563, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1476, "step": 10589 }, { "epoch": 15.853293413173652, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1415, "step": 10590 }, { "epoch": 15.854790419161677, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1436, "step": 10591 }, { "epoch": 15.8562874251497, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1389, "step": 10592 }, { "epoch": 15.857784431137725, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.149, "step": 10593 }, { "epoch": 15.859281437125748, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1458, "step": 10594 }, { "epoch": 15.860778443113773, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1417, "step": 10595 }, { "epoch": 15.862275449101796, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1472, "step": 10596 }, { "epoch": 15.863772455089821, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1382, "step": 10597 }, { "epoch": 15.865269461077844, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1476, "step": 10598 }, { "epoch": 15.866766467065869, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1405, "step": 10599 }, { "epoch": 15.868263473053892, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.145, "step": 10600 }, { "epoch": 15.869760479041917, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1487, "step": 10601 }, { "epoch": 15.87125748502994, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1436, "step": 10602 }, { "epoch": 15.872754491017965, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1502, "step": 10603 }, { "epoch": 15.874251497005988, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1456, "step": 10604 }, { "epoch": 15.875748502994012, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1485, "step": 10605 }, { "epoch": 15.877245508982035, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1384, "step": 10606 }, { "epoch": 15.87874251497006, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1451, "step": 10607 }, { "epoch": 15.880239520958083, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1447, "step": 10608 }, { "epoch": 15.881736526946108, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1429, "step": 10609 }, { "epoch": 15.883233532934131, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1382, "step": 10610 }, { "epoch": 15.884730538922156, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1435, "step": 10611 }, { "epoch": 15.886227544910179, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1397, "step": 10612 }, { "epoch": 15.887724550898204, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1476, "step": 10613 }, { "epoch": 15.889221556886227, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1374, "step": 10614 }, { "epoch": 15.890718562874252, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1405, "step": 10615 }, { "epoch": 15.892215568862275, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1447, "step": 10616 }, { "epoch": 15.8937125748503, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1406, "step": 10617 }, { "epoch": 15.895209580838323, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1403, "step": 10618 }, { "epoch": 15.896706586826348, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1463, "step": 10619 }, { "epoch": 15.89820359281437, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1433, "step": 10620 }, { "epoch": 15.899700598802395, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1395, "step": 10621 }, { "epoch": 15.901197604790418, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1409, "step": 10622 }, { "epoch": 15.902694610778443, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1541, "step": 10623 }, { "epoch": 15.904191616766466, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1415, "step": 10624 }, { "epoch": 15.905688622754491, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1414, "step": 10625 }, { "epoch": 15.907185628742514, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1463, "step": 10626 }, { "epoch": 15.908682634730539, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1441, "step": 10627 }, { "epoch": 15.910179640718562, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1447, "step": 10628 }, { "epoch": 15.911676646706587, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1458, "step": 10629 }, { "epoch": 15.91317365269461, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1417, "step": 10630 }, { "epoch": 15.914670658682635, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1402, "step": 10631 }, { "epoch": 15.91616766467066, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1455, "step": 10632 }, { "epoch": 15.917664670658683, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1464, "step": 10633 }, { "epoch": 15.919161676646706, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1459, "step": 10634 }, { "epoch": 15.92065868263473, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1429, "step": 10635 }, { "epoch": 15.922155688622755, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1394, "step": 10636 }, { "epoch": 15.923652694610778, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1403, "step": 10637 }, { "epoch": 15.925149700598803, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1402, "step": 10638 }, { "epoch": 15.926646706586826, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1415, "step": 10639 }, { "epoch": 15.928143712574851, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1451, "step": 10640 }, { "epoch": 15.929640718562874, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1415, "step": 10641 }, { "epoch": 15.931137724550899, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1486, "step": 10642 }, { "epoch": 15.932634730538922, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.142, "step": 10643 }, { "epoch": 15.934131736526947, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1456, "step": 10644 }, { "epoch": 15.93562874251497, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1415, "step": 10645 }, { "epoch": 15.937125748502995, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1392, "step": 10646 }, { "epoch": 15.938622754491018, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.145, "step": 10647 }, { "epoch": 15.940119760479043, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1437, "step": 10648 }, { "epoch": 15.941616766467066, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.146, "step": 10649 }, { "epoch": 15.94311377245509, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1403, "step": 10650 }, { "epoch": 15.944610778443113, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1396, "step": 10651 }, { "epoch": 15.946107784431138, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1418, "step": 10652 }, { "epoch": 15.947604790419161, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1378, "step": 10653 }, { "epoch": 15.949101796407186, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1401, "step": 10654 }, { "epoch": 15.95059880239521, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1404, "step": 10655 }, { "epoch": 15.952095808383234, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1467, "step": 10656 }, { "epoch": 15.953592814371257, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1413, "step": 10657 }, { "epoch": 15.955089820359282, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1416, "step": 10658 }, { "epoch": 15.956586826347305, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1448, "step": 10659 }, { "epoch": 15.95808383233533, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1424, "step": 10660 }, { "epoch": 15.959580838323353, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1465, "step": 10661 }, { "epoch": 15.961077844311378, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1386, "step": 10662 }, { "epoch": 15.9625748502994, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1406, "step": 10663 }, { "epoch": 15.964071856287426, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1465, "step": 10664 }, { "epoch": 15.965568862275449, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.15, "step": 10665 }, { "epoch": 15.967065868263473, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1451, "step": 10666 }, { "epoch": 15.968562874251496, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1533, "step": 10667 }, { "epoch": 15.970059880239521, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1398, "step": 10668 }, { "epoch": 15.971556886227544, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1452, "step": 10669 }, { "epoch": 15.97305389221557, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1424, "step": 10670 }, { "epoch": 15.974550898203592, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1396, "step": 10671 }, { "epoch": 15.976047904191617, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1436, "step": 10672 }, { "epoch": 15.97754491017964, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1422, "step": 10673 }, { "epoch": 15.979041916167665, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1449, "step": 10674 }, { "epoch": 15.980538922155688, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1474, "step": 10675 }, { "epoch": 15.982035928143713, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1429, "step": 10676 }, { "epoch": 15.983532934131736, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1468, "step": 10677 }, { "epoch": 15.98502994011976, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1336, "step": 10678 }, { "epoch": 15.986526946107784, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1439, "step": 10679 }, { "epoch": 15.988023952095809, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1412, "step": 10680 }, { "epoch": 15.989520958083832, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1386, "step": 10681 }, { "epoch": 15.991017964071856, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1407, "step": 10682 }, { "epoch": 15.99251497005988, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1413, "step": 10683 }, { "epoch": 15.994011976047904, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1432, "step": 10684 }, { "epoch": 15.995508982035929, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1387, "step": 10685 }, { "epoch": 15.997005988023952, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1435, "step": 10686 }, { "epoch": 15.998502994011975, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1474, "step": 10687 }, { "epoch": 16.0, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1321, "step": 10688 }, { "epoch": 16.001497005988025, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1448, "step": 10689 }, { "epoch": 16.00299401197605, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1479, "step": 10690 }, { "epoch": 16.00449101796407, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1438, "step": 10691 }, { "epoch": 16.005988023952096, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1398, "step": 10692 }, { "epoch": 16.00748502994012, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1416, "step": 10693 }, { "epoch": 16.008982035928145, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1441, "step": 10694 }, { "epoch": 16.010479041916167, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1349, "step": 10695 }, { "epoch": 16.01197604790419, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1458, "step": 10696 }, { "epoch": 16.013473053892216, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1412, "step": 10697 }, { "epoch": 16.01497005988024, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1404, "step": 10698 }, { "epoch": 16.016467065868262, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1421, "step": 10699 }, { "epoch": 16.017964071856287, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1451, "step": 10700 }, { "epoch": 16.019461077844312, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1326, "step": 10701 }, { "epoch": 16.020958083832337, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1461, "step": 10702 }, { "epoch": 16.022455089820358, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1412, "step": 10703 }, { "epoch": 16.023952095808383, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1389, "step": 10704 }, { "epoch": 16.025449101796408, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1423, "step": 10705 }, { "epoch": 16.026946107784433, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1381, "step": 10706 }, { "epoch": 16.028443113772454, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1456, "step": 10707 }, { "epoch": 16.02994011976048, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1443, "step": 10708 }, { "epoch": 16.031437125748504, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1365, "step": 10709 }, { "epoch": 16.03293413173653, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1433, "step": 10710 }, { "epoch": 16.03443113772455, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1467, "step": 10711 }, { "epoch": 16.035928143712574, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1368, "step": 10712 }, { "epoch": 16.0374251497006, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1381, "step": 10713 }, { "epoch": 16.038922155688624, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1377, "step": 10714 }, { "epoch": 16.040419161676645, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1397, "step": 10715 }, { "epoch": 16.04191616766467, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1478, "step": 10716 }, { "epoch": 16.043413173652695, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1367, "step": 10717 }, { "epoch": 16.04491017964072, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1365, "step": 10718 }, { "epoch": 16.04640718562874, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1407, "step": 10719 }, { "epoch": 16.047904191616766, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1429, "step": 10720 }, { "epoch": 16.04940119760479, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1515, "step": 10721 }, { "epoch": 16.050898203592816, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1311, "step": 10722 }, { "epoch": 16.052395209580837, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1478, "step": 10723 }, { "epoch": 16.05389221556886, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.144, "step": 10724 }, { "epoch": 16.055389221556887, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1446, "step": 10725 }, { "epoch": 16.05688622754491, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.14, "step": 10726 }, { "epoch": 16.058383233532933, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1355, "step": 10727 }, { "epoch": 16.059880239520957, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1506, "step": 10728 }, { "epoch": 16.061377245508982, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1406, "step": 10729 }, { "epoch": 16.062874251497007, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1409, "step": 10730 }, { "epoch": 16.06437125748503, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.138, "step": 10731 }, { "epoch": 16.065868263473053, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1374, "step": 10732 }, { "epoch": 16.067365269461078, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1445, "step": 10733 }, { "epoch": 16.068862275449103, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1405, "step": 10734 }, { "epoch": 16.070359281437124, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1448, "step": 10735 }, { "epoch": 16.07185628742515, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1468, "step": 10736 }, { "epoch": 16.073353293413174, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1402, "step": 10737 }, { "epoch": 16.0748502994012, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1392, "step": 10738 }, { "epoch": 16.07634730538922, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1448, "step": 10739 }, { "epoch": 16.077844311377245, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1392, "step": 10740 }, { "epoch": 16.07934131736527, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1496, "step": 10741 }, { "epoch": 16.080838323353294, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1402, "step": 10742 }, { "epoch": 16.082335329341316, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1374, "step": 10743 }, { "epoch": 16.08383233532934, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1439, "step": 10744 }, { "epoch": 16.085329341317365, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1489, "step": 10745 }, { "epoch": 16.08682634730539, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.138, "step": 10746 }, { "epoch": 16.088323353293415, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1431, "step": 10747 }, { "epoch": 16.089820359281436, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1441, "step": 10748 }, { "epoch": 16.09131736526946, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1386, "step": 10749 }, { "epoch": 16.092814371257486, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.144, "step": 10750 }, { "epoch": 16.09431137724551, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1432, "step": 10751 }, { "epoch": 16.095808383233532, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1427, "step": 10752 }, { "epoch": 16.097305389221557, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1472, "step": 10753 }, { "epoch": 16.09880239520958, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.139, "step": 10754 }, { "epoch": 16.100299401197606, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1392, "step": 10755 }, { "epoch": 16.101796407185628, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.141, "step": 10756 }, { "epoch": 16.103293413173652, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1406, "step": 10757 }, { "epoch": 16.104790419161677, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1428, "step": 10758 }, { "epoch": 16.106287425149702, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1381, "step": 10759 }, { "epoch": 16.107784431137723, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1398, "step": 10760 }, { "epoch": 16.10928143712575, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1377, "step": 10761 }, { "epoch": 16.110778443113773, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1336, "step": 10762 }, { "epoch": 16.112275449101798, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.1359, "step": 10763 }, { "epoch": 16.11377245508982, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1391, "step": 10764 }, { "epoch": 16.115269461077844, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1373, "step": 10765 }, { "epoch": 16.11676646706587, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1375, "step": 10766 }, { "epoch": 16.118263473053894, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1413, "step": 10767 }, { "epoch": 16.119760479041915, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.138, "step": 10768 }, { "epoch": 16.12125748502994, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1433, "step": 10769 }, { "epoch": 16.122754491017965, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1449, "step": 10770 }, { "epoch": 16.12425149700599, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.144, "step": 10771 }, { "epoch": 16.12574850299401, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.138, "step": 10772 }, { "epoch": 16.127245508982035, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.147, "step": 10773 }, { "epoch": 16.12874251497006, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1412, "step": 10774 }, { "epoch": 16.130239520958085, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1498, "step": 10775 }, { "epoch": 16.131736526946106, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1453, "step": 10776 }, { "epoch": 16.13323353293413, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1427, "step": 10777 }, { "epoch": 16.134730538922156, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1375, "step": 10778 }, { "epoch": 16.13622754491018, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1417, "step": 10779 }, { "epoch": 16.137724550898202, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1387, "step": 10780 }, { "epoch": 16.139221556886227, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1418, "step": 10781 }, { "epoch": 16.14071856287425, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1442, "step": 10782 }, { "epoch": 16.142215568862277, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1365, "step": 10783 }, { "epoch": 16.143712574850298, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1415, "step": 10784 }, { "epoch": 16.145209580838323, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1421, "step": 10785 }, { "epoch": 16.146706586826348, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1426, "step": 10786 }, { "epoch": 16.148203592814372, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1434, "step": 10787 }, { "epoch": 16.149700598802394, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1425, "step": 10788 }, { "epoch": 16.15119760479042, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1405, "step": 10789 }, { "epoch": 16.152694610778443, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1359, "step": 10790 }, { "epoch": 16.154191616766468, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1385, "step": 10791 }, { "epoch": 16.15568862275449, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.149, "step": 10792 }, { "epoch": 16.157185628742514, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1431, "step": 10793 }, { "epoch": 16.15868263473054, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1379, "step": 10794 }, { "epoch": 16.160179640718564, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1367, "step": 10795 }, { "epoch": 16.161676646706585, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.137, "step": 10796 }, { "epoch": 16.16317365269461, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1392, "step": 10797 }, { "epoch": 16.164670658682635, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1403, "step": 10798 }, { "epoch": 16.16616766467066, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1461, "step": 10799 }, { "epoch": 16.16766467065868, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.144, "step": 10800 }, { "epoch": 16.169161676646706, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1455, "step": 10801 }, { "epoch": 16.17065868263473, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1462, "step": 10802 }, { "epoch": 16.172155688622755, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1431, "step": 10803 }, { "epoch": 16.17365269461078, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1401, "step": 10804 }, { "epoch": 16.1751497005988, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1355, "step": 10805 }, { "epoch": 16.176646706586826, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1411, "step": 10806 }, { "epoch": 16.17814371257485, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1404, "step": 10807 }, { "epoch": 16.179640718562876, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1462, "step": 10808 }, { "epoch": 16.181137724550897, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1365, "step": 10809 }, { "epoch": 16.182634730538922, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1456, "step": 10810 }, { "epoch": 16.184131736526947, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1396, "step": 10811 }, { "epoch": 16.18562874251497, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1328, "step": 10812 }, { "epoch": 16.187125748502993, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1422, "step": 10813 }, { "epoch": 16.188622754491018, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1459, "step": 10814 }, { "epoch": 16.190119760479043, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1377, "step": 10815 }, { "epoch": 16.191616766467067, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1434, "step": 10816 }, { "epoch": 16.19311377245509, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1438, "step": 10817 }, { "epoch": 16.194610778443113, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1421, "step": 10818 }, { "epoch": 16.19610778443114, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1432, "step": 10819 }, { "epoch": 16.197604790419163, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1334, "step": 10820 }, { "epoch": 16.199101796407184, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1375, "step": 10821 }, { "epoch": 16.20059880239521, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1469, "step": 10822 }, { "epoch": 16.202095808383234, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1396, "step": 10823 }, { "epoch": 16.20359281437126, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1413, "step": 10824 }, { "epoch": 16.20508982035928, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1357, "step": 10825 }, { "epoch": 16.206586826347305, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1424, "step": 10826 }, { "epoch": 16.20808383233533, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1381, "step": 10827 }, { "epoch": 16.209580838323355, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1425, "step": 10828 }, { "epoch": 16.211077844311376, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1414, "step": 10829 }, { "epoch": 16.2125748502994, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.141, "step": 10830 }, { "epoch": 16.214071856287426, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1378, "step": 10831 }, { "epoch": 16.21556886227545, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1392, "step": 10832 }, { "epoch": 16.21706586826347, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1394, "step": 10833 }, { "epoch": 16.218562874251496, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1386, "step": 10834 }, { "epoch": 16.22005988023952, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1452, "step": 10835 }, { "epoch": 16.221556886227546, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1364, "step": 10836 }, { "epoch": 16.223053892215567, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1362, "step": 10837 }, { "epoch": 16.224550898203592, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1399, "step": 10838 }, { "epoch": 16.226047904191617, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1431, "step": 10839 }, { "epoch": 16.227544910179642, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1415, "step": 10840 }, { "epoch": 16.229041916167663, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1398, "step": 10841 }, { "epoch": 16.230538922155688, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1439, "step": 10842 }, { "epoch": 16.232035928143713, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1399, "step": 10843 }, { "epoch": 16.233532934131738, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1428, "step": 10844 }, { "epoch": 16.23502994011976, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1428, "step": 10845 }, { "epoch": 16.236526946107784, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1442, "step": 10846 }, { "epoch": 16.23802395209581, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1382, "step": 10847 }, { "epoch": 16.239520958083833, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1424, "step": 10848 }, { "epoch": 16.241017964071855, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1383, "step": 10849 }, { "epoch": 16.24251497005988, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1428, "step": 10850 }, { "epoch": 16.244011976047904, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1427, "step": 10851 }, { "epoch": 16.24550898203593, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1366, "step": 10852 }, { "epoch": 16.24700598802395, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1395, "step": 10853 }, { "epoch": 16.248502994011975, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.142, "step": 10854 }, { "epoch": 16.25, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1414, "step": 10855 }, { "epoch": 16.251497005988025, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1407, "step": 10856 }, { "epoch": 16.25299401197605, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.143, "step": 10857 }, { "epoch": 16.25449101796407, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1436, "step": 10858 }, { "epoch": 16.255988023952096, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1441, "step": 10859 }, { "epoch": 16.25748502994012, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1414, "step": 10860 }, { "epoch": 16.258982035928145, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1419, "step": 10861 }, { "epoch": 16.260479041916167, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1445, "step": 10862 }, { "epoch": 16.26197604790419, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1452, "step": 10863 }, { "epoch": 16.263473053892216, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1428, "step": 10864 }, { "epoch": 16.26497005988024, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1461, "step": 10865 }, { "epoch": 16.266467065868262, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1398, "step": 10866 }, { "epoch": 16.267964071856287, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1433, "step": 10867 }, { "epoch": 16.269461077844312, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1384, "step": 10868 }, { "epoch": 16.270958083832337, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1386, "step": 10869 }, { "epoch": 16.272455089820358, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1432, "step": 10870 }, { "epoch": 16.273952095808383, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1323, "step": 10871 }, { "epoch": 16.275449101796408, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.139, "step": 10872 }, { "epoch": 16.276946107784433, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1387, "step": 10873 }, { "epoch": 16.278443113772454, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1439, "step": 10874 }, { "epoch": 16.27994011976048, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1432, "step": 10875 }, { "epoch": 16.281437125748504, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1375, "step": 10876 }, { "epoch": 16.28293413173653, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.143, "step": 10877 }, { "epoch": 16.28443113772455, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1443, "step": 10878 }, { "epoch": 16.285928143712574, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1379, "step": 10879 }, { "epoch": 16.2874251497006, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1429, "step": 10880 }, { "epoch": 16.288922155688624, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1369, "step": 10881 }, { "epoch": 16.290419161676645, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1431, "step": 10882 }, { "epoch": 16.29191616766467, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1411, "step": 10883 }, { "epoch": 16.293413173652695, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1387, "step": 10884 }, { "epoch": 16.29491017964072, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1463, "step": 10885 }, { "epoch": 16.29640718562874, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1456, "step": 10886 }, { "epoch": 16.297904191616766, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1412, "step": 10887 }, { "epoch": 16.29940119760479, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1405, "step": 10888 }, { "epoch": 16.300898203592816, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1389, "step": 10889 }, { "epoch": 16.302395209580837, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1403, "step": 10890 }, { "epoch": 16.30389221556886, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.136, "step": 10891 }, { "epoch": 16.305389221556887, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1428, "step": 10892 }, { "epoch": 16.30688622754491, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.139, "step": 10893 }, { "epoch": 16.308383233532933, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1366, "step": 10894 }, { "epoch": 16.309880239520957, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1415, "step": 10895 }, { "epoch": 16.311377245508982, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1383, "step": 10896 }, { "epoch": 16.312874251497007, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1418, "step": 10897 }, { "epoch": 16.31437125748503, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1361, "step": 10898 }, { "epoch": 16.315868263473053, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1465, "step": 10899 }, { "epoch": 16.317365269461078, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1322, "step": 10900 }, { "epoch": 16.318862275449103, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1456, "step": 10901 }, { "epoch": 16.320359281437124, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1425, "step": 10902 }, { "epoch": 16.32185628742515, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1415, "step": 10903 }, { "epoch": 16.323353293413174, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1386, "step": 10904 }, { "epoch": 16.3248502994012, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.1431, "step": 10905 }, { "epoch": 16.32634730538922, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1446, "step": 10906 }, { "epoch": 16.327844311377245, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1373, "step": 10907 }, { "epoch": 16.32934131736527, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1389, "step": 10908 }, { "epoch": 16.330838323353294, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1437, "step": 10909 }, { "epoch": 16.33233532934132, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1382, "step": 10910 }, { "epoch": 16.33383233532934, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1418, "step": 10911 }, { "epoch": 16.335329341317365, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1402, "step": 10912 }, { "epoch": 16.33682634730539, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1366, "step": 10913 }, { "epoch": 16.338323353293415, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1353, "step": 10914 }, { "epoch": 16.339820359281436, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1389, "step": 10915 }, { "epoch": 16.34131736526946, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1375, "step": 10916 }, { "epoch": 16.342814371257486, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1444, "step": 10917 }, { "epoch": 16.34431137724551, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1398, "step": 10918 }, { "epoch": 16.345808383233532, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1431, "step": 10919 }, { "epoch": 16.347305389221557, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1371, "step": 10920 }, { "epoch": 16.34880239520958, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1441, "step": 10921 }, { "epoch": 16.350299401197606, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1459, "step": 10922 }, { "epoch": 16.351796407185628, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1415, "step": 10923 }, { "epoch": 16.353293413173652, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1375, "step": 10924 }, { "epoch": 16.354790419161677, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1433, "step": 10925 }, { "epoch": 16.356287425149702, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1402, "step": 10926 }, { "epoch": 16.357784431137723, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1357, "step": 10927 }, { "epoch": 16.35928143712575, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1507, "step": 10928 }, { "epoch": 16.360778443113773, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1341, "step": 10929 }, { "epoch": 16.362275449101798, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1432, "step": 10930 }, { "epoch": 16.36377245508982, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1373, "step": 10931 }, { "epoch": 16.365269461077844, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1403, "step": 10932 }, { "epoch": 16.36676646706587, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1397, "step": 10933 }, { "epoch": 16.368263473053894, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1324, "step": 10934 }, { "epoch": 16.369760479041915, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1408, "step": 10935 }, { "epoch": 16.37125748502994, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1402, "step": 10936 }, { "epoch": 16.372754491017965, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1441, "step": 10937 }, { "epoch": 16.37425149700599, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1475, "step": 10938 }, { "epoch": 16.37574850299401, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1335, "step": 10939 }, { "epoch": 16.377245508982035, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1383, "step": 10940 }, { "epoch": 16.37874251497006, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1394, "step": 10941 }, { "epoch": 16.380239520958085, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1458, "step": 10942 }, { "epoch": 16.381736526946106, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1424, "step": 10943 }, { "epoch": 16.38323353293413, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1408, "step": 10944 }, { "epoch": 16.384730538922156, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.137, "step": 10945 }, { "epoch": 16.38622754491018, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1411, "step": 10946 }, { "epoch": 16.387724550898202, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1342, "step": 10947 }, { "epoch": 16.389221556886227, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1427, "step": 10948 }, { "epoch": 16.39071856287425, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1422, "step": 10949 }, { "epoch": 16.392215568862277, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1422, "step": 10950 }, { "epoch": 16.393712574850298, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.137, "step": 10951 }, { "epoch": 16.395209580838323, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.138, "step": 10952 }, { "epoch": 16.396706586826348, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.135, "step": 10953 }, { "epoch": 16.398203592814372, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1416, "step": 10954 }, { "epoch": 16.399700598802394, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1368, "step": 10955 }, { "epoch": 16.40119760479042, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1437, "step": 10956 }, { "epoch": 16.402694610778443, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1473, "step": 10957 }, { "epoch": 16.404191616766468, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1425, "step": 10958 }, { "epoch": 16.40568862275449, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1449, "step": 10959 }, { "epoch": 16.407185628742514, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1343, "step": 10960 }, { "epoch": 16.40868263473054, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1388, "step": 10961 }, { "epoch": 16.410179640718564, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1428, "step": 10962 }, { "epoch": 16.411676646706585, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1391, "step": 10963 }, { "epoch": 16.41317365269461, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1457, "step": 10964 }, { "epoch": 16.414670658682635, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.136, "step": 10965 }, { "epoch": 16.41616766467066, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1456, "step": 10966 }, { "epoch": 16.41766467065868, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1467, "step": 10967 }, { "epoch": 16.419161676646706, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1455, "step": 10968 }, { "epoch": 16.42065868263473, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1359, "step": 10969 }, { "epoch": 16.422155688622755, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.148, "step": 10970 }, { "epoch": 16.42365269461078, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1379, "step": 10971 }, { "epoch": 16.4251497005988, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.135, "step": 10972 }, { "epoch": 16.426646706586826, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1339, "step": 10973 }, { "epoch": 16.42814371257485, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1383, "step": 10974 }, { "epoch": 16.429640718562876, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1362, "step": 10975 }, { "epoch": 16.431137724550897, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1423, "step": 10976 }, { "epoch": 16.432634730538922, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1449, "step": 10977 }, { "epoch": 16.434131736526947, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1464, "step": 10978 }, { "epoch": 16.43562874251497, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1439, "step": 10979 }, { "epoch": 16.437125748502993, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1469, "step": 10980 }, { "epoch": 16.438622754491018, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.142, "step": 10981 }, { "epoch": 16.440119760479043, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1417, "step": 10982 }, { "epoch": 16.441616766467067, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1452, "step": 10983 }, { "epoch": 16.44311377245509, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1384, "step": 10984 }, { "epoch": 16.444610778443113, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.14, "step": 10985 }, { "epoch": 16.44610778443114, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1406, "step": 10986 }, { "epoch": 16.447604790419163, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1361, "step": 10987 }, { "epoch": 16.449101796407184, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1381, "step": 10988 }, { "epoch": 16.45059880239521, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1428, "step": 10989 }, { "epoch": 16.452095808383234, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1384, "step": 10990 }, { "epoch": 16.45359281437126, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1423, "step": 10991 }, { "epoch": 16.45508982035928, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.145, "step": 10992 }, { "epoch": 16.456586826347305, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1333, "step": 10993 }, { "epoch": 16.45808383233533, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1336, "step": 10994 }, { "epoch": 16.459580838323355, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1362, "step": 10995 }, { "epoch": 16.461077844311376, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1435, "step": 10996 }, { "epoch": 16.4625748502994, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1343, "step": 10997 }, { "epoch": 16.464071856287426, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1448, "step": 10998 }, { "epoch": 16.46556886227545, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1449, "step": 10999 }, { "epoch": 16.46706586826347, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1373, "step": 11000 }, { "epoch": 16.468562874251496, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1537, "step": 11001 }, { "epoch": 16.47005988023952, "grad_norm": 0.5078125, "learning_rate": 0.0008, "loss": 1.2339, "step": 11002 }, { "epoch": 16.471556886227546, "grad_norm": 0.8203125, "learning_rate": 0.0008, "loss": 1.2478, "step": 11003 }, { "epoch": 16.473053892215567, "grad_norm": 0.69921875, "learning_rate": 0.0008, "loss": 1.2196, "step": 11004 }, { "epoch": 16.474550898203592, "grad_norm": 1.0390625, "learning_rate": 0.0008, "loss": 1.2444, "step": 11005 }, { "epoch": 16.476047904191617, "grad_norm": 0.9453125, "learning_rate": 0.0008, "loss": 1.2377, "step": 11006 }, { "epoch": 16.477544910179642, "grad_norm": 0.90234375, "learning_rate": 0.0008, "loss": 1.2631, "step": 11007 }, { "epoch": 16.479041916167663, "grad_norm": 0.73046875, "learning_rate": 0.0008, "loss": 1.2588, "step": 11008 }, { "epoch": 16.480538922155688, "grad_norm": 0.9921875, "learning_rate": 0.0008, "loss": 1.283, "step": 11009 }, { "epoch": 16.482035928143713, "grad_norm": 0.86328125, "learning_rate": 0.0008, "loss": 1.2862, "step": 11010 }, { "epoch": 16.483532934131738, "grad_norm": 0.45703125, "learning_rate": 0.0008, "loss": 1.2847, "step": 11011 }, { "epoch": 16.48502994011976, "grad_norm": 0.54296875, "learning_rate": 0.0008, "loss": 1.2827, "step": 11012 }, { "epoch": 16.486526946107784, "grad_norm": 0.90234375, "learning_rate": 0.0008, "loss": 1.2914, "step": 11013 }, { "epoch": 16.48802395209581, "grad_norm": 0.84375, "learning_rate": 0.0008, "loss": 1.2901, "step": 11014 }, { "epoch": 16.489520958083833, "grad_norm": 0.76171875, "learning_rate": 0.0008, "loss": 1.2746, "step": 11015 }, { "epoch": 16.491017964071855, "grad_norm": 0.703125, "learning_rate": 0.0008, "loss": 1.2701, "step": 11016 }, { "epoch": 16.49251497005988, "grad_norm": 1.015625, "learning_rate": 0.0008, "loss": 1.2886, "step": 11017 }, { "epoch": 16.494011976047904, "grad_norm": 1.4609375, "learning_rate": 0.0008, "loss": 1.3552, "step": 11018 }, { "epoch": 16.49550898203593, "grad_norm": 0.72265625, "learning_rate": 0.0008, "loss": 1.2829, "step": 11019 }, { "epoch": 16.49700598802395, "grad_norm": 1.0078125, "learning_rate": 0.0008, "loss": 1.2804, "step": 11020 }, { "epoch": 16.498502994011975, "grad_norm": 0.77734375, "learning_rate": 0.0008, "loss": 1.276, "step": 11021 }, { "epoch": 16.5, "grad_norm": 0.57421875, "learning_rate": 0.0008, "loss": 1.2676, "step": 11022 }, { "epoch": 16.501497005988025, "grad_norm": 0.671875, "learning_rate": 0.0008, "loss": 1.2626, "step": 11023 }, { "epoch": 16.50299401197605, "grad_norm": 0.4140625, "learning_rate": 0.0008, "loss": 1.2453, "step": 11024 }, { "epoch": 16.50449101796407, "grad_norm": 0.484375, "learning_rate": 0.0008, "loss": 1.248, "step": 11025 }, { "epoch": 16.505988023952096, "grad_norm": 0.287109375, "learning_rate": 0.0008, "loss": 1.2363, "step": 11026 }, { "epoch": 16.50748502994012, "grad_norm": 0.27734375, "learning_rate": 0.0008, "loss": 1.236, "step": 11027 }, { "epoch": 16.508982035928145, "grad_norm": 0.26171875, "learning_rate": 0.0008, "loss": 1.228, "step": 11028 }, { "epoch": 16.510479041916167, "grad_norm": 0.283203125, "learning_rate": 0.0008, "loss": 1.2281, "step": 11029 }, { "epoch": 16.51197604790419, "grad_norm": 0.2421875, "learning_rate": 0.0008, "loss": 1.2243, "step": 11030 }, { "epoch": 16.513473053892216, "grad_norm": 0.259765625, "learning_rate": 0.0008, "loss": 1.2231, "step": 11031 }, { "epoch": 16.51497005988024, "grad_norm": 0.203125, "learning_rate": 0.0008, "loss": 1.2208, "step": 11032 }, { "epoch": 16.516467065868262, "grad_norm": 0.29296875, "learning_rate": 0.0008, "loss": 1.2153, "step": 11033 }, { "epoch": 16.517964071856287, "grad_norm": 0.396484375, "learning_rate": 0.0008, "loss": 1.2161, "step": 11034 }, { "epoch": 16.519461077844312, "grad_norm": 0.53515625, "learning_rate": 0.0008, "loss": 1.2174, "step": 11035 }, { "epoch": 16.520958083832337, "grad_norm": 0.84375, "learning_rate": 0.0008, "loss": 1.2215, "step": 11036 }, { "epoch": 16.522455089820358, "grad_norm": 0.921875, "learning_rate": 0.0008, "loss": 1.2288, "step": 11037 }, { "epoch": 16.523952095808383, "grad_norm": 0.5703125, "learning_rate": 0.0008, "loss": 1.2081, "step": 11038 }, { "epoch": 16.525449101796408, "grad_norm": 0.3046875, "learning_rate": 0.0008, "loss": 1.2025, "step": 11039 }, { "epoch": 16.526946107784433, "grad_norm": 0.67578125, "learning_rate": 0.0008, "loss": 1.2166, "step": 11040 }, { "epoch": 16.528443113772454, "grad_norm": 0.408203125, "learning_rate": 0.0008, "loss": 1.2124, "step": 11041 }, { "epoch": 16.52994011976048, "grad_norm": 0.408203125, "learning_rate": 0.0008, "loss": 1.2155, "step": 11042 }, { "epoch": 16.531437125748504, "grad_norm": 0.32421875, "learning_rate": 0.0008, "loss": 1.2082, "step": 11043 }, { "epoch": 16.53293413173653, "grad_norm": 0.248046875, "learning_rate": 0.0008, "loss": 1.1996, "step": 11044 }, { "epoch": 16.53443113772455, "grad_norm": 0.26171875, "learning_rate": 0.0008, "loss": 1.191, "step": 11045 }, { "epoch": 16.535928143712574, "grad_norm": 0.2314453125, "learning_rate": 0.0008, "loss": 1.1928, "step": 11046 }, { "epoch": 16.5374251497006, "grad_norm": 0.197265625, "learning_rate": 0.0008, "loss": 1.1906, "step": 11047 }, { "epoch": 16.538922155688624, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.1906, "step": 11048 }, { "epoch": 16.540419161676645, "grad_norm": 0.16796875, "learning_rate": 0.0008, "loss": 1.1896, "step": 11049 }, { "epoch": 16.54191616766467, "grad_norm": 0.162109375, "learning_rate": 0.0008, "loss": 1.1849, "step": 11050 }, { "epoch": 16.543413173652695, "grad_norm": 0.16796875, "learning_rate": 0.0008, "loss": 1.1869, "step": 11051 }, { "epoch": 16.54491017964072, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.1876, "step": 11052 }, { "epoch": 16.54640718562874, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1822, "step": 11053 }, { "epoch": 16.547904191616766, "grad_norm": 0.1513671875, "learning_rate": 0.0008, "loss": 1.1865, "step": 11054 }, { "epoch": 16.54940119760479, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.1717, "step": 11055 }, { "epoch": 16.550898203592816, "grad_norm": 0.173828125, "learning_rate": 0.0008, "loss": 1.1808, "step": 11056 }, { "epoch": 16.552395209580837, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.175, "step": 11057 }, { "epoch": 16.55389221556886, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1804, "step": 11058 }, { "epoch": 16.555389221556887, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1745, "step": 11059 }, { "epoch": 16.55688622754491, "grad_norm": 0.2021484375, "learning_rate": 0.0008, "loss": 1.1694, "step": 11060 }, { "epoch": 16.558383233532933, "grad_norm": 0.1806640625, "learning_rate": 0.0008, "loss": 1.1768, "step": 11061 }, { "epoch": 16.559880239520957, "grad_norm": 0.205078125, "learning_rate": 0.0008, "loss": 1.171, "step": 11062 }, { "epoch": 16.561377245508982, "grad_norm": 0.291015625, "learning_rate": 0.0008, "loss": 1.1725, "step": 11063 }, { "epoch": 16.562874251497007, "grad_norm": 0.48046875, "learning_rate": 0.0008, "loss": 1.1779, "step": 11064 }, { "epoch": 16.56437125748503, "grad_norm": 0.79296875, "learning_rate": 0.0008, "loss": 1.1795, "step": 11065 }, { "epoch": 16.565868263473053, "grad_norm": 0.8515625, "learning_rate": 0.0008, "loss": 1.1998, "step": 11066 }, { "epoch": 16.567365269461078, "grad_norm": 0.314453125, "learning_rate": 0.0008, "loss": 1.1738, "step": 11067 }, { "epoch": 16.568862275449103, "grad_norm": 0.4375, "learning_rate": 0.0008, "loss": 1.1762, "step": 11068 }, { "epoch": 16.570359281437124, "grad_norm": 0.443359375, "learning_rate": 0.0008, "loss": 1.1845, "step": 11069 }, { "epoch": 16.57185628742515, "grad_norm": 0.330078125, "learning_rate": 0.0008, "loss": 1.1764, "step": 11070 }, { "epoch": 16.573353293413174, "grad_norm": 0.283203125, "learning_rate": 0.0008, "loss": 1.1823, "step": 11071 }, { "epoch": 16.5748502994012, "grad_norm": 0.208984375, "learning_rate": 0.0008, "loss": 1.1664, "step": 11072 }, { "epoch": 16.57634730538922, "grad_norm": 0.2294921875, "learning_rate": 0.0008, "loss": 1.1715, "step": 11073 }, { "epoch": 16.577844311377245, "grad_norm": 0.162109375, "learning_rate": 0.0008, "loss": 1.1694, "step": 11074 }, { "epoch": 16.57934131736527, "grad_norm": 0.2255859375, "learning_rate": 0.0008, "loss": 1.1633, "step": 11075 }, { "epoch": 16.580838323353294, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.1594, "step": 11076 }, { "epoch": 16.58233532934132, "grad_norm": 0.2490234375, "learning_rate": 0.0008, "loss": 1.1658, "step": 11077 }, { "epoch": 16.58383233532934, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1665, "step": 11078 }, { "epoch": 16.585329341317365, "grad_norm": 0.1904296875, "learning_rate": 0.0008, "loss": 1.1682, "step": 11079 }, { "epoch": 16.58682634730539, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1667, "step": 11080 }, { "epoch": 16.58832335329341, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1605, "step": 11081 }, { "epoch": 16.589820359281436, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1633, "step": 11082 }, { "epoch": 16.59131736526946, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1626, "step": 11083 }, { "epoch": 16.592814371257486, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1626, "step": 11084 }, { "epoch": 16.59431137724551, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1606, "step": 11085 }, { "epoch": 16.595808383233532, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1572, "step": 11086 }, { "epoch": 16.597305389221557, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1589, "step": 11087 }, { "epoch": 16.59880239520958, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1535, "step": 11088 }, { "epoch": 16.600299401197606, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1563, "step": 11089 }, { "epoch": 16.601796407185628, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1593, "step": 11090 }, { "epoch": 16.603293413173652, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1585, "step": 11091 }, { "epoch": 16.604790419161677, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1593, "step": 11092 }, { "epoch": 16.606287425149702, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.163, "step": 11093 }, { "epoch": 16.607784431137723, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1582, "step": 11094 }, { "epoch": 16.60928143712575, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1586, "step": 11095 }, { "epoch": 16.610778443113773, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1556, "step": 11096 }, { "epoch": 16.612275449101798, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.155, "step": 11097 }, { "epoch": 16.61377245508982, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1638, "step": 11098 }, { "epoch": 16.615269461077844, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1563, "step": 11099 }, { "epoch": 16.61676646706587, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1597, "step": 11100 }, { "epoch": 16.618263473053894, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1577, "step": 11101 }, { "epoch": 16.619760479041915, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1585, "step": 11102 }, { "epoch": 16.62125748502994, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1597, "step": 11103 }, { "epoch": 16.622754491017965, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1537, "step": 11104 }, { "epoch": 16.62425149700599, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1552, "step": 11105 }, { "epoch": 16.62574850299401, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1551, "step": 11106 }, { "epoch": 16.627245508982035, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1477, "step": 11107 }, { "epoch": 16.62874251497006, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1582, "step": 11108 }, { "epoch": 16.630239520958085, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.157, "step": 11109 }, { "epoch": 16.631736526946106, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1535, "step": 11110 }, { "epoch": 16.63323353293413, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1575, "step": 11111 }, { "epoch": 16.634730538922156, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1559, "step": 11112 }, { "epoch": 16.63622754491018, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1557, "step": 11113 }, { "epoch": 16.637724550898202, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1539, "step": 11114 }, { "epoch": 16.639221556886227, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1605, "step": 11115 }, { "epoch": 16.64071856287425, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1561, "step": 11116 }, { "epoch": 16.642215568862277, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1545, "step": 11117 }, { "epoch": 16.643712574850298, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1518, "step": 11118 }, { "epoch": 16.645209580838323, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1506, "step": 11119 }, { "epoch": 16.646706586826348, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1522, "step": 11120 }, { "epoch": 16.648203592814372, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1531, "step": 11121 }, { "epoch": 16.649700598802394, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1532, "step": 11122 }, { "epoch": 16.65119760479042, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1592, "step": 11123 }, { "epoch": 16.652694610778443, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1598, "step": 11124 }, { "epoch": 16.654191616766468, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1529, "step": 11125 }, { "epoch": 16.65568862275449, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1574, "step": 11126 }, { "epoch": 16.657185628742514, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1462, "step": 11127 }, { "epoch": 16.65868263473054, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1524, "step": 11128 }, { "epoch": 16.660179640718564, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1567, "step": 11129 }, { "epoch": 16.66167664670659, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1529, "step": 11130 }, { "epoch": 16.66317365269461, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1506, "step": 11131 }, { "epoch": 16.664670658682635, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1445, "step": 11132 }, { "epoch": 16.66616766467066, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1503, "step": 11133 }, { "epoch": 16.66766467065868, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1453, "step": 11134 }, { "epoch": 16.669161676646706, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1542, "step": 11135 }, { "epoch": 16.67065868263473, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1519, "step": 11136 }, { "epoch": 16.672155688622755, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1602, "step": 11137 }, { "epoch": 16.67365269461078, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1489, "step": 11138 }, { "epoch": 16.6751497005988, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1504, "step": 11139 }, { "epoch": 16.676646706586826, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1483, "step": 11140 }, { "epoch": 16.67814371257485, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.153, "step": 11141 }, { "epoch": 16.679640718562876, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1476, "step": 11142 }, { "epoch": 16.681137724550897, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1546, "step": 11143 }, { "epoch": 16.682634730538922, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1521, "step": 11144 }, { "epoch": 16.684131736526947, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1565, "step": 11145 }, { "epoch": 16.68562874251497, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1555, "step": 11146 }, { "epoch": 16.687125748502993, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1517, "step": 11147 }, { "epoch": 16.688622754491018, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1492, "step": 11148 }, { "epoch": 16.690119760479043, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1494, "step": 11149 }, { "epoch": 16.691616766467067, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1535, "step": 11150 }, { "epoch": 16.69311377245509, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1471, "step": 11151 }, { "epoch": 16.694610778443113, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1564, "step": 11152 }, { "epoch": 16.69610778443114, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1487, "step": 11153 }, { "epoch": 16.697604790419163, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1528, "step": 11154 }, { "epoch": 16.699101796407184, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1518, "step": 11155 }, { "epoch": 16.70059880239521, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1453, "step": 11156 }, { "epoch": 16.702095808383234, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.149, "step": 11157 }, { "epoch": 16.70359281437126, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1488, "step": 11158 }, { "epoch": 16.70508982035928, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1471, "step": 11159 }, { "epoch": 16.706586826347305, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1465, "step": 11160 }, { "epoch": 16.70808383233533, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1531, "step": 11161 }, { "epoch": 16.709580838323355, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1481, "step": 11162 }, { "epoch": 16.711077844311376, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.147, "step": 11163 }, { "epoch": 16.7125748502994, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1505, "step": 11164 }, { "epoch": 16.714071856287426, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.145, "step": 11165 }, { "epoch": 16.71556886227545, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1524, "step": 11166 }, { "epoch": 16.71706586826347, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1476, "step": 11167 }, { "epoch": 16.718562874251496, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1575, "step": 11168 }, { "epoch": 16.72005988023952, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1608, "step": 11169 }, { "epoch": 16.721556886227546, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1537, "step": 11170 }, { "epoch": 16.723053892215567, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1419, "step": 11171 }, { "epoch": 16.724550898203592, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1466, "step": 11172 }, { "epoch": 16.726047904191617, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1478, "step": 11173 }, { "epoch": 16.727544910179642, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1502, "step": 11174 }, { "epoch": 16.729041916167663, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1529, "step": 11175 }, { "epoch": 16.730538922155688, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1506, "step": 11176 }, { "epoch": 16.732035928143713, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1526, "step": 11177 }, { "epoch": 16.733532934131738, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1471, "step": 11178 }, { "epoch": 16.73502994011976, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.153, "step": 11179 }, { "epoch": 16.736526946107784, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1528, "step": 11180 }, { "epoch": 16.73802395209581, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1478, "step": 11181 }, { "epoch": 16.739520958083833, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1463, "step": 11182 }, { "epoch": 16.741017964071855, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1509, "step": 11183 }, { "epoch": 16.74251497005988, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1508, "step": 11184 }, { "epoch": 16.744011976047904, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1459, "step": 11185 }, { "epoch": 16.74550898203593, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1425, "step": 11186 }, { "epoch": 16.74700598802395, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1492, "step": 11187 }, { "epoch": 16.748502994011975, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1478, "step": 11188 }, { "epoch": 16.75, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1528, "step": 11189 }, { "epoch": 16.751497005988025, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1529, "step": 11190 }, { "epoch": 16.75299401197605, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1525, "step": 11191 }, { "epoch": 16.75449101796407, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1474, "step": 11192 }, { "epoch": 16.755988023952096, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.147, "step": 11193 }, { "epoch": 16.75748502994012, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1531, "step": 11194 }, { "epoch": 16.758982035928145, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1559, "step": 11195 }, { "epoch": 16.760479041916167, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1571, "step": 11196 }, { "epoch": 16.76197604790419, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1498, "step": 11197 }, { "epoch": 16.763473053892216, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1517, "step": 11198 }, { "epoch": 16.76497005988024, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1444, "step": 11199 }, { "epoch": 16.766467065868262, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1451, "step": 11200 }, { "epoch": 16.767964071856287, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1509, "step": 11201 }, { "epoch": 16.769461077844312, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1484, "step": 11202 }, { "epoch": 16.770958083832337, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.152, "step": 11203 }, { "epoch": 16.772455089820358, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1491, "step": 11204 }, { "epoch": 16.773952095808383, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1484, "step": 11205 }, { "epoch": 16.775449101796408, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1542, "step": 11206 }, { "epoch": 16.776946107784433, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1435, "step": 11207 }, { "epoch": 16.778443113772454, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1429, "step": 11208 }, { "epoch": 16.77994011976048, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.145, "step": 11209 }, { "epoch": 16.781437125748504, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1484, "step": 11210 }, { "epoch": 16.78293413173653, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1487, "step": 11211 }, { "epoch": 16.78443113772455, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1456, "step": 11212 }, { "epoch": 16.785928143712574, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1449, "step": 11213 }, { "epoch": 16.7874251497006, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1424, "step": 11214 }, { "epoch": 16.788922155688624, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1531, "step": 11215 }, { "epoch": 16.790419161676645, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1436, "step": 11216 }, { "epoch": 16.79191616766467, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.149, "step": 11217 }, { "epoch": 16.793413173652695, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1478, "step": 11218 }, { "epoch": 16.79491017964072, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1505, "step": 11219 }, { "epoch": 16.79640718562874, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1433, "step": 11220 }, { "epoch": 16.797904191616766, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1485, "step": 11221 }, { "epoch": 16.79940119760479, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1436, "step": 11222 }, { "epoch": 16.800898203592816, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1495, "step": 11223 }, { "epoch": 16.802395209580837, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1493, "step": 11224 }, { "epoch": 16.80389221556886, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1457, "step": 11225 }, { "epoch": 16.805389221556887, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1476, "step": 11226 }, { "epoch": 16.80688622754491, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1419, "step": 11227 }, { "epoch": 16.808383233532933, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1507, "step": 11228 }, { "epoch": 16.809880239520957, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1493, "step": 11229 }, { "epoch": 16.811377245508982, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1504, "step": 11230 }, { "epoch": 16.812874251497007, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1458, "step": 11231 }, { "epoch": 16.81437125748503, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1539, "step": 11232 }, { "epoch": 16.815868263473053, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1446, "step": 11233 }, { "epoch": 16.817365269461078, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1474, "step": 11234 }, { "epoch": 16.818862275449103, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1456, "step": 11235 }, { "epoch": 16.820359281437124, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1461, "step": 11236 }, { "epoch": 16.82185628742515, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1512, "step": 11237 }, { "epoch": 16.823353293413174, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1448, "step": 11238 }, { "epoch": 16.8248502994012, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1449, "step": 11239 }, { "epoch": 16.82634730538922, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1424, "step": 11240 }, { "epoch": 16.827844311377245, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1427, "step": 11241 }, { "epoch": 16.82934131736527, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1468, "step": 11242 }, { "epoch": 16.830838323353294, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1461, "step": 11243 }, { "epoch": 16.83233532934132, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1524, "step": 11244 }, { "epoch": 16.83383233532934, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1468, "step": 11245 }, { "epoch": 16.835329341317365, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1479, "step": 11246 }, { "epoch": 16.83682634730539, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1439, "step": 11247 }, { "epoch": 16.83832335329341, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1467, "step": 11248 }, { "epoch": 16.839820359281436, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1446, "step": 11249 }, { "epoch": 16.84131736526946, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1484, "step": 11250 }, { "epoch": 16.842814371257486, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1464, "step": 11251 }, { "epoch": 16.84431137724551, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1446, "step": 11252 }, { "epoch": 16.845808383233532, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1432, "step": 11253 }, { "epoch": 16.847305389221557, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.149, "step": 11254 }, { "epoch": 16.84880239520958, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1431, "step": 11255 }, { "epoch": 16.850299401197606, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1398, "step": 11256 }, { "epoch": 16.851796407185628, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1449, "step": 11257 }, { "epoch": 16.853293413173652, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1494, "step": 11258 }, { "epoch": 16.854790419161677, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1467, "step": 11259 }, { "epoch": 16.856287425149702, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1436, "step": 11260 }, { "epoch": 16.857784431137723, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1465, "step": 11261 }, { "epoch": 16.85928143712575, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1444, "step": 11262 }, { "epoch": 16.860778443113773, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1486, "step": 11263 }, { "epoch": 16.862275449101798, "grad_norm": 0.1728515625, "learning_rate": 0.0008, "loss": 1.1518, "step": 11264 }, { "epoch": 16.86377245508982, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1393, "step": 11265 }, { "epoch": 16.865269461077844, "grad_norm": 0.15625, "learning_rate": 0.0008, "loss": 1.148, "step": 11266 }, { "epoch": 16.86676646706587, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1459, "step": 11267 }, { "epoch": 16.868263473053894, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1498, "step": 11268 }, { "epoch": 16.869760479041915, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1425, "step": 11269 }, { "epoch": 16.87125748502994, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1456, "step": 11270 }, { "epoch": 16.872754491017965, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1509, "step": 11271 }, { "epoch": 16.87425149700599, "grad_norm": 0.1845703125, "learning_rate": 0.0008, "loss": 1.1422, "step": 11272 }, { "epoch": 16.87574850299401, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1411, "step": 11273 }, { "epoch": 16.877245508982035, "grad_norm": 0.1953125, "learning_rate": 0.0008, "loss": 1.143, "step": 11274 }, { "epoch": 16.87874251497006, "grad_norm": 0.16796875, "learning_rate": 0.0008, "loss": 1.1429, "step": 11275 }, { "epoch": 16.880239520958085, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1466, "step": 11276 }, { "epoch": 16.881736526946106, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1496, "step": 11277 }, { "epoch": 16.88323353293413, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1468, "step": 11278 }, { "epoch": 16.884730538922156, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1445, "step": 11279 }, { "epoch": 16.88622754491018, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1487, "step": 11280 }, { "epoch": 16.887724550898202, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1458, "step": 11281 }, { "epoch": 16.889221556886227, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1482, "step": 11282 }, { "epoch": 16.89071856287425, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1458, "step": 11283 }, { "epoch": 16.892215568862277, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1422, "step": 11284 }, { "epoch": 16.893712574850298, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1462, "step": 11285 }, { "epoch": 16.895209580838323, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1491, "step": 11286 }, { "epoch": 16.896706586826348, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1524, "step": 11287 }, { "epoch": 16.898203592814372, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1512, "step": 11288 }, { "epoch": 16.899700598802394, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1432, "step": 11289 }, { "epoch": 16.90119760479042, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1506, "step": 11290 }, { "epoch": 16.902694610778443, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1422, "step": 11291 }, { "epoch": 16.904191616766468, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1445, "step": 11292 }, { "epoch": 16.90568862275449, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1445, "step": 11293 }, { "epoch": 16.907185628742514, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1508, "step": 11294 }, { "epoch": 16.90868263473054, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1524, "step": 11295 }, { "epoch": 16.910179640718564, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1459, "step": 11296 }, { "epoch": 16.91167664670659, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1465, "step": 11297 }, { "epoch": 16.91317365269461, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.143, "step": 11298 }, { "epoch": 16.914670658682635, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1433, "step": 11299 }, { "epoch": 16.91616766467066, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1471, "step": 11300 }, { "epoch": 16.91766467065868, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1412, "step": 11301 }, { "epoch": 16.919161676646706, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1435, "step": 11302 }, { "epoch": 16.92065868263473, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1458, "step": 11303 }, { "epoch": 16.922155688622755, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1416, "step": 11304 }, { "epoch": 16.92365269461078, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1398, "step": 11305 }, { "epoch": 16.9251497005988, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1488, "step": 11306 }, { "epoch": 16.926646706586826, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1461, "step": 11307 }, { "epoch": 16.92814371257485, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1397, "step": 11308 }, { "epoch": 16.929640718562876, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1462, "step": 11309 }, { "epoch": 16.931137724550897, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1495, "step": 11310 }, { "epoch": 16.932634730538922, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1491, "step": 11311 }, { "epoch": 16.934131736526947, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1457, "step": 11312 }, { "epoch": 16.93562874251497, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1431, "step": 11313 }, { "epoch": 16.937125748502993, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1447, "step": 11314 }, { "epoch": 16.938622754491018, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1422, "step": 11315 }, { "epoch": 16.940119760479043, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1453, "step": 11316 }, { "epoch": 16.941616766467067, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1533, "step": 11317 }, { "epoch": 16.94311377245509, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1452, "step": 11318 }, { "epoch": 16.944610778443113, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1473, "step": 11319 }, { "epoch": 16.94610778443114, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1475, "step": 11320 }, { "epoch": 16.947604790419163, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.147, "step": 11321 }, { "epoch": 16.949101796407184, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1459, "step": 11322 }, { "epoch": 16.95059880239521, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1453, "step": 11323 }, { "epoch": 16.952095808383234, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1378, "step": 11324 }, { "epoch": 16.95359281437126, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1416, "step": 11325 }, { "epoch": 16.95508982035928, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1418, "step": 11326 }, { "epoch": 16.956586826347305, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1424, "step": 11327 }, { "epoch": 16.95808383233533, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1419, "step": 11328 }, { "epoch": 16.959580838323355, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1409, "step": 11329 }, { "epoch": 16.961077844311376, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1411, "step": 11330 }, { "epoch": 16.9625748502994, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.149, "step": 11331 }, { "epoch": 16.964071856287426, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1442, "step": 11332 }, { "epoch": 16.96556886227545, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1432, "step": 11333 }, { "epoch": 16.96706586826347, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1507, "step": 11334 }, { "epoch": 16.968562874251496, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1437, "step": 11335 }, { "epoch": 16.97005988023952, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1459, "step": 11336 }, { "epoch": 16.971556886227546, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1466, "step": 11337 }, { "epoch": 16.973053892215567, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1386, "step": 11338 }, { "epoch": 16.974550898203592, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1506, "step": 11339 }, { "epoch": 16.976047904191617, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1466, "step": 11340 }, { "epoch": 16.977544910179642, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1467, "step": 11341 }, { "epoch": 16.979041916167663, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1459, "step": 11342 }, { "epoch": 16.980538922155688, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.144, "step": 11343 }, { "epoch": 16.982035928143713, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1409, "step": 11344 }, { "epoch": 16.983532934131738, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1483, "step": 11345 }, { "epoch": 16.98502994011976, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1467, "step": 11346 }, { "epoch": 16.986526946107784, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1493, "step": 11347 }, { "epoch": 16.98802395209581, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1338, "step": 11348 }, { "epoch": 16.989520958083833, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1453, "step": 11349 }, { "epoch": 16.991017964071855, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1452, "step": 11350 }, { "epoch": 16.99251497005988, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1419, "step": 11351 }, { "epoch": 16.994011976047904, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1516, "step": 11352 }, { "epoch": 16.99550898203593, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1423, "step": 11353 }, { "epoch": 16.99700598802395, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1444, "step": 11354 }, { "epoch": 16.998502994011975, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1388, "step": 11355 }, { "epoch": 17.0, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1399, "step": 11356 }, { "epoch": 17.001497005988025, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1538, "step": 11357 }, { "epoch": 17.00299401197605, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1396, "step": 11358 }, { "epoch": 17.00449101796407, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1425, "step": 11359 }, { "epoch": 17.005988023952096, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1519, "step": 11360 }, { "epoch": 17.00748502994012, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1427, "step": 11361 }, { "epoch": 17.008982035928145, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1473, "step": 11362 }, { "epoch": 17.010479041916167, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1442, "step": 11363 }, { "epoch": 17.01197604790419, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1489, "step": 11364 }, { "epoch": 17.013473053892216, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1472, "step": 11365 }, { "epoch": 17.01497005988024, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1414, "step": 11366 }, { "epoch": 17.016467065868262, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1466, "step": 11367 }, { "epoch": 17.017964071856287, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1392, "step": 11368 }, { "epoch": 17.019461077844312, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1482, "step": 11369 }, { "epoch": 17.020958083832337, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1505, "step": 11370 }, { "epoch": 17.022455089820358, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1381, "step": 11371 }, { "epoch": 17.023952095808383, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1438, "step": 11372 }, { "epoch": 17.025449101796408, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1431, "step": 11373 }, { "epoch": 17.026946107784433, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1469, "step": 11374 }, { "epoch": 17.028443113772454, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1388, "step": 11375 }, { "epoch": 17.02994011976048, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1483, "step": 11376 }, { "epoch": 17.031437125748504, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1466, "step": 11377 }, { "epoch": 17.03293413173653, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1397, "step": 11378 }, { "epoch": 17.03443113772455, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.1482, "step": 11379 }, { "epoch": 17.035928143712574, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.141, "step": 11380 }, { "epoch": 17.0374251497006, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.146, "step": 11381 }, { "epoch": 17.038922155688624, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1474, "step": 11382 }, { "epoch": 17.040419161676645, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1398, "step": 11383 }, { "epoch": 17.04191616766467, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1442, "step": 11384 }, { "epoch": 17.043413173652695, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1484, "step": 11385 }, { "epoch": 17.04491017964072, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1421, "step": 11386 }, { "epoch": 17.04640718562874, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1474, "step": 11387 }, { "epoch": 17.047904191616766, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1446, "step": 11388 }, { "epoch": 17.04940119760479, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1424, "step": 11389 }, { "epoch": 17.050898203592816, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1436, "step": 11390 }, { "epoch": 17.052395209580837, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1481, "step": 11391 }, { "epoch": 17.05389221556886, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1428, "step": 11392 }, { "epoch": 17.055389221556887, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1457, "step": 11393 }, { "epoch": 17.05688622754491, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1391, "step": 11394 }, { "epoch": 17.058383233532933, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1416, "step": 11395 }, { "epoch": 17.059880239520957, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1469, "step": 11396 }, { "epoch": 17.061377245508982, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1495, "step": 11397 }, { "epoch": 17.062874251497007, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1405, "step": 11398 }, { "epoch": 17.06437125748503, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1394, "step": 11399 }, { "epoch": 17.065868263473053, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.144, "step": 11400 }, { "epoch": 17.067365269461078, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1394, "step": 11401 }, { "epoch": 17.068862275449103, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1446, "step": 11402 }, { "epoch": 17.070359281437124, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1422, "step": 11403 }, { "epoch": 17.07185628742515, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.138, "step": 11404 }, { "epoch": 17.073353293413174, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.143, "step": 11405 }, { "epoch": 17.0748502994012, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1447, "step": 11406 }, { "epoch": 17.07634730538922, "grad_norm": 0.1201171875, "learning_rate": 0.0008, "loss": 1.1341, "step": 11407 }, { "epoch": 17.077844311377245, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.136, "step": 11408 }, { "epoch": 17.07934131736527, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1488, "step": 11409 }, { "epoch": 17.080838323353294, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1499, "step": 11410 }, { "epoch": 17.082335329341316, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1437, "step": 11411 }, { "epoch": 17.08383233532934, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1475, "step": 11412 }, { "epoch": 17.085329341317365, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1447, "step": 11413 }, { "epoch": 17.08682634730539, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.1433, "step": 11414 }, { "epoch": 17.088323353293415, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.147, "step": 11415 }, { "epoch": 17.089820359281436, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1433, "step": 11416 }, { "epoch": 17.09131736526946, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1469, "step": 11417 }, { "epoch": 17.092814371257486, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1478, "step": 11418 }, { "epoch": 17.09431137724551, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1415, "step": 11419 }, { "epoch": 17.095808383233532, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1461, "step": 11420 }, { "epoch": 17.097305389221557, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1418, "step": 11421 }, { "epoch": 17.09880239520958, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1336, "step": 11422 }, { "epoch": 17.100299401197606, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1369, "step": 11423 }, { "epoch": 17.101796407185628, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1433, "step": 11424 }, { "epoch": 17.103293413173652, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.15, "step": 11425 }, { "epoch": 17.104790419161677, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1447, "step": 11426 }, { "epoch": 17.106287425149702, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1381, "step": 11427 }, { "epoch": 17.107784431137723, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.145, "step": 11428 }, { "epoch": 17.10928143712575, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1414, "step": 11429 }, { "epoch": 17.110778443113773, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1458, "step": 11430 }, { "epoch": 17.112275449101798, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1451, "step": 11431 }, { "epoch": 17.11377245508982, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1433, "step": 11432 }, { "epoch": 17.115269461077844, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1415, "step": 11433 }, { "epoch": 17.11676646706587, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.15, "step": 11434 }, { "epoch": 17.118263473053894, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1393, "step": 11435 }, { "epoch": 17.119760479041915, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1401, "step": 11436 }, { "epoch": 17.12125748502994, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1448, "step": 11437 }, { "epoch": 17.122754491017965, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1447, "step": 11438 }, { "epoch": 17.12425149700599, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1395, "step": 11439 }, { "epoch": 17.12574850299401, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1415, "step": 11440 }, { "epoch": 17.127245508982035, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1456, "step": 11441 }, { "epoch": 17.12874251497006, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1412, "step": 11442 }, { "epoch": 17.130239520958085, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.15, "step": 11443 }, { "epoch": 17.131736526946106, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1458, "step": 11444 }, { "epoch": 17.13323353293413, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1403, "step": 11445 }, { "epoch": 17.134730538922156, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1398, "step": 11446 }, { "epoch": 17.13622754491018, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1432, "step": 11447 }, { "epoch": 17.137724550898202, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1432, "step": 11448 }, { "epoch": 17.139221556886227, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1443, "step": 11449 }, { "epoch": 17.14071856287425, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1463, "step": 11450 }, { "epoch": 17.142215568862277, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1405, "step": 11451 }, { "epoch": 17.143712574850298, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1407, "step": 11452 }, { "epoch": 17.145209580838323, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1424, "step": 11453 }, { "epoch": 17.146706586826348, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1414, "step": 11454 }, { "epoch": 17.148203592814372, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.143, "step": 11455 }, { "epoch": 17.149700598802394, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1389, "step": 11456 }, { "epoch": 17.15119760479042, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1384, "step": 11457 }, { "epoch": 17.152694610778443, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1361, "step": 11458 }, { "epoch": 17.154191616766468, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1387, "step": 11459 }, { "epoch": 17.15568862275449, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1453, "step": 11460 }, { "epoch": 17.157185628742514, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1358, "step": 11461 }, { "epoch": 17.15868263473054, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1484, "step": 11462 }, { "epoch": 17.160179640718564, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1515, "step": 11463 }, { "epoch": 17.161676646706585, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1465, "step": 11464 }, { "epoch": 17.16317365269461, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1406, "step": 11465 }, { "epoch": 17.164670658682635, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1382, "step": 11466 }, { "epoch": 17.16616766467066, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1379, "step": 11467 }, { "epoch": 17.16766467065868, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1491, "step": 11468 }, { "epoch": 17.169161676646706, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1411, "step": 11469 }, { "epoch": 17.17065868263473, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1414, "step": 11470 }, { "epoch": 17.172155688622755, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1452, "step": 11471 }, { "epoch": 17.17365269461078, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1416, "step": 11472 }, { "epoch": 17.1751497005988, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.14, "step": 11473 }, { "epoch": 17.176646706586826, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1447, "step": 11474 }, { "epoch": 17.17814371257485, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1348, "step": 11475 }, { "epoch": 17.179640718562876, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1374, "step": 11476 }, { "epoch": 17.181137724550897, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1356, "step": 11477 }, { "epoch": 17.182634730538922, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1406, "step": 11478 }, { "epoch": 17.184131736526947, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1405, "step": 11479 }, { "epoch": 17.18562874251497, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1432, "step": 11480 }, { "epoch": 17.187125748502993, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1387, "step": 11481 }, { "epoch": 17.188622754491018, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1467, "step": 11482 }, { "epoch": 17.190119760479043, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1462, "step": 11483 }, { "epoch": 17.191616766467067, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1467, "step": 11484 }, { "epoch": 17.19311377245509, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1421, "step": 11485 }, { "epoch": 17.194610778443113, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1403, "step": 11486 }, { "epoch": 17.19610778443114, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.142, "step": 11487 }, { "epoch": 17.197604790419163, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1372, "step": 11488 }, { "epoch": 17.199101796407184, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1405, "step": 11489 }, { "epoch": 17.20059880239521, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1456, "step": 11490 }, { "epoch": 17.202095808383234, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.139, "step": 11491 }, { "epoch": 17.20359281437126, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1372, "step": 11492 }, { "epoch": 17.20508982035928, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1336, "step": 11493 }, { "epoch": 17.206586826347305, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1468, "step": 11494 }, { "epoch": 17.20808383233533, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1402, "step": 11495 }, { "epoch": 17.209580838323355, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1452, "step": 11496 }, { "epoch": 17.211077844311376, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1386, "step": 11497 }, { "epoch": 17.2125748502994, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1437, "step": 11498 }, { "epoch": 17.214071856287426, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.143, "step": 11499 }, { "epoch": 17.21556886227545, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1481, "step": 11500 }, { "epoch": 17.21706586826347, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1451, "step": 11501 }, { "epoch": 17.218562874251496, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1463, "step": 11502 }, { "epoch": 17.22005988023952, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1445, "step": 11503 }, { "epoch": 17.221556886227546, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.141, "step": 11504 }, { "epoch": 17.223053892215567, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1438, "step": 11505 }, { "epoch": 17.224550898203592, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1374, "step": 11506 }, { "epoch": 17.226047904191617, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1394, "step": 11507 }, { "epoch": 17.227544910179642, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1414, "step": 11508 }, { "epoch": 17.229041916167663, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1476, "step": 11509 }, { "epoch": 17.230538922155688, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1388, "step": 11510 }, { "epoch": 17.232035928143713, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1468, "step": 11511 }, { "epoch": 17.233532934131738, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1422, "step": 11512 }, { "epoch": 17.23502994011976, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1322, "step": 11513 }, { "epoch": 17.236526946107784, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1394, "step": 11514 }, { "epoch": 17.23802395209581, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1445, "step": 11515 }, { "epoch": 17.239520958083833, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1449, "step": 11516 }, { "epoch": 17.241017964071855, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.134, "step": 11517 }, { "epoch": 17.24251497005988, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1474, "step": 11518 }, { "epoch": 17.244011976047904, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.139, "step": 11519 }, { "epoch": 17.24550898203593, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1461, "step": 11520 }, { "epoch": 17.24700598802395, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1444, "step": 11521 }, { "epoch": 17.248502994011975, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1369, "step": 11522 }, { "epoch": 17.25, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1421, "step": 11523 }, { "epoch": 17.251497005988025, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1441, "step": 11524 }, { "epoch": 17.25299401197605, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1386, "step": 11525 }, { "epoch": 17.25449101796407, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1386, "step": 11526 }, { "epoch": 17.255988023952096, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1424, "step": 11527 }, { "epoch": 17.25748502994012, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1417, "step": 11528 }, { "epoch": 17.258982035928145, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1456, "step": 11529 }, { "epoch": 17.260479041916167, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1435, "step": 11530 }, { "epoch": 17.26197604790419, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1435, "step": 11531 }, { "epoch": 17.263473053892216, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1443, "step": 11532 }, { "epoch": 17.26497005988024, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1437, "step": 11533 }, { "epoch": 17.266467065868262, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1445, "step": 11534 }, { "epoch": 17.267964071856287, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1428, "step": 11535 }, { "epoch": 17.269461077844312, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1448, "step": 11536 }, { "epoch": 17.270958083832337, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1407, "step": 11537 }, { "epoch": 17.272455089820358, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1463, "step": 11538 }, { "epoch": 17.273952095808383, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1421, "step": 11539 }, { "epoch": 17.275449101796408, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1449, "step": 11540 }, { "epoch": 17.276946107784433, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1418, "step": 11541 }, { "epoch": 17.278443113772454, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1402, "step": 11542 }, { "epoch": 17.27994011976048, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1364, "step": 11543 }, { "epoch": 17.281437125748504, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1423, "step": 11544 }, { "epoch": 17.28293413173653, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1444, "step": 11545 }, { "epoch": 17.28443113772455, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.137, "step": 11546 }, { "epoch": 17.285928143712574, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1472, "step": 11547 }, { "epoch": 17.2874251497006, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1432, "step": 11548 }, { "epoch": 17.288922155688624, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1405, "step": 11549 }, { "epoch": 17.290419161676645, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1319, "step": 11550 }, { "epoch": 17.29191616766467, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1334, "step": 11551 }, { "epoch": 17.293413173652695, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1431, "step": 11552 }, { "epoch": 17.29491017964072, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1405, "step": 11553 }, { "epoch": 17.29640718562874, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1412, "step": 11554 }, { "epoch": 17.297904191616766, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1441, "step": 11555 }, { "epoch": 17.29940119760479, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1407, "step": 11556 }, { "epoch": 17.300898203592816, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1394, "step": 11557 }, { "epoch": 17.302395209580837, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1404, "step": 11558 }, { "epoch": 17.30389221556886, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1374, "step": 11559 }, { "epoch": 17.305389221556887, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1357, "step": 11560 }, { "epoch": 17.30688622754491, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1421, "step": 11561 }, { "epoch": 17.308383233532933, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1426, "step": 11562 }, { "epoch": 17.309880239520957, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1374, "step": 11563 }, { "epoch": 17.311377245508982, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1467, "step": 11564 }, { "epoch": 17.312874251497007, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1426, "step": 11565 }, { "epoch": 17.31437125748503, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1393, "step": 11566 }, { "epoch": 17.315868263473053, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1401, "step": 11567 }, { "epoch": 17.317365269461078, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.14, "step": 11568 }, { "epoch": 17.318862275449103, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1464, "step": 11569 }, { "epoch": 17.320359281437124, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1411, "step": 11570 }, { "epoch": 17.32185628742515, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1439, "step": 11571 }, { "epoch": 17.323353293413174, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1433, "step": 11572 }, { "epoch": 17.3248502994012, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1494, "step": 11573 }, { "epoch": 17.32634730538922, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1481, "step": 11574 }, { "epoch": 17.327844311377245, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.141, "step": 11575 }, { "epoch": 17.32934131736527, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1417, "step": 11576 }, { "epoch": 17.330838323353294, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1459, "step": 11577 }, { "epoch": 17.33233532934132, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1462, "step": 11578 }, { "epoch": 17.33383233532934, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1467, "step": 11579 }, { "epoch": 17.335329341317365, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.145, "step": 11580 }, { "epoch": 17.33682634730539, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1418, "step": 11581 }, { "epoch": 17.338323353293415, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1467, "step": 11582 }, { "epoch": 17.339820359281436, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1374, "step": 11583 }, { "epoch": 17.34131736526946, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1397, "step": 11584 }, { "epoch": 17.342814371257486, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1418, "step": 11585 }, { "epoch": 17.34431137724551, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1328, "step": 11586 }, { "epoch": 17.345808383233532, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1394, "step": 11587 }, { "epoch": 17.347305389221557, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1384, "step": 11588 }, { "epoch": 17.34880239520958, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1465, "step": 11589 }, { "epoch": 17.350299401197606, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1432, "step": 11590 }, { "epoch": 17.351796407185628, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1445, "step": 11591 }, { "epoch": 17.353293413173652, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1463, "step": 11592 }, { "epoch": 17.354790419161677, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1454, "step": 11593 }, { "epoch": 17.356287425149702, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1442, "step": 11594 }, { "epoch": 17.357784431137723, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1446, "step": 11595 }, { "epoch": 17.35928143712575, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1466, "step": 11596 }, { "epoch": 17.360778443113773, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1404, "step": 11597 }, { "epoch": 17.362275449101798, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1437, "step": 11598 }, { "epoch": 17.36377245508982, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1394, "step": 11599 }, { "epoch": 17.365269461077844, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1402, "step": 11600 }, { "epoch": 17.36676646706587, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1327, "step": 11601 }, { "epoch": 17.368263473053894, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1371, "step": 11602 }, { "epoch": 17.369760479041915, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1426, "step": 11603 }, { "epoch": 17.37125748502994, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1409, "step": 11604 }, { "epoch": 17.372754491017965, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1456, "step": 11605 }, { "epoch": 17.37425149700599, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1492, "step": 11606 }, { "epoch": 17.37574850299401, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1424, "step": 11607 }, { "epoch": 17.377245508982035, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1457, "step": 11608 }, { "epoch": 17.37874251497006, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1371, "step": 11609 }, { "epoch": 17.380239520958085, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1463, "step": 11610 }, { "epoch": 17.381736526946106, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1455, "step": 11611 }, { "epoch": 17.38323353293413, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1362, "step": 11612 }, { "epoch": 17.384730538922156, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1436, "step": 11613 }, { "epoch": 17.38622754491018, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1441, "step": 11614 }, { "epoch": 17.387724550898202, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1425, "step": 11615 }, { "epoch": 17.389221556886227, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1441, "step": 11616 }, { "epoch": 17.39071856287425, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1436, "step": 11617 }, { "epoch": 17.392215568862277, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1418, "step": 11618 }, { "epoch": 17.393712574850298, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1458, "step": 11619 }, { "epoch": 17.395209580838323, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1471, "step": 11620 }, { "epoch": 17.396706586826348, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1375, "step": 11621 }, { "epoch": 17.398203592814372, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1431, "step": 11622 }, { "epoch": 17.399700598802394, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1414, "step": 11623 }, { "epoch": 17.40119760479042, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1419, "step": 11624 }, { "epoch": 17.402694610778443, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1458, "step": 11625 }, { "epoch": 17.404191616766468, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1448, "step": 11626 }, { "epoch": 17.40568862275449, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1432, "step": 11627 }, { "epoch": 17.407185628742514, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1416, "step": 11628 }, { "epoch": 17.40868263473054, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1344, "step": 11629 }, { "epoch": 17.410179640718564, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1398, "step": 11630 }, { "epoch": 17.411676646706585, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1404, "step": 11631 }, { "epoch": 17.41317365269461, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1413, "step": 11632 }, { "epoch": 17.414670658682635, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1462, "step": 11633 }, { "epoch": 17.41616766467066, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1414, "step": 11634 }, { "epoch": 17.41766467065868, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1318, "step": 11635 }, { "epoch": 17.419161676646706, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1388, "step": 11636 }, { "epoch": 17.42065868263473, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1468, "step": 11637 }, { "epoch": 17.422155688622755, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1433, "step": 11638 }, { "epoch": 17.42365269461078, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1397, "step": 11639 }, { "epoch": 17.4251497005988, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1433, "step": 11640 }, { "epoch": 17.426646706586826, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1398, "step": 11641 }, { "epoch": 17.42814371257485, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1437, "step": 11642 }, { "epoch": 17.429640718562876, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1423, "step": 11643 }, { "epoch": 17.431137724550897, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1402, "step": 11644 }, { "epoch": 17.432634730538922, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1479, "step": 11645 }, { "epoch": 17.434131736526947, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1357, "step": 11646 }, { "epoch": 17.43562874251497, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.139, "step": 11647 }, { "epoch": 17.437125748502993, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1476, "step": 11648 }, { "epoch": 17.438622754491018, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1427, "step": 11649 }, { "epoch": 17.440119760479043, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1485, "step": 11650 }, { "epoch": 17.441616766467067, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1331, "step": 11651 }, { "epoch": 17.44311377245509, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1435, "step": 11652 }, { "epoch": 17.444610778443113, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1416, "step": 11653 }, { "epoch": 17.44610778443114, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1427, "step": 11654 }, { "epoch": 17.447604790419163, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1459, "step": 11655 }, { "epoch": 17.449101796407184, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1379, "step": 11656 }, { "epoch": 17.45059880239521, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1441, "step": 11657 }, { "epoch": 17.452095808383234, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1427, "step": 11658 }, { "epoch": 17.45359281437126, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1378, "step": 11659 }, { "epoch": 17.45508982035928, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1371, "step": 11660 }, { "epoch": 17.456586826347305, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1462, "step": 11661 }, { "epoch": 17.45808383233533, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1395, "step": 11662 }, { "epoch": 17.459580838323355, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.137, "step": 11663 }, { "epoch": 17.461077844311376, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.144, "step": 11664 }, { "epoch": 17.4625748502994, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1407, "step": 11665 }, { "epoch": 17.464071856287426, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1449, "step": 11666 }, { "epoch": 17.46556886227545, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.133, "step": 11667 }, { "epoch": 17.46706586826347, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1422, "step": 11668 }, { "epoch": 17.468562874251496, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1432, "step": 11669 }, { "epoch": 17.47005988023952, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1456, "step": 11670 }, { "epoch": 17.471556886227546, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1379, "step": 11671 }, { "epoch": 17.473053892215567, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1359, "step": 11672 }, { "epoch": 17.474550898203592, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1394, "step": 11673 }, { "epoch": 17.476047904191617, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1443, "step": 11674 }, { "epoch": 17.477544910179642, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1387, "step": 11675 }, { "epoch": 17.479041916167663, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1435, "step": 11676 }, { "epoch": 17.480538922155688, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1427, "step": 11677 }, { "epoch": 17.482035928143713, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1383, "step": 11678 }, { "epoch": 17.483532934131738, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1392, "step": 11679 }, { "epoch": 17.48502994011976, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1412, "step": 11680 }, { "epoch": 17.486526946107784, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.1433, "step": 11681 }, { "epoch": 17.48802395209581, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1455, "step": 11682 }, { "epoch": 17.489520958083833, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1405, "step": 11683 }, { "epoch": 17.491017964071855, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1393, "step": 11684 }, { "epoch": 17.49251497005988, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1357, "step": 11685 }, { "epoch": 17.494011976047904, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.144, "step": 11686 }, { "epoch": 17.49550898203593, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1452, "step": 11687 }, { "epoch": 17.49700598802395, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1385, "step": 11688 }, { "epoch": 17.498502994011975, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1431, "step": 11689 }, { "epoch": 17.5, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1373, "step": 11690 }, { "epoch": 17.501497005988025, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.138, "step": 11691 }, { "epoch": 17.50299401197605, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1384, "step": 11692 }, { "epoch": 17.50449101796407, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.141, "step": 11693 }, { "epoch": 17.505988023952096, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1401, "step": 11694 }, { "epoch": 17.50748502994012, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1441, "step": 11695 }, { "epoch": 17.508982035928145, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1442, "step": 11696 }, { "epoch": 17.510479041916167, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1423, "step": 11697 }, { "epoch": 17.51197604790419, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1378, "step": 11698 }, { "epoch": 17.513473053892216, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1423, "step": 11699 }, { "epoch": 17.51497005988024, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.139, "step": 11700 }, { "epoch": 17.516467065868262, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1389, "step": 11701 }, { "epoch": 17.517964071856287, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1341, "step": 11702 }, { "epoch": 17.519461077844312, "grad_norm": 0.1748046875, "learning_rate": 0.0008, "loss": 1.1485, "step": 11703 }, { "epoch": 17.520958083832337, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1405, "step": 11704 }, { "epoch": 17.522455089820358, "grad_norm": 0.1904296875, "learning_rate": 0.0008, "loss": 1.1368, "step": 11705 }, { "epoch": 17.523952095808383, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1447, "step": 11706 }, { "epoch": 17.525449101796408, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1414, "step": 11707 }, { "epoch": 17.526946107784433, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1456, "step": 11708 }, { "epoch": 17.528443113772454, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1384, "step": 11709 }, { "epoch": 17.52994011976048, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.14, "step": 11710 }, { "epoch": 17.531437125748504, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1357, "step": 11711 }, { "epoch": 17.53293413173653, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1425, "step": 11712 }, { "epoch": 17.53443113772455, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1429, "step": 11713 }, { "epoch": 17.535928143712574, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1435, "step": 11714 }, { "epoch": 17.5374251497006, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1443, "step": 11715 }, { "epoch": 17.538922155688624, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1411, "step": 11716 }, { "epoch": 17.540419161676645, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1442, "step": 11717 }, { "epoch": 17.54191616766467, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1409, "step": 11718 }, { "epoch": 17.543413173652695, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1448, "step": 11719 }, { "epoch": 17.54491017964072, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.143, "step": 11720 }, { "epoch": 17.54640718562874, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1525, "step": 11721 }, { "epoch": 17.547904191616766, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1354, "step": 11722 }, { "epoch": 17.54940119760479, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1451, "step": 11723 }, { "epoch": 17.550898203592816, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1436, "step": 11724 }, { "epoch": 17.552395209580837, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1402, "step": 11725 }, { "epoch": 17.55389221556886, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1417, "step": 11726 }, { "epoch": 17.555389221556887, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1335, "step": 11727 }, { "epoch": 17.55688622754491, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.134, "step": 11728 }, { "epoch": 17.558383233532933, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1426, "step": 11729 }, { "epoch": 17.559880239520957, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1323, "step": 11730 }, { "epoch": 17.561377245508982, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1483, "step": 11731 }, { "epoch": 17.562874251497007, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1402, "step": 11732 }, { "epoch": 17.56437125748503, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1425, "step": 11733 }, { "epoch": 17.565868263473053, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1403, "step": 11734 }, { "epoch": 17.567365269461078, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1455, "step": 11735 }, { "epoch": 17.568862275449103, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1377, "step": 11736 }, { "epoch": 17.570359281437124, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1472, "step": 11737 }, { "epoch": 17.57185628742515, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1457, "step": 11738 }, { "epoch": 17.573353293413174, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.141, "step": 11739 }, { "epoch": 17.5748502994012, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1395, "step": 11740 }, { "epoch": 17.57634730538922, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1363, "step": 11741 }, { "epoch": 17.577844311377245, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1415, "step": 11742 }, { "epoch": 17.57934131736527, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1367, "step": 11743 }, { "epoch": 17.580838323353294, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1433, "step": 11744 }, { "epoch": 17.58233532934132, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1393, "step": 11745 }, { "epoch": 17.58383233532934, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1433, "step": 11746 }, { "epoch": 17.585329341317365, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1425, "step": 11747 }, { "epoch": 17.58682634730539, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1417, "step": 11748 }, { "epoch": 17.58832335329341, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.144, "step": 11749 }, { "epoch": 17.589820359281436, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1365, "step": 11750 }, { "epoch": 17.59131736526946, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.142, "step": 11751 }, { "epoch": 17.592814371257486, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1437, "step": 11752 }, { "epoch": 17.59431137724551, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1374, "step": 11753 }, { "epoch": 17.595808383233532, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1412, "step": 11754 }, { "epoch": 17.597305389221557, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.1377, "step": 11755 }, { "epoch": 17.59880239520958, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.145, "step": 11756 }, { "epoch": 17.600299401197606, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1444, "step": 11757 }, { "epoch": 17.601796407185628, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1424, "step": 11758 }, { "epoch": 17.603293413173652, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1433, "step": 11759 }, { "epoch": 17.604790419161677, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1449, "step": 11760 }, { "epoch": 17.606287425149702, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1445, "step": 11761 }, { "epoch": 17.607784431137723, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1439, "step": 11762 }, { "epoch": 17.60928143712575, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1406, "step": 11763 }, { "epoch": 17.610778443113773, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1419, "step": 11764 }, { "epoch": 17.612275449101798, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1445, "step": 11765 }, { "epoch": 17.61377245508982, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1397, "step": 11766 }, { "epoch": 17.615269461077844, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1462, "step": 11767 }, { "epoch": 17.61676646706587, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1399, "step": 11768 }, { "epoch": 17.618263473053894, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1403, "step": 11769 }, { "epoch": 17.619760479041915, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.142, "step": 11770 }, { "epoch": 17.62125748502994, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1379, "step": 11771 }, { "epoch": 17.622754491017965, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1453, "step": 11772 }, { "epoch": 17.62425149700599, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1414, "step": 11773 }, { "epoch": 17.62574850299401, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1484, "step": 11774 }, { "epoch": 17.627245508982035, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1365, "step": 11775 }, { "epoch": 17.62874251497006, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1423, "step": 11776 }, { "epoch": 17.630239520958085, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1426, "step": 11777 }, { "epoch": 17.631736526946106, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1474, "step": 11778 }, { "epoch": 17.63323353293413, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1347, "step": 11779 }, { "epoch": 17.634730538922156, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1436, "step": 11780 }, { "epoch": 17.63622754491018, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1447, "step": 11781 }, { "epoch": 17.637724550898202, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.142, "step": 11782 }, { "epoch": 17.639221556886227, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1345, "step": 11783 }, { "epoch": 17.64071856287425, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1316, "step": 11784 }, { "epoch": 17.642215568862277, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1351, "step": 11785 }, { "epoch": 17.643712574850298, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1361, "step": 11786 }, { "epoch": 17.645209580838323, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1397, "step": 11787 }, { "epoch": 17.646706586826348, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1386, "step": 11788 }, { "epoch": 17.648203592814372, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.138, "step": 11789 }, { "epoch": 17.649700598802394, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1397, "step": 11790 }, { "epoch": 17.65119760479042, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1479, "step": 11791 }, { "epoch": 17.652694610778443, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1419, "step": 11792 }, { "epoch": 17.654191616766468, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.136, "step": 11793 }, { "epoch": 17.65568862275449, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.133, "step": 11794 }, { "epoch": 17.657185628742514, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1366, "step": 11795 }, { "epoch": 17.65868263473054, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1419, "step": 11796 }, { "epoch": 17.660179640718564, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1401, "step": 11797 }, { "epoch": 17.66167664670659, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1343, "step": 11798 }, { "epoch": 17.66317365269461, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.141, "step": 11799 }, { "epoch": 17.664670658682635, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1377, "step": 11800 }, { "epoch": 17.66616766467066, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1403, "step": 11801 }, { "epoch": 17.66766467065868, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1435, "step": 11802 }, { "epoch": 17.669161676646706, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.143, "step": 11803 }, { "epoch": 17.67065868263473, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1429, "step": 11804 }, { "epoch": 17.672155688622755, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1374, "step": 11805 }, { "epoch": 17.67365269461078, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.143, "step": 11806 }, { "epoch": 17.6751497005988, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1367, "step": 11807 }, { "epoch": 17.676646706586826, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1452, "step": 11808 }, { "epoch": 17.67814371257485, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1485, "step": 11809 }, { "epoch": 17.679640718562876, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1377, "step": 11810 }, { "epoch": 17.681137724550897, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.141, "step": 11811 }, { "epoch": 17.682634730538922, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1413, "step": 11812 }, { "epoch": 17.684131736526947, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1412, "step": 11813 }, { "epoch": 17.68562874251497, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1405, "step": 11814 }, { "epoch": 17.687125748502993, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1441, "step": 11815 }, { "epoch": 17.688622754491018, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1371, "step": 11816 }, { "epoch": 17.690119760479043, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1413, "step": 11817 }, { "epoch": 17.691616766467067, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.139, "step": 11818 }, { "epoch": 17.69311377245509, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1386, "step": 11819 }, { "epoch": 17.694610778443113, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1365, "step": 11820 }, { "epoch": 17.69610778443114, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1422, "step": 11821 }, { "epoch": 17.697604790419163, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1483, "step": 11822 }, { "epoch": 17.699101796407184, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1472, "step": 11823 }, { "epoch": 17.70059880239521, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1424, "step": 11824 }, { "epoch": 17.702095808383234, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1385, "step": 11825 }, { "epoch": 17.70359281437126, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1431, "step": 11826 }, { "epoch": 17.70508982035928, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1363, "step": 11827 }, { "epoch": 17.706586826347305, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1442, "step": 11828 }, { "epoch": 17.70808383233533, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1396, "step": 11829 }, { "epoch": 17.709580838323355, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1435, "step": 11830 }, { "epoch": 17.711077844311376, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1449, "step": 11831 }, { "epoch": 17.7125748502994, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1408, "step": 11832 }, { "epoch": 17.714071856287426, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1403, "step": 11833 }, { "epoch": 17.71556886227545, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1418, "step": 11834 }, { "epoch": 17.71706586826347, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1461, "step": 11835 }, { "epoch": 17.718562874251496, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1414, "step": 11836 }, { "epoch": 17.72005988023952, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1428, "step": 11837 }, { "epoch": 17.721556886227546, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1432, "step": 11838 }, { "epoch": 17.723053892215567, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1523, "step": 11839 }, { "epoch": 17.724550898203592, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1335, "step": 11840 }, { "epoch": 17.726047904191617, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1419, "step": 11841 }, { "epoch": 17.727544910179642, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1323, "step": 11842 }, { "epoch": 17.729041916167663, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1363, "step": 11843 }, { "epoch": 17.730538922155688, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1448, "step": 11844 }, { "epoch": 17.732035928143713, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1399, "step": 11845 }, { "epoch": 17.733532934131738, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1402, "step": 11846 }, { "epoch": 17.73502994011976, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1455, "step": 11847 }, { "epoch": 17.736526946107784, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1331, "step": 11848 }, { "epoch": 17.73802395209581, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1397, "step": 11849 }, { "epoch": 17.739520958083833, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1407, "step": 11850 }, { "epoch": 17.741017964071855, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1408, "step": 11851 }, { "epoch": 17.74251497005988, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1386, "step": 11852 }, { "epoch": 17.744011976047904, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1383, "step": 11853 }, { "epoch": 17.74550898203593, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1421, "step": 11854 }, { "epoch": 17.74700598802395, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1396, "step": 11855 }, { "epoch": 17.748502994011975, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1388, "step": 11856 }, { "epoch": 17.75, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.143, "step": 11857 }, { "epoch": 17.751497005988025, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1376, "step": 11858 }, { "epoch": 17.75299401197605, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1403, "step": 11859 }, { "epoch": 17.75449101796407, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1482, "step": 11860 }, { "epoch": 17.755988023952096, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1351, "step": 11861 }, { "epoch": 17.75748502994012, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1365, "step": 11862 }, { "epoch": 17.758982035928145, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1381, "step": 11863 }, { "epoch": 17.760479041916167, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1433, "step": 11864 }, { "epoch": 17.76197604790419, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1449, "step": 11865 }, { "epoch": 17.763473053892216, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1351, "step": 11866 }, { "epoch": 17.76497005988024, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1396, "step": 11867 }, { "epoch": 17.766467065868262, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1434, "step": 11868 }, { "epoch": 17.767964071856287, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1394, "step": 11869 }, { "epoch": 17.769461077844312, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1479, "step": 11870 }, { "epoch": 17.770958083832337, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1407, "step": 11871 }, { "epoch": 17.772455089820358, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1424, "step": 11872 }, { "epoch": 17.773952095808383, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1383, "step": 11873 }, { "epoch": 17.775449101796408, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1451, "step": 11874 }, { "epoch": 17.776946107784433, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1344, "step": 11875 }, { "epoch": 17.778443113772454, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1374, "step": 11876 }, { "epoch": 17.77994011976048, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1359, "step": 11877 }, { "epoch": 17.781437125748504, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1454, "step": 11878 }, { "epoch": 17.78293413173653, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1428, "step": 11879 }, { "epoch": 17.78443113772455, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1405, "step": 11880 }, { "epoch": 17.785928143712574, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1443, "step": 11881 }, { "epoch": 17.7874251497006, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1411, "step": 11882 }, { "epoch": 17.788922155688624, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1424, "step": 11883 }, { "epoch": 17.790419161676645, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.136, "step": 11884 }, { "epoch": 17.79191616766467, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1442, "step": 11885 }, { "epoch": 17.793413173652695, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.14, "step": 11886 }, { "epoch": 17.79491017964072, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1357, "step": 11887 }, { "epoch": 17.79640718562874, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1382, "step": 11888 }, { "epoch": 17.797904191616766, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1375, "step": 11889 }, { "epoch": 17.79940119760479, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1409, "step": 11890 }, { "epoch": 17.800898203592816, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1447, "step": 11891 }, { "epoch": 17.802395209580837, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1361, "step": 11892 }, { "epoch": 17.80389221556886, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1367, "step": 11893 }, { "epoch": 17.805389221556887, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1398, "step": 11894 }, { "epoch": 17.80688622754491, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1443, "step": 11895 }, { "epoch": 17.808383233532933, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.143, "step": 11896 }, { "epoch": 17.809880239520957, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.138, "step": 11897 }, { "epoch": 17.811377245508982, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.1393, "step": 11898 }, { "epoch": 17.812874251497007, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.132, "step": 11899 }, { "epoch": 17.81437125748503, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1441, "step": 11900 }, { "epoch": 17.815868263473053, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1398, "step": 11901 }, { "epoch": 17.817365269461078, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.141, "step": 11902 }, { "epoch": 17.818862275449103, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1403, "step": 11903 }, { "epoch": 17.820359281437124, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1435, "step": 11904 }, { "epoch": 17.82185628742515, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1504, "step": 11905 }, { "epoch": 17.823353293413174, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1364, "step": 11906 }, { "epoch": 17.8248502994012, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1391, "step": 11907 }, { "epoch": 17.82634730538922, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1454, "step": 11908 }, { "epoch": 17.827844311377245, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1405, "step": 11909 }, { "epoch": 17.82934131736527, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1475, "step": 11910 }, { "epoch": 17.830838323353294, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1416, "step": 11911 }, { "epoch": 17.83233532934132, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1377, "step": 11912 }, { "epoch": 17.83383233532934, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1416, "step": 11913 }, { "epoch": 17.835329341317365, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1419, "step": 11914 }, { "epoch": 17.83682634730539, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1389, "step": 11915 }, { "epoch": 17.83832335329341, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1429, "step": 11916 }, { "epoch": 17.839820359281436, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1432, "step": 11917 }, { "epoch": 17.84131736526946, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1372, "step": 11918 }, { "epoch": 17.842814371257486, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1403, "step": 11919 }, { "epoch": 17.84431137724551, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1423, "step": 11920 }, { "epoch": 17.845808383233532, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1408, "step": 11921 }, { "epoch": 17.847305389221557, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1412, "step": 11922 }, { "epoch": 17.84880239520958, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1345, "step": 11923 }, { "epoch": 17.850299401197606, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1328, "step": 11924 }, { "epoch": 17.851796407185628, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1372, "step": 11925 }, { "epoch": 17.853293413173652, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1406, "step": 11926 }, { "epoch": 17.854790419161677, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1409, "step": 11927 }, { "epoch": 17.856287425149702, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1355, "step": 11928 }, { "epoch": 17.857784431137723, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1431, "step": 11929 }, { "epoch": 17.85928143712575, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.137, "step": 11930 }, { "epoch": 17.860778443113773, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1442, "step": 11931 }, { "epoch": 17.862275449101798, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1418, "step": 11932 }, { "epoch": 17.86377245508982, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1404, "step": 11933 }, { "epoch": 17.865269461077844, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1341, "step": 11934 }, { "epoch": 17.86676646706587, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1477, "step": 11935 }, { "epoch": 17.868263473053894, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1425, "step": 11936 }, { "epoch": 17.869760479041915, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1375, "step": 11937 }, { "epoch": 17.87125748502994, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.138, "step": 11938 }, { "epoch": 17.872754491017965, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.141, "step": 11939 }, { "epoch": 17.87425149700599, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1345, "step": 11940 }, { "epoch": 17.87574850299401, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1448, "step": 11941 }, { "epoch": 17.877245508982035, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1425, "step": 11942 }, { "epoch": 17.87874251497006, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1412, "step": 11943 }, { "epoch": 17.880239520958085, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.151, "step": 11944 }, { "epoch": 17.881736526946106, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1405, "step": 11945 }, { "epoch": 17.88323353293413, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1365, "step": 11946 }, { "epoch": 17.884730538922156, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1374, "step": 11947 }, { "epoch": 17.88622754491018, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1439, "step": 11948 }, { "epoch": 17.887724550898202, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1369, "step": 11949 }, { "epoch": 17.889221556886227, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1411, "step": 11950 }, { "epoch": 17.89071856287425, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1458, "step": 11951 }, { "epoch": 17.892215568862277, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1404, "step": 11952 }, { "epoch": 17.893712574850298, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.139, "step": 11953 }, { "epoch": 17.895209580838323, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1475, "step": 11954 }, { "epoch": 17.896706586826348, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1438, "step": 11955 }, { "epoch": 17.898203592814372, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.139, "step": 11956 }, { "epoch": 17.899700598802394, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1386, "step": 11957 }, { "epoch": 17.90119760479042, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1386, "step": 11958 }, { "epoch": 17.902694610778443, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1453, "step": 11959 }, { "epoch": 17.904191616766468, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.137, "step": 11960 }, { "epoch": 17.90568862275449, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1391, "step": 11961 }, { "epoch": 17.907185628742514, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1449, "step": 11962 }, { "epoch": 17.90868263473054, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1401, "step": 11963 }, { "epoch": 17.910179640718564, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1325, "step": 11964 }, { "epoch": 17.91167664670659, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.143, "step": 11965 }, { "epoch": 17.91317365269461, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1376, "step": 11966 }, { "epoch": 17.914670658682635, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1461, "step": 11967 }, { "epoch": 17.91616766467066, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1405, "step": 11968 }, { "epoch": 17.91766467065868, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.144, "step": 11969 }, { "epoch": 17.919161676646706, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1474, "step": 11970 }, { "epoch": 17.92065868263473, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1355, "step": 11971 }, { "epoch": 17.922155688622755, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1409, "step": 11972 }, { "epoch": 17.92365269461078, "grad_norm": 0.1787109375, "learning_rate": 0.0008, "loss": 1.1415, "step": 11973 }, { "epoch": 17.9251497005988, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1452, "step": 11974 }, { "epoch": 17.926646706586826, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1364, "step": 11975 }, { "epoch": 17.92814371257485, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1411, "step": 11976 }, { "epoch": 17.929640718562876, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1461, "step": 11977 }, { "epoch": 17.931137724550897, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1497, "step": 11978 }, { "epoch": 17.932634730538922, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1381, "step": 11979 }, { "epoch": 17.934131736526947, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1409, "step": 11980 }, { "epoch": 17.93562874251497, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1367, "step": 11981 }, { "epoch": 17.937125748502993, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1451, "step": 11982 }, { "epoch": 17.938622754491018, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1413, "step": 11983 }, { "epoch": 17.940119760479043, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1455, "step": 11984 }, { "epoch": 17.941616766467067, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1467, "step": 11985 }, { "epoch": 17.94311377245509, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1368, "step": 11986 }, { "epoch": 17.944610778443113, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1354, "step": 11987 }, { "epoch": 17.94610778443114, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1455, "step": 11988 }, { "epoch": 17.947604790419163, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1361, "step": 11989 }, { "epoch": 17.949101796407184, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.138, "step": 11990 }, { "epoch": 17.95059880239521, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1428, "step": 11991 }, { "epoch": 17.952095808383234, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1367, "step": 11992 }, { "epoch": 17.95359281437126, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1364, "step": 11993 }, { "epoch": 17.95508982035928, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1439, "step": 11994 }, { "epoch": 17.956586826347305, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1414, "step": 11995 }, { "epoch": 17.95808383233533, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.147, "step": 11996 }, { "epoch": 17.959580838323355, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1399, "step": 11997 }, { "epoch": 17.961077844311376, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1385, "step": 11998 }, { "epoch": 17.9625748502994, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1406, "step": 11999 }, { "epoch": 17.964071856287426, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.142, "step": 12000 }, { "epoch": 17.96556886227545, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1469, "step": 12001 }, { "epoch": 17.96706586826347, "grad_norm": 0.2314453125, "learning_rate": 0.0008, "loss": 1.1497, "step": 12002 }, { "epoch": 17.968562874251496, "grad_norm": 0.251953125, "learning_rate": 0.0008, "loss": 1.1484, "step": 12003 }, { "epoch": 17.97005988023952, "grad_norm": 0.337890625, "learning_rate": 0.0008, "loss": 1.1496, "step": 12004 }, { "epoch": 17.971556886227546, "grad_norm": 0.283203125, "learning_rate": 0.0008, "loss": 1.1488, "step": 12005 }, { "epoch": 17.973053892215567, "grad_norm": 0.2265625, "learning_rate": 0.0008, "loss": 1.1412, "step": 12006 }, { "epoch": 17.974550898203592, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1473, "step": 12007 }, { "epoch": 17.976047904191617, "grad_norm": 0.2197265625, "learning_rate": 0.0008, "loss": 1.1396, "step": 12008 }, { "epoch": 17.977544910179642, "grad_norm": 0.2265625, "learning_rate": 0.0008, "loss": 1.1439, "step": 12009 }, { "epoch": 17.979041916167663, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.142, "step": 12010 }, { "epoch": 17.980538922155688, "grad_norm": 0.2421875, "learning_rate": 0.0008, "loss": 1.1466, "step": 12011 }, { "epoch": 17.982035928143713, "grad_norm": 0.201171875, "learning_rate": 0.0008, "loss": 1.1401, "step": 12012 }, { "epoch": 17.983532934131738, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1367, "step": 12013 }, { "epoch": 17.98502994011976, "grad_norm": 0.189453125, "learning_rate": 0.0008, "loss": 1.1391, "step": 12014 }, { "epoch": 17.986526946107784, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1421, "step": 12015 }, { "epoch": 17.98802395209581, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1431, "step": 12016 }, { "epoch": 17.989520958083833, "grad_norm": 0.2060546875, "learning_rate": 0.0008, "loss": 1.1421, "step": 12017 }, { "epoch": 17.991017964071855, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1448, "step": 12018 }, { "epoch": 17.99251497005988, "grad_norm": 0.154296875, "learning_rate": 0.0008, "loss": 1.1355, "step": 12019 }, { "epoch": 17.994011976047904, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1412, "step": 12020 }, { "epoch": 17.99550898203593, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.1451, "step": 12021 }, { "epoch": 17.99700598802395, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1428, "step": 12022 }, { "epoch": 17.998502994011975, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1452, "step": 12023 }, { "epoch": 18.0, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1394, "step": 12024 }, { "epoch": 18.001497005988025, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1368, "step": 12025 }, { "epoch": 18.00299401197605, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1372, "step": 12026 }, { "epoch": 18.00449101796407, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1339, "step": 12027 }, { "epoch": 18.005988023952096, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1422, "step": 12028 }, { "epoch": 18.00748502994012, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1406, "step": 12029 }, { "epoch": 18.008982035928145, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1355, "step": 12030 }, { "epoch": 18.010479041916167, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1426, "step": 12031 }, { "epoch": 18.01197604790419, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.14, "step": 12032 }, { "epoch": 18.013473053892216, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1356, "step": 12033 }, { "epoch": 18.01497005988024, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1406, "step": 12034 }, { "epoch": 18.016467065868262, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1407, "step": 12035 }, { "epoch": 18.017964071856287, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1364, "step": 12036 }, { "epoch": 18.019461077844312, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.141, "step": 12037 }, { "epoch": 18.020958083832337, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1366, "step": 12038 }, { "epoch": 18.022455089820358, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1398, "step": 12039 }, { "epoch": 18.023952095808383, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.139, "step": 12040 }, { "epoch": 18.025449101796408, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.137, "step": 12041 }, { "epoch": 18.026946107784433, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1414, "step": 12042 }, { "epoch": 18.028443113772454, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1476, "step": 12043 }, { "epoch": 18.02994011976048, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1429, "step": 12044 }, { "epoch": 18.031437125748504, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1356, "step": 12045 }, { "epoch": 18.03293413173653, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1427, "step": 12046 }, { "epoch": 18.03443113772455, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1393, "step": 12047 }, { "epoch": 18.035928143712574, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1379, "step": 12048 }, { "epoch": 18.0374251497006, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1355, "step": 12049 }, { "epoch": 18.038922155688624, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1358, "step": 12050 }, { "epoch": 18.040419161676645, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1424, "step": 12051 }, { "epoch": 18.04191616766467, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1419, "step": 12052 }, { "epoch": 18.043413173652695, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1399, "step": 12053 }, { "epoch": 18.04491017964072, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1433, "step": 12054 }, { "epoch": 18.04640718562874, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1385, "step": 12055 }, { "epoch": 18.047904191616766, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1396, "step": 12056 }, { "epoch": 18.04940119760479, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.141, "step": 12057 }, { "epoch": 18.050898203592816, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1454, "step": 12058 }, { "epoch": 18.052395209580837, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1404, "step": 12059 }, { "epoch": 18.05389221556886, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1385, "step": 12060 }, { "epoch": 18.055389221556887, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1414, "step": 12061 }, { "epoch": 18.05688622754491, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1398, "step": 12062 }, { "epoch": 18.058383233532933, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1418, "step": 12063 }, { "epoch": 18.059880239520957, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1348, "step": 12064 }, { "epoch": 18.061377245508982, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1368, "step": 12065 }, { "epoch": 18.062874251497007, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1433, "step": 12066 }, { "epoch": 18.06437125748503, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1396, "step": 12067 }, { "epoch": 18.065868263473053, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1358, "step": 12068 }, { "epoch": 18.067365269461078, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1419, "step": 12069 }, { "epoch": 18.068862275449103, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1376, "step": 12070 }, { "epoch": 18.070359281437124, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1371, "step": 12071 }, { "epoch": 18.07185628742515, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1449, "step": 12072 }, { "epoch": 18.073353293413174, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1457, "step": 12073 }, { "epoch": 18.0748502994012, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1335, "step": 12074 }, { "epoch": 18.07634730538922, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1328, "step": 12075 }, { "epoch": 18.077844311377245, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1344, "step": 12076 }, { "epoch": 18.07934131736527, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1406, "step": 12077 }, { "epoch": 18.080838323353294, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1464, "step": 12078 }, { "epoch": 18.082335329341316, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.142, "step": 12079 }, { "epoch": 18.08383233532934, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1406, "step": 12080 }, { "epoch": 18.085329341317365, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1425, "step": 12081 }, { "epoch": 18.08682634730539, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1456, "step": 12082 }, { "epoch": 18.088323353293415, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1407, "step": 12083 }, { "epoch": 18.089820359281436, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1448, "step": 12084 }, { "epoch": 18.09131736526946, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1364, "step": 12085 }, { "epoch": 18.092814371257486, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.145, "step": 12086 }, { "epoch": 18.09431137724551, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1419, "step": 12087 }, { "epoch": 18.095808383233532, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1455, "step": 12088 }, { "epoch": 18.097305389221557, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1411, "step": 12089 }, { "epoch": 18.09880239520958, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1504, "step": 12090 }, { "epoch": 18.100299401197606, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1392, "step": 12091 }, { "epoch": 18.101796407185628, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1395, "step": 12092 }, { "epoch": 18.103293413173652, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1332, "step": 12093 }, { "epoch": 18.104790419161677, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1369, "step": 12094 }, { "epoch": 18.106287425149702, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1398, "step": 12095 }, { "epoch": 18.107784431137723, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1369, "step": 12096 }, { "epoch": 18.10928143712575, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1362, "step": 12097 }, { "epoch": 18.110778443113773, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1383, "step": 12098 }, { "epoch": 18.112275449101798, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1367, "step": 12099 }, { "epoch": 18.11377245508982, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1411, "step": 12100 }, { "epoch": 18.115269461077844, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1378, "step": 12101 }, { "epoch": 18.11676646706587, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1348, "step": 12102 }, { "epoch": 18.118263473053894, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1358, "step": 12103 }, { "epoch": 18.119760479041915, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1419, "step": 12104 }, { "epoch": 18.12125748502994, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1442, "step": 12105 }, { "epoch": 18.122754491017965, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.134, "step": 12106 }, { "epoch": 18.12425149700599, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1381, "step": 12107 }, { "epoch": 18.12574850299401, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1442, "step": 12108 }, { "epoch": 18.127245508982035, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1359, "step": 12109 }, { "epoch": 18.12874251497006, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1359, "step": 12110 }, { "epoch": 18.130239520958085, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.14, "step": 12111 }, { "epoch": 18.131736526946106, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1382, "step": 12112 }, { "epoch": 18.13323353293413, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.14, "step": 12113 }, { "epoch": 18.134730538922156, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1388, "step": 12114 }, { "epoch": 18.13622754491018, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1437, "step": 12115 }, { "epoch": 18.137724550898202, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1407, "step": 12116 }, { "epoch": 18.139221556886227, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1467, "step": 12117 }, { "epoch": 18.14071856287425, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1397, "step": 12118 }, { "epoch": 18.142215568862277, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.136, "step": 12119 }, { "epoch": 18.143712574850298, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1436, "step": 12120 }, { "epoch": 18.145209580838323, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1411, "step": 12121 }, { "epoch": 18.146706586826348, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1374, "step": 12122 }, { "epoch": 18.148203592814372, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1395, "step": 12123 }, { "epoch": 18.149700598802394, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1356, "step": 12124 }, { "epoch": 18.15119760479042, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1362, "step": 12125 }, { "epoch": 18.152694610778443, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.137, "step": 12126 }, { "epoch": 18.154191616766468, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1366, "step": 12127 }, { "epoch": 18.15568862275449, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1444, "step": 12128 }, { "epoch": 18.157185628742514, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1381, "step": 12129 }, { "epoch": 18.15868263473054, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1337, "step": 12130 }, { "epoch": 18.160179640718564, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1397, "step": 12131 }, { "epoch": 18.161676646706585, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.137, "step": 12132 }, { "epoch": 18.16317365269461, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1381, "step": 12133 }, { "epoch": 18.164670658682635, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1362, "step": 12134 }, { "epoch": 18.16616766467066, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1467, "step": 12135 }, { "epoch": 18.16766467065868, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1398, "step": 12136 }, { "epoch": 18.169161676646706, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1384, "step": 12137 }, { "epoch": 18.17065868263473, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1383, "step": 12138 }, { "epoch": 18.172155688622755, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1386, "step": 12139 }, { "epoch": 18.17365269461078, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1381, "step": 12140 }, { "epoch": 18.1751497005988, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1378, "step": 12141 }, { "epoch": 18.176646706586826, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.143, "step": 12142 }, { "epoch": 18.17814371257485, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1396, "step": 12143 }, { "epoch": 18.179640718562876, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1411, "step": 12144 }, { "epoch": 18.181137724550897, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1388, "step": 12145 }, { "epoch": 18.182634730538922, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1402, "step": 12146 }, { "epoch": 18.184131736526947, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1385, "step": 12147 }, { "epoch": 18.18562874251497, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1359, "step": 12148 }, { "epoch": 18.187125748502993, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1369, "step": 12149 }, { "epoch": 18.188622754491018, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1357, "step": 12150 }, { "epoch": 18.190119760479043, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1323, "step": 12151 }, { "epoch": 18.191616766467067, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1394, "step": 12152 }, { "epoch": 18.19311377245509, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1432, "step": 12153 }, { "epoch": 18.194610778443113, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1442, "step": 12154 }, { "epoch": 18.19610778443114, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1355, "step": 12155 }, { "epoch": 18.197604790419163, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.143, "step": 12156 }, { "epoch": 18.199101796407184, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1407, "step": 12157 }, { "epoch": 18.20059880239521, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1424, "step": 12158 }, { "epoch": 18.202095808383234, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1346, "step": 12159 }, { "epoch": 18.20359281437126, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1398, "step": 12160 }, { "epoch": 18.20508982035928, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1436, "step": 12161 }, { "epoch": 18.206586826347305, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1482, "step": 12162 }, { "epoch": 18.20808383233533, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1398, "step": 12163 }, { "epoch": 18.209580838323355, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1387, "step": 12164 }, { "epoch": 18.211077844311376, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1397, "step": 12165 }, { "epoch": 18.2125748502994, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1417, "step": 12166 }, { "epoch": 18.214071856287426, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1341, "step": 12167 }, { "epoch": 18.21556886227545, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1302, "step": 12168 }, { "epoch": 18.21706586826347, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1421, "step": 12169 }, { "epoch": 18.218562874251496, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1416, "step": 12170 }, { "epoch": 18.22005988023952, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1408, "step": 12171 }, { "epoch": 18.221556886227546, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1438, "step": 12172 }, { "epoch": 18.223053892215567, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.149, "step": 12173 }, { "epoch": 18.224550898203592, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1396, "step": 12174 }, { "epoch": 18.226047904191617, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1426, "step": 12175 }, { "epoch": 18.227544910179642, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1388, "step": 12176 }, { "epoch": 18.229041916167663, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1394, "step": 12177 }, { "epoch": 18.230538922155688, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1406, "step": 12178 }, { "epoch": 18.232035928143713, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.137, "step": 12179 }, { "epoch": 18.233532934131738, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1409, "step": 12180 }, { "epoch": 18.23502994011976, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1486, "step": 12181 }, { "epoch": 18.236526946107784, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1332, "step": 12182 }, { "epoch": 18.23802395209581, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1437, "step": 12183 }, { "epoch": 18.239520958083833, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1444, "step": 12184 }, { "epoch": 18.241017964071855, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1369, "step": 12185 }, { "epoch": 18.24251497005988, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1407, "step": 12186 }, { "epoch": 18.244011976047904, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.14, "step": 12187 }, { "epoch": 18.24550898203593, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1442, "step": 12188 }, { "epoch": 18.24700598802395, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1401, "step": 12189 }, { "epoch": 18.248502994011975, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1315, "step": 12190 }, { "epoch": 18.25, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1415, "step": 12191 }, { "epoch": 18.251497005988025, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.136, "step": 12192 }, { "epoch": 18.25299401197605, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1454, "step": 12193 }, { "epoch": 18.25449101796407, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1307, "step": 12194 }, { "epoch": 18.255988023952096, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1358, "step": 12195 }, { "epoch": 18.25748502994012, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1415, "step": 12196 }, { "epoch": 18.258982035928145, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1426, "step": 12197 }, { "epoch": 18.260479041916167, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1419, "step": 12198 }, { "epoch": 18.26197604790419, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1491, "step": 12199 }, { "epoch": 18.263473053892216, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1406, "step": 12200 }, { "epoch": 18.26497005988024, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1447, "step": 12201 }, { "epoch": 18.266467065868262, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1439, "step": 12202 }, { "epoch": 18.267964071856287, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1387, "step": 12203 }, { "epoch": 18.269461077844312, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1403, "step": 12204 }, { "epoch": 18.270958083832337, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1398, "step": 12205 }, { "epoch": 18.272455089820358, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1366, "step": 12206 }, { "epoch": 18.273952095808383, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1428, "step": 12207 }, { "epoch": 18.275449101796408, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1382, "step": 12208 }, { "epoch": 18.276946107784433, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1429, "step": 12209 }, { "epoch": 18.278443113772454, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1388, "step": 12210 }, { "epoch": 18.27994011976048, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1305, "step": 12211 }, { "epoch": 18.281437125748504, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1359, "step": 12212 }, { "epoch": 18.28293413173653, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1344, "step": 12213 }, { "epoch": 18.28443113772455, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1374, "step": 12214 }, { "epoch": 18.285928143712574, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1363, "step": 12215 }, { "epoch": 18.2874251497006, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1416, "step": 12216 }, { "epoch": 18.288922155688624, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1424, "step": 12217 }, { "epoch": 18.290419161676645, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1369, "step": 12218 }, { "epoch": 18.29191616766467, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1392, "step": 12219 }, { "epoch": 18.293413173652695, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1407, "step": 12220 }, { "epoch": 18.29491017964072, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1425, "step": 12221 }, { "epoch": 18.29640718562874, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1389, "step": 12222 }, { "epoch": 18.297904191616766, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1418, "step": 12223 }, { "epoch": 18.29940119760479, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1375, "step": 12224 }, { "epoch": 18.300898203592816, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1429, "step": 12225 }, { "epoch": 18.302395209580837, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1361, "step": 12226 }, { "epoch": 18.30389221556886, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1462, "step": 12227 }, { "epoch": 18.305389221556887, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.137, "step": 12228 }, { "epoch": 18.30688622754491, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.138, "step": 12229 }, { "epoch": 18.308383233532933, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1348, "step": 12230 }, { "epoch": 18.309880239520957, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1347, "step": 12231 }, { "epoch": 18.311377245508982, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1361, "step": 12232 }, { "epoch": 18.312874251497007, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.139, "step": 12233 }, { "epoch": 18.31437125748503, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1372, "step": 12234 }, { "epoch": 18.315868263473053, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1423, "step": 12235 }, { "epoch": 18.317365269461078, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1355, "step": 12236 }, { "epoch": 18.318862275449103, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1356, "step": 12237 }, { "epoch": 18.320359281437124, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1394, "step": 12238 }, { "epoch": 18.32185628742515, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1414, "step": 12239 }, { "epoch": 18.323353293413174, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1376, "step": 12240 }, { "epoch": 18.3248502994012, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1374, "step": 12241 }, { "epoch": 18.32634730538922, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1414, "step": 12242 }, { "epoch": 18.327844311377245, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.136, "step": 12243 }, { "epoch": 18.32934131736527, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.146, "step": 12244 }, { "epoch": 18.330838323353294, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1415, "step": 12245 }, { "epoch": 18.33233532934132, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1413, "step": 12246 }, { "epoch": 18.33383233532934, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1393, "step": 12247 }, { "epoch": 18.335329341317365, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1331, "step": 12248 }, { "epoch": 18.33682634730539, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1374, "step": 12249 }, { "epoch": 18.338323353293415, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1385, "step": 12250 }, { "epoch": 18.339820359281436, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.138, "step": 12251 }, { "epoch": 18.34131736526946, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1368, "step": 12252 }, { "epoch": 18.342814371257486, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1351, "step": 12253 }, { "epoch": 18.34431137724551, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1359, "step": 12254 }, { "epoch": 18.345808383233532, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1381, "step": 12255 }, { "epoch": 18.347305389221557, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1407, "step": 12256 }, { "epoch": 18.34880239520958, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1418, "step": 12257 }, { "epoch": 18.350299401197606, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1333, "step": 12258 }, { "epoch": 18.351796407185628, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1356, "step": 12259 }, { "epoch": 18.353293413173652, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1424, "step": 12260 }, { "epoch": 18.354790419161677, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1407, "step": 12261 }, { "epoch": 18.356287425149702, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1344, "step": 12262 }, { "epoch": 18.357784431137723, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1406, "step": 12263 }, { "epoch": 18.35928143712575, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1444, "step": 12264 }, { "epoch": 18.360778443113773, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1372, "step": 12265 }, { "epoch": 18.362275449101798, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1421, "step": 12266 }, { "epoch": 18.36377245508982, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.138, "step": 12267 }, { "epoch": 18.365269461077844, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.138, "step": 12268 }, { "epoch": 18.36676646706587, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1353, "step": 12269 }, { "epoch": 18.368263473053894, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1341, "step": 12270 }, { "epoch": 18.369760479041915, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1368, "step": 12271 }, { "epoch": 18.37125748502994, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1391, "step": 12272 }, { "epoch": 18.372754491017965, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1442, "step": 12273 }, { "epoch": 18.37425149700599, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.132, "step": 12274 }, { "epoch": 18.37574850299401, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1354, "step": 12275 }, { "epoch": 18.377245508982035, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1351, "step": 12276 }, { "epoch": 18.37874251497006, "grad_norm": 0.1591796875, "learning_rate": 0.0008, "loss": 1.1424, "step": 12277 }, { "epoch": 18.380239520958085, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1438, "step": 12278 }, { "epoch": 18.381736526946106, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1345, "step": 12279 }, { "epoch": 18.38323353293413, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1439, "step": 12280 }, { "epoch": 18.384730538922156, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1347, "step": 12281 }, { "epoch": 18.38622754491018, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1374, "step": 12282 }, { "epoch": 18.387724550898202, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1396, "step": 12283 }, { "epoch": 18.389221556886227, "grad_norm": 0.1640625, "learning_rate": 0.0008, "loss": 1.1306, "step": 12284 }, { "epoch": 18.39071856287425, "grad_norm": 0.12255859375, "learning_rate": 0.0008, "loss": 1.1385, "step": 12285 }, { "epoch": 18.392215568862277, "grad_norm": 0.1611328125, "learning_rate": 0.0008, "loss": 1.1415, "step": 12286 }, { "epoch": 18.393712574850298, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1379, "step": 12287 }, { "epoch": 18.395209580838323, "grad_norm": 0.1484375, "learning_rate": 0.0008, "loss": 1.1376, "step": 12288 }, { "epoch": 18.396706586826348, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.142, "step": 12289 }, { "epoch": 18.398203592814372, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1409, "step": 12290 }, { "epoch": 18.399700598802394, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1399, "step": 12291 }, { "epoch": 18.40119760479042, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1474, "step": 12292 }, { "epoch": 18.402694610778443, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1399, "step": 12293 }, { "epoch": 18.404191616766468, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.145, "step": 12294 }, { "epoch": 18.40568862275449, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1422, "step": 12295 }, { "epoch": 18.407185628742514, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.1342, "step": 12296 }, { "epoch": 18.40868263473054, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1411, "step": 12297 }, { "epoch": 18.410179640718564, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1404, "step": 12298 }, { "epoch": 18.411676646706585, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1435, "step": 12299 }, { "epoch": 18.41317365269461, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1361, "step": 12300 }, { "epoch": 18.414670658682635, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1332, "step": 12301 }, { "epoch": 18.41616766467066, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1384, "step": 12302 }, { "epoch": 18.41766467065868, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1423, "step": 12303 }, { "epoch": 18.419161676646706, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.136, "step": 12304 }, { "epoch": 18.42065868263473, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1423, "step": 12305 }, { "epoch": 18.422155688622755, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1435, "step": 12306 }, { "epoch": 18.42365269461078, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1339, "step": 12307 }, { "epoch": 18.4251497005988, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1395, "step": 12308 }, { "epoch": 18.426646706586826, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1447, "step": 12309 }, { "epoch": 18.42814371257485, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1365, "step": 12310 }, { "epoch": 18.429640718562876, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1436, "step": 12311 }, { "epoch": 18.431137724550897, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1375, "step": 12312 }, { "epoch": 18.432634730538922, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1393, "step": 12313 }, { "epoch": 18.434131736526947, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1382, "step": 12314 }, { "epoch": 18.43562874251497, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1334, "step": 12315 }, { "epoch": 18.437125748502993, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1314, "step": 12316 }, { "epoch": 18.438622754491018, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1438, "step": 12317 }, { "epoch": 18.440119760479043, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1406, "step": 12318 }, { "epoch": 18.441616766467067, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1376, "step": 12319 }, { "epoch": 18.44311377245509, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.14, "step": 12320 }, { "epoch": 18.444610778443113, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1371, "step": 12321 }, { "epoch": 18.44610778443114, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1313, "step": 12322 }, { "epoch": 18.447604790419163, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1365, "step": 12323 }, { "epoch": 18.449101796407184, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1386, "step": 12324 }, { "epoch": 18.45059880239521, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1348, "step": 12325 }, { "epoch": 18.452095808383234, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1323, "step": 12326 }, { "epoch": 18.45359281437126, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1431, "step": 12327 }, { "epoch": 18.45508982035928, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1334, "step": 12328 }, { "epoch": 18.456586826347305, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1384, "step": 12329 }, { "epoch": 18.45808383233533, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1387, "step": 12330 }, { "epoch": 18.459580838323355, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1429, "step": 12331 }, { "epoch": 18.461077844311376, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1315, "step": 12332 }, { "epoch": 18.4625748502994, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.133, "step": 12333 }, { "epoch": 18.464071856287426, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.137, "step": 12334 }, { "epoch": 18.46556886227545, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1418, "step": 12335 }, { "epoch": 18.46706586826347, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1409, "step": 12336 }, { "epoch": 18.468562874251496, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1393, "step": 12337 }, { "epoch": 18.47005988023952, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1481, "step": 12338 }, { "epoch": 18.471556886227546, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1356, "step": 12339 }, { "epoch": 18.473053892215567, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1347, "step": 12340 }, { "epoch": 18.474550898203592, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1446, "step": 12341 }, { "epoch": 18.476047904191617, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1411, "step": 12342 }, { "epoch": 18.477544910179642, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1351, "step": 12343 }, { "epoch": 18.479041916167663, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1333, "step": 12344 }, { "epoch": 18.480538922155688, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1407, "step": 12345 }, { "epoch": 18.482035928143713, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1368, "step": 12346 }, { "epoch": 18.483532934131738, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1384, "step": 12347 }, { "epoch": 18.48502994011976, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1417, "step": 12348 }, { "epoch": 18.486526946107784, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1377, "step": 12349 }, { "epoch": 18.48802395209581, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1351, "step": 12350 }, { "epoch": 18.489520958083833, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1387, "step": 12351 }, { "epoch": 18.491017964071855, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1399, "step": 12352 }, { "epoch": 18.49251497005988, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1389, "step": 12353 }, { "epoch": 18.494011976047904, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1404, "step": 12354 }, { "epoch": 18.49550898203593, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.137, "step": 12355 }, { "epoch": 18.49700598802395, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1412, "step": 12356 }, { "epoch": 18.498502994011975, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1407, "step": 12357 }, { "epoch": 18.5, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1447, "step": 12358 }, { "epoch": 18.501497005988025, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1414, "step": 12359 }, { "epoch": 18.50299401197605, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1506, "step": 12360 }, { "epoch": 18.50449101796407, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1373, "step": 12361 }, { "epoch": 18.505988023952096, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1368, "step": 12362 }, { "epoch": 18.50748502994012, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1385, "step": 12363 }, { "epoch": 18.508982035928145, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.142, "step": 12364 }, { "epoch": 18.510479041916167, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.134, "step": 12365 }, { "epoch": 18.51197604790419, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1456, "step": 12366 }, { "epoch": 18.513473053892216, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1398, "step": 12367 }, { "epoch": 18.51497005988024, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1355, "step": 12368 }, { "epoch": 18.516467065868262, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1408, "step": 12369 }, { "epoch": 18.517964071856287, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1432, "step": 12370 }, { "epoch": 18.519461077844312, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1357, "step": 12371 }, { "epoch": 18.520958083832337, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1435, "step": 12372 }, { "epoch": 18.522455089820358, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1382, "step": 12373 }, { "epoch": 18.523952095808383, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1357, "step": 12374 }, { "epoch": 18.525449101796408, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1407, "step": 12375 }, { "epoch": 18.526946107784433, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1403, "step": 12376 }, { "epoch": 18.528443113772454, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1371, "step": 12377 }, { "epoch": 18.52994011976048, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1381, "step": 12378 }, { "epoch": 18.531437125748504, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1358, "step": 12379 }, { "epoch": 18.53293413173653, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1362, "step": 12380 }, { "epoch": 18.53443113772455, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.14, "step": 12381 }, { "epoch": 18.535928143712574, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1373, "step": 12382 }, { "epoch": 18.5374251497006, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1433, "step": 12383 }, { "epoch": 18.538922155688624, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1377, "step": 12384 }, { "epoch": 18.540419161676645, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1343, "step": 12385 }, { "epoch": 18.54191616766467, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1456, "step": 12386 }, { "epoch": 18.543413173652695, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.136, "step": 12387 }, { "epoch": 18.54491017964072, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1373, "step": 12388 }, { "epoch": 18.54640718562874, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1329, "step": 12389 }, { "epoch": 18.547904191616766, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1401, "step": 12390 }, { "epoch": 18.54940119760479, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1395, "step": 12391 }, { "epoch": 18.550898203592816, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1347, "step": 12392 }, { "epoch": 18.552395209580837, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.137, "step": 12393 }, { "epoch": 18.55389221556886, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1462, "step": 12394 }, { "epoch": 18.555389221556887, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1389, "step": 12395 }, { "epoch": 18.55688622754491, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1419, "step": 12396 }, { "epoch": 18.558383233532933, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1436, "step": 12397 }, { "epoch": 18.559880239520957, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1374, "step": 12398 }, { "epoch": 18.561377245508982, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1377, "step": 12399 }, { "epoch": 18.562874251497007, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1364, "step": 12400 }, { "epoch": 18.56437125748503, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1379, "step": 12401 }, { "epoch": 18.565868263473053, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1425, "step": 12402 }, { "epoch": 18.567365269461078, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1423, "step": 12403 }, { "epoch": 18.568862275449103, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1378, "step": 12404 }, { "epoch": 18.570359281437124, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1354, "step": 12405 }, { "epoch": 18.57185628742515, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1381, "step": 12406 }, { "epoch": 18.573353293413174, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1465, "step": 12407 }, { "epoch": 18.5748502994012, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1293, "step": 12408 }, { "epoch": 18.57634730538922, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.137, "step": 12409 }, { "epoch": 18.577844311377245, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1396, "step": 12410 }, { "epoch": 18.57934131736527, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1351, "step": 12411 }, { "epoch": 18.580838323353294, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1423, "step": 12412 }, { "epoch": 18.58233532934132, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1408, "step": 12413 }, { "epoch": 18.58383233532934, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.138, "step": 12414 }, { "epoch": 18.585329341317365, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.133, "step": 12415 }, { "epoch": 18.58682634730539, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1381, "step": 12416 }, { "epoch": 18.58832335329341, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1402, "step": 12417 }, { "epoch": 18.589820359281436, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1353, "step": 12418 }, { "epoch": 18.59131736526946, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.142, "step": 12419 }, { "epoch": 18.592814371257486, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1389, "step": 12420 }, { "epoch": 18.59431137724551, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1394, "step": 12421 }, { "epoch": 18.595808383233532, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1412, "step": 12422 }, { "epoch": 18.597305389221557, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1323, "step": 12423 }, { "epoch": 18.59880239520958, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1358, "step": 12424 }, { "epoch": 18.600299401197606, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1399, "step": 12425 }, { "epoch": 18.601796407185628, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1423, "step": 12426 }, { "epoch": 18.603293413173652, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1402, "step": 12427 }, { "epoch": 18.604790419161677, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1319, "step": 12428 }, { "epoch": 18.606287425149702, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1383, "step": 12429 }, { "epoch": 18.607784431137723, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1396, "step": 12430 }, { "epoch": 18.60928143712575, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1355, "step": 12431 }, { "epoch": 18.610778443113773, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1307, "step": 12432 }, { "epoch": 18.612275449101798, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1423, "step": 12433 }, { "epoch": 18.61377245508982, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.142, "step": 12434 }, { "epoch": 18.615269461077844, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1443, "step": 12435 }, { "epoch": 18.61676646706587, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1384, "step": 12436 }, { "epoch": 18.618263473053894, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1384, "step": 12437 }, { "epoch": 18.619760479041915, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1414, "step": 12438 }, { "epoch": 18.62125748502994, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1402, "step": 12439 }, { "epoch": 18.622754491017965, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1334, "step": 12440 }, { "epoch": 18.62425149700599, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.136, "step": 12441 }, { "epoch": 18.62574850299401, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1399, "step": 12442 }, { "epoch": 18.627245508982035, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1422, "step": 12443 }, { "epoch": 18.62874251497006, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.138, "step": 12444 }, { "epoch": 18.630239520958085, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1391, "step": 12445 }, { "epoch": 18.631736526946106, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1332, "step": 12446 }, { "epoch": 18.63323353293413, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1333, "step": 12447 }, { "epoch": 18.634730538922156, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1396, "step": 12448 }, { "epoch": 18.63622754491018, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1355, "step": 12449 }, { "epoch": 18.637724550898202, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1399, "step": 12450 }, { "epoch": 18.639221556886227, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1432, "step": 12451 }, { "epoch": 18.64071856287425, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1403, "step": 12452 }, { "epoch": 18.642215568862277, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1366, "step": 12453 }, { "epoch": 18.643712574850298, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1429, "step": 12454 }, { "epoch": 18.645209580838323, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1398, "step": 12455 }, { "epoch": 18.646706586826348, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1444, "step": 12456 }, { "epoch": 18.648203592814372, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1387, "step": 12457 }, { "epoch": 18.649700598802394, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1345, "step": 12458 }, { "epoch": 18.65119760479042, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1365, "step": 12459 }, { "epoch": 18.652694610778443, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1364, "step": 12460 }, { "epoch": 18.654191616766468, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1405, "step": 12461 }, { "epoch": 18.65568862275449, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1408, "step": 12462 }, { "epoch": 18.657185628742514, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1325, "step": 12463 }, { "epoch": 18.65868263473054, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1365, "step": 12464 }, { "epoch": 18.660179640718564, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1378, "step": 12465 }, { "epoch": 18.66167664670659, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1362, "step": 12466 }, { "epoch": 18.66317365269461, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1363, "step": 12467 }, { "epoch": 18.664670658682635, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1317, "step": 12468 }, { "epoch": 18.66616766467066, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.142, "step": 12469 }, { "epoch": 18.66766467065868, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1445, "step": 12470 }, { "epoch": 18.669161676646706, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1374, "step": 12471 }, { "epoch": 18.67065868263473, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1423, "step": 12472 }, { "epoch": 18.672155688622755, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.134, "step": 12473 }, { "epoch": 18.67365269461078, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1393, "step": 12474 }, { "epoch": 18.6751497005988, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1383, "step": 12475 }, { "epoch": 18.676646706586826, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1358, "step": 12476 }, { "epoch": 18.67814371257485, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1314, "step": 12477 }, { "epoch": 18.679640718562876, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1412, "step": 12478 }, { "epoch": 18.681137724550897, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1409, "step": 12479 }, { "epoch": 18.682634730538922, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1294, "step": 12480 }, { "epoch": 18.684131736526947, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1412, "step": 12481 }, { "epoch": 18.68562874251497, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1401, "step": 12482 }, { "epoch": 18.687125748502993, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1393, "step": 12483 }, { "epoch": 18.688622754491018, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1392, "step": 12484 }, { "epoch": 18.690119760479043, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1333, "step": 12485 }, { "epoch": 18.691616766467067, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1422, "step": 12486 }, { "epoch": 18.69311377245509, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1368, "step": 12487 }, { "epoch": 18.694610778443113, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1369, "step": 12488 }, { "epoch": 18.69610778443114, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1398, "step": 12489 }, { "epoch": 18.697604790419163, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1375, "step": 12490 }, { "epoch": 18.699101796407184, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1333, "step": 12491 }, { "epoch": 18.70059880239521, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1364, "step": 12492 }, { "epoch": 18.702095808383234, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1363, "step": 12493 }, { "epoch": 18.70359281437126, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1358, "step": 12494 }, { "epoch": 18.70508982035928, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1399, "step": 12495 }, { "epoch": 18.706586826347305, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.142, "step": 12496 }, { "epoch": 18.70808383233533, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1386, "step": 12497 }, { "epoch": 18.709580838323355, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1385, "step": 12498 }, { "epoch": 18.711077844311376, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1372, "step": 12499 }, { "epoch": 18.7125748502994, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1452, "step": 12500 }, { "epoch": 18.714071856287426, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1319, "step": 12501 }, { "epoch": 18.71556886227545, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1358, "step": 12502 }, { "epoch": 18.71706586826347, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1414, "step": 12503 }, { "epoch": 18.718562874251496, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1378, "step": 12504 }, { "epoch": 18.72005988023952, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1425, "step": 12505 }, { "epoch": 18.721556886227546, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1495, "step": 12506 }, { "epoch": 18.723053892215567, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1402, "step": 12507 }, { "epoch": 18.724550898203592, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1364, "step": 12508 }, { "epoch": 18.726047904191617, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1412, "step": 12509 }, { "epoch": 18.727544910179642, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1415, "step": 12510 }, { "epoch": 18.729041916167663, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1382, "step": 12511 }, { "epoch": 18.730538922155688, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1422, "step": 12512 }, { "epoch": 18.732035928143713, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1397, "step": 12513 }, { "epoch": 18.733532934131738, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1359, "step": 12514 }, { "epoch": 18.73502994011976, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.14, "step": 12515 }, { "epoch": 18.736526946107784, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1386, "step": 12516 }, { "epoch": 18.73802395209581, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1329, "step": 12517 }, { "epoch": 18.739520958083833, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1421, "step": 12518 }, { "epoch": 18.741017964071855, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1397, "step": 12519 }, { "epoch": 18.74251497005988, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1448, "step": 12520 }, { "epoch": 18.744011976047904, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1403, "step": 12521 }, { "epoch": 18.74550898203593, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1392, "step": 12522 }, { "epoch": 18.74700598802395, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1343, "step": 12523 }, { "epoch": 18.748502994011975, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1431, "step": 12524 }, { "epoch": 18.75, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1365, "step": 12525 }, { "epoch": 18.751497005988025, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1433, "step": 12526 }, { "epoch": 18.75299401197605, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1388, "step": 12527 }, { "epoch": 18.75449101796407, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1375, "step": 12528 }, { "epoch": 18.755988023952096, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.142, "step": 12529 }, { "epoch": 18.75748502994012, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1375, "step": 12530 }, { "epoch": 18.758982035928145, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.142, "step": 12531 }, { "epoch": 18.760479041916167, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1347, "step": 12532 }, { "epoch": 18.76197604790419, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1423, "step": 12533 }, { "epoch": 18.763473053892216, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1374, "step": 12534 }, { "epoch": 18.76497005988024, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1405, "step": 12535 }, { "epoch": 18.766467065868262, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1347, "step": 12536 }, { "epoch": 18.767964071856287, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1387, "step": 12537 }, { "epoch": 18.769461077844312, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1353, "step": 12538 }, { "epoch": 18.770958083832337, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1399, "step": 12539 }, { "epoch": 18.772455089820358, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1359, "step": 12540 }, { "epoch": 18.773952095808383, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1422, "step": 12541 }, { "epoch": 18.775449101796408, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1371, "step": 12542 }, { "epoch": 18.776946107784433, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.143, "step": 12543 }, { "epoch": 18.778443113772454, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1412, "step": 12544 }, { "epoch": 18.77994011976048, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1382, "step": 12545 }, { "epoch": 18.781437125748504, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1366, "step": 12546 }, { "epoch": 18.78293413173653, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1395, "step": 12547 }, { "epoch": 18.78443113772455, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.14, "step": 12548 }, { "epoch": 18.785928143712574, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1396, "step": 12549 }, { "epoch": 18.7874251497006, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1478, "step": 12550 }, { "epoch": 18.788922155688624, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1444, "step": 12551 }, { "epoch": 18.790419161676645, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1365, "step": 12552 }, { "epoch": 18.79191616766467, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1336, "step": 12553 }, { "epoch": 18.793413173652695, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1391, "step": 12554 }, { "epoch": 18.79491017964072, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1373, "step": 12555 }, { "epoch": 18.79640718562874, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1368, "step": 12556 }, { "epoch": 18.797904191616766, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1387, "step": 12557 }, { "epoch": 18.79940119760479, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1384, "step": 12558 }, { "epoch": 18.800898203592816, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1445, "step": 12559 }, { "epoch": 18.802395209580837, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.141, "step": 12560 }, { "epoch": 18.80389221556886, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1441, "step": 12561 }, { "epoch": 18.805389221556887, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1445, "step": 12562 }, { "epoch": 18.80688622754491, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1362, "step": 12563 }, { "epoch": 18.808383233532933, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1352, "step": 12564 }, { "epoch": 18.809880239520957, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1488, "step": 12565 }, { "epoch": 18.811377245508982, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1394, "step": 12566 }, { "epoch": 18.812874251497007, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1359, "step": 12567 }, { "epoch": 18.81437125748503, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1345, "step": 12568 }, { "epoch": 18.815868263473053, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1317, "step": 12569 }, { "epoch": 18.817365269461078, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1396, "step": 12570 }, { "epoch": 18.818862275449103, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1334, "step": 12571 }, { "epoch": 18.820359281437124, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1398, "step": 12572 }, { "epoch": 18.82185628742515, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1361, "step": 12573 }, { "epoch": 18.823353293413174, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 12574 }, { "epoch": 18.8248502994012, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1392, "step": 12575 }, { "epoch": 18.82634730538922, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1364, "step": 12576 }, { "epoch": 18.827844311377245, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1426, "step": 12577 }, { "epoch": 18.82934131736527, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1369, "step": 12578 }, { "epoch": 18.830838323353294, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1391, "step": 12579 }, { "epoch": 18.83233532934132, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1406, "step": 12580 }, { "epoch": 18.83383233532934, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1459, "step": 12581 }, { "epoch": 18.835329341317365, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1379, "step": 12582 }, { "epoch": 18.83682634730539, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1346, "step": 12583 }, { "epoch": 18.83832335329341, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1383, "step": 12584 }, { "epoch": 18.839820359281436, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1327, "step": 12585 }, { "epoch": 18.84131736526946, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1327, "step": 12586 }, { "epoch": 18.842814371257486, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1444, "step": 12587 }, { "epoch": 18.84431137724551, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1335, "step": 12588 }, { "epoch": 18.845808383233532, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1431, "step": 12589 }, { "epoch": 18.847305389221557, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1292, "step": 12590 }, { "epoch": 18.84880239520958, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1429, "step": 12591 }, { "epoch": 18.850299401197606, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1395, "step": 12592 }, { "epoch": 18.851796407185628, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1348, "step": 12593 }, { "epoch": 18.853293413173652, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1337, "step": 12594 }, { "epoch": 18.854790419161677, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1456, "step": 12595 }, { "epoch": 18.856287425149702, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1366, "step": 12596 }, { "epoch": 18.857784431137723, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1374, "step": 12597 }, { "epoch": 18.85928143712575, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1385, "step": 12598 }, { "epoch": 18.860778443113773, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1398, "step": 12599 }, { "epoch": 18.862275449101798, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1336, "step": 12600 }, { "epoch": 18.86377245508982, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1402, "step": 12601 }, { "epoch": 18.865269461077844, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1313, "step": 12602 }, { "epoch": 18.86676646706587, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1413, "step": 12603 }, { "epoch": 18.868263473053894, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1344, "step": 12604 }, { "epoch": 18.869760479041915, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.132, "step": 12605 }, { "epoch": 18.87125748502994, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1423, "step": 12606 }, { "epoch": 18.872754491017965, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1418, "step": 12607 }, { "epoch": 18.87425149700599, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1396, "step": 12608 }, { "epoch": 18.87574850299401, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1354, "step": 12609 }, { "epoch": 18.877245508982035, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1359, "step": 12610 }, { "epoch": 18.87874251497006, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1353, "step": 12611 }, { "epoch": 18.880239520958085, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1385, "step": 12612 }, { "epoch": 18.881736526946106, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1357, "step": 12613 }, { "epoch": 18.88323353293413, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1461, "step": 12614 }, { "epoch": 18.884730538922156, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1348, "step": 12615 }, { "epoch": 18.88622754491018, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1371, "step": 12616 }, { "epoch": 18.887724550898202, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1408, "step": 12617 }, { "epoch": 18.889221556886227, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.133, "step": 12618 }, { "epoch": 18.89071856287425, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1378, "step": 12619 }, { "epoch": 18.892215568862277, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1448, "step": 12620 }, { "epoch": 18.893712574850298, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.139, "step": 12621 }, { "epoch": 18.895209580838323, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1378, "step": 12622 }, { "epoch": 18.896706586826348, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1385, "step": 12623 }, { "epoch": 18.898203592814372, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1362, "step": 12624 }, { "epoch": 18.899700598802394, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1354, "step": 12625 }, { "epoch": 18.90119760479042, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1411, "step": 12626 }, { "epoch": 18.902694610778443, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1399, "step": 12627 }, { "epoch": 18.904191616766468, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1296, "step": 12628 }, { "epoch": 18.90568862275449, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1343, "step": 12629 }, { "epoch": 18.907185628742514, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1392, "step": 12630 }, { "epoch": 18.90868263473054, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1379, "step": 12631 }, { "epoch": 18.910179640718564, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1377, "step": 12632 }, { "epoch": 18.91167664670659, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1409, "step": 12633 }, { "epoch": 18.91317365269461, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1373, "step": 12634 }, { "epoch": 18.914670658682635, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1351, "step": 12635 }, { "epoch": 18.91616766467066, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1368, "step": 12636 }, { "epoch": 18.91766467065868, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1377, "step": 12637 }, { "epoch": 18.919161676646706, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1439, "step": 12638 }, { "epoch": 18.92065868263473, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1394, "step": 12639 }, { "epoch": 18.922155688622755, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1383, "step": 12640 }, { "epoch": 18.92365269461078, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1384, "step": 12641 }, { "epoch": 18.9251497005988, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1425, "step": 12642 }, { "epoch": 18.926646706586826, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1379, "step": 12643 }, { "epoch": 18.92814371257485, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1407, "step": 12644 }, { "epoch": 18.929640718562876, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1399, "step": 12645 }, { "epoch": 18.931137724550897, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1368, "step": 12646 }, { "epoch": 18.932634730538922, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1345, "step": 12647 }, { "epoch": 18.934131736526947, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.144, "step": 12648 }, { "epoch": 18.93562874251497, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1379, "step": 12649 }, { "epoch": 18.937125748502993, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1399, "step": 12650 }, { "epoch": 18.938622754491018, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1402, "step": 12651 }, { "epoch": 18.940119760479043, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1366, "step": 12652 }, { "epoch": 18.941616766467067, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1398, "step": 12653 }, { "epoch": 18.94311377245509, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1438, "step": 12654 }, { "epoch": 18.944610778443113, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.135, "step": 12655 }, { "epoch": 18.94610778443114, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1381, "step": 12656 }, { "epoch": 18.947604790419163, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1388, "step": 12657 }, { "epoch": 18.949101796407184, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1348, "step": 12658 }, { "epoch": 18.95059880239521, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1378, "step": 12659 }, { "epoch": 18.952095808383234, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1403, "step": 12660 }, { "epoch": 18.95359281437126, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1392, "step": 12661 }, { "epoch": 18.95508982035928, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1376, "step": 12662 }, { "epoch": 18.956586826347305, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1406, "step": 12663 }, { "epoch": 18.95808383233533, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1374, "step": 12664 }, { "epoch": 18.959580838323355, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1342, "step": 12665 }, { "epoch": 18.961077844311376, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.137, "step": 12666 }, { "epoch": 18.9625748502994, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1344, "step": 12667 }, { "epoch": 18.964071856287426, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1369, "step": 12668 }, { "epoch": 18.96556886227545, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1376, "step": 12669 }, { "epoch": 18.96706586826347, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1397, "step": 12670 }, { "epoch": 18.968562874251496, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.14, "step": 12671 }, { "epoch": 18.97005988023952, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1435, "step": 12672 }, { "epoch": 18.971556886227546, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1317, "step": 12673 }, { "epoch": 18.973053892215567, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1408, "step": 12674 }, { "epoch": 18.974550898203592, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1361, "step": 12675 }, { "epoch": 18.976047904191617, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1359, "step": 12676 }, { "epoch": 18.977544910179642, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1315, "step": 12677 }, { "epoch": 18.979041916167663, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1398, "step": 12678 }, { "epoch": 18.980538922155688, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1362, "step": 12679 }, { "epoch": 18.982035928143713, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1385, "step": 12680 }, { "epoch": 18.983532934131738, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1338, "step": 12681 }, { "epoch": 18.98502994011976, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1391, "step": 12682 }, { "epoch": 18.986526946107784, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1404, "step": 12683 }, { "epoch": 18.98802395209581, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1378, "step": 12684 }, { "epoch": 18.989520958083833, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.14, "step": 12685 }, { "epoch": 18.991017964071855, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1371, "step": 12686 }, { "epoch": 18.99251497005988, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1408, "step": 12687 }, { "epoch": 18.994011976047904, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1363, "step": 12688 }, { "epoch": 18.99550898203593, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1381, "step": 12689 }, { "epoch": 18.99700598802395, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1354, "step": 12690 }, { "epoch": 18.998502994011975, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1448, "step": 12691 }, { "epoch": 19.0, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.136, "step": 12692 }, { "epoch": 19.001497005988025, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1364, "step": 12693 }, { "epoch": 19.00299401197605, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1417, "step": 12694 }, { "epoch": 19.00449101796407, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1391, "step": 12695 }, { "epoch": 19.005988023952096, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1348, "step": 12696 }, { "epoch": 19.00748502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1401, "step": 12697 }, { "epoch": 19.008982035928145, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1413, "step": 12698 }, { "epoch": 19.010479041916167, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1412, "step": 12699 }, { "epoch": 19.01197604790419, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1359, "step": 12700 }, { "epoch": 19.013473053892216, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1388, "step": 12701 }, { "epoch": 19.01497005988024, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1365, "step": 12702 }, { "epoch": 19.016467065868262, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1345, "step": 12703 }, { "epoch": 19.017964071856287, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1309, "step": 12704 }, { "epoch": 19.019461077844312, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1364, "step": 12705 }, { "epoch": 19.020958083832337, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1362, "step": 12706 }, { "epoch": 19.022455089820358, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1342, "step": 12707 }, { "epoch": 19.023952095808383, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1406, "step": 12708 }, { "epoch": 19.025449101796408, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1389, "step": 12709 }, { "epoch": 19.026946107784433, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1393, "step": 12710 }, { "epoch": 19.028443113772454, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1334, "step": 12711 }, { "epoch": 19.02994011976048, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1373, "step": 12712 }, { "epoch": 19.031437125748504, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1351, "step": 12713 }, { "epoch": 19.03293413173653, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1329, "step": 12714 }, { "epoch": 19.03443113772455, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1385, "step": 12715 }, { "epoch": 19.035928143712574, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1331, "step": 12716 }, { "epoch": 19.0374251497006, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1321, "step": 12717 }, { "epoch": 19.038922155688624, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.139, "step": 12718 }, { "epoch": 19.040419161676645, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1374, "step": 12719 }, { "epoch": 19.04191616766467, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1347, "step": 12720 }, { "epoch": 19.043413173652695, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1321, "step": 12721 }, { "epoch": 19.04491017964072, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1403, "step": 12722 }, { "epoch": 19.04640718562874, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1429, "step": 12723 }, { "epoch": 19.047904191616766, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1372, "step": 12724 }, { "epoch": 19.04940119760479, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1381, "step": 12725 }, { "epoch": 19.050898203592816, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1353, "step": 12726 }, { "epoch": 19.052395209580837, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1349, "step": 12727 }, { "epoch": 19.05389221556886, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1418, "step": 12728 }, { "epoch": 19.055389221556887, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1365, "step": 12729 }, { "epoch": 19.05688622754491, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1351, "step": 12730 }, { "epoch": 19.058383233532933, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1392, "step": 12731 }, { "epoch": 19.059880239520957, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1373, "step": 12732 }, { "epoch": 19.061377245508982, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.137, "step": 12733 }, { "epoch": 19.062874251497007, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1426, "step": 12734 }, { "epoch": 19.06437125748503, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1445, "step": 12735 }, { "epoch": 19.065868263473053, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1313, "step": 12736 }, { "epoch": 19.067365269461078, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1398, "step": 12737 }, { "epoch": 19.068862275449103, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1417, "step": 12738 }, { "epoch": 19.070359281437124, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1472, "step": 12739 }, { "epoch": 19.07185628742515, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1315, "step": 12740 }, { "epoch": 19.073353293413174, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1376, "step": 12741 }, { "epoch": 19.0748502994012, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1362, "step": 12742 }, { "epoch": 19.07634730538922, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1367, "step": 12743 }, { "epoch": 19.077844311377245, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.14, "step": 12744 }, { "epoch": 19.07934131736527, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.142, "step": 12745 }, { "epoch": 19.080838323353294, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.138, "step": 12746 }, { "epoch": 19.082335329341316, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1347, "step": 12747 }, { "epoch": 19.08383233532934, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1349, "step": 12748 }, { "epoch": 19.085329341317365, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1446, "step": 12749 }, { "epoch": 19.08682634730539, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1431, "step": 12750 }, { "epoch": 19.088323353293415, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1408, "step": 12751 }, { "epoch": 19.089820359281436, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1403, "step": 12752 }, { "epoch": 19.09131736526946, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1417, "step": 12753 }, { "epoch": 19.092814371257486, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1417, "step": 12754 }, { "epoch": 19.09431137724551, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1368, "step": 12755 }, { "epoch": 19.095808383233532, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1398, "step": 12756 }, { "epoch": 19.097305389221557, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1296, "step": 12757 }, { "epoch": 19.09880239520958, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1474, "step": 12758 }, { "epoch": 19.100299401197606, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1361, "step": 12759 }, { "epoch": 19.101796407185628, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1438, "step": 12760 }, { "epoch": 19.103293413173652, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1421, "step": 12761 }, { "epoch": 19.104790419161677, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.141, "step": 12762 }, { "epoch": 19.106287425149702, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1375, "step": 12763 }, { "epoch": 19.107784431137723, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.132, "step": 12764 }, { "epoch": 19.10928143712575, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1426, "step": 12765 }, { "epoch": 19.110778443113773, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1439, "step": 12766 }, { "epoch": 19.112275449101798, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1346, "step": 12767 }, { "epoch": 19.11377245508982, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1383, "step": 12768 }, { "epoch": 19.115269461077844, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.132, "step": 12769 }, { "epoch": 19.11676646706587, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1365, "step": 12770 }, { "epoch": 19.118263473053894, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1334, "step": 12771 }, { "epoch": 19.119760479041915, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1397, "step": 12772 }, { "epoch": 19.12125748502994, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1396, "step": 12773 }, { "epoch": 19.122754491017965, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1378, "step": 12774 }, { "epoch": 19.12425149700599, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1337, "step": 12775 }, { "epoch": 19.12574850299401, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1386, "step": 12776 }, { "epoch": 19.127245508982035, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1387, "step": 12777 }, { "epoch": 19.12874251497006, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1334, "step": 12778 }, { "epoch": 19.130239520958085, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.137, "step": 12779 }, { "epoch": 19.131736526946106, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1375, "step": 12780 }, { "epoch": 19.13323353293413, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1366, "step": 12781 }, { "epoch": 19.134730538922156, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.132, "step": 12782 }, { "epoch": 19.13622754491018, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1378, "step": 12783 }, { "epoch": 19.137724550898202, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1291, "step": 12784 }, { "epoch": 19.139221556886227, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1356, "step": 12785 }, { "epoch": 19.14071856287425, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1371, "step": 12786 }, { "epoch": 19.142215568862277, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1386, "step": 12787 }, { "epoch": 19.143712574850298, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1375, "step": 12788 }, { "epoch": 19.145209580838323, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1373, "step": 12789 }, { "epoch": 19.146706586826348, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1349, "step": 12790 }, { "epoch": 19.148203592814372, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1402, "step": 12791 }, { "epoch": 19.149700598802394, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1396, "step": 12792 }, { "epoch": 19.15119760479042, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1336, "step": 12793 }, { "epoch": 19.152694610778443, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1286, "step": 12794 }, { "epoch": 19.154191616766468, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1401, "step": 12795 }, { "epoch": 19.15568862275449, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1355, "step": 12796 }, { "epoch": 19.157185628742514, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1358, "step": 12797 }, { "epoch": 19.15868263473054, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1445, "step": 12798 }, { "epoch": 19.160179640718564, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1429, "step": 12799 }, { "epoch": 19.161676646706585, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1361, "step": 12800 }, { "epoch": 19.16317365269461, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1369, "step": 12801 }, { "epoch": 19.164670658682635, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.127, "step": 12802 }, { "epoch": 19.16616766467066, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1312, "step": 12803 }, { "epoch": 19.16766467065868, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1376, "step": 12804 }, { "epoch": 19.169161676646706, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1325, "step": 12805 }, { "epoch": 19.17065868263473, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1364, "step": 12806 }, { "epoch": 19.172155688622755, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.136, "step": 12807 }, { "epoch": 19.17365269461078, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1336, "step": 12808 }, { "epoch": 19.1751497005988, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1363, "step": 12809 }, { "epoch": 19.176646706586826, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1401, "step": 12810 }, { "epoch": 19.17814371257485, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1329, "step": 12811 }, { "epoch": 19.179640718562876, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1385, "step": 12812 }, { "epoch": 19.181137724550897, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1409, "step": 12813 }, { "epoch": 19.182634730538922, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1431, "step": 12814 }, { "epoch": 19.184131736526947, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1379, "step": 12815 }, { "epoch": 19.18562874251497, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1342, "step": 12816 }, { "epoch": 19.187125748502993, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1403, "step": 12817 }, { "epoch": 19.188622754491018, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1396, "step": 12818 }, { "epoch": 19.190119760479043, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1389, "step": 12819 }, { "epoch": 19.191616766467067, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1407, "step": 12820 }, { "epoch": 19.19311377245509, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1374, "step": 12821 }, { "epoch": 19.194610778443113, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1308, "step": 12822 }, { "epoch": 19.19610778443114, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1353, "step": 12823 }, { "epoch": 19.197604790419163, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1387, "step": 12824 }, { "epoch": 19.199101796407184, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1389, "step": 12825 }, { "epoch": 19.20059880239521, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1452, "step": 12826 }, { "epoch": 19.202095808383234, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1378, "step": 12827 }, { "epoch": 19.20359281437126, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1342, "step": 12828 }, { "epoch": 19.20508982035928, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1352, "step": 12829 }, { "epoch": 19.206586826347305, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1414, "step": 12830 }, { "epoch": 19.20808383233533, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1445, "step": 12831 }, { "epoch": 19.209580838323355, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1319, "step": 12832 }, { "epoch": 19.211077844311376, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1394, "step": 12833 }, { "epoch": 19.2125748502994, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1439, "step": 12834 }, { "epoch": 19.214071856287426, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1367, "step": 12835 }, { "epoch": 19.21556886227545, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1304, "step": 12836 }, { "epoch": 19.21706586826347, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1363, "step": 12837 }, { "epoch": 19.218562874251496, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1401, "step": 12838 }, { "epoch": 19.22005988023952, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1371, "step": 12839 }, { "epoch": 19.221556886227546, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1364, "step": 12840 }, { "epoch": 19.223053892215567, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1379, "step": 12841 }, { "epoch": 19.224550898203592, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1376, "step": 12842 }, { "epoch": 19.226047904191617, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1392, "step": 12843 }, { "epoch": 19.227544910179642, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1373, "step": 12844 }, { "epoch": 19.229041916167663, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1289, "step": 12845 }, { "epoch": 19.230538922155688, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1312, "step": 12846 }, { "epoch": 19.232035928143713, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1374, "step": 12847 }, { "epoch": 19.233532934131738, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1357, "step": 12848 }, { "epoch": 19.23502994011976, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1346, "step": 12849 }, { "epoch": 19.236526946107784, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1372, "step": 12850 }, { "epoch": 19.23802395209581, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1414, "step": 12851 }, { "epoch": 19.239520958083833, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1332, "step": 12852 }, { "epoch": 19.241017964071855, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.135, "step": 12853 }, { "epoch": 19.24251497005988, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1342, "step": 12854 }, { "epoch": 19.244011976047904, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1382, "step": 12855 }, { "epoch": 19.24550898203593, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1384, "step": 12856 }, { "epoch": 19.24700598802395, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1262, "step": 12857 }, { "epoch": 19.248502994011975, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1312, "step": 12858 }, { "epoch": 19.25, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1383, "step": 12859 }, { "epoch": 19.251497005988025, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1367, "step": 12860 }, { "epoch": 19.25299401197605, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1422, "step": 12861 }, { "epoch": 19.25449101796407, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1341, "step": 12862 }, { "epoch": 19.255988023952096, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1373, "step": 12863 }, { "epoch": 19.25748502994012, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1369, "step": 12864 }, { "epoch": 19.258982035928145, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1335, "step": 12865 }, { "epoch": 19.260479041916167, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.134, "step": 12866 }, { "epoch": 19.26197604790419, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1322, "step": 12867 }, { "epoch": 19.263473053892216, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1383, "step": 12868 }, { "epoch": 19.26497005988024, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1388, "step": 12869 }, { "epoch": 19.266467065868262, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1364, "step": 12870 }, { "epoch": 19.267964071856287, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.145, "step": 12871 }, { "epoch": 19.269461077844312, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1376, "step": 12872 }, { "epoch": 19.270958083832337, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.129, "step": 12873 }, { "epoch": 19.272455089820358, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1341, "step": 12874 }, { "epoch": 19.273952095808383, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1406, "step": 12875 }, { "epoch": 19.275449101796408, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1277, "step": 12876 }, { "epoch": 19.276946107784433, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1333, "step": 12877 }, { "epoch": 19.278443113772454, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.1434, "step": 12878 }, { "epoch": 19.27994011976048, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1386, "step": 12879 }, { "epoch": 19.281437125748504, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1316, "step": 12880 }, { "epoch": 19.28293413173653, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1361, "step": 12881 }, { "epoch": 19.28443113772455, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1371, "step": 12882 }, { "epoch": 19.285928143712574, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1359, "step": 12883 }, { "epoch": 19.2874251497006, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1382, "step": 12884 }, { "epoch": 19.288922155688624, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1421, "step": 12885 }, { "epoch": 19.290419161676645, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1377, "step": 12886 }, { "epoch": 19.29191616766467, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1357, "step": 12887 }, { "epoch": 19.293413173652695, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1362, "step": 12888 }, { "epoch": 19.29491017964072, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1331, "step": 12889 }, { "epoch": 19.29640718562874, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1463, "step": 12890 }, { "epoch": 19.297904191616766, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1434, "step": 12891 }, { "epoch": 19.29940119760479, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1402, "step": 12892 }, { "epoch": 19.300898203592816, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1392, "step": 12893 }, { "epoch": 19.302395209580837, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1415, "step": 12894 }, { "epoch": 19.30389221556886, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1362, "step": 12895 }, { "epoch": 19.305389221556887, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.141, "step": 12896 }, { "epoch": 19.30688622754491, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.134, "step": 12897 }, { "epoch": 19.308383233532933, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1454, "step": 12898 }, { "epoch": 19.309880239520957, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.134, "step": 12899 }, { "epoch": 19.311377245508982, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1376, "step": 12900 }, { "epoch": 19.312874251497007, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1446, "step": 12901 }, { "epoch": 19.31437125748503, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1372, "step": 12902 }, { "epoch": 19.315868263473053, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1331, "step": 12903 }, { "epoch": 19.317365269461078, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1394, "step": 12904 }, { "epoch": 19.318862275449103, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1404, "step": 12905 }, { "epoch": 19.320359281437124, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1394, "step": 12906 }, { "epoch": 19.32185628742515, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1396, "step": 12907 }, { "epoch": 19.323353293413174, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1411, "step": 12908 }, { "epoch": 19.3248502994012, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.136, "step": 12909 }, { "epoch": 19.32634730538922, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1378, "step": 12910 }, { "epoch": 19.327844311377245, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.133, "step": 12911 }, { "epoch": 19.32934131736527, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1387, "step": 12912 }, { "epoch": 19.330838323353294, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1266, "step": 12913 }, { "epoch": 19.33233532934132, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1405, "step": 12914 }, { "epoch": 19.33383233532934, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1359, "step": 12915 }, { "epoch": 19.335329341317365, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1397, "step": 12916 }, { "epoch": 19.33682634730539, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1341, "step": 12917 }, { "epoch": 19.338323353293415, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1312, "step": 12918 }, { "epoch": 19.339820359281436, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1379, "step": 12919 }, { "epoch": 19.34131736526946, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.137, "step": 12920 }, { "epoch": 19.342814371257486, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1371, "step": 12921 }, { "epoch": 19.34431137724551, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1387, "step": 12922 }, { "epoch": 19.345808383233532, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1388, "step": 12923 }, { "epoch": 19.347305389221557, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1377, "step": 12924 }, { "epoch": 19.34880239520958, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1404, "step": 12925 }, { "epoch": 19.350299401197606, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1408, "step": 12926 }, { "epoch": 19.351796407185628, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1364, "step": 12927 }, { "epoch": 19.353293413173652, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.141, "step": 12928 }, { "epoch": 19.354790419161677, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1349, "step": 12929 }, { "epoch": 19.356287425149702, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1401, "step": 12930 }, { "epoch": 19.357784431137723, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1373, "step": 12931 }, { "epoch": 19.35928143712575, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1389, "step": 12932 }, { "epoch": 19.360778443113773, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1371, "step": 12933 }, { "epoch": 19.362275449101798, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1369, "step": 12934 }, { "epoch": 19.36377245508982, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1368, "step": 12935 }, { "epoch": 19.365269461077844, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1336, "step": 12936 }, { "epoch": 19.36676646706587, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1336, "step": 12937 }, { "epoch": 19.368263473053894, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1431, "step": 12938 }, { "epoch": 19.369760479041915, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1378, "step": 12939 }, { "epoch": 19.37125748502994, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1369, "step": 12940 }, { "epoch": 19.372754491017965, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.139, "step": 12941 }, { "epoch": 19.37425149700599, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1417, "step": 12942 }, { "epoch": 19.37574850299401, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1356, "step": 12943 }, { "epoch": 19.377245508982035, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.135, "step": 12944 }, { "epoch": 19.37874251497006, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1343, "step": 12945 }, { "epoch": 19.380239520958085, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1367, "step": 12946 }, { "epoch": 19.381736526946106, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.141, "step": 12947 }, { "epoch": 19.38323353293413, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1461, "step": 12948 }, { "epoch": 19.384730538922156, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1404, "step": 12949 }, { "epoch": 19.38622754491018, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1311, "step": 12950 }, { "epoch": 19.387724550898202, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1437, "step": 12951 }, { "epoch": 19.389221556886227, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1384, "step": 12952 }, { "epoch": 19.39071856287425, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1408, "step": 12953 }, { "epoch": 19.392215568862277, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1393, "step": 12954 }, { "epoch": 19.393712574850298, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1405, "step": 12955 }, { "epoch": 19.395209580838323, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1389, "step": 12956 }, { "epoch": 19.396706586826348, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.134, "step": 12957 }, { "epoch": 19.398203592814372, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1331, "step": 12958 }, { "epoch": 19.399700598802394, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1368, "step": 12959 }, { "epoch": 19.40119760479042, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1328, "step": 12960 }, { "epoch": 19.402694610778443, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1458, "step": 12961 }, { "epoch": 19.404191616766468, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1418, "step": 12962 }, { "epoch": 19.40568862275449, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1361, "step": 12963 }, { "epoch": 19.407185628742514, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1343, "step": 12964 }, { "epoch": 19.40868263473054, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1425, "step": 12965 }, { "epoch": 19.410179640718564, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1323, "step": 12966 }, { "epoch": 19.411676646706585, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1365, "step": 12967 }, { "epoch": 19.41317365269461, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1372, "step": 12968 }, { "epoch": 19.414670658682635, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1469, "step": 12969 }, { "epoch": 19.41616766467066, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1364, "step": 12970 }, { "epoch": 19.41766467065868, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1402, "step": 12971 }, { "epoch": 19.419161676646706, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.137, "step": 12972 }, { "epoch": 19.42065868263473, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1302, "step": 12973 }, { "epoch": 19.422155688622755, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1397, "step": 12974 }, { "epoch": 19.42365269461078, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 12975 }, { "epoch": 19.4251497005988, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1404, "step": 12976 }, { "epoch": 19.426646706586826, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1344, "step": 12977 }, { "epoch": 19.42814371257485, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1354, "step": 12978 }, { "epoch": 19.429640718562876, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1338, "step": 12979 }, { "epoch": 19.431137724550897, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1398, "step": 12980 }, { "epoch": 19.432634730538922, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1393, "step": 12981 }, { "epoch": 19.434131736526947, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1395, "step": 12982 }, { "epoch": 19.43562874251497, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1431, "step": 12983 }, { "epoch": 19.437125748502993, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.14, "step": 12984 }, { "epoch": 19.438622754491018, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1405, "step": 12985 }, { "epoch": 19.440119760479043, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1347, "step": 12986 }, { "epoch": 19.441616766467067, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1354, "step": 12987 }, { "epoch": 19.44311377245509, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1404, "step": 12988 }, { "epoch": 19.444610778443113, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1392, "step": 12989 }, { "epoch": 19.44610778443114, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1336, "step": 12990 }, { "epoch": 19.447604790419163, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.131, "step": 12991 }, { "epoch": 19.449101796407184, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1363, "step": 12992 }, { "epoch": 19.45059880239521, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1356, "step": 12993 }, { "epoch": 19.452095808383234, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1414, "step": 12994 }, { "epoch": 19.45359281437126, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1437, "step": 12995 }, { "epoch": 19.45508982035928, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1347, "step": 12996 }, { "epoch": 19.456586826347305, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1349, "step": 12997 }, { "epoch": 19.45808383233533, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1314, "step": 12998 }, { "epoch": 19.459580838323355, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1362, "step": 12999 }, { "epoch": 19.461077844311376, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1371, "step": 13000 }, { "epoch": 19.4625748502994, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1384, "step": 13001 }, { "epoch": 19.464071856287426, "grad_norm": 0.1376953125, "learning_rate": 0.0008, "loss": 1.1414, "step": 13002 }, { "epoch": 19.46556886227545, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1386, "step": 13003 }, { "epoch": 19.46706586826347, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1371, "step": 13004 }, { "epoch": 19.468562874251496, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1314, "step": 13005 }, { "epoch": 19.47005988023952, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1388, "step": 13006 }, { "epoch": 19.471556886227546, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1409, "step": 13007 }, { "epoch": 19.473053892215567, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1328, "step": 13008 }, { "epoch": 19.474550898203592, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1416, "step": 13009 }, { "epoch": 19.476047904191617, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1387, "step": 13010 }, { "epoch": 19.477544910179642, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1391, "step": 13011 }, { "epoch": 19.479041916167663, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1372, "step": 13012 }, { "epoch": 19.480538922155688, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1303, "step": 13013 }, { "epoch": 19.482035928143713, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1347, "step": 13014 }, { "epoch": 19.483532934131738, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1417, "step": 13015 }, { "epoch": 19.48502994011976, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1365, "step": 13016 }, { "epoch": 19.486526946107784, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1371, "step": 13017 }, { "epoch": 19.48802395209581, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.135, "step": 13018 }, { "epoch": 19.489520958083833, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1366, "step": 13019 }, { "epoch": 19.491017964071855, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1293, "step": 13020 }, { "epoch": 19.49251497005988, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1392, "step": 13021 }, { "epoch": 19.494011976047904, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1396, "step": 13022 }, { "epoch": 19.49550898203593, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1334, "step": 13023 }, { "epoch": 19.49700598802395, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1369, "step": 13024 }, { "epoch": 19.498502994011975, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1428, "step": 13025 }, { "epoch": 19.5, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1361, "step": 13026 }, { "epoch": 19.501497005988025, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1398, "step": 13027 }, { "epoch": 19.50299401197605, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1323, "step": 13028 }, { "epoch": 19.50449101796407, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1376, "step": 13029 }, { "epoch": 19.505988023952096, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.14, "step": 13030 }, { "epoch": 19.50748502994012, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1315, "step": 13031 }, { "epoch": 19.508982035928145, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1286, "step": 13032 }, { "epoch": 19.510479041916167, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1345, "step": 13033 }, { "epoch": 19.51197604790419, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1339, "step": 13034 }, { "epoch": 19.513473053892216, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1382, "step": 13035 }, { "epoch": 19.51497005988024, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.142, "step": 13036 }, { "epoch": 19.516467065868262, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1397, "step": 13037 }, { "epoch": 19.517964071856287, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1405, "step": 13038 }, { "epoch": 19.519461077844312, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1409, "step": 13039 }, { "epoch": 19.520958083832337, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1329, "step": 13040 }, { "epoch": 19.522455089820358, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1374, "step": 13041 }, { "epoch": 19.523952095808383, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1364, "step": 13042 }, { "epoch": 19.525449101796408, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1451, "step": 13043 }, { "epoch": 19.526946107784433, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1428, "step": 13044 }, { "epoch": 19.528443113772454, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1289, "step": 13045 }, { "epoch": 19.52994011976048, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1357, "step": 13046 }, { "epoch": 19.531437125748504, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1376, "step": 13047 }, { "epoch": 19.53293413173653, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1375, "step": 13048 }, { "epoch": 19.53443113772455, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1363, "step": 13049 }, { "epoch": 19.535928143712574, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1339, "step": 13050 }, { "epoch": 19.5374251497006, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1401, "step": 13051 }, { "epoch": 19.538922155688624, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1399, "step": 13052 }, { "epoch": 19.540419161676645, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.142, "step": 13053 }, { "epoch": 19.54191616766467, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.135, "step": 13054 }, { "epoch": 19.543413173652695, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1332, "step": 13055 }, { "epoch": 19.54491017964072, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1392, "step": 13056 }, { "epoch": 19.54640718562874, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.144, "step": 13057 }, { "epoch": 19.547904191616766, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.131, "step": 13058 }, { "epoch": 19.54940119760479, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1367, "step": 13059 }, { "epoch": 19.550898203592816, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1378, "step": 13060 }, { "epoch": 19.552395209580837, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1435, "step": 13061 }, { "epoch": 19.55389221556886, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1412, "step": 13062 }, { "epoch": 19.555389221556887, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.134, "step": 13063 }, { "epoch": 19.55688622754491, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1332, "step": 13064 }, { "epoch": 19.558383233532933, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1368, "step": 13065 }, { "epoch": 19.559880239520957, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1399, "step": 13066 }, { "epoch": 19.561377245508982, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1386, "step": 13067 }, { "epoch": 19.562874251497007, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1381, "step": 13068 }, { "epoch": 19.56437125748503, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1338, "step": 13069 }, { "epoch": 19.565868263473053, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.138, "step": 13070 }, { "epoch": 19.567365269461078, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1324, "step": 13071 }, { "epoch": 19.568862275449103, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1394, "step": 13072 }, { "epoch": 19.570359281437124, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1372, "step": 13073 }, { "epoch": 19.57185628742515, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1348, "step": 13074 }, { "epoch": 19.573353293413174, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1313, "step": 13075 }, { "epoch": 19.5748502994012, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1395, "step": 13076 }, { "epoch": 19.57634730538922, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1307, "step": 13077 }, { "epoch": 19.577844311377245, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1398, "step": 13078 }, { "epoch": 19.57934131736527, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1406, "step": 13079 }, { "epoch": 19.580838323353294, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1377, "step": 13080 }, { "epoch": 19.58233532934132, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1352, "step": 13081 }, { "epoch": 19.58383233532934, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1322, "step": 13082 }, { "epoch": 19.585329341317365, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1382, "step": 13083 }, { "epoch": 19.58682634730539, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1369, "step": 13084 }, { "epoch": 19.58832335329341, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1363, "step": 13085 }, { "epoch": 19.589820359281436, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1335, "step": 13086 }, { "epoch": 19.59131736526946, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1336, "step": 13087 }, { "epoch": 19.592814371257486, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1319, "step": 13088 }, { "epoch": 19.59431137724551, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1388, "step": 13089 }, { "epoch": 19.595808383233532, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1364, "step": 13090 }, { "epoch": 19.597305389221557, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1378, "step": 13091 }, { "epoch": 19.59880239520958, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1351, "step": 13092 }, { "epoch": 19.600299401197606, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1323, "step": 13093 }, { "epoch": 19.601796407185628, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.133, "step": 13094 }, { "epoch": 19.603293413173652, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1384, "step": 13095 }, { "epoch": 19.604790419161677, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1414, "step": 13096 }, { "epoch": 19.606287425149702, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1336, "step": 13097 }, { "epoch": 19.607784431137723, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1338, "step": 13098 }, { "epoch": 19.60928143712575, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1389, "step": 13099 }, { "epoch": 19.610778443113773, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1384, "step": 13100 }, { "epoch": 19.612275449101798, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1397, "step": 13101 }, { "epoch": 19.61377245508982, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1385, "step": 13102 }, { "epoch": 19.615269461077844, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1403, "step": 13103 }, { "epoch": 19.61676646706587, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1367, "step": 13104 }, { "epoch": 19.618263473053894, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.137, "step": 13105 }, { "epoch": 19.619760479041915, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1338, "step": 13106 }, { "epoch": 19.62125748502994, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.138, "step": 13107 }, { "epoch": 19.622754491017965, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1362, "step": 13108 }, { "epoch": 19.62425149700599, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.14, "step": 13109 }, { "epoch": 19.62574850299401, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1359, "step": 13110 }, { "epoch": 19.627245508982035, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1366, "step": 13111 }, { "epoch": 19.62874251497006, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1337, "step": 13112 }, { "epoch": 19.630239520958085, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1354, "step": 13113 }, { "epoch": 19.631736526946106, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1344, "step": 13114 }, { "epoch": 19.63323353293413, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1383, "step": 13115 }, { "epoch": 19.634730538922156, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1382, "step": 13116 }, { "epoch": 19.63622754491018, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1414, "step": 13117 }, { "epoch": 19.637724550898202, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1421, "step": 13118 }, { "epoch": 19.639221556886227, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1353, "step": 13119 }, { "epoch": 19.64071856287425, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1397, "step": 13120 }, { "epoch": 19.642215568862277, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1454, "step": 13121 }, { "epoch": 19.643712574850298, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1369, "step": 13122 }, { "epoch": 19.645209580838323, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1436, "step": 13123 }, { "epoch": 19.646706586826348, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1349, "step": 13124 }, { "epoch": 19.648203592814372, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1333, "step": 13125 }, { "epoch": 19.649700598802394, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1368, "step": 13126 }, { "epoch": 19.65119760479042, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1352, "step": 13127 }, { "epoch": 19.652694610778443, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1369, "step": 13128 }, { "epoch": 19.654191616766468, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1363, "step": 13129 }, { "epoch": 19.65568862275449, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1405, "step": 13130 }, { "epoch": 19.657185628742514, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1376, "step": 13131 }, { "epoch": 19.65868263473054, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1422, "step": 13132 }, { "epoch": 19.660179640718564, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.139, "step": 13133 }, { "epoch": 19.66167664670659, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1342, "step": 13134 }, { "epoch": 19.66317365269461, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1422, "step": 13135 }, { "epoch": 19.664670658682635, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1434, "step": 13136 }, { "epoch": 19.66616766467066, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1429, "step": 13137 }, { "epoch": 19.66766467065868, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1374, "step": 13138 }, { "epoch": 19.669161676646706, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1414, "step": 13139 }, { "epoch": 19.67065868263473, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1332, "step": 13140 }, { "epoch": 19.672155688622755, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1354, "step": 13141 }, { "epoch": 19.67365269461078, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1424, "step": 13142 }, { "epoch": 19.6751497005988, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1389, "step": 13143 }, { "epoch": 19.676646706586826, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1352, "step": 13144 }, { "epoch": 19.67814371257485, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1304, "step": 13145 }, { "epoch": 19.679640718562876, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1484, "step": 13146 }, { "epoch": 19.681137724550897, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1421, "step": 13147 }, { "epoch": 19.682634730538922, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1358, "step": 13148 }, { "epoch": 19.684131736526947, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1361, "step": 13149 }, { "epoch": 19.68562874251497, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1381, "step": 13150 }, { "epoch": 19.687125748502993, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1438, "step": 13151 }, { "epoch": 19.688622754491018, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1327, "step": 13152 }, { "epoch": 19.690119760479043, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1439, "step": 13153 }, { "epoch": 19.691616766467067, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1361, "step": 13154 }, { "epoch": 19.69311377245509, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.139, "step": 13155 }, { "epoch": 19.694610778443113, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1366, "step": 13156 }, { "epoch": 19.69610778443114, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1409, "step": 13157 }, { "epoch": 19.697604790419163, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1325, "step": 13158 }, { "epoch": 19.699101796407184, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1409, "step": 13159 }, { "epoch": 19.70059880239521, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1439, "step": 13160 }, { "epoch": 19.702095808383234, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1374, "step": 13161 }, { "epoch": 19.70359281437126, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.14, "step": 13162 }, { "epoch": 19.70508982035928, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.134, "step": 13163 }, { "epoch": 19.706586826347305, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1345, "step": 13164 }, { "epoch": 19.70808383233533, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1325, "step": 13165 }, { "epoch": 19.709580838323355, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1413, "step": 13166 }, { "epoch": 19.711077844311376, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1456, "step": 13167 }, { "epoch": 19.7125748502994, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1321, "step": 13168 }, { "epoch": 19.714071856287426, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1437, "step": 13169 }, { "epoch": 19.71556886227545, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1365, "step": 13170 }, { "epoch": 19.71706586826347, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1412, "step": 13171 }, { "epoch": 19.718562874251496, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.138, "step": 13172 }, { "epoch": 19.72005988023952, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1352, "step": 13173 }, { "epoch": 19.721556886227546, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1365, "step": 13174 }, { "epoch": 19.723053892215567, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1375, "step": 13175 }, { "epoch": 19.724550898203592, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1353, "step": 13176 }, { "epoch": 19.726047904191617, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1293, "step": 13177 }, { "epoch": 19.727544910179642, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 13178 }, { "epoch": 19.729041916167663, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1343, "step": 13179 }, { "epoch": 19.730538922155688, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1362, "step": 13180 }, { "epoch": 19.732035928143713, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1388, "step": 13181 }, { "epoch": 19.733532934131738, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1326, "step": 13182 }, { "epoch": 19.73502994011976, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1403, "step": 13183 }, { "epoch": 19.736526946107784, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1373, "step": 13184 }, { "epoch": 19.73802395209581, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1411, "step": 13185 }, { "epoch": 19.739520958083833, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1368, "step": 13186 }, { "epoch": 19.741017964071855, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1356, "step": 13187 }, { "epoch": 19.74251497005988, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.137, "step": 13188 }, { "epoch": 19.744011976047904, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1366, "step": 13189 }, { "epoch": 19.74550898203593, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.138, "step": 13190 }, { "epoch": 19.74700598802395, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1361, "step": 13191 }, { "epoch": 19.748502994011975, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1332, "step": 13192 }, { "epoch": 19.75, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1381, "step": 13193 }, { "epoch": 19.751497005988025, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1345, "step": 13194 }, { "epoch": 19.75299401197605, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1383, "step": 13195 }, { "epoch": 19.75449101796407, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1385, "step": 13196 }, { "epoch": 19.755988023952096, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1381, "step": 13197 }, { "epoch": 19.75748502994012, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1398, "step": 13198 }, { "epoch": 19.758982035928145, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1463, "step": 13199 }, { "epoch": 19.760479041916167, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1422, "step": 13200 }, { "epoch": 19.76197604790419, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1364, "step": 13201 }, { "epoch": 19.763473053892216, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1395, "step": 13202 }, { "epoch": 19.76497005988024, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 13203 }, { "epoch": 19.766467065868262, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.137, "step": 13204 }, { "epoch": 19.767964071856287, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1366, "step": 13205 }, { "epoch": 19.769461077844312, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.149, "step": 13206 }, { "epoch": 19.770958083832337, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.135, "step": 13207 }, { "epoch": 19.772455089820358, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1329, "step": 13208 }, { "epoch": 19.773952095808383, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1394, "step": 13209 }, { "epoch": 19.775449101796408, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.139, "step": 13210 }, { "epoch": 19.776946107784433, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1387, "step": 13211 }, { "epoch": 19.778443113772454, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.138, "step": 13212 }, { "epoch": 19.77994011976048, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1374, "step": 13213 }, { "epoch": 19.781437125748504, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1308, "step": 13214 }, { "epoch": 19.78293413173653, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1376, "step": 13215 }, { "epoch": 19.78443113772455, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1304, "step": 13216 }, { "epoch": 19.785928143712574, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1373, "step": 13217 }, { "epoch": 19.7874251497006, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1445, "step": 13218 }, { "epoch": 19.788922155688624, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1369, "step": 13219 }, { "epoch": 19.790419161676645, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1343, "step": 13220 }, { "epoch": 19.79191616766467, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1307, "step": 13221 }, { "epoch": 19.793413173652695, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.136, "step": 13222 }, { "epoch": 19.79491017964072, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.136, "step": 13223 }, { "epoch": 19.79640718562874, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1431, "step": 13224 }, { "epoch": 19.797904191616766, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1393, "step": 13225 }, { "epoch": 19.79940119760479, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1386, "step": 13226 }, { "epoch": 19.800898203592816, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.141, "step": 13227 }, { "epoch": 19.802395209580837, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1394, "step": 13228 }, { "epoch": 19.80389221556886, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1372, "step": 13229 }, { "epoch": 19.805389221556887, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1375, "step": 13230 }, { "epoch": 19.80688622754491, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1396, "step": 13231 }, { "epoch": 19.808383233532933, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1336, "step": 13232 }, { "epoch": 19.809880239520957, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1383, "step": 13233 }, { "epoch": 19.811377245508982, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1389, "step": 13234 }, { "epoch": 19.812874251497007, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1372, "step": 13235 }, { "epoch": 19.81437125748503, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1342, "step": 13236 }, { "epoch": 19.815868263473053, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1375, "step": 13237 }, { "epoch": 19.817365269461078, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1339, "step": 13238 }, { "epoch": 19.818862275449103, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1433, "step": 13239 }, { "epoch": 19.820359281437124, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1327, "step": 13240 }, { "epoch": 19.82185628742515, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1367, "step": 13241 }, { "epoch": 19.823353293413174, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1424, "step": 13242 }, { "epoch": 19.8248502994012, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1407, "step": 13243 }, { "epoch": 19.82634730538922, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1347, "step": 13244 }, { "epoch": 19.827844311377245, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1388, "step": 13245 }, { "epoch": 19.82934131736527, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1415, "step": 13246 }, { "epoch": 19.830838323353294, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1388, "step": 13247 }, { "epoch": 19.83233532934132, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1412, "step": 13248 }, { "epoch": 19.83383233532934, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1374, "step": 13249 }, { "epoch": 19.835329341317365, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1319, "step": 13250 }, { "epoch": 19.83682634730539, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1387, "step": 13251 }, { "epoch": 19.83832335329341, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.149, "step": 13252 }, { "epoch": 19.839820359281436, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1374, "step": 13253 }, { "epoch": 19.84131736526946, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 13254 }, { "epoch": 19.842814371257486, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1417, "step": 13255 }, { "epoch": 19.84431137724551, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1323, "step": 13256 }, { "epoch": 19.845808383233532, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1396, "step": 13257 }, { "epoch": 19.847305389221557, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1384, "step": 13258 }, { "epoch": 19.84880239520958, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1364, "step": 13259 }, { "epoch": 19.850299401197606, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1384, "step": 13260 }, { "epoch": 19.851796407185628, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.15, "step": 13261 }, { "epoch": 19.853293413173652, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.129, "step": 13262 }, { "epoch": 19.854790419161677, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1386, "step": 13263 }, { "epoch": 19.856287425149702, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1361, "step": 13264 }, { "epoch": 19.857784431137723, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1395, "step": 13265 }, { "epoch": 19.85928143712575, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1353, "step": 13266 }, { "epoch": 19.860778443113773, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1348, "step": 13267 }, { "epoch": 19.862275449101798, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1404, "step": 13268 }, { "epoch": 19.86377245508982, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1345, "step": 13269 }, { "epoch": 19.865269461077844, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1369, "step": 13270 }, { "epoch": 19.86676646706587, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1365, "step": 13271 }, { "epoch": 19.868263473053894, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1321, "step": 13272 }, { "epoch": 19.869760479041915, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1441, "step": 13273 }, { "epoch": 19.87125748502994, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1387, "step": 13274 }, { "epoch": 19.872754491017965, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1398, "step": 13275 }, { "epoch": 19.87425149700599, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1359, "step": 13276 }, { "epoch": 19.87574850299401, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1424, "step": 13277 }, { "epoch": 19.877245508982035, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1375, "step": 13278 }, { "epoch": 19.87874251497006, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1375, "step": 13279 }, { "epoch": 19.880239520958085, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1321, "step": 13280 }, { "epoch": 19.881736526946106, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1432, "step": 13281 }, { "epoch": 19.88323353293413, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1359, "step": 13282 }, { "epoch": 19.884730538922156, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1336, "step": 13283 }, { "epoch": 19.88622754491018, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1385, "step": 13284 }, { "epoch": 19.887724550898202, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.137, "step": 13285 }, { "epoch": 19.889221556886227, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1368, "step": 13286 }, { "epoch": 19.89071856287425, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1391, "step": 13287 }, { "epoch": 19.892215568862277, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1374, "step": 13288 }, { "epoch": 19.893712574850298, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1328, "step": 13289 }, { "epoch": 19.895209580838323, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1357, "step": 13290 }, { "epoch": 19.896706586826348, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1327, "step": 13291 }, { "epoch": 19.898203592814372, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.128, "step": 13292 }, { "epoch": 19.899700598802394, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1305, "step": 13293 }, { "epoch": 19.90119760479042, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1411, "step": 13294 }, { "epoch": 19.902694610778443, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1384, "step": 13295 }, { "epoch": 19.904191616766468, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1366, "step": 13296 }, { "epoch": 19.90568862275449, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1394, "step": 13297 }, { "epoch": 19.907185628742514, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1377, "step": 13298 }, { "epoch": 19.90868263473054, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1416, "step": 13299 }, { "epoch": 19.910179640718564, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1367, "step": 13300 }, { "epoch": 19.91167664670659, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1344, "step": 13301 }, { "epoch": 19.91317365269461, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1384, "step": 13302 }, { "epoch": 19.914670658682635, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1296, "step": 13303 }, { "epoch": 19.91616766467066, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1393, "step": 13304 }, { "epoch": 19.91766467065868, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1474, "step": 13305 }, { "epoch": 19.919161676646706, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1326, "step": 13306 }, { "epoch": 19.92065868263473, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.137, "step": 13307 }, { "epoch": 19.922155688622755, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1404, "step": 13308 }, { "epoch": 19.92365269461078, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.132, "step": 13309 }, { "epoch": 19.9251497005988, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1394, "step": 13310 }, { "epoch": 19.926646706586826, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1398, "step": 13311 }, { "epoch": 19.92814371257485, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1283, "step": 13312 }, { "epoch": 19.929640718562876, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1308, "step": 13313 }, { "epoch": 19.931137724550897, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1339, "step": 13314 }, { "epoch": 19.932634730538922, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.137, "step": 13315 }, { "epoch": 19.934131736526947, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1347, "step": 13316 }, { "epoch": 19.93562874251497, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1388, "step": 13317 }, { "epoch": 19.937125748502993, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1367, "step": 13318 }, { "epoch": 19.938622754491018, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1436, "step": 13319 }, { "epoch": 19.940119760479043, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1293, "step": 13320 }, { "epoch": 19.941616766467067, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1377, "step": 13321 }, { "epoch": 19.94311377245509, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.133, "step": 13322 }, { "epoch": 19.944610778443113, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1379, "step": 13323 }, { "epoch": 19.94610778443114, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1369, "step": 13324 }, { "epoch": 19.947604790419163, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1407, "step": 13325 }, { "epoch": 19.949101796407184, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1352, "step": 13326 }, { "epoch": 19.95059880239521, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1368, "step": 13327 }, { "epoch": 19.952095808383234, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1325, "step": 13328 }, { "epoch": 19.95359281437126, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.138, "step": 13329 }, { "epoch": 19.95508982035928, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1353, "step": 13330 }, { "epoch": 19.956586826347305, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1379, "step": 13331 }, { "epoch": 19.95808383233533, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1393, "step": 13332 }, { "epoch": 19.959580838323355, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.139, "step": 13333 }, { "epoch": 19.961077844311376, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1379, "step": 13334 }, { "epoch": 19.9625748502994, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1342, "step": 13335 }, { "epoch": 19.964071856287426, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1349, "step": 13336 }, { "epoch": 19.96556886227545, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1288, "step": 13337 }, { "epoch": 19.96706586826347, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.136, "step": 13338 }, { "epoch": 19.968562874251496, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1323, "step": 13339 }, { "epoch": 19.97005988023952, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1365, "step": 13340 }, { "epoch": 19.971556886227546, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1409, "step": 13341 }, { "epoch": 19.973053892215567, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1401, "step": 13342 }, { "epoch": 19.974550898203592, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1369, "step": 13343 }, { "epoch": 19.976047904191617, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1358, "step": 13344 }, { "epoch": 19.977544910179642, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1387, "step": 13345 }, { "epoch": 19.979041916167663, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1391, "step": 13346 }, { "epoch": 19.980538922155688, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1385, "step": 13347 }, { "epoch": 19.982035928143713, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1384, "step": 13348 }, { "epoch": 19.983532934131738, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1372, "step": 13349 }, { "epoch": 19.98502994011976, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1265, "step": 13350 }, { "epoch": 19.986526946107784, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1384, "step": 13351 }, { "epoch": 19.98802395209581, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1367, "step": 13352 }, { "epoch": 19.989520958083833, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.137, "step": 13353 }, { "epoch": 19.991017964071855, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1381, "step": 13354 }, { "epoch": 19.99251497005988, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1375, "step": 13355 }, { "epoch": 19.994011976047904, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1342, "step": 13356 }, { "epoch": 19.99550898203593, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1384, "step": 13357 }, { "epoch": 19.99700598802395, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.142, "step": 13358 }, { "epoch": 19.998502994011975, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1367, "step": 13359 }, { "epoch": 20.0, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1418, "step": 13360 }, { "epoch": 20.001497005988025, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1359, "step": 13361 }, { "epoch": 20.00299401197605, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1374, "step": 13362 }, { "epoch": 20.00449101796407, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1342, "step": 13363 }, { "epoch": 20.005988023952096, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1338, "step": 13364 }, { "epoch": 20.00748502994012, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1325, "step": 13365 }, { "epoch": 20.008982035928145, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1362, "step": 13366 }, { "epoch": 20.010479041916167, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1339, "step": 13367 }, { "epoch": 20.01197604790419, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1364, "step": 13368 }, { "epoch": 20.013473053892216, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1396, "step": 13369 }, { "epoch": 20.01497005988024, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1278, "step": 13370 }, { "epoch": 20.016467065868262, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1351, "step": 13371 }, { "epoch": 20.017964071856287, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1417, "step": 13372 }, { "epoch": 20.019461077844312, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1354, "step": 13373 }, { "epoch": 20.020958083832337, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1314, "step": 13374 }, { "epoch": 20.022455089820358, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1337, "step": 13375 }, { "epoch": 20.023952095808383, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1379, "step": 13376 }, { "epoch": 20.025449101796408, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1341, "step": 13377 }, { "epoch": 20.026946107784433, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1353, "step": 13378 }, { "epoch": 20.028443113772454, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1315, "step": 13379 }, { "epoch": 20.02994011976048, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1369, "step": 13380 }, { "epoch": 20.031437125748504, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.14, "step": 13381 }, { "epoch": 20.03293413173653, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.131, "step": 13382 }, { "epoch": 20.03443113772455, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1349, "step": 13383 }, { "epoch": 20.035928143712574, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1341, "step": 13384 }, { "epoch": 20.0374251497006, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1401, "step": 13385 }, { "epoch": 20.038922155688624, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1409, "step": 13386 }, { "epoch": 20.040419161676645, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1353, "step": 13387 }, { "epoch": 20.04191616766467, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1313, "step": 13388 }, { "epoch": 20.043413173652695, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1376, "step": 13389 }, { "epoch": 20.04491017964072, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1474, "step": 13390 }, { "epoch": 20.04640718562874, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1397, "step": 13391 }, { "epoch": 20.047904191616766, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1384, "step": 13392 }, { "epoch": 20.04940119760479, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1407, "step": 13393 }, { "epoch": 20.050898203592816, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1336, "step": 13394 }, { "epoch": 20.052395209580837, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1431, "step": 13395 }, { "epoch": 20.05389221556886, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1386, "step": 13396 }, { "epoch": 20.055389221556887, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1329, "step": 13397 }, { "epoch": 20.05688622754491, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1415, "step": 13398 }, { "epoch": 20.058383233532933, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1342, "step": 13399 }, { "epoch": 20.059880239520957, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1306, "step": 13400 }, { "epoch": 20.061377245508982, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1368, "step": 13401 }, { "epoch": 20.062874251497007, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1381, "step": 13402 }, { "epoch": 20.06437125748503, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1321, "step": 13403 }, { "epoch": 20.065868263473053, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.134, "step": 13404 }, { "epoch": 20.067365269461078, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1347, "step": 13405 }, { "epoch": 20.068862275449103, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.136, "step": 13406 }, { "epoch": 20.070359281437124, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1406, "step": 13407 }, { "epoch": 20.07185628742515, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.144, "step": 13408 }, { "epoch": 20.073353293413174, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1288, "step": 13409 }, { "epoch": 20.0748502994012, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1472, "step": 13410 }, { "epoch": 20.07634730538922, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1329, "step": 13411 }, { "epoch": 20.077844311377245, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1337, "step": 13412 }, { "epoch": 20.07934131736527, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1386, "step": 13413 }, { "epoch": 20.080838323353294, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1366, "step": 13414 }, { "epoch": 20.082335329341316, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1325, "step": 13415 }, { "epoch": 20.08383233532934, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1373, "step": 13416 }, { "epoch": 20.085329341317365, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1358, "step": 13417 }, { "epoch": 20.08682634730539, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1467, "step": 13418 }, { "epoch": 20.088323353293415, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1367, "step": 13419 }, { "epoch": 20.089820359281436, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1331, "step": 13420 }, { "epoch": 20.09131736526946, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1378, "step": 13421 }, { "epoch": 20.092814371257486, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.133, "step": 13422 }, { "epoch": 20.09431137724551, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1353, "step": 13423 }, { "epoch": 20.095808383233532, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1367, "step": 13424 }, { "epoch": 20.097305389221557, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1389, "step": 13425 }, { "epoch": 20.09880239520958, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1284, "step": 13426 }, { "epoch": 20.100299401197606, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1406, "step": 13427 }, { "epoch": 20.101796407185628, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1389, "step": 13428 }, { "epoch": 20.103293413173652, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1352, "step": 13429 }, { "epoch": 20.104790419161677, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1323, "step": 13430 }, { "epoch": 20.106287425149702, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1361, "step": 13431 }, { "epoch": 20.107784431137723, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1414, "step": 13432 }, { "epoch": 20.10928143712575, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1338, "step": 13433 }, { "epoch": 20.110778443113773, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1415, "step": 13434 }, { "epoch": 20.112275449101798, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1456, "step": 13435 }, { "epoch": 20.11377245508982, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1323, "step": 13436 }, { "epoch": 20.115269461077844, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1391, "step": 13437 }, { "epoch": 20.11676646706587, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.139, "step": 13438 }, { "epoch": 20.118263473053894, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1387, "step": 13439 }, { "epoch": 20.119760479041915, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1347, "step": 13440 }, { "epoch": 20.12125748502994, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1324, "step": 13441 }, { "epoch": 20.122754491017965, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1327, "step": 13442 }, { "epoch": 20.12425149700599, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1313, "step": 13443 }, { "epoch": 20.12574850299401, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1349, "step": 13444 }, { "epoch": 20.127245508982035, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1367, "step": 13445 }, { "epoch": 20.12874251497006, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1276, "step": 13446 }, { "epoch": 20.130239520958085, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.133, "step": 13447 }, { "epoch": 20.131736526946106, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.142, "step": 13448 }, { "epoch": 20.13323353293413, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1377, "step": 13449 }, { "epoch": 20.134730538922156, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1342, "step": 13450 }, { "epoch": 20.13622754491018, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1383, "step": 13451 }, { "epoch": 20.137724550898202, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1353, "step": 13452 }, { "epoch": 20.139221556886227, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1373, "step": 13453 }, { "epoch": 20.14071856287425, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.138, "step": 13454 }, { "epoch": 20.142215568862277, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1352, "step": 13455 }, { "epoch": 20.143712574850298, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1364, "step": 13456 }, { "epoch": 20.145209580838323, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1404, "step": 13457 }, { "epoch": 20.146706586826348, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.135, "step": 13458 }, { "epoch": 20.148203592814372, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1347, "step": 13459 }, { "epoch": 20.149700598802394, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1407, "step": 13460 }, { "epoch": 20.15119760479042, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1354, "step": 13461 }, { "epoch": 20.152694610778443, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1369, "step": 13462 }, { "epoch": 20.154191616766468, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1357, "step": 13463 }, { "epoch": 20.15568862275449, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1411, "step": 13464 }, { "epoch": 20.157185628742514, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1359, "step": 13465 }, { "epoch": 20.15868263473054, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1323, "step": 13466 }, { "epoch": 20.160179640718564, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1331, "step": 13467 }, { "epoch": 20.161676646706585, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1371, "step": 13468 }, { "epoch": 20.16317365269461, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1406, "step": 13469 }, { "epoch": 20.164670658682635, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1359, "step": 13470 }, { "epoch": 20.16616766467066, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1335, "step": 13471 }, { "epoch": 20.16766467065868, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1319, "step": 13472 }, { "epoch": 20.169161676646706, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1367, "step": 13473 }, { "epoch": 20.17065868263473, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1386, "step": 13474 }, { "epoch": 20.172155688622755, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1367, "step": 13475 }, { "epoch": 20.17365269461078, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1362, "step": 13476 }, { "epoch": 20.1751497005988, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1301, "step": 13477 }, { "epoch": 20.176646706586826, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1391, "step": 13478 }, { "epoch": 20.17814371257485, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1299, "step": 13479 }, { "epoch": 20.179640718562876, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1387, "step": 13480 }, { "epoch": 20.181137724550897, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1342, "step": 13481 }, { "epoch": 20.182634730538922, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1383, "step": 13482 }, { "epoch": 20.184131736526947, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1389, "step": 13483 }, { "epoch": 20.18562874251497, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1387, "step": 13484 }, { "epoch": 20.187125748502993, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1333, "step": 13485 }, { "epoch": 20.188622754491018, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1365, "step": 13486 }, { "epoch": 20.190119760479043, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1403, "step": 13487 }, { "epoch": 20.191616766467067, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1386, "step": 13488 }, { "epoch": 20.19311377245509, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1469, "step": 13489 }, { "epoch": 20.194610778443113, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1374, "step": 13490 }, { "epoch": 20.19610778443114, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1381, "step": 13491 }, { "epoch": 20.197604790419163, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.137, "step": 13492 }, { "epoch": 20.199101796407184, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1362, "step": 13493 }, { "epoch": 20.20059880239521, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1352, "step": 13494 }, { "epoch": 20.202095808383234, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1335, "step": 13495 }, { "epoch": 20.20359281437126, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1348, "step": 13496 }, { "epoch": 20.20508982035928, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.139, "step": 13497 }, { "epoch": 20.206586826347305, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1381, "step": 13498 }, { "epoch": 20.20808383233533, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1397, "step": 13499 }, { "epoch": 20.209580838323355, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1425, "step": 13500 }, { "epoch": 20.211077844311376, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1469, "step": 13501 }, { "epoch": 20.2125748502994, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1368, "step": 13502 }, { "epoch": 20.214071856287426, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1348, "step": 13503 }, { "epoch": 20.21556886227545, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1368, "step": 13504 }, { "epoch": 20.21706586826347, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1395, "step": 13505 }, { "epoch": 20.218562874251496, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1354, "step": 13506 }, { "epoch": 20.22005988023952, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1358, "step": 13507 }, { "epoch": 20.221556886227546, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.134, "step": 13508 }, { "epoch": 20.223053892215567, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1371, "step": 13509 }, { "epoch": 20.224550898203592, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1338, "step": 13510 }, { "epoch": 20.226047904191617, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1368, "step": 13511 }, { "epoch": 20.227544910179642, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1312, "step": 13512 }, { "epoch": 20.229041916167663, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1371, "step": 13513 }, { "epoch": 20.230538922155688, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1342, "step": 13514 }, { "epoch": 20.232035928143713, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1357, "step": 13515 }, { "epoch": 20.233532934131738, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.137, "step": 13516 }, { "epoch": 20.23502994011976, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1322, "step": 13517 }, { "epoch": 20.236526946107784, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1381, "step": 13518 }, { "epoch": 20.23802395209581, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1422, "step": 13519 }, { "epoch": 20.239520958083833, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1385, "step": 13520 }, { "epoch": 20.241017964071855, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1325, "step": 13521 }, { "epoch": 20.24251497005988, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1336, "step": 13522 }, { "epoch": 20.244011976047904, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1361, "step": 13523 }, { "epoch": 20.24550898203593, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1312, "step": 13524 }, { "epoch": 20.24700598802395, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1336, "step": 13525 }, { "epoch": 20.248502994011975, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1277, "step": 13526 }, { "epoch": 20.25, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1312, "step": 13527 }, { "epoch": 20.251497005988025, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.133, "step": 13528 }, { "epoch": 20.25299401197605, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.138, "step": 13529 }, { "epoch": 20.25449101796407, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1379, "step": 13530 }, { "epoch": 20.255988023952096, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1368, "step": 13531 }, { "epoch": 20.25748502994012, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1333, "step": 13532 }, { "epoch": 20.258982035928145, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1304, "step": 13533 }, { "epoch": 20.260479041916167, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1352, "step": 13534 }, { "epoch": 20.26197604790419, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1385, "step": 13535 }, { "epoch": 20.263473053892216, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1347, "step": 13536 }, { "epoch": 20.26497005988024, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1339, "step": 13537 }, { "epoch": 20.266467065868262, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.133, "step": 13538 }, { "epoch": 20.267964071856287, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.142, "step": 13539 }, { "epoch": 20.269461077844312, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1354, "step": 13540 }, { "epoch": 20.270958083832337, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1286, "step": 13541 }, { "epoch": 20.272455089820358, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1347, "step": 13542 }, { "epoch": 20.273952095808383, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1371, "step": 13543 }, { "epoch": 20.275449101796408, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1315, "step": 13544 }, { "epoch": 20.276946107784433, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1335, "step": 13545 }, { "epoch": 20.278443113772454, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1331, "step": 13546 }, { "epoch": 20.27994011976048, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1329, "step": 13547 }, { "epoch": 20.281437125748504, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1376, "step": 13548 }, { "epoch": 20.28293413173653, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1374, "step": 13549 }, { "epoch": 20.28443113772455, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1374, "step": 13550 }, { "epoch": 20.285928143712574, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1358, "step": 13551 }, { "epoch": 20.2874251497006, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.138, "step": 13552 }, { "epoch": 20.288922155688624, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.136, "step": 13553 }, { "epoch": 20.290419161676645, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1361, "step": 13554 }, { "epoch": 20.29191616766467, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1402, "step": 13555 }, { "epoch": 20.293413173652695, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1422, "step": 13556 }, { "epoch": 20.29491017964072, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1348, "step": 13557 }, { "epoch": 20.29640718562874, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1366, "step": 13558 }, { "epoch": 20.297904191616766, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.136, "step": 13559 }, { "epoch": 20.29940119760479, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1364, "step": 13560 }, { "epoch": 20.300898203592816, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1351, "step": 13561 }, { "epoch": 20.302395209580837, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1396, "step": 13562 }, { "epoch": 20.30389221556886, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1336, "step": 13563 }, { "epoch": 20.305389221556887, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1339, "step": 13564 }, { "epoch": 20.30688622754491, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1334, "step": 13565 }, { "epoch": 20.308383233532933, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1296, "step": 13566 }, { "epoch": 20.309880239520957, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1374, "step": 13567 }, { "epoch": 20.311377245508982, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.133, "step": 13568 }, { "epoch": 20.312874251497007, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1382, "step": 13569 }, { "epoch": 20.31437125748503, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1378, "step": 13570 }, { "epoch": 20.315868263473053, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1439, "step": 13571 }, { "epoch": 20.317365269461078, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.136, "step": 13572 }, { "epoch": 20.318862275449103, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.136, "step": 13573 }, { "epoch": 20.320359281437124, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1388, "step": 13574 }, { "epoch": 20.32185628742515, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1403, "step": 13575 }, { "epoch": 20.323353293413174, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1394, "step": 13576 }, { "epoch": 20.3248502994012, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1302, "step": 13577 }, { "epoch": 20.32634730538922, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1373, "step": 13578 }, { "epoch": 20.327844311377245, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1348, "step": 13579 }, { "epoch": 20.32934131736527, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1322, "step": 13580 }, { "epoch": 20.330838323353294, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1315, "step": 13581 }, { "epoch": 20.33233532934132, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1322, "step": 13582 }, { "epoch": 20.33383233532934, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1368, "step": 13583 }, { "epoch": 20.335329341317365, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1374, "step": 13584 }, { "epoch": 20.33682634730539, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1364, "step": 13585 }, { "epoch": 20.338323353293415, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1342, "step": 13586 }, { "epoch": 20.339820359281436, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.143, "step": 13587 }, { "epoch": 20.34131736526946, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1283, "step": 13588 }, { "epoch": 20.342814371257486, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1438, "step": 13589 }, { "epoch": 20.34431137724551, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1293, "step": 13590 }, { "epoch": 20.345808383233532, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1297, "step": 13591 }, { "epoch": 20.347305389221557, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1426, "step": 13592 }, { "epoch": 20.34880239520958, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1378, "step": 13593 }, { "epoch": 20.350299401197606, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1376, "step": 13594 }, { "epoch": 20.351796407185628, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.139, "step": 13595 }, { "epoch": 20.353293413173652, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1355, "step": 13596 }, { "epoch": 20.354790419161677, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1357, "step": 13597 }, { "epoch": 20.356287425149702, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1374, "step": 13598 }, { "epoch": 20.357784431137723, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1375, "step": 13599 }, { "epoch": 20.35928143712575, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.131, "step": 13600 }, { "epoch": 20.360778443113773, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1366, "step": 13601 }, { "epoch": 20.362275449101798, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1383, "step": 13602 }, { "epoch": 20.36377245508982, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1371, "step": 13603 }, { "epoch": 20.365269461077844, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.142, "step": 13604 }, { "epoch": 20.36676646706587, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1369, "step": 13605 }, { "epoch": 20.368263473053894, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1385, "step": 13606 }, { "epoch": 20.369760479041915, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1359, "step": 13607 }, { "epoch": 20.37125748502994, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.136, "step": 13608 }, { "epoch": 20.372754491017965, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.138, "step": 13609 }, { "epoch": 20.37425149700599, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1338, "step": 13610 }, { "epoch": 20.37574850299401, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1416, "step": 13611 }, { "epoch": 20.377245508982035, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1373, "step": 13612 }, { "epoch": 20.37874251497006, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1337, "step": 13613 }, { "epoch": 20.380239520958085, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1383, "step": 13614 }, { "epoch": 20.381736526946106, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1378, "step": 13615 }, { "epoch": 20.38323353293413, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1286, "step": 13616 }, { "epoch": 20.384730538922156, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1468, "step": 13617 }, { "epoch": 20.38622754491018, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1398, "step": 13618 }, { "epoch": 20.387724550898202, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1346, "step": 13619 }, { "epoch": 20.389221556886227, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1418, "step": 13620 }, { "epoch": 20.39071856287425, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1353, "step": 13621 }, { "epoch": 20.392215568862277, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.135, "step": 13622 }, { "epoch": 20.393712574850298, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1393, "step": 13623 }, { "epoch": 20.395209580838323, "grad_norm": 0.146484375, "learning_rate": 0.0008, "loss": 1.1363, "step": 13624 }, { "epoch": 20.396706586826348, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1367, "step": 13625 }, { "epoch": 20.398203592814372, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1375, "step": 13626 }, { "epoch": 20.399700598802394, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1448, "step": 13627 }, { "epoch": 20.40119760479042, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1372, "step": 13628 }, { "epoch": 20.402694610778443, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1357, "step": 13629 }, { "epoch": 20.404191616766468, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1414, "step": 13630 }, { "epoch": 20.40568862275449, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1383, "step": 13631 }, { "epoch": 20.407185628742514, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1308, "step": 13632 }, { "epoch": 20.40868263473054, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.136, "step": 13633 }, { "epoch": 20.410179640718564, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1246, "step": 13634 }, { "epoch": 20.411676646706585, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1272, "step": 13635 }, { "epoch": 20.41317365269461, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1378, "step": 13636 }, { "epoch": 20.414670658682635, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1384, "step": 13637 }, { "epoch": 20.41616766467066, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1352, "step": 13638 }, { "epoch": 20.41766467065868, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1381, "step": 13639 }, { "epoch": 20.419161676646706, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1351, "step": 13640 }, { "epoch": 20.42065868263473, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1388, "step": 13641 }, { "epoch": 20.422155688622755, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.13, "step": 13642 }, { "epoch": 20.42365269461078, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1316, "step": 13643 }, { "epoch": 20.4251497005988, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1435, "step": 13644 }, { "epoch": 20.426646706586826, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1362, "step": 13645 }, { "epoch": 20.42814371257485, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1389, "step": 13646 }, { "epoch": 20.429640718562876, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1437, "step": 13647 }, { "epoch": 20.431137724550897, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1371, "step": 13648 }, { "epoch": 20.432634730538922, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1428, "step": 13649 }, { "epoch": 20.434131736526947, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1364, "step": 13650 }, { "epoch": 20.43562874251497, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1413, "step": 13651 }, { "epoch": 20.437125748502993, "grad_norm": 0.1357421875, "learning_rate": 0.0008, "loss": 1.1375, "step": 13652 }, { "epoch": 20.438622754491018, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1375, "step": 13653 }, { "epoch": 20.440119760479043, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.138, "step": 13654 }, { "epoch": 20.441616766467067, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1412, "step": 13655 }, { "epoch": 20.44311377245509, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1371, "step": 13656 }, { "epoch": 20.444610778443113, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1377, "step": 13657 }, { "epoch": 20.44610778443114, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1322, "step": 13658 }, { "epoch": 20.447604790419163, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1386, "step": 13659 }, { "epoch": 20.449101796407184, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1342, "step": 13660 }, { "epoch": 20.45059880239521, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1363, "step": 13661 }, { "epoch": 20.452095808383234, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1364, "step": 13662 }, { "epoch": 20.45359281437126, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1334, "step": 13663 }, { "epoch": 20.45508982035928, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1417, "step": 13664 }, { "epoch": 20.456586826347305, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1412, "step": 13665 }, { "epoch": 20.45808383233533, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1359, "step": 13666 }, { "epoch": 20.459580838323355, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1368, "step": 13667 }, { "epoch": 20.461077844311376, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1328, "step": 13668 }, { "epoch": 20.4625748502994, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1326, "step": 13669 }, { "epoch": 20.464071856287426, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.14, "step": 13670 }, { "epoch": 20.46556886227545, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1383, "step": 13671 }, { "epoch": 20.46706586826347, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1353, "step": 13672 }, { "epoch": 20.468562874251496, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1328, "step": 13673 }, { "epoch": 20.47005988023952, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1319, "step": 13674 }, { "epoch": 20.471556886227546, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1344, "step": 13675 }, { "epoch": 20.473053892215567, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1372, "step": 13676 }, { "epoch": 20.474550898203592, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1362, "step": 13677 }, { "epoch": 20.476047904191617, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.134, "step": 13678 }, { "epoch": 20.477544910179642, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1363, "step": 13679 }, { "epoch": 20.479041916167663, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1408, "step": 13680 }, { "epoch": 20.480538922155688, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1334, "step": 13681 }, { "epoch": 20.482035928143713, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1355, "step": 13682 }, { "epoch": 20.483532934131738, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1363, "step": 13683 }, { "epoch": 20.48502994011976, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1328, "step": 13684 }, { "epoch": 20.486526946107784, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1368, "step": 13685 }, { "epoch": 20.48802395209581, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.128, "step": 13686 }, { "epoch": 20.489520958083833, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1269, "step": 13687 }, { "epoch": 20.491017964071855, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.136, "step": 13688 }, { "epoch": 20.49251497005988, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1414, "step": 13689 }, { "epoch": 20.494011976047904, "grad_norm": 0.25390625, "learning_rate": 0.0008, "loss": 1.1311, "step": 13690 }, { "epoch": 20.49550898203593, "grad_norm": 0.1796875, "learning_rate": 0.0008, "loss": 1.1351, "step": 13691 }, { "epoch": 20.49700598802395, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1397, "step": 13692 }, { "epoch": 20.498502994011975, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1382, "step": 13693 }, { "epoch": 20.5, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1342, "step": 13694 }, { "epoch": 20.501497005988025, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1369, "step": 13695 }, { "epoch": 20.50299401197605, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1344, "step": 13696 }, { "epoch": 20.50449101796407, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1366, "step": 13697 }, { "epoch": 20.505988023952096, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1422, "step": 13698 }, { "epoch": 20.50748502994012, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1347, "step": 13699 }, { "epoch": 20.508982035928145, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1437, "step": 13700 }, { "epoch": 20.510479041916167, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1406, "step": 13701 }, { "epoch": 20.51197604790419, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1372, "step": 13702 }, { "epoch": 20.513473053892216, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1335, "step": 13703 }, { "epoch": 20.51497005988024, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1394, "step": 13704 }, { "epoch": 20.516467065868262, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1359, "step": 13705 }, { "epoch": 20.517964071856287, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1273, "step": 13706 }, { "epoch": 20.519461077844312, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.138, "step": 13707 }, { "epoch": 20.520958083832337, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.137, "step": 13708 }, { "epoch": 20.522455089820358, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1341, "step": 13709 }, { "epoch": 20.523952095808383, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1361, "step": 13710 }, { "epoch": 20.525449101796408, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1415, "step": 13711 }, { "epoch": 20.526946107784433, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1352, "step": 13712 }, { "epoch": 20.528443113772454, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1347, "step": 13713 }, { "epoch": 20.52994011976048, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1347, "step": 13714 }, { "epoch": 20.531437125748504, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1372, "step": 13715 }, { "epoch": 20.53293413173653, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1329, "step": 13716 }, { "epoch": 20.53443113772455, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1374, "step": 13717 }, { "epoch": 20.535928143712574, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.136, "step": 13718 }, { "epoch": 20.5374251497006, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1404, "step": 13719 }, { "epoch": 20.538922155688624, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1351, "step": 13720 }, { "epoch": 20.540419161676645, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1305, "step": 13721 }, { "epoch": 20.54191616766467, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1357, "step": 13722 }, { "epoch": 20.543413173652695, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1345, "step": 13723 }, { "epoch": 20.54491017964072, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1252, "step": 13724 }, { "epoch": 20.54640718562874, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1378, "step": 13725 }, { "epoch": 20.547904191616766, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1295, "step": 13726 }, { "epoch": 20.54940119760479, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1355, "step": 13727 }, { "epoch": 20.550898203592816, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1339, "step": 13728 }, { "epoch": 20.552395209580837, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1367, "step": 13729 }, { "epoch": 20.55389221556886, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1363, "step": 13730 }, { "epoch": 20.555389221556887, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1431, "step": 13731 }, { "epoch": 20.55688622754491, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1373, "step": 13732 }, { "epoch": 20.558383233532933, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1404, "step": 13733 }, { "epoch": 20.559880239520957, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1404, "step": 13734 }, { "epoch": 20.561377245508982, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1358, "step": 13735 }, { "epoch": 20.562874251497007, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1348, "step": 13736 }, { "epoch": 20.56437125748503, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1439, "step": 13737 }, { "epoch": 20.565868263473053, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1356, "step": 13738 }, { "epoch": 20.567365269461078, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.135, "step": 13739 }, { "epoch": 20.568862275449103, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1404, "step": 13740 }, { "epoch": 20.570359281437124, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1305, "step": 13741 }, { "epoch": 20.57185628742515, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.142, "step": 13742 }, { "epoch": 20.573353293413174, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1367, "step": 13743 }, { "epoch": 20.5748502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.142, "step": 13744 }, { "epoch": 20.57634730538922, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1352, "step": 13745 }, { "epoch": 20.577844311377245, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.136, "step": 13746 }, { "epoch": 20.57934131736527, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.135, "step": 13747 }, { "epoch": 20.580838323353294, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1371, "step": 13748 }, { "epoch": 20.58233532934132, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1304, "step": 13749 }, { "epoch": 20.58383233532934, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1361, "step": 13750 }, { "epoch": 20.585329341317365, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1255, "step": 13751 }, { "epoch": 20.58682634730539, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1376, "step": 13752 }, { "epoch": 20.58832335329341, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1317, "step": 13753 }, { "epoch": 20.589820359281436, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1289, "step": 13754 }, { "epoch": 20.59131736526946, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1354, "step": 13755 }, { "epoch": 20.592814371257486, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1326, "step": 13756 }, { "epoch": 20.59431137724551, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1342, "step": 13757 }, { "epoch": 20.595808383233532, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1369, "step": 13758 }, { "epoch": 20.597305389221557, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1388, "step": 13759 }, { "epoch": 20.59880239520958, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.131, "step": 13760 }, { "epoch": 20.600299401197606, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1357, "step": 13761 }, { "epoch": 20.601796407185628, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1289, "step": 13762 }, { "epoch": 20.603293413173652, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1387, "step": 13763 }, { "epoch": 20.604790419161677, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1367, "step": 13764 }, { "epoch": 20.606287425149702, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1393, "step": 13765 }, { "epoch": 20.607784431137723, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1372, "step": 13766 }, { "epoch": 20.60928143712575, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1368, "step": 13767 }, { "epoch": 20.610778443113773, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1314, "step": 13768 }, { "epoch": 20.612275449101798, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1363, "step": 13769 }, { "epoch": 20.61377245508982, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1267, "step": 13770 }, { "epoch": 20.615269461077844, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1371, "step": 13771 }, { "epoch": 20.61676646706587, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1409, "step": 13772 }, { "epoch": 20.618263473053894, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1305, "step": 13773 }, { "epoch": 20.619760479041915, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1361, "step": 13774 }, { "epoch": 20.62125748502994, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1346, "step": 13775 }, { "epoch": 20.622754491017965, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1307, "step": 13776 }, { "epoch": 20.62425149700599, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1404, "step": 13777 }, { "epoch": 20.62574850299401, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1356, "step": 13778 }, { "epoch": 20.627245508982035, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1299, "step": 13779 }, { "epoch": 20.62874251497006, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1409, "step": 13780 }, { "epoch": 20.630239520958085, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1448, "step": 13781 }, { "epoch": 20.631736526946106, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1303, "step": 13782 }, { "epoch": 20.63323353293413, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1363, "step": 13783 }, { "epoch": 20.634730538922156, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1326, "step": 13784 }, { "epoch": 20.63622754491018, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1368, "step": 13785 }, { "epoch": 20.637724550898202, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1305, "step": 13786 }, { "epoch": 20.639221556886227, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1341, "step": 13787 }, { "epoch": 20.64071856287425, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.14, "step": 13788 }, { "epoch": 20.642215568862277, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1401, "step": 13789 }, { "epoch": 20.643712574850298, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1357, "step": 13790 }, { "epoch": 20.645209580838323, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1345, "step": 13791 }, { "epoch": 20.646706586826348, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1368, "step": 13792 }, { "epoch": 20.648203592814372, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1335, "step": 13793 }, { "epoch": 20.649700598802394, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1299, "step": 13794 }, { "epoch": 20.65119760479042, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1466, "step": 13795 }, { "epoch": 20.652694610778443, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1365, "step": 13796 }, { "epoch": 20.654191616766468, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1335, "step": 13797 }, { "epoch": 20.65568862275449, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1335, "step": 13798 }, { "epoch": 20.657185628742514, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1322, "step": 13799 }, { "epoch": 20.65868263473054, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1423, "step": 13800 }, { "epoch": 20.660179640718564, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1432, "step": 13801 }, { "epoch": 20.66167664670659, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.137, "step": 13802 }, { "epoch": 20.66317365269461, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1299, "step": 13803 }, { "epoch": 20.664670658682635, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1361, "step": 13804 }, { "epoch": 20.66616766467066, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1308, "step": 13805 }, { "epoch": 20.66766467065868, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1372, "step": 13806 }, { "epoch": 20.669161676646706, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1352, "step": 13807 }, { "epoch": 20.67065868263473, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1321, "step": 13808 }, { "epoch": 20.672155688622755, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1373, "step": 13809 }, { "epoch": 20.67365269461078, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1385, "step": 13810 }, { "epoch": 20.6751497005988, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1377, "step": 13811 }, { "epoch": 20.676646706586826, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1316, "step": 13812 }, { "epoch": 20.67814371257485, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1399, "step": 13813 }, { "epoch": 20.679640718562876, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1422, "step": 13814 }, { "epoch": 20.681137724550897, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.133, "step": 13815 }, { "epoch": 20.682634730538922, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1345, "step": 13816 }, { "epoch": 20.684131736526947, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1384, "step": 13817 }, { "epoch": 20.68562874251497, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1377, "step": 13818 }, { "epoch": 20.687125748502993, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1281, "step": 13819 }, { "epoch": 20.688622754491018, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1382, "step": 13820 }, { "epoch": 20.690119760479043, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1385, "step": 13821 }, { "epoch": 20.691616766467067, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1388, "step": 13822 }, { "epoch": 20.69311377245509, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1393, "step": 13823 }, { "epoch": 20.694610778443113, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1411, "step": 13824 }, { "epoch": 20.69610778443114, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1346, "step": 13825 }, { "epoch": 20.697604790419163, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1335, "step": 13826 }, { "epoch": 20.699101796407184, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1348, "step": 13827 }, { "epoch": 20.70059880239521, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1367, "step": 13828 }, { "epoch": 20.702095808383234, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1376, "step": 13829 }, { "epoch": 20.70359281437126, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1337, "step": 13830 }, { "epoch": 20.70508982035928, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1387, "step": 13831 }, { "epoch": 20.706586826347305, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1365, "step": 13832 }, { "epoch": 20.70808383233533, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1314, "step": 13833 }, { "epoch": 20.709580838323355, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1348, "step": 13834 }, { "epoch": 20.711077844311376, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1308, "step": 13835 }, { "epoch": 20.7125748502994, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1367, "step": 13836 }, { "epoch": 20.714071856287426, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1408, "step": 13837 }, { "epoch": 20.71556886227545, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1383, "step": 13838 }, { "epoch": 20.71706586826347, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1396, "step": 13839 }, { "epoch": 20.718562874251496, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1321, "step": 13840 }, { "epoch": 20.72005988023952, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1385, "step": 13841 }, { "epoch": 20.721556886227546, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1371, "step": 13842 }, { "epoch": 20.723053892215567, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1354, "step": 13843 }, { "epoch": 20.724550898203592, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1308, "step": 13844 }, { "epoch": 20.726047904191617, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1389, "step": 13845 }, { "epoch": 20.727544910179642, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1322, "step": 13846 }, { "epoch": 20.729041916167663, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1321, "step": 13847 }, { "epoch": 20.730538922155688, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1398, "step": 13848 }, { "epoch": 20.732035928143713, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1379, "step": 13849 }, { "epoch": 20.733532934131738, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1384, "step": 13850 }, { "epoch": 20.73502994011976, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1443, "step": 13851 }, { "epoch": 20.736526946107784, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1369, "step": 13852 }, { "epoch": 20.73802395209581, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1317, "step": 13853 }, { "epoch": 20.739520958083833, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1406, "step": 13854 }, { "epoch": 20.741017964071855, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1247, "step": 13855 }, { "epoch": 20.74251497005988, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1406, "step": 13856 }, { "epoch": 20.744011976047904, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1329, "step": 13857 }, { "epoch": 20.74550898203593, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1344, "step": 13858 }, { "epoch": 20.74700598802395, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1361, "step": 13859 }, { "epoch": 20.748502994011975, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1415, "step": 13860 }, { "epoch": 20.75, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1393, "step": 13861 }, { "epoch": 20.751497005988025, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1388, "step": 13862 }, { "epoch": 20.75299401197605, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1339, "step": 13863 }, { "epoch": 20.75449101796407, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.134, "step": 13864 }, { "epoch": 20.755988023952096, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1296, "step": 13865 }, { "epoch": 20.75748502994012, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1362, "step": 13866 }, { "epoch": 20.758982035928145, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1325, "step": 13867 }, { "epoch": 20.760479041916167, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1372, "step": 13868 }, { "epoch": 20.76197604790419, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.144, "step": 13869 }, { "epoch": 20.763473053892216, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1388, "step": 13870 }, { "epoch": 20.76497005988024, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1402, "step": 13871 }, { "epoch": 20.766467065868262, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1362, "step": 13872 }, { "epoch": 20.767964071856287, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1373, "step": 13873 }, { "epoch": 20.769461077844312, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1388, "step": 13874 }, { "epoch": 20.770958083832337, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1338, "step": 13875 }, { "epoch": 20.772455089820358, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.133, "step": 13876 }, { "epoch": 20.773952095808383, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1363, "step": 13877 }, { "epoch": 20.775449101796408, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1368, "step": 13878 }, { "epoch": 20.776946107784433, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1379, "step": 13879 }, { "epoch": 20.778443113772454, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1355, "step": 13880 }, { "epoch": 20.77994011976048, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1389, "step": 13881 }, { "epoch": 20.781437125748504, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1377, "step": 13882 }, { "epoch": 20.78293413173653, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1362, "step": 13883 }, { "epoch": 20.78443113772455, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1354, "step": 13884 }, { "epoch": 20.785928143712574, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1407, "step": 13885 }, { "epoch": 20.7874251497006, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1376, "step": 13886 }, { "epoch": 20.788922155688624, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.139, "step": 13887 }, { "epoch": 20.790419161676645, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1357, "step": 13888 }, { "epoch": 20.79191616766467, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1341, "step": 13889 }, { "epoch": 20.793413173652695, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1308, "step": 13890 }, { "epoch": 20.79491017964072, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1373, "step": 13891 }, { "epoch": 20.79640718562874, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1402, "step": 13892 }, { "epoch": 20.797904191616766, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1308, "step": 13893 }, { "epoch": 20.79940119760479, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1371, "step": 13894 }, { "epoch": 20.800898203592816, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1327, "step": 13895 }, { "epoch": 20.802395209580837, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1429, "step": 13896 }, { "epoch": 20.80389221556886, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1416, "step": 13897 }, { "epoch": 20.805389221556887, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1319, "step": 13898 }, { "epoch": 20.80688622754491, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1479, "step": 13899 }, { "epoch": 20.808383233532933, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1385, "step": 13900 }, { "epoch": 20.809880239520957, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1309, "step": 13901 }, { "epoch": 20.811377245508982, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1452, "step": 13902 }, { "epoch": 20.812874251497007, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1308, "step": 13903 }, { "epoch": 20.81437125748503, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.133, "step": 13904 }, { "epoch": 20.815868263473053, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1335, "step": 13905 }, { "epoch": 20.817365269461078, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1361, "step": 13906 }, { "epoch": 20.818862275449103, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1378, "step": 13907 }, { "epoch": 20.820359281437124, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1395, "step": 13908 }, { "epoch": 20.82185628742515, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1305, "step": 13909 }, { "epoch": 20.823353293413174, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1381, "step": 13910 }, { "epoch": 20.8248502994012, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.138, "step": 13911 }, { "epoch": 20.82634730538922, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1333, "step": 13912 }, { "epoch": 20.827844311377245, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1358, "step": 13913 }, { "epoch": 20.82934131736527, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1381, "step": 13914 }, { "epoch": 20.830838323353294, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1363, "step": 13915 }, { "epoch": 20.83233532934132, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1397, "step": 13916 }, { "epoch": 20.83383233532934, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1354, "step": 13917 }, { "epoch": 20.835329341317365, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1394, "step": 13918 }, { "epoch": 20.83682634730539, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1305, "step": 13919 }, { "epoch": 20.83832335329341, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1368, "step": 13920 }, { "epoch": 20.839820359281436, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1359, "step": 13921 }, { "epoch": 20.84131736526946, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1419, "step": 13922 }, { "epoch": 20.842814371257486, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1367, "step": 13923 }, { "epoch": 20.84431137724551, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1372, "step": 13924 }, { "epoch": 20.845808383233532, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1338, "step": 13925 }, { "epoch": 20.847305389221557, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1352, "step": 13926 }, { "epoch": 20.84880239520958, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1412, "step": 13927 }, { "epoch": 20.850299401197606, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.131, "step": 13928 }, { "epoch": 20.851796407185628, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1384, "step": 13929 }, { "epoch": 20.853293413173652, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1318, "step": 13930 }, { "epoch": 20.854790419161677, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1368, "step": 13931 }, { "epoch": 20.856287425149702, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1374, "step": 13932 }, { "epoch": 20.857784431137723, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1429, "step": 13933 }, { "epoch": 20.85928143712575, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1259, "step": 13934 }, { "epoch": 20.860778443113773, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1374, "step": 13935 }, { "epoch": 20.862275449101798, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1428, "step": 13936 }, { "epoch": 20.86377245508982, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.13, "step": 13937 }, { "epoch": 20.865269461077844, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1351, "step": 13938 }, { "epoch": 20.86676646706587, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1368, "step": 13939 }, { "epoch": 20.868263473053894, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1423, "step": 13940 }, { "epoch": 20.869760479041915, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.135, "step": 13941 }, { "epoch": 20.87125748502994, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1383, "step": 13942 }, { "epoch": 20.872754491017965, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1361, "step": 13943 }, { "epoch": 20.87425149700599, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1346, "step": 13944 }, { "epoch": 20.87574850299401, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1342, "step": 13945 }, { "epoch": 20.877245508982035, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1361, "step": 13946 }, { "epoch": 20.87874251497006, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1368, "step": 13947 }, { "epoch": 20.880239520958085, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1438, "step": 13948 }, { "epoch": 20.881736526946106, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.138, "step": 13949 }, { "epoch": 20.88323353293413, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1375, "step": 13950 }, { "epoch": 20.884730538922156, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.132, "step": 13951 }, { "epoch": 20.88622754491018, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1412, "step": 13952 }, { "epoch": 20.887724550898202, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1321, "step": 13953 }, { "epoch": 20.889221556886227, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1359, "step": 13954 }, { "epoch": 20.89071856287425, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1329, "step": 13955 }, { "epoch": 20.892215568862277, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1344, "step": 13956 }, { "epoch": 20.893712574850298, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1316, "step": 13957 }, { "epoch": 20.895209580838323, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1387, "step": 13958 }, { "epoch": 20.896706586826348, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1282, "step": 13959 }, { "epoch": 20.898203592814372, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1374, "step": 13960 }, { "epoch": 20.899700598802394, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1352, "step": 13961 }, { "epoch": 20.90119760479042, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1374, "step": 13962 }, { "epoch": 20.902694610778443, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1317, "step": 13963 }, { "epoch": 20.904191616766468, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1355, "step": 13964 }, { "epoch": 20.90568862275449, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.137, "step": 13965 }, { "epoch": 20.907185628742514, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1291, "step": 13966 }, { "epoch": 20.90868263473054, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1398, "step": 13967 }, { "epoch": 20.910179640718564, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1305, "step": 13968 }, { "epoch": 20.91167664670659, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1379, "step": 13969 }, { "epoch": 20.91317365269461, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1308, "step": 13970 }, { "epoch": 20.914670658682635, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1332, "step": 13971 }, { "epoch": 20.91616766467066, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1347, "step": 13972 }, { "epoch": 20.91766467065868, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1311, "step": 13973 }, { "epoch": 20.919161676646706, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1362, "step": 13974 }, { "epoch": 20.92065868263473, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1342, "step": 13975 }, { "epoch": 20.922155688622755, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.131, "step": 13976 }, { "epoch": 20.92365269461078, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.141, "step": 13977 }, { "epoch": 20.9251497005988, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1333, "step": 13978 }, { "epoch": 20.926646706586826, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1363, "step": 13979 }, { "epoch": 20.92814371257485, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1339, "step": 13980 }, { "epoch": 20.929640718562876, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1409, "step": 13981 }, { "epoch": 20.931137724550897, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1414, "step": 13982 }, { "epoch": 20.932634730538922, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1368, "step": 13983 }, { "epoch": 20.934131736526947, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.135, "step": 13984 }, { "epoch": 20.93562874251497, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1395, "step": 13985 }, { "epoch": 20.937125748502993, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1389, "step": 13986 }, { "epoch": 20.938622754491018, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1321, "step": 13987 }, { "epoch": 20.940119760479043, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1397, "step": 13988 }, { "epoch": 20.941616766467067, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1353, "step": 13989 }, { "epoch": 20.94311377245509, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1352, "step": 13990 }, { "epoch": 20.944610778443113, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1435, "step": 13991 }, { "epoch": 20.94610778443114, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1348, "step": 13992 }, { "epoch": 20.947604790419163, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1332, "step": 13993 }, { "epoch": 20.949101796407184, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1334, "step": 13994 }, { "epoch": 20.95059880239521, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1354, "step": 13995 }, { "epoch": 20.952095808383234, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.137, "step": 13996 }, { "epoch": 20.95359281437126, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1361, "step": 13997 }, { "epoch": 20.95508982035928, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1334, "step": 13998 }, { "epoch": 20.956586826347305, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1323, "step": 13999 }, { "epoch": 20.95808383233533, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1366, "step": 14000 }, { "epoch": 20.959580838323355, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1351, "step": 14001 }, { "epoch": 20.961077844311376, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.135, "step": 14002 }, { "epoch": 20.9625748502994, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1436, "step": 14003 }, { "epoch": 20.964071856287426, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1404, "step": 14004 }, { "epoch": 20.96556886227545, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1407, "step": 14005 }, { "epoch": 20.96706586826347, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1418, "step": 14006 }, { "epoch": 20.968562874251496, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1322, "step": 14007 }, { "epoch": 20.97005988023952, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1395, "step": 14008 }, { "epoch": 20.971556886227546, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1378, "step": 14009 }, { "epoch": 20.973053892215567, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1317, "step": 14010 }, { "epoch": 20.974550898203592, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.141, "step": 14011 }, { "epoch": 20.976047904191617, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1386, "step": 14012 }, { "epoch": 20.977544910179642, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1393, "step": 14013 }, { "epoch": 20.979041916167663, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1339, "step": 14014 }, { "epoch": 20.980538922155688, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1372, "step": 14015 }, { "epoch": 20.982035928143713, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1341, "step": 14016 }, { "epoch": 20.983532934131738, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1353, "step": 14017 }, { "epoch": 20.98502994011976, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1377, "step": 14018 }, { "epoch": 20.986526946107784, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1292, "step": 14019 }, { "epoch": 20.98802395209581, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1332, "step": 14020 }, { "epoch": 20.989520958083833, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1355, "step": 14021 }, { "epoch": 20.991017964071855, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1379, "step": 14022 }, { "epoch": 20.99251497005988, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1379, "step": 14023 }, { "epoch": 20.994011976047904, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1392, "step": 14024 }, { "epoch": 20.99550898203593, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1344, "step": 14025 }, { "epoch": 20.99700598802395, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1295, "step": 14026 }, { "epoch": 20.998502994011975, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.139, "step": 14027 }, { "epoch": 21.0, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1373, "step": 14028 }, { "epoch": 21.001497005988025, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1394, "step": 14029 }, { "epoch": 21.00299401197605, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1381, "step": 14030 }, { "epoch": 21.00449101796407, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1362, "step": 14031 }, { "epoch": 21.005988023952096, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1422, "step": 14032 }, { "epoch": 21.00748502994012, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1379, "step": 14033 }, { "epoch": 21.008982035928145, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1311, "step": 14034 }, { "epoch": 21.010479041916167, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1383, "step": 14035 }, { "epoch": 21.01197604790419, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.138, "step": 14036 }, { "epoch": 21.013473053892216, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1312, "step": 14037 }, { "epoch": 21.01497005988024, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1308, "step": 14038 }, { "epoch": 21.016467065868262, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1337, "step": 14039 }, { "epoch": 21.017964071856287, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1326, "step": 14040 }, { "epoch": 21.019461077844312, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1348, "step": 14041 }, { "epoch": 21.020958083832337, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.136, "step": 14042 }, { "epoch": 21.022455089820358, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1426, "step": 14043 }, { "epoch": 21.023952095808383, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1281, "step": 14044 }, { "epoch": 21.025449101796408, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1342, "step": 14045 }, { "epoch": 21.026946107784433, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1471, "step": 14046 }, { "epoch": 21.028443113772454, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1416, "step": 14047 }, { "epoch": 21.02994011976048, "grad_norm": 0.1396484375, "learning_rate": 0.0008, "loss": 1.1393, "step": 14048 }, { "epoch": 21.031437125748504, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1327, "step": 14049 }, { "epoch": 21.03293413173653, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1324, "step": 14050 }, { "epoch": 21.03443113772455, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.129, "step": 14051 }, { "epoch": 21.035928143712574, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1377, "step": 14052 }, { "epoch": 21.0374251497006, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1367, "step": 14053 }, { "epoch": 21.038922155688624, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1346, "step": 14054 }, { "epoch": 21.040419161676645, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1321, "step": 14055 }, { "epoch": 21.04191616766467, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.131, "step": 14056 }, { "epoch": 21.043413173652695, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1384, "step": 14057 }, { "epoch": 21.04491017964072, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1358, "step": 14058 }, { "epoch": 21.04640718562874, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1378, "step": 14059 }, { "epoch": 21.047904191616766, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1352, "step": 14060 }, { "epoch": 21.04940119760479, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1408, "step": 14061 }, { "epoch": 21.050898203592816, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1421, "step": 14062 }, { "epoch": 21.052395209580837, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1374, "step": 14063 }, { "epoch": 21.05389221556886, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1375, "step": 14064 }, { "epoch": 21.055389221556887, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.141, "step": 14065 }, { "epoch": 21.05688622754491, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1399, "step": 14066 }, { "epoch": 21.058383233532933, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1298, "step": 14067 }, { "epoch": 21.059880239520957, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1389, "step": 14068 }, { "epoch": 21.061377245508982, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1365, "step": 14069 }, { "epoch": 21.062874251497007, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1358, "step": 14070 }, { "epoch": 21.06437125748503, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1342, "step": 14071 }, { "epoch": 21.065868263473053, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.132, "step": 14072 }, { "epoch": 21.067365269461078, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1326, "step": 14073 }, { "epoch": 21.068862275449103, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1346, "step": 14074 }, { "epoch": 21.070359281437124, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1339, "step": 14075 }, { "epoch": 21.07185628742515, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1319, "step": 14076 }, { "epoch": 21.073353293413174, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.139, "step": 14077 }, { "epoch": 21.0748502994012, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.136, "step": 14078 }, { "epoch": 21.07634730538922, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1309, "step": 14079 }, { "epoch": 21.077844311377245, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1313, "step": 14080 }, { "epoch": 21.07934131736527, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1375, "step": 14081 }, { "epoch": 21.080838323353294, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1388, "step": 14082 }, { "epoch": 21.082335329341316, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1388, "step": 14083 }, { "epoch": 21.08383233532934, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1281, "step": 14084 }, { "epoch": 21.085329341317365, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.135, "step": 14085 }, { "epoch": 21.08682634730539, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1317, "step": 14086 }, { "epoch": 21.088323353293415, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1384, "step": 14087 }, { "epoch": 21.089820359281436, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1349, "step": 14088 }, { "epoch": 21.09131736526946, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1318, "step": 14089 }, { "epoch": 21.092814371257486, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.134, "step": 14090 }, { "epoch": 21.09431137724551, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1274, "step": 14091 }, { "epoch": 21.095808383233532, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1364, "step": 14092 }, { "epoch": 21.097305389221557, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1339, "step": 14093 }, { "epoch": 21.09880239520958, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1344, "step": 14094 }, { "epoch": 21.100299401197606, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1291, "step": 14095 }, { "epoch": 21.101796407185628, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1323, "step": 14096 }, { "epoch": 21.103293413173652, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1371, "step": 14097 }, { "epoch": 21.104790419161677, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1323, "step": 14098 }, { "epoch": 21.106287425149702, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1359, "step": 14099 }, { "epoch": 21.107784431137723, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1316, "step": 14100 }, { "epoch": 21.10928143712575, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.134, "step": 14101 }, { "epoch": 21.110778443113773, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1385, "step": 14102 }, { "epoch": 21.112275449101798, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1316, "step": 14103 }, { "epoch": 21.11377245508982, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1377, "step": 14104 }, { "epoch": 21.115269461077844, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1329, "step": 14105 }, { "epoch": 21.11676646706587, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1377, "step": 14106 }, { "epoch": 21.118263473053894, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.136, "step": 14107 }, { "epoch": 21.119760479041915, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1388, "step": 14108 }, { "epoch": 21.12125748502994, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.127, "step": 14109 }, { "epoch": 21.122754491017965, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1349, "step": 14110 }, { "epoch": 21.12425149700599, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1396, "step": 14111 }, { "epoch": 21.12574850299401, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1437, "step": 14112 }, { "epoch": 21.127245508982035, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1399, "step": 14113 }, { "epoch": 21.12874251497006, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1369, "step": 14114 }, { "epoch": 21.130239520958085, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1417, "step": 14115 }, { "epoch": 21.131736526946106, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1427, "step": 14116 }, { "epoch": 21.13323353293413, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1355, "step": 14117 }, { "epoch": 21.134730538922156, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1383, "step": 14118 }, { "epoch": 21.13622754491018, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1379, "step": 14119 }, { "epoch": 21.137724550898202, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1376, "step": 14120 }, { "epoch": 21.139221556886227, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1315, "step": 14121 }, { "epoch": 21.14071856287425, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1326, "step": 14122 }, { "epoch": 21.142215568862277, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1346, "step": 14123 }, { "epoch": 21.143712574850298, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.138, "step": 14124 }, { "epoch": 21.145209580838323, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1384, "step": 14125 }, { "epoch": 21.146706586826348, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1293, "step": 14126 }, { "epoch": 21.148203592814372, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.135, "step": 14127 }, { "epoch": 21.149700598802394, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1299, "step": 14128 }, { "epoch": 21.15119760479042, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1391, "step": 14129 }, { "epoch": 21.152694610778443, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1354, "step": 14130 }, { "epoch": 21.154191616766468, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1372, "step": 14131 }, { "epoch": 21.15568862275449, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1337, "step": 14132 }, { "epoch": 21.157185628742514, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1323, "step": 14133 }, { "epoch": 21.15868263473054, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1344, "step": 14134 }, { "epoch": 21.160179640718564, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1235, "step": 14135 }, { "epoch": 21.161676646706585, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.141, "step": 14136 }, { "epoch": 21.16317365269461, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1365, "step": 14137 }, { "epoch": 21.164670658682635, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1392, "step": 14138 }, { "epoch": 21.16616766467066, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1359, "step": 14139 }, { "epoch": 21.16766467065868, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1325, "step": 14140 }, { "epoch": 21.169161676646706, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1324, "step": 14141 }, { "epoch": 21.17065868263473, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1416, "step": 14142 }, { "epoch": 21.172155688622755, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1274, "step": 14143 }, { "epoch": 21.17365269461078, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1357, "step": 14144 }, { "epoch": 21.1751497005988, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.133, "step": 14145 }, { "epoch": 21.176646706586826, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1361, "step": 14146 }, { "epoch": 21.17814371257485, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1394, "step": 14147 }, { "epoch": 21.179640718562876, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1312, "step": 14148 }, { "epoch": 21.181137724550897, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.136, "step": 14149 }, { "epoch": 21.182634730538922, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1365, "step": 14150 }, { "epoch": 21.184131736526947, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1363, "step": 14151 }, { "epoch": 21.18562874251497, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1341, "step": 14152 }, { "epoch": 21.187125748502993, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1367, "step": 14153 }, { "epoch": 21.188622754491018, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1383, "step": 14154 }, { "epoch": 21.190119760479043, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1482, "step": 14155 }, { "epoch": 21.191616766467067, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1347, "step": 14156 }, { "epoch": 21.19311377245509, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1423, "step": 14157 }, { "epoch": 21.194610778443113, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1343, "step": 14158 }, { "epoch": 21.19610778443114, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1318, "step": 14159 }, { "epoch": 21.197604790419163, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1301, "step": 14160 }, { "epoch": 21.199101796407184, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1301, "step": 14161 }, { "epoch": 21.20059880239521, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1386, "step": 14162 }, { "epoch": 21.202095808383234, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1361, "step": 14163 }, { "epoch": 21.20359281437126, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.128, "step": 14164 }, { "epoch": 21.20508982035928, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1317, "step": 14165 }, { "epoch": 21.206586826347305, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1387, "step": 14166 }, { "epoch": 21.20808383233533, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1328, "step": 14167 }, { "epoch": 21.209580838323355, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1341, "step": 14168 }, { "epoch": 21.211077844311376, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1314, "step": 14169 }, { "epoch": 21.2125748502994, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1302, "step": 14170 }, { "epoch": 21.214071856287426, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1384, "step": 14171 }, { "epoch": 21.21556886227545, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1379, "step": 14172 }, { "epoch": 21.21706586826347, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1386, "step": 14173 }, { "epoch": 21.218562874251496, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1302, "step": 14174 }, { "epoch": 21.22005988023952, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1362, "step": 14175 }, { "epoch": 21.221556886227546, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1372, "step": 14176 }, { "epoch": 21.223053892215567, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1411, "step": 14177 }, { "epoch": 21.224550898203592, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1347, "step": 14178 }, { "epoch": 21.226047904191617, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1332, "step": 14179 }, { "epoch": 21.227544910179642, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1425, "step": 14180 }, { "epoch": 21.229041916167663, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1351, "step": 14181 }, { "epoch": 21.230538922155688, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1337, "step": 14182 }, { "epoch": 21.232035928143713, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1322, "step": 14183 }, { "epoch": 21.233532934131738, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1398, "step": 14184 }, { "epoch": 21.23502994011976, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1313, "step": 14185 }, { "epoch": 21.236526946107784, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1324, "step": 14186 }, { "epoch": 21.23802395209581, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1313, "step": 14187 }, { "epoch": 21.239520958083833, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1355, "step": 14188 }, { "epoch": 21.241017964071855, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1401, "step": 14189 }, { "epoch": 21.24251497005988, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1374, "step": 14190 }, { "epoch": 21.244011976047904, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.14, "step": 14191 }, { "epoch": 21.24550898203593, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1329, "step": 14192 }, { "epoch": 21.24700598802395, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1261, "step": 14193 }, { "epoch": 21.248502994011975, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1311, "step": 14194 }, { "epoch": 21.25, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1349, "step": 14195 }, { "epoch": 21.251497005988025, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1355, "step": 14196 }, { "epoch": 21.25299401197605, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1386, "step": 14197 }, { "epoch": 21.25449101796407, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1375, "step": 14198 }, { "epoch": 21.255988023952096, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1367, "step": 14199 }, { "epoch": 21.25748502994012, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.14, "step": 14200 }, { "epoch": 21.258982035928145, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.141, "step": 14201 }, { "epoch": 21.260479041916167, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1401, "step": 14202 }, { "epoch": 21.26197604790419, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1276, "step": 14203 }, { "epoch": 21.263473053892216, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1337, "step": 14204 }, { "epoch": 21.26497005988024, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1362, "step": 14205 }, { "epoch": 21.266467065868262, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1313, "step": 14206 }, { "epoch": 21.267964071856287, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 14207 }, { "epoch": 21.269461077844312, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1317, "step": 14208 }, { "epoch": 21.270958083832337, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1351, "step": 14209 }, { "epoch": 21.272455089820358, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1319, "step": 14210 }, { "epoch": 21.273952095808383, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.14, "step": 14211 }, { "epoch": 21.275449101796408, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1407, "step": 14212 }, { "epoch": 21.276946107784433, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1318, "step": 14213 }, { "epoch": 21.278443113772454, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1306, "step": 14214 }, { "epoch": 21.27994011976048, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1315, "step": 14215 }, { "epoch": 21.281437125748504, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1387, "step": 14216 }, { "epoch": 21.28293413173653, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1316, "step": 14217 }, { "epoch": 21.28443113772455, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1382, "step": 14218 }, { "epoch": 21.285928143712574, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1373, "step": 14219 }, { "epoch": 21.2874251497006, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1374, "step": 14220 }, { "epoch": 21.288922155688624, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.138, "step": 14221 }, { "epoch": 21.290419161676645, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1415, "step": 14222 }, { "epoch": 21.29191616766467, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1326, "step": 14223 }, { "epoch": 21.293413173652695, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1386, "step": 14224 }, { "epoch": 21.29491017964072, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1423, "step": 14225 }, { "epoch": 21.29640718562874, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1367, "step": 14226 }, { "epoch": 21.297904191616766, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1347, "step": 14227 }, { "epoch": 21.29940119760479, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1341, "step": 14228 }, { "epoch": 21.300898203592816, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1302, "step": 14229 }, { "epoch": 21.302395209580837, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1311, "step": 14230 }, { "epoch": 21.30389221556886, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1326, "step": 14231 }, { "epoch": 21.305389221556887, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1409, "step": 14232 }, { "epoch": 21.30688622754491, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.134, "step": 14233 }, { "epoch": 21.308383233532933, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1357, "step": 14234 }, { "epoch": 21.309880239520957, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1376, "step": 14235 }, { "epoch": 21.311377245508982, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1354, "step": 14236 }, { "epoch": 21.312874251497007, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1302, "step": 14237 }, { "epoch": 21.31437125748503, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1388, "step": 14238 }, { "epoch": 21.315868263473053, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1293, "step": 14239 }, { "epoch": 21.317365269461078, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1378, "step": 14240 }, { "epoch": 21.318862275449103, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1315, "step": 14241 }, { "epoch": 21.320359281437124, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1329, "step": 14242 }, { "epoch": 21.32185628742515, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1374, "step": 14243 }, { "epoch": 21.323353293413174, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.139, "step": 14244 }, { "epoch": 21.3248502994012, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1315, "step": 14245 }, { "epoch": 21.32634730538922, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1322, "step": 14246 }, { "epoch": 21.327844311377245, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1332, "step": 14247 }, { "epoch": 21.32934131736527, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1303, "step": 14248 }, { "epoch": 21.330838323353294, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1313, "step": 14249 }, { "epoch": 21.33233532934132, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1362, "step": 14250 }, { "epoch": 21.33383233532934, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1381, "step": 14251 }, { "epoch": 21.335329341317365, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1371, "step": 14252 }, { "epoch": 21.33682634730539, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1373, "step": 14253 }, { "epoch": 21.338323353293415, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1387, "step": 14254 }, { "epoch": 21.339820359281436, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1423, "step": 14255 }, { "epoch": 21.34131736526946, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1372, "step": 14256 }, { "epoch": 21.342814371257486, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1401, "step": 14257 }, { "epoch": 21.34431137724551, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1413, "step": 14258 }, { "epoch": 21.345808383233532, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1379, "step": 14259 }, { "epoch": 21.347305389221557, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1382, "step": 14260 }, { "epoch": 21.34880239520958, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1306, "step": 14261 }, { "epoch": 21.350299401197606, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1367, "step": 14262 }, { "epoch": 21.351796407185628, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1312, "step": 14263 }, { "epoch": 21.353293413173652, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1406, "step": 14264 }, { "epoch": 21.354790419161677, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1362, "step": 14265 }, { "epoch": 21.356287425149702, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1343, "step": 14266 }, { "epoch": 21.357784431137723, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1344, "step": 14267 }, { "epoch": 21.35928143712575, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1338, "step": 14268 }, { "epoch": 21.360778443113773, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1333, "step": 14269 }, { "epoch": 21.362275449101798, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1383, "step": 14270 }, { "epoch": 21.36377245508982, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1341, "step": 14271 }, { "epoch": 21.365269461077844, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.138, "step": 14272 }, { "epoch": 21.36676646706587, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1434, "step": 14273 }, { "epoch": 21.368263473053894, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1322, "step": 14274 }, { "epoch": 21.369760479041915, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1348, "step": 14275 }, { "epoch": 21.37125748502994, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.134, "step": 14276 }, { "epoch": 21.372754491017965, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1313, "step": 14277 }, { "epoch": 21.37425149700599, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.142, "step": 14278 }, { "epoch": 21.37574850299401, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1398, "step": 14279 }, { "epoch": 21.377245508982035, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1431, "step": 14280 }, { "epoch": 21.37874251497006, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1346, "step": 14281 }, { "epoch": 21.380239520958085, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1335, "step": 14282 }, { "epoch": 21.381736526946106, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1345, "step": 14283 }, { "epoch": 21.38323353293413, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 14284 }, { "epoch": 21.384730538922156, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1398, "step": 14285 }, { "epoch": 21.38622754491018, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1301, "step": 14286 }, { "epoch": 21.387724550898202, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1394, "step": 14287 }, { "epoch": 21.389221556886227, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1306, "step": 14288 }, { "epoch": 21.39071856287425, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1304, "step": 14289 }, { "epoch": 21.392215568862277, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1342, "step": 14290 }, { "epoch": 21.393712574850298, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.139, "step": 14291 }, { "epoch": 21.395209580838323, "grad_norm": 0.05078125, "learning_rate": 0.0008, "loss": 1.1369, "step": 14292 }, { "epoch": 21.396706586826348, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1329, "step": 14293 }, { "epoch": 21.398203592814372, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1383, "step": 14294 }, { "epoch": 21.399700598802394, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1366, "step": 14295 }, { "epoch": 21.40119760479042, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1337, "step": 14296 }, { "epoch": 21.402694610778443, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1408, "step": 14297 }, { "epoch": 21.404191616766468, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1315, "step": 14298 }, { "epoch": 21.40568862275449, "grad_norm": 0.1240234375, "learning_rate": 0.0008, "loss": 1.1333, "step": 14299 }, { "epoch": 21.407185628742514, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1323, "step": 14300 }, { "epoch": 21.40868263473054, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1349, "step": 14301 }, { "epoch": 21.410179640718564, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1338, "step": 14302 }, { "epoch": 21.411676646706585, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1407, "step": 14303 }, { "epoch": 21.41317365269461, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1364, "step": 14304 }, { "epoch": 21.414670658682635, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1398, "step": 14305 }, { "epoch": 21.41616766467066, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1354, "step": 14306 }, { "epoch": 21.41766467065868, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1373, "step": 14307 }, { "epoch": 21.419161676646706, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1353, "step": 14308 }, { "epoch": 21.42065868263473, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.129, "step": 14309 }, { "epoch": 21.422155688622755, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1368, "step": 14310 }, { "epoch": 21.42365269461078, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1312, "step": 14311 }, { "epoch": 21.4251497005988, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1356, "step": 14312 }, { "epoch": 21.426646706586826, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1329, "step": 14313 }, { "epoch": 21.42814371257485, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1336, "step": 14314 }, { "epoch": 21.429640718562876, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.134, "step": 14315 }, { "epoch": 21.431137724550897, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1354, "step": 14316 }, { "epoch": 21.432634730538922, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1398, "step": 14317 }, { "epoch": 21.434131736526947, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1323, "step": 14318 }, { "epoch": 21.43562874251497, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1284, "step": 14319 }, { "epoch": 21.437125748502993, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1362, "step": 14320 }, { "epoch": 21.438622754491018, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1374, "step": 14321 }, { "epoch": 21.440119760479043, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1323, "step": 14322 }, { "epoch": 21.441616766467067, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1342, "step": 14323 }, { "epoch": 21.44311377245509, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1308, "step": 14324 }, { "epoch": 21.444610778443113, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1313, "step": 14325 }, { "epoch": 21.44610778443114, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.136, "step": 14326 }, { "epoch": 21.447604790419163, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1352, "step": 14327 }, { "epoch": 21.449101796407184, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1333, "step": 14328 }, { "epoch": 21.45059880239521, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1327, "step": 14329 }, { "epoch": 21.452095808383234, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1361, "step": 14330 }, { "epoch": 21.45359281437126, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1324, "step": 14331 }, { "epoch": 21.45508982035928, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1376, "step": 14332 }, { "epoch": 21.456586826347305, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1285, "step": 14333 }, { "epoch": 21.45808383233533, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1332, "step": 14334 }, { "epoch": 21.459580838323355, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1394, "step": 14335 }, { "epoch": 21.461077844311376, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1385, "step": 14336 }, { "epoch": 21.4625748502994, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1378, "step": 14337 }, { "epoch": 21.464071856287426, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1292, "step": 14338 }, { "epoch": 21.46556886227545, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1375, "step": 14339 }, { "epoch": 21.46706586826347, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1296, "step": 14340 }, { "epoch": 21.468562874251496, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1338, "step": 14341 }, { "epoch": 21.47005988023952, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1409, "step": 14342 }, { "epoch": 21.471556886227546, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1353, "step": 14343 }, { "epoch": 21.473053892215567, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1362, "step": 14344 }, { "epoch": 21.474550898203592, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1294, "step": 14345 }, { "epoch": 21.476047904191617, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1312, "step": 14346 }, { "epoch": 21.477544910179642, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1423, "step": 14347 }, { "epoch": 21.479041916167663, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1356, "step": 14348 }, { "epoch": 21.480538922155688, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1341, "step": 14349 }, { "epoch": 21.482035928143713, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1364, "step": 14350 }, { "epoch": 21.483532934131738, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1296, "step": 14351 }, { "epoch": 21.48502994011976, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1286, "step": 14352 }, { "epoch": 21.486526946107784, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1315, "step": 14353 }, { "epoch": 21.48802395209581, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1273, "step": 14354 }, { "epoch": 21.489520958083833, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1374, "step": 14355 }, { "epoch": 21.491017964071855, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1333, "step": 14356 }, { "epoch": 21.49251497005988, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1312, "step": 14357 }, { "epoch": 21.494011976047904, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1365, "step": 14358 }, { "epoch": 21.49550898203593, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1317, "step": 14359 }, { "epoch": 21.49700598802395, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1378, "step": 14360 }, { "epoch": 21.498502994011975, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1359, "step": 14361 }, { "epoch": 21.5, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1374, "step": 14362 }, { "epoch": 21.501497005988025, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1349, "step": 14363 }, { "epoch": 21.50299401197605, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1269, "step": 14364 }, { "epoch": 21.50449101796407, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1392, "step": 14365 }, { "epoch": 21.505988023952096, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.143, "step": 14366 }, { "epoch": 21.50748502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1382, "step": 14367 }, { "epoch": 21.508982035928145, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.137, "step": 14368 }, { "epoch": 21.510479041916167, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1366, "step": 14369 }, { "epoch": 21.51197604790419, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.127, "step": 14370 }, { "epoch": 21.513473053892216, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1368, "step": 14371 }, { "epoch": 21.51497005988024, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1322, "step": 14372 }, { "epoch": 21.516467065868262, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1315, "step": 14373 }, { "epoch": 21.517964071856287, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1345, "step": 14374 }, { "epoch": 21.519461077844312, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.13, "step": 14375 }, { "epoch": 21.520958083832337, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1387, "step": 14376 }, { "epoch": 21.522455089820358, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1315, "step": 14377 }, { "epoch": 21.523952095808383, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1317, "step": 14378 }, { "epoch": 21.525449101796408, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1302, "step": 14379 }, { "epoch": 21.526946107784433, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1363, "step": 14380 }, { "epoch": 21.528443113772454, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1296, "step": 14381 }, { "epoch": 21.52994011976048, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1365, "step": 14382 }, { "epoch": 21.531437125748504, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1356, "step": 14383 }, { "epoch": 21.53293413173653, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.13, "step": 14384 }, { "epoch": 21.53443113772455, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1344, "step": 14385 }, { "epoch": 21.535928143712574, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1337, "step": 14386 }, { "epoch": 21.5374251497006, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1349, "step": 14387 }, { "epoch": 21.538922155688624, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1283, "step": 14388 }, { "epoch": 21.540419161676645, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1376, "step": 14389 }, { "epoch": 21.54191616766467, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1365, "step": 14390 }, { "epoch": 21.543413173652695, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1299, "step": 14391 }, { "epoch": 21.54491017964072, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1394, "step": 14392 }, { "epoch": 21.54640718562874, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1301, "step": 14393 }, { "epoch": 21.547904191616766, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1315, "step": 14394 }, { "epoch": 21.54940119760479, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1311, "step": 14395 }, { "epoch": 21.550898203592816, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1259, "step": 14396 }, { "epoch": 21.552395209580837, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1353, "step": 14397 }, { "epoch": 21.55389221556886, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1263, "step": 14398 }, { "epoch": 21.555389221556887, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1383, "step": 14399 }, { "epoch": 21.55688622754491, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1373, "step": 14400 }, { "epoch": 21.558383233532933, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1309, "step": 14401 }, { "epoch": 21.559880239520957, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1351, "step": 14402 }, { "epoch": 21.561377245508982, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1329, "step": 14403 }, { "epoch": 21.562874251497007, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1315, "step": 14404 }, { "epoch": 21.56437125748503, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1388, "step": 14405 }, { "epoch": 21.565868263473053, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1384, "step": 14406 }, { "epoch": 21.567365269461078, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1283, "step": 14407 }, { "epoch": 21.568862275449103, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1307, "step": 14408 }, { "epoch": 21.570359281437124, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1397, "step": 14409 }, { "epoch": 21.57185628742515, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1384, "step": 14410 }, { "epoch": 21.573353293413174, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1338, "step": 14411 }, { "epoch": 21.5748502994012, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1357, "step": 14412 }, { "epoch": 21.57634730538922, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1377, "step": 14413 }, { "epoch": 21.577844311377245, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1378, "step": 14414 }, { "epoch": 21.57934131736527, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1332, "step": 14415 }, { "epoch": 21.580838323353294, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1372, "step": 14416 }, { "epoch": 21.58233532934132, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1401, "step": 14417 }, { "epoch": 21.58383233532934, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1363, "step": 14418 }, { "epoch": 21.585329341317365, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1284, "step": 14419 }, { "epoch": 21.58682634730539, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.136, "step": 14420 }, { "epoch": 21.58832335329341, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1368, "step": 14421 }, { "epoch": 21.589820359281436, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1294, "step": 14422 }, { "epoch": 21.59131736526946, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1376, "step": 14423 }, { "epoch": 21.592814371257486, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1319, "step": 14424 }, { "epoch": 21.59431137724551, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1338, "step": 14425 }, { "epoch": 21.595808383233532, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1389, "step": 14426 }, { "epoch": 21.597305389221557, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1392, "step": 14427 }, { "epoch": 21.59880239520958, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1379, "step": 14428 }, { "epoch": 21.600299401197606, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1361, "step": 14429 }, { "epoch": 21.601796407185628, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.136, "step": 14430 }, { "epoch": 21.603293413173652, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1406, "step": 14431 }, { "epoch": 21.604790419161677, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.131, "step": 14432 }, { "epoch": 21.606287425149702, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1303, "step": 14433 }, { "epoch": 21.607784431137723, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1402, "step": 14434 }, { "epoch": 21.60928143712575, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1348, "step": 14435 }, { "epoch": 21.610778443113773, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1351, "step": 14436 }, { "epoch": 21.612275449101798, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1268, "step": 14437 }, { "epoch": 21.61377245508982, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1359, "step": 14438 }, { "epoch": 21.615269461077844, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1398, "step": 14439 }, { "epoch": 21.61676646706587, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1351, "step": 14440 }, { "epoch": 21.618263473053894, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1342, "step": 14441 }, { "epoch": 21.619760479041915, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1339, "step": 14442 }, { "epoch": 21.62125748502994, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1338, "step": 14443 }, { "epoch": 21.622754491017965, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1367, "step": 14444 }, { "epoch": 21.62425149700599, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1381, "step": 14445 }, { "epoch": 21.62574850299401, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1335, "step": 14446 }, { "epoch": 21.627245508982035, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.134, "step": 14447 }, { "epoch": 21.62874251497006, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1328, "step": 14448 }, { "epoch": 21.630239520958085, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1354, "step": 14449 }, { "epoch": 21.631736526946106, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1363, "step": 14450 }, { "epoch": 21.63323353293413, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1336, "step": 14451 }, { "epoch": 21.634730538922156, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.129, "step": 14452 }, { "epoch": 21.63622754491018, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1339, "step": 14453 }, { "epoch": 21.637724550898202, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1338, "step": 14454 }, { "epoch": 21.639221556886227, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1329, "step": 14455 }, { "epoch": 21.64071856287425, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1317, "step": 14456 }, { "epoch": 21.642215568862277, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1272, "step": 14457 }, { "epoch": 21.643712574850298, "grad_norm": 0.1162109375, "learning_rate": 0.0008, "loss": 1.1305, "step": 14458 }, { "epoch": 21.645209580838323, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1353, "step": 14459 }, { "epoch": 21.646706586826348, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1354, "step": 14460 }, { "epoch": 21.648203592814372, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.137, "step": 14461 }, { "epoch": 21.649700598802394, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1361, "step": 14462 }, { "epoch": 21.65119760479042, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1337, "step": 14463 }, { "epoch": 21.652694610778443, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1419, "step": 14464 }, { "epoch": 21.654191616766468, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1333, "step": 14465 }, { "epoch": 21.65568862275449, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1322, "step": 14466 }, { "epoch": 21.657185628742514, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1352, "step": 14467 }, { "epoch": 21.65868263473054, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 14468 }, { "epoch": 21.660179640718564, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1388, "step": 14469 }, { "epoch": 21.66167664670659, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1368, "step": 14470 }, { "epoch": 21.66317365269461, "grad_norm": 0.158203125, "learning_rate": 0.0008, "loss": 1.1296, "step": 14471 }, { "epoch": 21.664670658682635, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.128, "step": 14472 }, { "epoch": 21.66616766467066, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1354, "step": 14473 }, { "epoch": 21.66766467065868, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1361, "step": 14474 }, { "epoch": 21.669161676646706, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1376, "step": 14475 }, { "epoch": 21.67065868263473, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.129, "step": 14476 }, { "epoch": 21.672155688622755, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1415, "step": 14477 }, { "epoch": 21.67365269461078, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.133, "step": 14478 }, { "epoch": 21.6751497005988, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1296, "step": 14479 }, { "epoch": 21.676646706586826, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1322, "step": 14480 }, { "epoch": 21.67814371257485, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1383, "step": 14481 }, { "epoch": 21.679640718562876, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1301, "step": 14482 }, { "epoch": 21.681137724550897, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1306, "step": 14483 }, { "epoch": 21.682634730538922, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1321, "step": 14484 }, { "epoch": 21.684131736526947, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1296, "step": 14485 }, { "epoch": 21.68562874251497, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1372, "step": 14486 }, { "epoch": 21.687125748502993, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1255, "step": 14487 }, { "epoch": 21.688622754491018, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1344, "step": 14488 }, { "epoch": 21.690119760479043, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1297, "step": 14489 }, { "epoch": 21.691616766467067, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1279, "step": 14490 }, { "epoch": 21.69311377245509, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1378, "step": 14491 }, { "epoch": 21.694610778443113, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1322, "step": 14492 }, { "epoch": 21.69610778443114, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1337, "step": 14493 }, { "epoch": 21.697604790419163, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1359, "step": 14494 }, { "epoch": 21.699101796407184, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1344, "step": 14495 }, { "epoch": 21.70059880239521, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.133, "step": 14496 }, { "epoch": 21.702095808383234, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1375, "step": 14497 }, { "epoch": 21.70359281437126, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1297, "step": 14498 }, { "epoch": 21.70508982035928, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1338, "step": 14499 }, { "epoch": 21.706586826347305, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1366, "step": 14500 }, { "epoch": 21.70808383233533, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1324, "step": 14501 }, { "epoch": 21.709580838323355, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1376, "step": 14502 }, { "epoch": 21.711077844311376, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1316, "step": 14503 }, { "epoch": 21.7125748502994, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1284, "step": 14504 }, { "epoch": 21.714071856287426, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1402, "step": 14505 }, { "epoch": 21.71556886227545, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1328, "step": 14506 }, { "epoch": 21.71706586826347, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.131, "step": 14507 }, { "epoch": 21.718562874251496, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1343, "step": 14508 }, { "epoch": 21.72005988023952, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.135, "step": 14509 }, { "epoch": 21.721556886227546, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1272, "step": 14510 }, { "epoch": 21.723053892215567, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1382, "step": 14511 }, { "epoch": 21.724550898203592, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.135, "step": 14512 }, { "epoch": 21.726047904191617, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1413, "step": 14513 }, { "epoch": 21.727544910179642, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1352, "step": 14514 }, { "epoch": 21.729041916167663, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1375, "step": 14515 }, { "epoch": 21.730538922155688, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.133, "step": 14516 }, { "epoch": 21.732035928143713, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1341, "step": 14517 }, { "epoch": 21.733532934131738, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1345, "step": 14518 }, { "epoch": 21.73502994011976, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.13, "step": 14519 }, { "epoch": 21.736526946107784, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1357, "step": 14520 }, { "epoch": 21.73802395209581, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1289, "step": 14521 }, { "epoch": 21.739520958083833, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1395, "step": 14522 }, { "epoch": 21.741017964071855, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1389, "step": 14523 }, { "epoch": 21.74251497005988, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1404, "step": 14524 }, { "epoch": 21.744011976047904, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1357, "step": 14525 }, { "epoch": 21.74550898203593, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.131, "step": 14526 }, { "epoch": 21.74700598802395, "grad_norm": 0.15234375, "learning_rate": 0.0008, "loss": 1.1349, "step": 14527 }, { "epoch": 21.748502994011975, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1339, "step": 14528 }, { "epoch": 21.75, "grad_norm": 0.150390625, "learning_rate": 0.0008, "loss": 1.1315, "step": 14529 }, { "epoch": 21.751497005988025, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1337, "step": 14530 }, { "epoch": 21.75299401197605, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1335, "step": 14531 }, { "epoch": 21.75449101796407, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1313, "step": 14532 }, { "epoch": 21.755988023952096, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1319, "step": 14533 }, { "epoch": 21.75748502994012, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1335, "step": 14534 }, { "epoch": 21.758982035928145, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.134, "step": 14535 }, { "epoch": 21.760479041916167, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1378, "step": 14536 }, { "epoch": 21.76197604790419, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1358, "step": 14537 }, { "epoch": 21.763473053892216, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.132, "step": 14538 }, { "epoch": 21.76497005988024, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1345, "step": 14539 }, { "epoch": 21.766467065868262, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1312, "step": 14540 }, { "epoch": 21.767964071856287, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1354, "step": 14541 }, { "epoch": 21.769461077844312, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1389, "step": 14542 }, { "epoch": 21.770958083832337, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1262, "step": 14543 }, { "epoch": 21.772455089820358, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1306, "step": 14544 }, { "epoch": 21.773952095808383, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1316, "step": 14545 }, { "epoch": 21.775449101796408, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1355, "step": 14546 }, { "epoch": 21.776946107784433, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1359, "step": 14547 }, { "epoch": 21.778443113772454, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1333, "step": 14548 }, { "epoch": 21.77994011976048, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1284, "step": 14549 }, { "epoch": 21.781437125748504, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1354, "step": 14550 }, { "epoch": 21.78293413173653, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1355, "step": 14551 }, { "epoch": 21.78443113772455, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1275, "step": 14552 }, { "epoch": 21.785928143712574, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1339, "step": 14553 }, { "epoch": 21.7874251497006, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1286, "step": 14554 }, { "epoch": 21.788922155688624, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1323, "step": 14555 }, { "epoch": 21.790419161676645, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.128, "step": 14556 }, { "epoch": 21.79191616766467, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1319, "step": 14557 }, { "epoch": 21.793413173652695, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1362, "step": 14558 }, { "epoch": 21.79491017964072, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1349, "step": 14559 }, { "epoch": 21.79640718562874, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1349, "step": 14560 }, { "epoch": 21.797904191616766, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1347, "step": 14561 }, { "epoch": 21.79940119760479, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1346, "step": 14562 }, { "epoch": 21.800898203592816, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1358, "step": 14563 }, { "epoch": 21.802395209580837, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1348, "step": 14564 }, { "epoch": 21.80389221556886, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.135, "step": 14565 }, { "epoch": 21.805389221556887, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1388, "step": 14566 }, { "epoch": 21.80688622754491, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1388, "step": 14567 }, { "epoch": 21.808383233532933, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1302, "step": 14568 }, { "epoch": 21.809880239520957, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1347, "step": 14569 }, { "epoch": 21.811377245508982, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1404, "step": 14570 }, { "epoch": 21.812874251497007, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1342, "step": 14571 }, { "epoch": 21.81437125748503, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1331, "step": 14572 }, { "epoch": 21.815868263473053, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.14, "step": 14573 }, { "epoch": 21.817365269461078, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1372, "step": 14574 }, { "epoch": 21.818862275449103, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.13, "step": 14575 }, { "epoch": 21.820359281437124, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.141, "step": 14576 }, { "epoch": 21.82185628742515, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1316, "step": 14577 }, { "epoch": 21.823353293413174, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.134, "step": 14578 }, { "epoch": 21.8248502994012, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1336, "step": 14579 }, { "epoch": 21.82634730538922, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1368, "step": 14580 }, { "epoch": 21.827844311377245, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1365, "step": 14581 }, { "epoch": 21.82934131736527, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1303, "step": 14582 }, { "epoch": 21.830838323353294, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1357, "step": 14583 }, { "epoch": 21.83233532934132, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1411, "step": 14584 }, { "epoch": 21.83383233532934, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1277, "step": 14585 }, { "epoch": 21.835329341317365, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1407, "step": 14586 }, { "epoch": 21.83682634730539, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1402, "step": 14587 }, { "epoch": 21.83832335329341, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.139, "step": 14588 }, { "epoch": 21.839820359281436, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1278, "step": 14589 }, { "epoch": 21.84131736526946, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1354, "step": 14590 }, { "epoch": 21.842814371257486, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1465, "step": 14591 }, { "epoch": 21.84431137724551, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1364, "step": 14592 }, { "epoch": 21.845808383233532, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1333, "step": 14593 }, { "epoch": 21.847305389221557, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1317, "step": 14594 }, { "epoch": 21.84880239520958, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.127, "step": 14595 }, { "epoch": 21.850299401197606, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1351, "step": 14596 }, { "epoch": 21.851796407185628, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1378, "step": 14597 }, { "epoch": 21.853293413173652, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1366, "step": 14598 }, { "epoch": 21.854790419161677, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.139, "step": 14599 }, { "epoch": 21.856287425149702, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1366, "step": 14600 }, { "epoch": 21.857784431137723, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1308, "step": 14601 }, { "epoch": 21.85928143712575, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1317, "step": 14602 }, { "epoch": 21.860778443113773, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1335, "step": 14603 }, { "epoch": 21.862275449101798, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1356, "step": 14604 }, { "epoch": 21.86377245508982, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1332, "step": 14605 }, { "epoch": 21.865269461077844, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1426, "step": 14606 }, { "epoch": 21.86676646706587, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1293, "step": 14607 }, { "epoch": 21.868263473053894, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1397, "step": 14608 }, { "epoch": 21.869760479041915, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1298, "step": 14609 }, { "epoch": 21.87125748502994, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1336, "step": 14610 }, { "epoch": 21.872754491017965, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1307, "step": 14611 }, { "epoch": 21.87425149700599, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1317, "step": 14612 }, { "epoch": 21.87574850299401, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1296, "step": 14613 }, { "epoch": 21.877245508982035, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1356, "step": 14614 }, { "epoch": 21.87874251497006, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1358, "step": 14615 }, { "epoch": 21.880239520958085, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1323, "step": 14616 }, { "epoch": 21.881736526946106, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1306, "step": 14617 }, { "epoch": 21.88323353293413, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1369, "step": 14618 }, { "epoch": 21.884730538922156, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1366, "step": 14619 }, { "epoch": 21.88622754491018, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1371, "step": 14620 }, { "epoch": 21.887724550898202, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1233, "step": 14621 }, { "epoch": 21.889221556886227, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1315, "step": 14622 }, { "epoch": 21.89071856287425, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1276, "step": 14623 }, { "epoch": 21.892215568862277, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1401, "step": 14624 }, { "epoch": 21.893712574850298, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1349, "step": 14625 }, { "epoch": 21.895209580838323, "grad_norm": 0.119140625, "learning_rate": 0.0008, "loss": 1.1355, "step": 14626 }, { "epoch": 21.896706586826348, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1313, "step": 14627 }, { "epoch": 21.898203592814372, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1266, "step": 14628 }, { "epoch": 21.899700598802394, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1324, "step": 14629 }, { "epoch": 21.90119760479042, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1376, "step": 14630 }, { "epoch": 21.902694610778443, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1338, "step": 14631 }, { "epoch": 21.904191616766468, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1342, "step": 14632 }, { "epoch": 21.90568862275449, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1352, "step": 14633 }, { "epoch": 21.907185628742514, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1338, "step": 14634 }, { "epoch": 21.90868263473054, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1363, "step": 14635 }, { "epoch": 21.910179640718564, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1287, "step": 14636 }, { "epoch": 21.91167664670659, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1347, "step": 14637 }, { "epoch": 21.91317365269461, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1353, "step": 14638 }, { "epoch": 21.914670658682635, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1348, "step": 14639 }, { "epoch": 21.91616766467066, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1306, "step": 14640 }, { "epoch": 21.91766467065868, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.133, "step": 14641 }, { "epoch": 21.919161676646706, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.138, "step": 14642 }, { "epoch": 21.92065868263473, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1336, "step": 14643 }, { "epoch": 21.922155688622755, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1372, "step": 14644 }, { "epoch": 21.92365269461078, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1291, "step": 14645 }, { "epoch": 21.9251497005988, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1312, "step": 14646 }, { "epoch": 21.926646706586826, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1341, "step": 14647 }, { "epoch": 21.92814371257485, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1385, "step": 14648 }, { "epoch": 21.929640718562876, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1343, "step": 14649 }, { "epoch": 21.931137724550897, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1343, "step": 14650 }, { "epoch": 21.932634730538922, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1416, "step": 14651 }, { "epoch": 21.934131736526947, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 14652 }, { "epoch": 21.93562874251497, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1334, "step": 14653 }, { "epoch": 21.937125748502993, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1369, "step": 14654 }, { "epoch": 21.938622754491018, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1319, "step": 14655 }, { "epoch": 21.940119760479043, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1308, "step": 14656 }, { "epoch": 21.941616766467067, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1277, "step": 14657 }, { "epoch": 21.94311377245509, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1428, "step": 14658 }, { "epoch": 21.944610778443113, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1409, "step": 14659 }, { "epoch": 21.94610778443114, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1302, "step": 14660 }, { "epoch": 21.947604790419163, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.135, "step": 14661 }, { "epoch": 21.949101796407184, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1445, "step": 14662 }, { "epoch": 21.95059880239521, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1323, "step": 14663 }, { "epoch": 21.952095808383234, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1341, "step": 14664 }, { "epoch": 21.95359281437126, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1294, "step": 14665 }, { "epoch": 21.95508982035928, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1353, "step": 14666 }, { "epoch": 21.956586826347305, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1378, "step": 14667 }, { "epoch": 21.95808383233533, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1304, "step": 14668 }, { "epoch": 21.959580838323355, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1281, "step": 14669 }, { "epoch": 21.961077844311376, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1252, "step": 14670 }, { "epoch": 21.9625748502994, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1377, "step": 14671 }, { "epoch": 21.964071856287426, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1305, "step": 14672 }, { "epoch": 21.96556886227545, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1285, "step": 14673 }, { "epoch": 21.96706586826347, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1287, "step": 14674 }, { "epoch": 21.968562874251496, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1179, "step": 14675 }, { "epoch": 21.97005988023952, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1339, "step": 14676 }, { "epoch": 21.971556886227546, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1376, "step": 14677 }, { "epoch": 21.973053892215567, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1301, "step": 14678 }, { "epoch": 21.974550898203592, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1356, "step": 14679 }, { "epoch": 21.976047904191617, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1382, "step": 14680 }, { "epoch": 21.977544910179642, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1334, "step": 14681 }, { "epoch": 21.979041916167663, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1293, "step": 14682 }, { "epoch": 21.980538922155688, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1388, "step": 14683 }, { "epoch": 21.982035928143713, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1317, "step": 14684 }, { "epoch": 21.983532934131738, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1269, "step": 14685 }, { "epoch": 21.98502994011976, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1309, "step": 14686 }, { "epoch": 21.986526946107784, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1351, "step": 14687 }, { "epoch": 21.98802395209581, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1326, "step": 14688 }, { "epoch": 21.989520958083833, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1295, "step": 14689 }, { "epoch": 21.991017964071855, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1406, "step": 14690 }, { "epoch": 21.99251497005988, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1346, "step": 14691 }, { "epoch": 21.994011976047904, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1265, "step": 14692 }, { "epoch": 21.99550898203593, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1339, "step": 14693 }, { "epoch": 21.99700598802395, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1311, "step": 14694 }, { "epoch": 21.998502994011975, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.134, "step": 14695 }, { "epoch": 22.0, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.131, "step": 14696 }, { "epoch": 22.001497005988025, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1381, "step": 14697 }, { "epoch": 22.00299401197605, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1292, "step": 14698 }, { "epoch": 22.00449101796407, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1362, "step": 14699 }, { "epoch": 22.005988023952096, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1319, "step": 14700 }, { "epoch": 22.00748502994012, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1281, "step": 14701 }, { "epoch": 22.008982035928145, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.135, "step": 14702 }, { "epoch": 22.010479041916167, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1384, "step": 14703 }, { "epoch": 22.01197604790419, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.136, "step": 14704 }, { "epoch": 22.013473053892216, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1327, "step": 14705 }, { "epoch": 22.01497005988024, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1327, "step": 14706 }, { "epoch": 22.016467065868262, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1374, "step": 14707 }, { "epoch": 22.017964071856287, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1373, "step": 14708 }, { "epoch": 22.019461077844312, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1365, "step": 14709 }, { "epoch": 22.020958083832337, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1339, "step": 14710 }, { "epoch": 22.022455089820358, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1294, "step": 14711 }, { "epoch": 22.023952095808383, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1338, "step": 14712 }, { "epoch": 22.025449101796408, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1318, "step": 14713 }, { "epoch": 22.026946107784433, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1372, "step": 14714 }, { "epoch": 22.028443113772454, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1329, "step": 14715 }, { "epoch": 22.02994011976048, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1281, "step": 14716 }, { "epoch": 22.031437125748504, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1354, "step": 14717 }, { "epoch": 22.03293413173653, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1321, "step": 14718 }, { "epoch": 22.03443113772455, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1333, "step": 14719 }, { "epoch": 22.035928143712574, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1346, "step": 14720 }, { "epoch": 22.0374251497006, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1379, "step": 14721 }, { "epoch": 22.038922155688624, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1374, "step": 14722 }, { "epoch": 22.040419161676645, "grad_norm": 0.173828125, "learning_rate": 0.0008, "loss": 1.1362, "step": 14723 }, { "epoch": 22.04191616766467, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1407, "step": 14724 }, { "epoch": 22.043413173652695, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1317, "step": 14725 }, { "epoch": 22.04491017964072, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1349, "step": 14726 }, { "epoch": 22.04640718562874, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1325, "step": 14727 }, { "epoch": 22.047904191616766, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1327, "step": 14728 }, { "epoch": 22.04940119760479, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1345, "step": 14729 }, { "epoch": 22.050898203592816, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.13, "step": 14730 }, { "epoch": 22.052395209580837, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1297, "step": 14731 }, { "epoch": 22.05389221556886, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.137, "step": 14732 }, { "epoch": 22.055389221556887, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1276, "step": 14733 }, { "epoch": 22.05688622754491, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1305, "step": 14734 }, { "epoch": 22.058383233532933, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1372, "step": 14735 }, { "epoch": 22.059880239520957, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.133, "step": 14736 }, { "epoch": 22.061377245508982, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1303, "step": 14737 }, { "epoch": 22.062874251497007, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1334, "step": 14738 }, { "epoch": 22.06437125748503, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1355, "step": 14739 }, { "epoch": 22.065868263473053, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1319, "step": 14740 }, { "epoch": 22.067365269461078, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1332, "step": 14741 }, { "epoch": 22.068862275449103, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1372, "step": 14742 }, { "epoch": 22.070359281437124, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1374, "step": 14743 }, { "epoch": 22.07185628742515, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1268, "step": 14744 }, { "epoch": 22.073353293413174, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1401, "step": 14745 }, { "epoch": 22.0748502994012, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1302, "step": 14746 }, { "epoch": 22.07634730538922, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1331, "step": 14747 }, { "epoch": 22.077844311377245, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1296, "step": 14748 }, { "epoch": 22.07934131736527, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1337, "step": 14749 }, { "epoch": 22.080838323353294, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1357, "step": 14750 }, { "epoch": 22.082335329341316, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1292, "step": 14751 }, { "epoch": 22.08383233532934, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1301, "step": 14752 }, { "epoch": 22.085329341317365, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1316, "step": 14753 }, { "epoch": 22.08682634730539, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1355, "step": 14754 }, { "epoch": 22.088323353293415, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1406, "step": 14755 }, { "epoch": 22.089820359281436, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1357, "step": 14756 }, { "epoch": 22.09131736526946, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1341, "step": 14757 }, { "epoch": 22.092814371257486, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1335, "step": 14758 }, { "epoch": 22.09431137724551, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1342, "step": 14759 }, { "epoch": 22.095808383233532, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1336, "step": 14760 }, { "epoch": 22.097305389221557, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1298, "step": 14761 }, { "epoch": 22.09880239520958, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.13, "step": 14762 }, { "epoch": 22.100299401197606, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1267, "step": 14763 }, { "epoch": 22.101796407185628, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1359, "step": 14764 }, { "epoch": 22.103293413173652, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.135, "step": 14765 }, { "epoch": 22.104790419161677, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1337, "step": 14766 }, { "epoch": 22.106287425149702, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1364, "step": 14767 }, { "epoch": 22.107784431137723, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.136, "step": 14768 }, { "epoch": 22.10928143712575, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1334, "step": 14769 }, { "epoch": 22.110778443113773, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1299, "step": 14770 }, { "epoch": 22.112275449101798, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1346, "step": 14771 }, { "epoch": 22.11377245508982, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1386, "step": 14772 }, { "epoch": 22.115269461077844, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1301, "step": 14773 }, { "epoch": 22.11676646706587, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1419, "step": 14774 }, { "epoch": 22.118263473053894, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1324, "step": 14775 }, { "epoch": 22.119760479041915, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.137, "step": 14776 }, { "epoch": 22.12125748502994, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1376, "step": 14777 }, { "epoch": 22.122754491017965, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1276, "step": 14778 }, { "epoch": 22.12425149700599, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1356, "step": 14779 }, { "epoch": 22.12574850299401, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1317, "step": 14780 }, { "epoch": 22.127245508982035, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1345, "step": 14781 }, { "epoch": 22.12874251497006, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1234, "step": 14782 }, { "epoch": 22.130239520958085, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1355, "step": 14783 }, { "epoch": 22.131736526946106, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1301, "step": 14784 }, { "epoch": 22.13323353293413, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1376, "step": 14785 }, { "epoch": 22.134730538922156, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1341, "step": 14786 }, { "epoch": 22.13622754491018, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1337, "step": 14787 }, { "epoch": 22.137724550898202, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1371, "step": 14788 }, { "epoch": 22.139221556886227, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1394, "step": 14789 }, { "epoch": 22.14071856287425, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1354, "step": 14790 }, { "epoch": 22.142215568862277, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.134, "step": 14791 }, { "epoch": 22.143712574850298, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1307, "step": 14792 }, { "epoch": 22.145209580838323, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1326, "step": 14793 }, { "epoch": 22.146706586826348, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1342, "step": 14794 }, { "epoch": 22.148203592814372, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1354, "step": 14795 }, { "epoch": 22.149700598802394, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.137, "step": 14796 }, { "epoch": 22.15119760479042, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.133, "step": 14797 }, { "epoch": 22.152694610778443, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1269, "step": 14798 }, { "epoch": 22.154191616766468, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.133, "step": 14799 }, { "epoch": 22.15568862275449, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.13, "step": 14800 }, { "epoch": 22.157185628742514, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.134, "step": 14801 }, { "epoch": 22.15868263473054, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1322, "step": 14802 }, { "epoch": 22.160179640718564, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1324, "step": 14803 }, { "epoch": 22.161676646706585, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1382, "step": 14804 }, { "epoch": 22.16317365269461, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1348, "step": 14805 }, { "epoch": 22.164670658682635, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1335, "step": 14806 }, { "epoch": 22.16616766467066, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1315, "step": 14807 }, { "epoch": 22.16766467065868, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1339, "step": 14808 }, { "epoch": 22.169161676646706, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1335, "step": 14809 }, { "epoch": 22.17065868263473, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1333, "step": 14810 }, { "epoch": 22.172155688622755, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1331, "step": 14811 }, { "epoch": 22.17365269461078, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1392, "step": 14812 }, { "epoch": 22.1751497005988, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1318, "step": 14813 }, { "epoch": 22.176646706586826, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.132, "step": 14814 }, { "epoch": 22.17814371257485, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1358, "step": 14815 }, { "epoch": 22.179640718562876, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1357, "step": 14816 }, { "epoch": 22.181137724550897, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1431, "step": 14817 }, { "epoch": 22.182634730538922, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1319, "step": 14818 }, { "epoch": 22.184131736526947, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1376, "step": 14819 }, { "epoch": 22.18562874251497, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1358, "step": 14820 }, { "epoch": 22.187125748502993, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1345, "step": 14821 }, { "epoch": 22.188622754491018, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1351, "step": 14822 }, { "epoch": 22.190119760479043, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1357, "step": 14823 }, { "epoch": 22.191616766467067, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1276, "step": 14824 }, { "epoch": 22.19311377245509, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1328, "step": 14825 }, { "epoch": 22.194610778443113, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1384, "step": 14826 }, { "epoch": 22.19610778443114, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1311, "step": 14827 }, { "epoch": 22.197604790419163, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1288, "step": 14828 }, { "epoch": 22.199101796407184, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1312, "step": 14829 }, { "epoch": 22.20059880239521, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1354, "step": 14830 }, { "epoch": 22.202095808383234, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1257, "step": 14831 }, { "epoch": 22.20359281437126, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.14, "step": 14832 }, { "epoch": 22.20508982035928, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1334, "step": 14833 }, { "epoch": 22.206586826347305, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1257, "step": 14834 }, { "epoch": 22.20808383233533, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1382, "step": 14835 }, { "epoch": 22.209580838323355, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.131, "step": 14836 }, { "epoch": 22.211077844311376, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1276, "step": 14837 }, { "epoch": 22.2125748502994, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1296, "step": 14838 }, { "epoch": 22.214071856287426, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1333, "step": 14839 }, { "epoch": 22.21556886227545, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1373, "step": 14840 }, { "epoch": 22.21706586826347, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1373, "step": 14841 }, { "epoch": 22.218562874251496, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1308, "step": 14842 }, { "epoch": 22.22005988023952, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1432, "step": 14843 }, { "epoch": 22.221556886227546, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.131, "step": 14844 }, { "epoch": 22.223053892215567, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1313, "step": 14845 }, { "epoch": 22.224550898203592, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1339, "step": 14846 }, { "epoch": 22.226047904191617, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1393, "step": 14847 }, { "epoch": 22.227544910179642, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.132, "step": 14848 }, { "epoch": 22.229041916167663, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1287, "step": 14849 }, { "epoch": 22.230538922155688, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.135, "step": 14850 }, { "epoch": 22.232035928143713, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1353, "step": 14851 }, { "epoch": 22.233532934131738, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1296, "step": 14852 }, { "epoch": 22.23502994011976, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1336, "step": 14853 }, { "epoch": 22.236526946107784, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1389, "step": 14854 }, { "epoch": 22.23802395209581, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1339, "step": 14855 }, { "epoch": 22.239520958083833, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1227, "step": 14856 }, { "epoch": 22.241017964071855, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 14857 }, { "epoch": 22.24251497005988, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1288, "step": 14858 }, { "epoch": 22.244011976047904, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1385, "step": 14859 }, { "epoch": 22.24550898203593, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1312, "step": 14860 }, { "epoch": 22.24700598802395, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1311, "step": 14861 }, { "epoch": 22.248502994011975, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1314, "step": 14862 }, { "epoch": 22.25, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1315, "step": 14863 }, { "epoch": 22.251497005988025, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.129, "step": 14864 }, { "epoch": 22.25299401197605, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1339, "step": 14865 }, { "epoch": 22.25449101796407, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1296, "step": 14866 }, { "epoch": 22.255988023952096, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1335, "step": 14867 }, { "epoch": 22.25748502994012, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1315, "step": 14868 }, { "epoch": 22.258982035928145, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1254, "step": 14869 }, { "epoch": 22.260479041916167, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1289, "step": 14870 }, { "epoch": 22.26197604790419, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1331, "step": 14871 }, { "epoch": 22.263473053892216, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1414, "step": 14872 }, { "epoch": 22.26497005988024, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1309, "step": 14873 }, { "epoch": 22.266467065868262, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1294, "step": 14874 }, { "epoch": 22.267964071856287, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1342, "step": 14875 }, { "epoch": 22.269461077844312, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1288, "step": 14876 }, { "epoch": 22.270958083832337, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1377, "step": 14877 }, { "epoch": 22.272455089820358, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.141, "step": 14878 }, { "epoch": 22.273952095808383, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1333, "step": 14879 }, { "epoch": 22.275449101796408, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1336, "step": 14880 }, { "epoch": 22.276946107784433, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1353, "step": 14881 }, { "epoch": 22.278443113772454, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1316, "step": 14882 }, { "epoch": 22.27994011976048, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.138, "step": 14883 }, { "epoch": 22.281437125748504, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1315, "step": 14884 }, { "epoch": 22.28293413173653, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1353, "step": 14885 }, { "epoch": 22.28443113772455, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1296, "step": 14886 }, { "epoch": 22.285928143712574, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.131, "step": 14887 }, { "epoch": 22.2874251497006, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 14888 }, { "epoch": 22.288922155688624, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1303, "step": 14889 }, { "epoch": 22.290419161676645, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1346, "step": 14890 }, { "epoch": 22.29191616766467, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1322, "step": 14891 }, { "epoch": 22.293413173652695, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1333, "step": 14892 }, { "epoch": 22.29491017964072, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1396, "step": 14893 }, { "epoch": 22.29640718562874, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.137, "step": 14894 }, { "epoch": 22.297904191616766, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.1353, "step": 14895 }, { "epoch": 22.29940119760479, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1344, "step": 14896 }, { "epoch": 22.300898203592816, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1386, "step": 14897 }, { "epoch": 22.302395209580837, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1339, "step": 14898 }, { "epoch": 22.30389221556886, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1266, "step": 14899 }, { "epoch": 22.305389221556887, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.133, "step": 14900 }, { "epoch": 22.30688622754491, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1362, "step": 14901 }, { "epoch": 22.308383233532933, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1353, "step": 14902 }, { "epoch": 22.309880239520957, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1341, "step": 14903 }, { "epoch": 22.311377245508982, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.143, "step": 14904 }, { "epoch": 22.312874251497007, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1317, "step": 14905 }, { "epoch": 22.31437125748503, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1407, "step": 14906 }, { "epoch": 22.315868263473053, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1279, "step": 14907 }, { "epoch": 22.317365269461078, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1299, "step": 14908 }, { "epoch": 22.318862275449103, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1282, "step": 14909 }, { "epoch": 22.320359281437124, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.134, "step": 14910 }, { "epoch": 22.32185628742515, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1299, "step": 14911 }, { "epoch": 22.323353293413174, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1371, "step": 14912 }, { "epoch": 22.3248502994012, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.132, "step": 14913 }, { "epoch": 22.32634730538922, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1345, "step": 14914 }, { "epoch": 22.327844311377245, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1284, "step": 14915 }, { "epoch": 22.32934131736527, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1321, "step": 14916 }, { "epoch": 22.330838323353294, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1369, "step": 14917 }, { "epoch": 22.33233532934132, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1402, "step": 14918 }, { "epoch": 22.33383233532934, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1353, "step": 14919 }, { "epoch": 22.335329341317365, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1292, "step": 14920 }, { "epoch": 22.33682634730539, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.136, "step": 14921 }, { "epoch": 22.338323353293415, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1377, "step": 14922 }, { "epoch": 22.339820359281436, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1284, "step": 14923 }, { "epoch": 22.34131736526946, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1272, "step": 14924 }, { "epoch": 22.342814371257486, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1268, "step": 14925 }, { "epoch": 22.34431137724551, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1334, "step": 14926 }, { "epoch": 22.345808383233532, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1423, "step": 14927 }, { "epoch": 22.347305389221557, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1349, "step": 14928 }, { "epoch": 22.34880239520958, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1323, "step": 14929 }, { "epoch": 22.350299401197606, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1289, "step": 14930 }, { "epoch": 22.351796407185628, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1347, "step": 14931 }, { "epoch": 22.353293413173652, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1303, "step": 14932 }, { "epoch": 22.354790419161677, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1316, "step": 14933 }, { "epoch": 22.356287425149702, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1371, "step": 14934 }, { "epoch": 22.357784431137723, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1305, "step": 14935 }, { "epoch": 22.35928143712575, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1354, "step": 14936 }, { "epoch": 22.360778443113773, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1355, "step": 14937 }, { "epoch": 22.362275449101798, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1303, "step": 14938 }, { "epoch": 22.36377245508982, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1335, "step": 14939 }, { "epoch": 22.365269461077844, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1316, "step": 14940 }, { "epoch": 22.36676646706587, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1389, "step": 14941 }, { "epoch": 22.368263473053894, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1366, "step": 14942 }, { "epoch": 22.369760479041915, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1315, "step": 14943 }, { "epoch": 22.37125748502994, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.143, "step": 14944 }, { "epoch": 22.372754491017965, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1315, "step": 14945 }, { "epoch": 22.37425149700599, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1385, "step": 14946 }, { "epoch": 22.37574850299401, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1361, "step": 14947 }, { "epoch": 22.377245508982035, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1333, "step": 14948 }, { "epoch": 22.37874251497006, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.13, "step": 14949 }, { "epoch": 22.380239520958085, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1313, "step": 14950 }, { "epoch": 22.381736526946106, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1345, "step": 14951 }, { "epoch": 22.38323353293413, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1382, "step": 14952 }, { "epoch": 22.384730538922156, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1256, "step": 14953 }, { "epoch": 22.38622754491018, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1338, "step": 14954 }, { "epoch": 22.387724550898202, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1395, "step": 14955 }, { "epoch": 22.389221556886227, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1312, "step": 14956 }, { "epoch": 22.39071856287425, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1311, "step": 14957 }, { "epoch": 22.392215568862277, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1347, "step": 14958 }, { "epoch": 22.393712574850298, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1306, "step": 14959 }, { "epoch": 22.395209580838323, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1338, "step": 14960 }, { "epoch": 22.396706586826348, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1336, "step": 14961 }, { "epoch": 22.398203592814372, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1315, "step": 14962 }, { "epoch": 22.399700598802394, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1339, "step": 14963 }, { "epoch": 22.40119760479042, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1309, "step": 14964 }, { "epoch": 22.402694610778443, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1349, "step": 14965 }, { "epoch": 22.404191616766468, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1368, "step": 14966 }, { "epoch": 22.40568862275449, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1294, "step": 14967 }, { "epoch": 22.407185628742514, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1299, "step": 14968 }, { "epoch": 22.40868263473054, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1369, "step": 14969 }, { "epoch": 22.410179640718564, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1342, "step": 14970 }, { "epoch": 22.411676646706585, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1343, "step": 14971 }, { "epoch": 22.41317365269461, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1352, "step": 14972 }, { "epoch": 22.414670658682635, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1261, "step": 14973 }, { "epoch": 22.41616766467066, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.132, "step": 14974 }, { "epoch": 22.41766467065868, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1312, "step": 14975 }, { "epoch": 22.419161676646706, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1257, "step": 14976 }, { "epoch": 22.42065868263473, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.135, "step": 14977 }, { "epoch": 22.422155688622755, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1213, "step": 14978 }, { "epoch": 22.42365269461078, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1316, "step": 14979 }, { "epoch": 22.4251497005988, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1331, "step": 14980 }, { "epoch": 22.426646706586826, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.136, "step": 14981 }, { "epoch": 22.42814371257485, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1358, "step": 14982 }, { "epoch": 22.429640718562876, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1344, "step": 14983 }, { "epoch": 22.431137724550897, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1277, "step": 14984 }, { "epoch": 22.432634730538922, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1361, "step": 14985 }, { "epoch": 22.434131736526947, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1282, "step": 14986 }, { "epoch": 22.43562874251497, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1309, "step": 14987 }, { "epoch": 22.437125748502993, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1381, "step": 14988 }, { "epoch": 22.438622754491018, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1325, "step": 14989 }, { "epoch": 22.440119760479043, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1343, "step": 14990 }, { "epoch": 22.441616766467067, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1356, "step": 14991 }, { "epoch": 22.44311377245509, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1322, "step": 14992 }, { "epoch": 22.444610778443113, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1349, "step": 14993 }, { "epoch": 22.44610778443114, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1276, "step": 14994 }, { "epoch": 22.447604790419163, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1305, "step": 14995 }, { "epoch": 22.449101796407184, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1375, "step": 14996 }, { "epoch": 22.45059880239521, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.131, "step": 14997 }, { "epoch": 22.452095808383234, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1289, "step": 14998 }, { "epoch": 22.45359281437126, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1377, "step": 14999 }, { "epoch": 22.45508982035928, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1386, "step": 15000 }, { "epoch": 22.456586826347305, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1304, "step": 15001 }, { "epoch": 22.45808383233533, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.132, "step": 15002 }, { "epoch": 22.459580838323355, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1414, "step": 15003 }, { "epoch": 22.461077844311376, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1383, "step": 15004 }, { "epoch": 22.4625748502994, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1327, "step": 15005 }, { "epoch": 22.464071856287426, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1351, "step": 15006 }, { "epoch": 22.46556886227545, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1382, "step": 15007 }, { "epoch": 22.46706586826347, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1358, "step": 15008 }, { "epoch": 22.468562874251496, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1333, "step": 15009 }, { "epoch": 22.47005988023952, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1362, "step": 15010 }, { "epoch": 22.471556886227546, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1388, "step": 15011 }, { "epoch": 22.473053892215567, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1334, "step": 15012 }, { "epoch": 22.474550898203592, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1287, "step": 15013 }, { "epoch": 22.476047904191617, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1321, "step": 15014 }, { "epoch": 22.477544910179642, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1344, "step": 15015 }, { "epoch": 22.479041916167663, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1379, "step": 15016 }, { "epoch": 22.480538922155688, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.133, "step": 15017 }, { "epoch": 22.482035928143713, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1342, "step": 15018 }, { "epoch": 22.483532934131738, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1268, "step": 15019 }, { "epoch": 22.48502994011976, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1276, "step": 15020 }, { "epoch": 22.486526946107784, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1295, "step": 15021 }, { "epoch": 22.48802395209581, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1324, "step": 15022 }, { "epoch": 22.489520958083833, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1313, "step": 15023 }, { "epoch": 22.491017964071855, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1324, "step": 15024 }, { "epoch": 22.49251497005988, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1261, "step": 15025 }, { "epoch": 22.494011976047904, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1343, "step": 15026 }, { "epoch": 22.49550898203593, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.135, "step": 15027 }, { "epoch": 22.49700598802395, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1371, "step": 15028 }, { "epoch": 22.498502994011975, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1335, "step": 15029 }, { "epoch": 22.5, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1396, "step": 15030 }, { "epoch": 22.501497005988025, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1307, "step": 15031 }, { "epoch": 22.50299401197605, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1303, "step": 15032 }, { "epoch": 22.50449101796407, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1267, "step": 15033 }, { "epoch": 22.505988023952096, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1371, "step": 15034 }, { "epoch": 22.50748502994012, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1319, "step": 15035 }, { "epoch": 22.508982035928145, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1303, "step": 15036 }, { "epoch": 22.510479041916167, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1333, "step": 15037 }, { "epoch": 22.51197604790419, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1319, "step": 15038 }, { "epoch": 22.513473053892216, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1301, "step": 15039 }, { "epoch": 22.51497005988024, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1304, "step": 15040 }, { "epoch": 22.516467065868262, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1317, "step": 15041 }, { "epoch": 22.517964071856287, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1315, "step": 15042 }, { "epoch": 22.519461077844312, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1301, "step": 15043 }, { "epoch": 22.520958083832337, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.129, "step": 15044 }, { "epoch": 22.522455089820358, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1308, "step": 15045 }, { "epoch": 22.523952095808383, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1312, "step": 15046 }, { "epoch": 22.525449101796408, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1284, "step": 15047 }, { "epoch": 22.526946107784433, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1301, "step": 15048 }, { "epoch": 22.528443113772454, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1295, "step": 15049 }, { "epoch": 22.52994011976048, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1306, "step": 15050 }, { "epoch": 22.531437125748504, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1297, "step": 15051 }, { "epoch": 22.53293413173653, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1341, "step": 15052 }, { "epoch": 22.53443113772455, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1396, "step": 15053 }, { "epoch": 22.535928143712574, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1393, "step": 15054 }, { "epoch": 22.5374251497006, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1286, "step": 15055 }, { "epoch": 22.538922155688624, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1337, "step": 15056 }, { "epoch": 22.540419161676645, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1321, "step": 15057 }, { "epoch": 22.54191616766467, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1376, "step": 15058 }, { "epoch": 22.543413173652695, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1301, "step": 15059 }, { "epoch": 22.54491017964072, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1344, "step": 15060 }, { "epoch": 22.54640718562874, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1278, "step": 15061 }, { "epoch": 22.547904191616766, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1399, "step": 15062 }, { "epoch": 22.54940119760479, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1305, "step": 15063 }, { "epoch": 22.550898203592816, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1341, "step": 15064 }, { "epoch": 22.552395209580837, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1372, "step": 15065 }, { "epoch": 22.55389221556886, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1346, "step": 15066 }, { "epoch": 22.555389221556887, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1345, "step": 15067 }, { "epoch": 22.55688622754491, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1335, "step": 15068 }, { "epoch": 22.558383233532933, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1287, "step": 15069 }, { "epoch": 22.559880239520957, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.131, "step": 15070 }, { "epoch": 22.561377245508982, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.133, "step": 15071 }, { "epoch": 22.562874251497007, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1308, "step": 15072 }, { "epoch": 22.56437125748503, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1352, "step": 15073 }, { "epoch": 22.565868263473053, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1331, "step": 15074 }, { "epoch": 22.567365269461078, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1362, "step": 15075 }, { "epoch": 22.568862275449103, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1297, "step": 15076 }, { "epoch": 22.570359281437124, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1358, "step": 15077 }, { "epoch": 22.57185628742515, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1327, "step": 15078 }, { "epoch": 22.573353293413174, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1312, "step": 15079 }, { "epoch": 22.5748502994012, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.132, "step": 15080 }, { "epoch": 22.57634730538922, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1371, "step": 15081 }, { "epoch": 22.577844311377245, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1391, "step": 15082 }, { "epoch": 22.57934131736527, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.133, "step": 15083 }, { "epoch": 22.580838323353294, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1296, "step": 15084 }, { "epoch": 22.58233532934132, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1311, "step": 15085 }, { "epoch": 22.58383233532934, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1297, "step": 15086 }, { "epoch": 22.585329341317365, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1308, "step": 15087 }, { "epoch": 22.58682634730539, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1337, "step": 15088 }, { "epoch": 22.58832335329341, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1356, "step": 15089 }, { "epoch": 22.589820359281436, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.131, "step": 15090 }, { "epoch": 22.59131736526946, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1357, "step": 15091 }, { "epoch": 22.592814371257486, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1388, "step": 15092 }, { "epoch": 22.59431137724551, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1339, "step": 15093 }, { "epoch": 22.595808383233532, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1239, "step": 15094 }, { "epoch": 22.597305389221557, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1298, "step": 15095 }, { "epoch": 22.59880239520958, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.13, "step": 15096 }, { "epoch": 22.600299401197606, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1323, "step": 15097 }, { "epoch": 22.601796407185628, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1314, "step": 15098 }, { "epoch": 22.603293413173652, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.137, "step": 15099 }, { "epoch": 22.604790419161677, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1411, "step": 15100 }, { "epoch": 22.606287425149702, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1265, "step": 15101 }, { "epoch": 22.607784431137723, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1304, "step": 15102 }, { "epoch": 22.60928143712575, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1364, "step": 15103 }, { "epoch": 22.610778443113773, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1269, "step": 15104 }, { "epoch": 22.612275449101798, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1289, "step": 15105 }, { "epoch": 22.61377245508982, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1304, "step": 15106 }, { "epoch": 22.615269461077844, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1347, "step": 15107 }, { "epoch": 22.61676646706587, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1295, "step": 15108 }, { "epoch": 22.618263473053894, "grad_norm": 0.14453125, "learning_rate": 0.0008, "loss": 1.1372, "step": 15109 }, { "epoch": 22.619760479041915, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1321, "step": 15110 }, { "epoch": 22.62125748502994, "grad_norm": 0.20703125, "learning_rate": 0.0008, "loss": 1.1378, "step": 15111 }, { "epoch": 22.622754491017965, "grad_norm": 0.138671875, "learning_rate": 0.0008, "loss": 1.1314, "step": 15112 }, { "epoch": 22.62425149700599, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1352, "step": 15113 }, { "epoch": 22.62574850299401, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1398, "step": 15114 }, { "epoch": 22.627245508982035, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1292, "step": 15115 }, { "epoch": 22.62874251497006, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1299, "step": 15116 }, { "epoch": 22.630239520958085, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1358, "step": 15117 }, { "epoch": 22.631736526946106, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1262, "step": 15118 }, { "epoch": 22.63323353293413, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1365, "step": 15119 }, { "epoch": 22.634730538922156, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1304, "step": 15120 }, { "epoch": 22.63622754491018, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1322, "step": 15121 }, { "epoch": 22.637724550898202, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1316, "step": 15122 }, { "epoch": 22.639221556886227, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.128, "step": 15123 }, { "epoch": 22.64071856287425, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1298, "step": 15124 }, { "epoch": 22.642215568862277, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.135, "step": 15125 }, { "epoch": 22.643712574850298, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1294, "step": 15126 }, { "epoch": 22.645209580838323, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.134, "step": 15127 }, { "epoch": 22.646706586826348, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1294, "step": 15128 }, { "epoch": 22.648203592814372, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1362, "step": 15129 }, { "epoch": 22.649700598802394, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1249, "step": 15130 }, { "epoch": 22.65119760479042, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1338, "step": 15131 }, { "epoch": 22.652694610778443, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1371, "step": 15132 }, { "epoch": 22.654191616766468, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1307, "step": 15133 }, { "epoch": 22.65568862275449, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1312, "step": 15134 }, { "epoch": 22.657185628742514, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1327, "step": 15135 }, { "epoch": 22.65868263473054, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1349, "step": 15136 }, { "epoch": 22.660179640718564, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1373, "step": 15137 }, { "epoch": 22.66167664670659, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1367, "step": 15138 }, { "epoch": 22.66317365269461, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1316, "step": 15139 }, { "epoch": 22.664670658682635, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1355, "step": 15140 }, { "epoch": 22.66616766467066, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1353, "step": 15141 }, { "epoch": 22.66766467065868, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1303, "step": 15142 }, { "epoch": 22.669161676646706, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1324, "step": 15143 }, { "epoch": 22.67065868263473, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1345, "step": 15144 }, { "epoch": 22.672155688622755, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1373, "step": 15145 }, { "epoch": 22.67365269461078, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1373, "step": 15146 }, { "epoch": 22.6751497005988, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1265, "step": 15147 }, { "epoch": 22.676646706586826, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1359, "step": 15148 }, { "epoch": 22.67814371257485, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1311, "step": 15149 }, { "epoch": 22.679640718562876, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1347, "step": 15150 }, { "epoch": 22.681137724550897, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1272, "step": 15151 }, { "epoch": 22.682634730538922, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1367, "step": 15152 }, { "epoch": 22.684131736526947, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1349, "step": 15153 }, { "epoch": 22.68562874251497, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1339, "step": 15154 }, { "epoch": 22.687125748502993, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1311, "step": 15155 }, { "epoch": 22.688622754491018, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1295, "step": 15156 }, { "epoch": 22.690119760479043, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1369, "step": 15157 }, { "epoch": 22.691616766467067, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1335, "step": 15158 }, { "epoch": 22.69311377245509, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1356, "step": 15159 }, { "epoch": 22.694610778443113, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1323, "step": 15160 }, { "epoch": 22.69610778443114, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1276, "step": 15161 }, { "epoch": 22.697604790419163, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1327, "step": 15162 }, { "epoch": 22.699101796407184, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1306, "step": 15163 }, { "epoch": 22.70059880239521, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1345, "step": 15164 }, { "epoch": 22.702095808383234, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1307, "step": 15165 }, { "epoch": 22.70359281437126, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1383, "step": 15166 }, { "epoch": 22.70508982035928, "grad_norm": 0.1435546875, "learning_rate": 0.0008, "loss": 1.1319, "step": 15167 }, { "epoch": 22.706586826347305, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1225, "step": 15168 }, { "epoch": 22.70808383233533, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1372, "step": 15169 }, { "epoch": 22.709580838323355, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1326, "step": 15170 }, { "epoch": 22.711077844311376, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1356, "step": 15171 }, { "epoch": 22.7125748502994, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1375, "step": 15172 }, { "epoch": 22.714071856287426, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1329, "step": 15173 }, { "epoch": 22.71556886227545, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1354, "step": 15174 }, { "epoch": 22.71706586826347, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1249, "step": 15175 }, { "epoch": 22.718562874251496, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1308, "step": 15176 }, { "epoch": 22.72005988023952, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1336, "step": 15177 }, { "epoch": 22.721556886227546, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1331, "step": 15178 }, { "epoch": 22.723053892215567, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1303, "step": 15179 }, { "epoch": 22.724550898203592, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1353, "step": 15180 }, { "epoch": 22.726047904191617, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1279, "step": 15181 }, { "epoch": 22.727544910179642, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1366, "step": 15182 }, { "epoch": 22.729041916167663, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1361, "step": 15183 }, { "epoch": 22.730538922155688, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1352, "step": 15184 }, { "epoch": 22.732035928143713, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.13, "step": 15185 }, { "epoch": 22.733532934131738, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1311, "step": 15186 }, { "epoch": 22.73502994011976, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1379, "step": 15187 }, { "epoch": 22.736526946107784, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1327, "step": 15188 }, { "epoch": 22.73802395209581, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1338, "step": 15189 }, { "epoch": 22.739520958083833, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1419, "step": 15190 }, { "epoch": 22.741017964071855, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1311, "step": 15191 }, { "epoch": 22.74251497005988, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1303, "step": 15192 }, { "epoch": 22.744011976047904, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1265, "step": 15193 }, { "epoch": 22.74550898203593, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1332, "step": 15194 }, { "epoch": 22.74700598802395, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1324, "step": 15195 }, { "epoch": 22.748502994011975, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1316, "step": 15196 }, { "epoch": 22.75, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1366, "step": 15197 }, { "epoch": 22.751497005988025, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1204, "step": 15198 }, { "epoch": 22.75299401197605, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1288, "step": 15199 }, { "epoch": 22.75449101796407, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1344, "step": 15200 }, { "epoch": 22.755988023952096, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1341, "step": 15201 }, { "epoch": 22.75748502994012, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.131, "step": 15202 }, { "epoch": 22.758982035928145, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1325, "step": 15203 }, { "epoch": 22.760479041916167, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1377, "step": 15204 }, { "epoch": 22.76197604790419, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1299, "step": 15205 }, { "epoch": 22.763473053892216, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1375, "step": 15206 }, { "epoch": 22.76497005988024, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1393, "step": 15207 }, { "epoch": 22.766467065868262, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1355, "step": 15208 }, { "epoch": 22.767964071856287, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1342, "step": 15209 }, { "epoch": 22.769461077844312, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1337, "step": 15210 }, { "epoch": 22.770958083832337, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.13, "step": 15211 }, { "epoch": 22.772455089820358, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1268, "step": 15212 }, { "epoch": 22.773952095808383, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.135, "step": 15213 }, { "epoch": 22.775449101796408, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1414, "step": 15214 }, { "epoch": 22.776946107784433, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1282, "step": 15215 }, { "epoch": 22.778443113772454, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1328, "step": 15216 }, { "epoch": 22.77994011976048, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1336, "step": 15217 }, { "epoch": 22.781437125748504, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1288, "step": 15218 }, { "epoch": 22.78293413173653, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1285, "step": 15219 }, { "epoch": 22.78443113772455, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.131, "step": 15220 }, { "epoch": 22.785928143712574, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1283, "step": 15221 }, { "epoch": 22.7874251497006, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1288, "step": 15222 }, { "epoch": 22.788922155688624, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1325, "step": 15223 }, { "epoch": 22.790419161676645, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1283, "step": 15224 }, { "epoch": 22.79191616766467, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1356, "step": 15225 }, { "epoch": 22.793413173652695, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1294, "step": 15226 }, { "epoch": 22.79491017964072, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1373, "step": 15227 }, { "epoch": 22.79640718562874, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1363, "step": 15228 }, { "epoch": 22.797904191616766, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1316, "step": 15229 }, { "epoch": 22.79940119760479, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1385, "step": 15230 }, { "epoch": 22.800898203592816, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.138, "step": 15231 }, { "epoch": 22.802395209580837, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1359, "step": 15232 }, { "epoch": 22.80389221556886, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1366, "step": 15233 }, { "epoch": 22.805389221556887, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1328, "step": 15234 }, { "epoch": 22.80688622754491, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1356, "step": 15235 }, { "epoch": 22.808383233532933, "grad_norm": 0.11572265625, "learning_rate": 0.0008, "loss": 1.1355, "step": 15236 }, { "epoch": 22.809880239520957, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1333, "step": 15237 }, { "epoch": 22.811377245508982, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1242, "step": 15238 }, { "epoch": 22.812874251497007, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1356, "step": 15239 }, { "epoch": 22.81437125748503, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1332, "step": 15240 }, { "epoch": 22.815868263473053, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1376, "step": 15241 }, { "epoch": 22.817365269461078, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1333, "step": 15242 }, { "epoch": 22.818862275449103, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1316, "step": 15243 }, { "epoch": 22.820359281437124, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1326, "step": 15244 }, { "epoch": 22.82185628742515, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1346, "step": 15245 }, { "epoch": 22.823353293413174, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1316, "step": 15246 }, { "epoch": 22.8248502994012, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1382, "step": 15247 }, { "epoch": 22.82634730538922, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1287, "step": 15248 }, { "epoch": 22.827844311377245, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1283, "step": 15249 }, { "epoch": 22.82934131736527, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1347, "step": 15250 }, { "epoch": 22.830838323353294, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1351, "step": 15251 }, { "epoch": 22.83233532934132, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1354, "step": 15252 }, { "epoch": 22.83383233532934, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1382, "step": 15253 }, { "epoch": 22.835329341317365, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1364, "step": 15254 }, { "epoch": 22.83682634730539, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1277, "step": 15255 }, { "epoch": 22.83832335329341, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1369, "step": 15256 }, { "epoch": 22.839820359281436, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1416, "step": 15257 }, { "epoch": 22.84131736526946, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1338, "step": 15258 }, { "epoch": 22.842814371257486, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1281, "step": 15259 }, { "epoch": 22.84431137724551, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1328, "step": 15260 }, { "epoch": 22.845808383233532, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1324, "step": 15261 }, { "epoch": 22.847305389221557, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1318, "step": 15262 }, { "epoch": 22.84880239520958, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1262, "step": 15263 }, { "epoch": 22.850299401197606, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1317, "step": 15264 }, { "epoch": 22.851796407185628, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1283, "step": 15265 }, { "epoch": 22.853293413173652, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1346, "step": 15266 }, { "epoch": 22.854790419161677, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1366, "step": 15267 }, { "epoch": 22.856287425149702, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1362, "step": 15268 }, { "epoch": 22.857784431137723, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1352, "step": 15269 }, { "epoch": 22.85928143712575, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1362, "step": 15270 }, { "epoch": 22.860778443113773, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1384, "step": 15271 }, { "epoch": 22.862275449101798, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1374, "step": 15272 }, { "epoch": 22.86377245508982, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1349, "step": 15273 }, { "epoch": 22.865269461077844, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.131, "step": 15274 }, { "epoch": 22.86676646706587, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.13, "step": 15275 }, { "epoch": 22.868263473053894, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1335, "step": 15276 }, { "epoch": 22.869760479041915, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1325, "step": 15277 }, { "epoch": 22.87125748502994, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1293, "step": 15278 }, { "epoch": 22.872754491017965, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1383, "step": 15279 }, { "epoch": 22.87425149700599, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1269, "step": 15280 }, { "epoch": 22.87574850299401, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1311, "step": 15281 }, { "epoch": 22.877245508982035, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1354, "step": 15282 }, { "epoch": 22.87874251497006, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.133, "step": 15283 }, { "epoch": 22.880239520958085, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1376, "step": 15284 }, { "epoch": 22.881736526946106, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1333, "step": 15285 }, { "epoch": 22.88323353293413, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1351, "step": 15286 }, { "epoch": 22.884730538922156, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1336, "step": 15287 }, { "epoch": 22.88622754491018, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1314, "step": 15288 }, { "epoch": 22.887724550898202, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1378, "step": 15289 }, { "epoch": 22.889221556886227, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1338, "step": 15290 }, { "epoch": 22.89071856287425, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1329, "step": 15291 }, { "epoch": 22.892215568862277, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1353, "step": 15292 }, { "epoch": 22.893712574850298, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1297, "step": 15293 }, { "epoch": 22.895209580838323, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1329, "step": 15294 }, { "epoch": 22.896706586826348, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1289, "step": 15295 }, { "epoch": 22.898203592814372, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1276, "step": 15296 }, { "epoch": 22.899700598802394, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1294, "step": 15297 }, { "epoch": 22.90119760479042, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1309, "step": 15298 }, { "epoch": 22.902694610778443, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1356, "step": 15299 }, { "epoch": 22.904191616766468, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1353, "step": 15300 }, { "epoch": 22.90568862275449, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1327, "step": 15301 }, { "epoch": 22.907185628742514, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1277, "step": 15302 }, { "epoch": 22.90868263473054, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1415, "step": 15303 }, { "epoch": 22.910179640718564, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1361, "step": 15304 }, { "epoch": 22.91167664670659, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1367, "step": 15305 }, { "epoch": 22.91317365269461, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1346, "step": 15306 }, { "epoch": 22.914670658682635, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1301, "step": 15307 }, { "epoch": 22.91616766467066, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1288, "step": 15308 }, { "epoch": 22.91766467065868, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1281, "step": 15309 }, { "epoch": 22.919161676646706, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1338, "step": 15310 }, { "epoch": 22.92065868263473, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1341, "step": 15311 }, { "epoch": 22.922155688622755, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1317, "step": 15312 }, { "epoch": 22.92365269461078, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1327, "step": 15313 }, { "epoch": 22.9251497005988, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1323, "step": 15314 }, { "epoch": 22.926646706586826, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1326, "step": 15315 }, { "epoch": 22.92814371257485, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1381, "step": 15316 }, { "epoch": 22.929640718562876, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1321, "step": 15317 }, { "epoch": 22.931137724550897, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1391, "step": 15318 }, { "epoch": 22.932634730538922, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.142, "step": 15319 }, { "epoch": 22.934131736526947, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.137, "step": 15320 }, { "epoch": 22.93562874251497, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1316, "step": 15321 }, { "epoch": 22.937125748502993, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1342, "step": 15322 }, { "epoch": 22.938622754491018, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.132, "step": 15323 }, { "epoch": 22.940119760479043, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1395, "step": 15324 }, { "epoch": 22.941616766467067, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1393, "step": 15325 }, { "epoch": 22.94311377245509, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1398, "step": 15326 }, { "epoch": 22.944610778443113, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1396, "step": 15327 }, { "epoch": 22.94610778443114, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1272, "step": 15328 }, { "epoch": 22.947604790419163, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1295, "step": 15329 }, { "epoch": 22.949101796407184, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1345, "step": 15330 }, { "epoch": 22.95059880239521, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1353, "step": 15331 }, { "epoch": 22.952095808383234, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1338, "step": 15332 }, { "epoch": 22.95359281437126, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1324, "step": 15333 }, { "epoch": 22.95508982035928, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1342, "step": 15334 }, { "epoch": 22.956586826347305, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1376, "step": 15335 }, { "epoch": 22.95808383233533, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.13, "step": 15336 }, { "epoch": 22.959580838323355, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.133, "step": 15337 }, { "epoch": 22.961077844311376, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1389, "step": 15338 }, { "epoch": 22.9625748502994, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1371, "step": 15339 }, { "epoch": 22.964071856287426, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1326, "step": 15340 }, { "epoch": 22.96556886227545, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1316, "step": 15341 }, { "epoch": 22.96706586826347, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1276, "step": 15342 }, { "epoch": 22.968562874251496, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1317, "step": 15343 }, { "epoch": 22.97005988023952, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1329, "step": 15344 }, { "epoch": 22.971556886227546, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1384, "step": 15345 }, { "epoch": 22.973053892215567, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.128, "step": 15346 }, { "epoch": 22.974550898203592, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1363, "step": 15347 }, { "epoch": 22.976047904191617, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1343, "step": 15348 }, { "epoch": 22.977544910179642, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1287, "step": 15349 }, { "epoch": 22.979041916167663, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1346, "step": 15350 }, { "epoch": 22.980538922155688, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.128, "step": 15351 }, { "epoch": 22.982035928143713, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1337, "step": 15352 }, { "epoch": 22.983532934131738, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1339, "step": 15353 }, { "epoch": 22.98502994011976, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1337, "step": 15354 }, { "epoch": 22.986526946107784, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1385, "step": 15355 }, { "epoch": 22.98802395209581, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1365, "step": 15356 }, { "epoch": 22.989520958083833, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1344, "step": 15357 }, { "epoch": 22.991017964071855, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1312, "step": 15358 }, { "epoch": 22.99251497005988, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1278, "step": 15359 }, { "epoch": 22.994011976047904, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1393, "step": 15360 }, { "epoch": 22.99550898203593, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1367, "step": 15361 }, { "epoch": 22.99700598802395, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1389, "step": 15362 }, { "epoch": 22.998502994011975, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1262, "step": 15363 }, { "epoch": 23.0, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1374, "step": 15364 }, { "epoch": 23.001497005988025, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1261, "step": 15365 }, { "epoch": 23.00299401197605, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1319, "step": 15366 }, { "epoch": 23.00449101796407, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.124, "step": 15367 }, { "epoch": 23.005988023952096, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1335, "step": 15368 }, { "epoch": 23.00748502994012, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.132, "step": 15369 }, { "epoch": 23.008982035928145, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1353, "step": 15370 }, { "epoch": 23.010479041916167, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1326, "step": 15371 }, { "epoch": 23.01197604790419, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1359, "step": 15372 }, { "epoch": 23.013473053892216, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1352, "step": 15373 }, { "epoch": 23.01497005988024, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1383, "step": 15374 }, { "epoch": 23.016467065868262, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1359, "step": 15375 }, { "epoch": 23.017964071856287, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1362, "step": 15376 }, { "epoch": 23.019461077844312, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1261, "step": 15377 }, { "epoch": 23.020958083832337, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1355, "step": 15378 }, { "epoch": 23.022455089820358, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1298, "step": 15379 }, { "epoch": 23.023952095808383, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1367, "step": 15380 }, { "epoch": 23.025449101796408, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1372, "step": 15381 }, { "epoch": 23.026946107784433, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1341, "step": 15382 }, { "epoch": 23.028443113772454, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1326, "step": 15383 }, { "epoch": 23.02994011976048, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1284, "step": 15384 }, { "epoch": 23.031437125748504, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1341, "step": 15385 }, { "epoch": 23.03293413173653, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1319, "step": 15386 }, { "epoch": 23.03443113772455, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1317, "step": 15387 }, { "epoch": 23.035928143712574, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.138, "step": 15388 }, { "epoch": 23.0374251497006, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1354, "step": 15389 }, { "epoch": 23.038922155688624, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1345, "step": 15390 }, { "epoch": 23.040419161676645, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1275, "step": 15391 }, { "epoch": 23.04191616766467, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1301, "step": 15392 }, { "epoch": 23.043413173652695, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1302, "step": 15393 }, { "epoch": 23.04491017964072, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1274, "step": 15394 }, { "epoch": 23.04640718562874, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1287, "step": 15395 }, { "epoch": 23.047904191616766, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.136, "step": 15396 }, { "epoch": 23.04940119760479, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1347, "step": 15397 }, { "epoch": 23.050898203592816, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1332, "step": 15398 }, { "epoch": 23.052395209580837, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1389, "step": 15399 }, { "epoch": 23.05389221556886, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1302, "step": 15400 }, { "epoch": 23.055389221556887, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1312, "step": 15401 }, { "epoch": 23.05688622754491, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1363, "step": 15402 }, { "epoch": 23.058383233532933, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1324, "step": 15403 }, { "epoch": 23.059880239520957, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1267, "step": 15404 }, { "epoch": 23.061377245508982, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1292, "step": 15405 }, { "epoch": 23.062874251497007, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1382, "step": 15406 }, { "epoch": 23.06437125748503, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1303, "step": 15407 }, { "epoch": 23.065868263473053, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1401, "step": 15408 }, { "epoch": 23.067365269461078, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1338, "step": 15409 }, { "epoch": 23.068862275449103, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1317, "step": 15410 }, { "epoch": 23.070359281437124, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1279, "step": 15411 }, { "epoch": 23.07185628742515, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1312, "step": 15412 }, { "epoch": 23.073353293413174, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1322, "step": 15413 }, { "epoch": 23.0748502994012, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1303, "step": 15414 }, { "epoch": 23.07634730538922, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1301, "step": 15415 }, { "epoch": 23.077844311377245, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1365, "step": 15416 }, { "epoch": 23.07934131736527, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1342, "step": 15417 }, { "epoch": 23.080838323353294, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1319, "step": 15418 }, { "epoch": 23.082335329341316, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 15419 }, { "epoch": 23.08383233532934, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1336, "step": 15420 }, { "epoch": 23.085329341317365, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1314, "step": 15421 }, { "epoch": 23.08682634730539, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1313, "step": 15422 }, { "epoch": 23.088323353293415, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1352, "step": 15423 }, { "epoch": 23.089820359281436, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1309, "step": 15424 }, { "epoch": 23.09131736526946, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1313, "step": 15425 }, { "epoch": 23.092814371257486, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1295, "step": 15426 }, { "epoch": 23.09431137724551, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1314, "step": 15427 }, { "epoch": 23.095808383233532, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1369, "step": 15428 }, { "epoch": 23.097305389221557, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1335, "step": 15429 }, { "epoch": 23.09880239520958, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1287, "step": 15430 }, { "epoch": 23.100299401197606, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1347, "step": 15431 }, { "epoch": 23.101796407185628, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1352, "step": 15432 }, { "epoch": 23.103293413173652, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1325, "step": 15433 }, { "epoch": 23.104790419161677, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1299, "step": 15434 }, { "epoch": 23.106287425149702, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1274, "step": 15435 }, { "epoch": 23.107784431137723, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 15436 }, { "epoch": 23.10928143712575, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1309, "step": 15437 }, { "epoch": 23.110778443113773, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1275, "step": 15438 }, { "epoch": 23.112275449101798, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1356, "step": 15439 }, { "epoch": 23.11377245508982, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1274, "step": 15440 }, { "epoch": 23.115269461077844, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.131, "step": 15441 }, { "epoch": 23.11676646706587, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1302, "step": 15442 }, { "epoch": 23.118263473053894, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1338, "step": 15443 }, { "epoch": 23.119760479041915, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1343, "step": 15444 }, { "epoch": 23.12125748502994, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1315, "step": 15445 }, { "epoch": 23.122754491017965, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.13, "step": 15446 }, { "epoch": 23.12425149700599, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1315, "step": 15447 }, { "epoch": 23.12574850299401, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1278, "step": 15448 }, { "epoch": 23.127245508982035, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1354, "step": 15449 }, { "epoch": 23.12874251497006, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1252, "step": 15450 }, { "epoch": 23.130239520958085, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.133, "step": 15451 }, { "epoch": 23.131736526946106, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1334, "step": 15452 }, { "epoch": 23.13323353293413, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1335, "step": 15453 }, { "epoch": 23.134730538922156, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1274, "step": 15454 }, { "epoch": 23.13622754491018, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1331, "step": 15455 }, { "epoch": 23.137724550898202, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1339, "step": 15456 }, { "epoch": 23.139221556886227, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1327, "step": 15457 }, { "epoch": 23.14071856287425, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.127, "step": 15458 }, { "epoch": 23.142215568862277, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1246, "step": 15459 }, { "epoch": 23.143712574850298, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1345, "step": 15460 }, { "epoch": 23.145209580838323, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1317, "step": 15461 }, { "epoch": 23.146706586826348, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1331, "step": 15462 }, { "epoch": 23.148203592814372, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1321, "step": 15463 }, { "epoch": 23.149700598802394, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1367, "step": 15464 }, { "epoch": 23.15119760479042, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1285, "step": 15465 }, { "epoch": 23.152694610778443, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1324, "step": 15466 }, { "epoch": 23.154191616766468, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1342, "step": 15467 }, { "epoch": 23.15568862275449, "grad_norm": 0.1484375, "learning_rate": 0.0008, "loss": 1.133, "step": 15468 }, { "epoch": 23.157185628742514, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1352, "step": 15469 }, { "epoch": 23.15868263473054, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1317, "step": 15470 }, { "epoch": 23.160179640718564, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1399, "step": 15471 }, { "epoch": 23.161676646706585, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1278, "step": 15472 }, { "epoch": 23.16317365269461, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1311, "step": 15473 }, { "epoch": 23.164670658682635, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1334, "step": 15474 }, { "epoch": 23.16616766467066, "grad_norm": 0.12158203125, "learning_rate": 0.0008, "loss": 1.1291, "step": 15475 }, { "epoch": 23.16766467065868, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1331, "step": 15476 }, { "epoch": 23.169161676646706, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1309, "step": 15477 }, { "epoch": 23.17065868263473, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1347, "step": 15478 }, { "epoch": 23.172155688622755, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1317, "step": 15479 }, { "epoch": 23.17365269461078, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1366, "step": 15480 }, { "epoch": 23.1751497005988, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1303, "step": 15481 }, { "epoch": 23.176646706586826, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1223, "step": 15482 }, { "epoch": 23.17814371257485, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1333, "step": 15483 }, { "epoch": 23.179640718562876, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1343, "step": 15484 }, { "epoch": 23.181137724550897, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1345, "step": 15485 }, { "epoch": 23.182634730538922, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1356, "step": 15486 }, { "epoch": 23.184131736526947, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1365, "step": 15487 }, { "epoch": 23.18562874251497, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1415, "step": 15488 }, { "epoch": 23.187125748502993, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1299, "step": 15489 }, { "epoch": 23.188622754491018, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1289, "step": 15490 }, { "epoch": 23.190119760479043, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.14, "step": 15491 }, { "epoch": 23.191616766467067, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1387, "step": 15492 }, { "epoch": 23.19311377245509, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1281, "step": 15493 }, { "epoch": 23.194610778443113, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1343, "step": 15494 }, { "epoch": 23.19610778443114, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1311, "step": 15495 }, { "epoch": 23.197604790419163, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1329, "step": 15496 }, { "epoch": 23.199101796407184, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1286, "step": 15497 }, { "epoch": 23.20059880239521, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1361, "step": 15498 }, { "epoch": 23.202095808383234, "grad_norm": 0.14453125, "learning_rate": 0.0008, "loss": 1.135, "step": 15499 }, { "epoch": 23.20359281437126, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1255, "step": 15500 }, { "epoch": 23.20508982035928, "grad_norm": 0.10888671875, "learning_rate": 0.0008, "loss": 1.1337, "step": 15501 }, { "epoch": 23.206586826347305, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1337, "step": 15502 }, { "epoch": 23.20808383233533, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1358, "step": 15503 }, { "epoch": 23.209580838323355, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1342, "step": 15504 }, { "epoch": 23.211077844311376, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1351, "step": 15505 }, { "epoch": 23.2125748502994, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1329, "step": 15506 }, { "epoch": 23.214071856287426, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1293, "step": 15507 }, { "epoch": 23.21556886227545, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1268, "step": 15508 }, { "epoch": 23.21706586826347, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1384, "step": 15509 }, { "epoch": 23.218562874251496, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1287, "step": 15510 }, { "epoch": 23.22005988023952, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1353, "step": 15511 }, { "epoch": 23.221556886227546, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1316, "step": 15512 }, { "epoch": 23.223053892215567, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1378, "step": 15513 }, { "epoch": 23.224550898203592, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1263, "step": 15514 }, { "epoch": 23.226047904191617, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1328, "step": 15515 }, { "epoch": 23.227544910179642, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1291, "step": 15516 }, { "epoch": 23.229041916167663, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1334, "step": 15517 }, { "epoch": 23.230538922155688, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1292, "step": 15518 }, { "epoch": 23.232035928143713, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1306, "step": 15519 }, { "epoch": 23.233532934131738, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1298, "step": 15520 }, { "epoch": 23.23502994011976, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.135, "step": 15521 }, { "epoch": 23.236526946107784, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1333, "step": 15522 }, { "epoch": 23.23802395209581, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.134, "step": 15523 }, { "epoch": 23.239520958083833, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.129, "step": 15524 }, { "epoch": 23.241017964071855, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1319, "step": 15525 }, { "epoch": 23.24251497005988, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1392, "step": 15526 }, { "epoch": 23.244011976047904, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1304, "step": 15527 }, { "epoch": 23.24550898203593, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1388, "step": 15528 }, { "epoch": 23.24700598802395, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1343, "step": 15529 }, { "epoch": 23.248502994011975, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 15530 }, { "epoch": 23.25, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1338, "step": 15531 }, { "epoch": 23.251497005988025, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1322, "step": 15532 }, { "epoch": 23.25299401197605, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1287, "step": 15533 }, { "epoch": 23.25449101796407, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1365, "step": 15534 }, { "epoch": 23.255988023952096, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1318, "step": 15535 }, { "epoch": 23.25748502994012, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.132, "step": 15536 }, { "epoch": 23.258982035928145, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1346, "step": 15537 }, { "epoch": 23.260479041916167, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1318, "step": 15538 }, { "epoch": 23.26197604790419, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1341, "step": 15539 }, { "epoch": 23.263473053892216, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1291, "step": 15540 }, { "epoch": 23.26497005988024, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1301, "step": 15541 }, { "epoch": 23.266467065868262, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1313, "step": 15542 }, { "epoch": 23.267964071856287, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 15543 }, { "epoch": 23.269461077844312, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1338, "step": 15544 }, { "epoch": 23.270958083832337, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1318, "step": 15545 }, { "epoch": 23.272455089820358, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.135, "step": 15546 }, { "epoch": 23.273952095808383, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1324, "step": 15547 }, { "epoch": 23.275449101796408, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1328, "step": 15548 }, { "epoch": 23.276946107784433, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1398, "step": 15549 }, { "epoch": 23.278443113772454, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1304, "step": 15550 }, { "epoch": 23.27994011976048, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1331, "step": 15551 }, { "epoch": 23.281437125748504, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1316, "step": 15552 }, { "epoch": 23.28293413173653, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1361, "step": 15553 }, { "epoch": 23.28443113772455, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1306, "step": 15554 }, { "epoch": 23.285928143712574, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1289, "step": 15555 }, { "epoch": 23.2874251497006, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1264, "step": 15556 }, { "epoch": 23.288922155688624, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1364, "step": 15557 }, { "epoch": 23.290419161676645, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.123, "step": 15558 }, { "epoch": 23.29191616766467, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1358, "step": 15559 }, { "epoch": 23.293413173652695, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1327, "step": 15560 }, { "epoch": 23.29491017964072, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1245, "step": 15561 }, { "epoch": 23.29640718562874, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1285, "step": 15562 }, { "epoch": 23.297904191616766, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1358, "step": 15563 }, { "epoch": 23.29940119760479, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1385, "step": 15564 }, { "epoch": 23.300898203592816, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1333, "step": 15565 }, { "epoch": 23.302395209580837, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1333, "step": 15566 }, { "epoch": 23.30389221556886, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1358, "step": 15567 }, { "epoch": 23.305389221556887, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1343, "step": 15568 }, { "epoch": 23.30688622754491, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1335, "step": 15569 }, { "epoch": 23.308383233532933, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1312, "step": 15570 }, { "epoch": 23.309880239520957, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1261, "step": 15571 }, { "epoch": 23.311377245508982, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1294, "step": 15572 }, { "epoch": 23.312874251497007, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1335, "step": 15573 }, { "epoch": 23.31437125748503, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1319, "step": 15574 }, { "epoch": 23.315868263473053, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.135, "step": 15575 }, { "epoch": 23.317365269461078, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1359, "step": 15576 }, { "epoch": 23.318862275449103, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1386, "step": 15577 }, { "epoch": 23.320359281437124, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1261, "step": 15578 }, { "epoch": 23.32185628742515, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.125, "step": 15579 }, { "epoch": 23.323353293413174, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1281, "step": 15580 }, { "epoch": 23.3248502994012, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1362, "step": 15581 }, { "epoch": 23.32634730538922, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1299, "step": 15582 }, { "epoch": 23.327844311377245, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.1334, "step": 15583 }, { "epoch": 23.32934131736527, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1325, "step": 15584 }, { "epoch": 23.330838323353294, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1277, "step": 15585 }, { "epoch": 23.33233532934132, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1378, "step": 15586 }, { "epoch": 23.33383233532934, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1281, "step": 15587 }, { "epoch": 23.335329341317365, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1399, "step": 15588 }, { "epoch": 23.33682634730539, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1301, "step": 15589 }, { "epoch": 23.338323353293415, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1282, "step": 15590 }, { "epoch": 23.339820359281436, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1339, "step": 15591 }, { "epoch": 23.34131736526946, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1308, "step": 15592 }, { "epoch": 23.342814371257486, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1358, "step": 15593 }, { "epoch": 23.34431137724551, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1289, "step": 15594 }, { "epoch": 23.345808383233532, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1274, "step": 15595 }, { "epoch": 23.347305389221557, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1314, "step": 15596 }, { "epoch": 23.34880239520958, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1277, "step": 15597 }, { "epoch": 23.350299401197606, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1386, "step": 15598 }, { "epoch": 23.351796407185628, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1366, "step": 15599 }, { "epoch": 23.353293413173652, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1317, "step": 15600 }, { "epoch": 23.354790419161677, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1275, "step": 15601 }, { "epoch": 23.356287425149702, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.127, "step": 15602 }, { "epoch": 23.357784431137723, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1338, "step": 15603 }, { "epoch": 23.35928143712575, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1329, "step": 15604 }, { "epoch": 23.360778443113773, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1277, "step": 15605 }, { "epoch": 23.362275449101798, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1311, "step": 15606 }, { "epoch": 23.36377245508982, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1327, "step": 15607 }, { "epoch": 23.365269461077844, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1323, "step": 15608 }, { "epoch": 23.36676646706587, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1315, "step": 15609 }, { "epoch": 23.368263473053894, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1398, "step": 15610 }, { "epoch": 23.369760479041915, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1298, "step": 15611 }, { "epoch": 23.37125748502994, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1311, "step": 15612 }, { "epoch": 23.372754491017965, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1295, "step": 15613 }, { "epoch": 23.37425149700599, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1311, "step": 15614 }, { "epoch": 23.37574850299401, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1349, "step": 15615 }, { "epoch": 23.377245508982035, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1326, "step": 15616 }, { "epoch": 23.37874251497006, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1301, "step": 15617 }, { "epoch": 23.380239520958085, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1323, "step": 15618 }, { "epoch": 23.381736526946106, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1292, "step": 15619 }, { "epoch": 23.38323353293413, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1296, "step": 15620 }, { "epoch": 23.384730538922156, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1343, "step": 15621 }, { "epoch": 23.38622754491018, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1329, "step": 15622 }, { "epoch": 23.387724550898202, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1308, "step": 15623 }, { "epoch": 23.389221556886227, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1371, "step": 15624 }, { "epoch": 23.39071856287425, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1302, "step": 15625 }, { "epoch": 23.392215568862277, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1363, "step": 15626 }, { "epoch": 23.393712574850298, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.137, "step": 15627 }, { "epoch": 23.395209580838323, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1337, "step": 15628 }, { "epoch": 23.396706586826348, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1389, "step": 15629 }, { "epoch": 23.398203592814372, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1357, "step": 15630 }, { "epoch": 23.399700598802394, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1298, "step": 15631 }, { "epoch": 23.40119760479042, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1276, "step": 15632 }, { "epoch": 23.402694610778443, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1377, "step": 15633 }, { "epoch": 23.404191616766468, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1353, "step": 15634 }, { "epoch": 23.40568862275449, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1318, "step": 15635 }, { "epoch": 23.407185628742514, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1316, "step": 15636 }, { "epoch": 23.40868263473054, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1369, "step": 15637 }, { "epoch": 23.410179640718564, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1335, "step": 15638 }, { "epoch": 23.411676646706585, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1413, "step": 15639 }, { "epoch": 23.41317365269461, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1293, "step": 15640 }, { "epoch": 23.414670658682635, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1313, "step": 15641 }, { "epoch": 23.41616766467066, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1239, "step": 15642 }, { "epoch": 23.41766467065868, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1396, "step": 15643 }, { "epoch": 23.419161676646706, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1355, "step": 15644 }, { "epoch": 23.42065868263473, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1335, "step": 15645 }, { "epoch": 23.422155688622755, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1312, "step": 15646 }, { "epoch": 23.42365269461078, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1301, "step": 15647 }, { "epoch": 23.4251497005988, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1317, "step": 15648 }, { "epoch": 23.426646706586826, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1359, "step": 15649 }, { "epoch": 23.42814371257485, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1241, "step": 15650 }, { "epoch": 23.429640718562876, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1286, "step": 15651 }, { "epoch": 23.431137724550897, "grad_norm": 0.125, "learning_rate": 0.0008, "loss": 1.1341, "step": 15652 }, { "epoch": 23.432634730538922, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1299, "step": 15653 }, { "epoch": 23.434131736526947, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1308, "step": 15654 }, { "epoch": 23.43562874251497, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1333, "step": 15655 }, { "epoch": 23.437125748502993, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1347, "step": 15656 }, { "epoch": 23.438622754491018, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1351, "step": 15657 }, { "epoch": 23.440119760479043, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1254, "step": 15658 }, { "epoch": 23.441616766467067, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1326, "step": 15659 }, { "epoch": 23.44311377245509, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1302, "step": 15660 }, { "epoch": 23.444610778443113, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1273, "step": 15661 }, { "epoch": 23.44610778443114, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1342, "step": 15662 }, { "epoch": 23.447604790419163, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1327, "step": 15663 }, { "epoch": 23.449101796407184, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1358, "step": 15664 }, { "epoch": 23.45059880239521, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1307, "step": 15665 }, { "epoch": 23.452095808383234, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.136, "step": 15666 }, { "epoch": 23.45359281437126, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1361, "step": 15667 }, { "epoch": 23.45508982035928, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.131, "step": 15668 }, { "epoch": 23.456586826347305, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1298, "step": 15669 }, { "epoch": 23.45808383233533, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.127, "step": 15670 }, { "epoch": 23.459580838323355, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1307, "step": 15671 }, { "epoch": 23.461077844311376, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1337, "step": 15672 }, { "epoch": 23.4625748502994, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1326, "step": 15673 }, { "epoch": 23.464071856287426, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1348, "step": 15674 }, { "epoch": 23.46556886227545, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1306, "step": 15675 }, { "epoch": 23.46706586826347, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1291, "step": 15676 }, { "epoch": 23.468562874251496, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1343, "step": 15677 }, { "epoch": 23.47005988023952, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1291, "step": 15678 }, { "epoch": 23.471556886227546, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1291, "step": 15679 }, { "epoch": 23.473053892215567, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1321, "step": 15680 }, { "epoch": 23.474550898203592, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1308, "step": 15681 }, { "epoch": 23.476047904191617, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1339, "step": 15682 }, { "epoch": 23.477544910179642, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.134, "step": 15683 }, { "epoch": 23.479041916167663, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1299, "step": 15684 }, { "epoch": 23.480538922155688, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1355, "step": 15685 }, { "epoch": 23.482035928143713, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1375, "step": 15686 }, { "epoch": 23.483532934131738, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1274, "step": 15687 }, { "epoch": 23.48502994011976, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1279, "step": 15688 }, { "epoch": 23.486526946107784, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1293, "step": 15689 }, { "epoch": 23.48802395209581, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.132, "step": 15690 }, { "epoch": 23.489520958083833, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1261, "step": 15691 }, { "epoch": 23.491017964071855, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1369, "step": 15692 }, { "epoch": 23.49251497005988, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1259, "step": 15693 }, { "epoch": 23.494011976047904, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1244, "step": 15694 }, { "epoch": 23.49550898203593, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1301, "step": 15695 }, { "epoch": 23.49700598802395, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.136, "step": 15696 }, { "epoch": 23.498502994011975, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1285, "step": 15697 }, { "epoch": 23.5, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1413, "step": 15698 }, { "epoch": 23.501497005988025, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1337, "step": 15699 }, { "epoch": 23.50299401197605, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1341, "step": 15700 }, { "epoch": 23.50449101796407, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1375, "step": 15701 }, { "epoch": 23.505988023952096, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1341, "step": 15702 }, { "epoch": 23.50748502994012, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1302, "step": 15703 }, { "epoch": 23.508982035928145, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1331, "step": 15704 }, { "epoch": 23.510479041916167, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1316, "step": 15705 }, { "epoch": 23.51197604790419, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1273, "step": 15706 }, { "epoch": 23.513473053892216, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1328, "step": 15707 }, { "epoch": 23.51497005988024, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1363, "step": 15708 }, { "epoch": 23.516467065868262, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1303, "step": 15709 }, { "epoch": 23.517964071856287, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1352, "step": 15710 }, { "epoch": 23.519461077844312, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1311, "step": 15711 }, { "epoch": 23.520958083832337, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1328, "step": 15712 }, { "epoch": 23.522455089820358, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1362, "step": 15713 }, { "epoch": 23.523952095808383, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1314, "step": 15714 }, { "epoch": 23.525449101796408, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1262, "step": 15715 }, { "epoch": 23.526946107784433, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.136, "step": 15716 }, { "epoch": 23.528443113772454, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1344, "step": 15717 }, { "epoch": 23.52994011976048, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1349, "step": 15718 }, { "epoch": 23.531437125748504, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1311, "step": 15719 }, { "epoch": 23.53293413173653, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1333, "step": 15720 }, { "epoch": 23.53443113772455, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1274, "step": 15721 }, { "epoch": 23.535928143712574, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1317, "step": 15722 }, { "epoch": 23.5374251497006, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1277, "step": 15723 }, { "epoch": 23.538922155688624, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1313, "step": 15724 }, { "epoch": 23.540419161676645, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1385, "step": 15725 }, { "epoch": 23.54191616766467, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1287, "step": 15726 }, { "epoch": 23.543413173652695, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1271, "step": 15727 }, { "epoch": 23.54491017964072, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1347, "step": 15728 }, { "epoch": 23.54640718562874, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1363, "step": 15729 }, { "epoch": 23.547904191616766, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1342, "step": 15730 }, { "epoch": 23.54940119760479, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1385, "step": 15731 }, { "epoch": 23.550898203592816, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1357, "step": 15732 }, { "epoch": 23.552395209580837, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.134, "step": 15733 }, { "epoch": 23.55389221556886, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1328, "step": 15734 }, { "epoch": 23.555389221556887, "grad_norm": 0.05029296875, "learning_rate": 0.0008, "loss": 1.1303, "step": 15735 }, { "epoch": 23.55688622754491, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1342, "step": 15736 }, { "epoch": 23.558383233532933, "grad_norm": 0.0498046875, "learning_rate": 0.0008, "loss": 1.1365, "step": 15737 }, { "epoch": 23.559880239520957, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1319, "step": 15738 }, { "epoch": 23.561377245508982, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1309, "step": 15739 }, { "epoch": 23.562874251497007, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1312, "step": 15740 }, { "epoch": 23.56437125748503, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1347, "step": 15741 }, { "epoch": 23.565868263473053, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1345, "step": 15742 }, { "epoch": 23.567365269461078, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1298, "step": 15743 }, { "epoch": 23.568862275449103, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1236, "step": 15744 }, { "epoch": 23.570359281437124, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1256, "step": 15745 }, { "epoch": 23.57185628742515, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1369, "step": 15746 }, { "epoch": 23.573353293413174, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1337, "step": 15747 }, { "epoch": 23.5748502994012, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1292, "step": 15748 }, { "epoch": 23.57634730538922, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1398, "step": 15749 }, { "epoch": 23.577844311377245, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1321, "step": 15750 }, { "epoch": 23.57934131736527, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1296, "step": 15751 }, { "epoch": 23.580838323353294, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1376, "step": 15752 }, { "epoch": 23.58233532934132, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1323, "step": 15753 }, { "epoch": 23.58383233532934, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.134, "step": 15754 }, { "epoch": 23.585329341317365, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.127, "step": 15755 }, { "epoch": 23.58682634730539, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1326, "step": 15756 }, { "epoch": 23.58832335329341, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1395, "step": 15757 }, { "epoch": 23.589820359281436, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1308, "step": 15758 }, { "epoch": 23.59131736526946, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1302, "step": 15759 }, { "epoch": 23.592814371257486, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.134, "step": 15760 }, { "epoch": 23.59431137724551, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1342, "step": 15761 }, { "epoch": 23.595808383233532, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1322, "step": 15762 }, { "epoch": 23.597305389221557, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1276, "step": 15763 }, { "epoch": 23.59880239520958, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.131, "step": 15764 }, { "epoch": 23.600299401197606, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.126, "step": 15765 }, { "epoch": 23.601796407185628, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1279, "step": 15766 }, { "epoch": 23.603293413173652, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1353, "step": 15767 }, { "epoch": 23.604790419161677, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1281, "step": 15768 }, { "epoch": 23.606287425149702, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1372, "step": 15769 }, { "epoch": 23.607784431137723, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1336, "step": 15770 }, { "epoch": 23.60928143712575, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1356, "step": 15771 }, { "epoch": 23.610778443113773, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1276, "step": 15772 }, { "epoch": 23.612275449101798, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1352, "step": 15773 }, { "epoch": 23.61377245508982, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1317, "step": 15774 }, { "epoch": 23.615269461077844, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1284, "step": 15775 }, { "epoch": 23.61676646706587, "grad_norm": 0.126953125, "learning_rate": 0.0008, "loss": 1.1327, "step": 15776 }, { "epoch": 23.618263473053894, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1307, "step": 15777 }, { "epoch": 23.619760479041915, "grad_norm": 0.134765625, "learning_rate": 0.0008, "loss": 1.1337, "step": 15778 }, { "epoch": 23.62125748502994, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1372, "step": 15779 }, { "epoch": 23.622754491017965, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1265, "step": 15780 }, { "epoch": 23.62425149700599, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1289, "step": 15781 }, { "epoch": 23.62574850299401, "grad_norm": 0.140625, "learning_rate": 0.0008, "loss": 1.1368, "step": 15782 }, { "epoch": 23.627245508982035, "grad_norm": 0.1328125, "learning_rate": 0.0008, "loss": 1.1395, "step": 15783 }, { "epoch": 23.62874251497006, "grad_norm": 0.287109375, "learning_rate": 0.0008, "loss": 1.1331, "step": 15784 }, { "epoch": 23.630239520958085, "grad_norm": 0.19921875, "learning_rate": 0.0008, "loss": 1.1357, "step": 15785 }, { "epoch": 23.631736526946106, "grad_norm": 0.11669921875, "learning_rate": 0.0008, "loss": 1.1308, "step": 15786 }, { "epoch": 23.63323353293413, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1364, "step": 15787 }, { "epoch": 23.634730538922156, "grad_norm": 0.1875, "learning_rate": 0.0008, "loss": 1.1334, "step": 15788 }, { "epoch": 23.63622754491018, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.127, "step": 15789 }, { "epoch": 23.637724550898202, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1243, "step": 15790 }, { "epoch": 23.639221556886227, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1314, "step": 15791 }, { "epoch": 23.64071856287425, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1366, "step": 15792 }, { "epoch": 23.642215568862277, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1288, "step": 15793 }, { "epoch": 23.643712574850298, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1387, "step": 15794 }, { "epoch": 23.645209580838323, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1368, "step": 15795 }, { "epoch": 23.646706586826348, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.134, "step": 15796 }, { "epoch": 23.648203592814372, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.14, "step": 15797 }, { "epoch": 23.649700598802394, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1385, "step": 15798 }, { "epoch": 23.65119760479042, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1263, "step": 15799 }, { "epoch": 23.652694610778443, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1407, "step": 15800 }, { "epoch": 23.654191616766468, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1337, "step": 15801 }, { "epoch": 23.65568862275449, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1291, "step": 15802 }, { "epoch": 23.657185628742514, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1332, "step": 15803 }, { "epoch": 23.65868263473054, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.131, "step": 15804 }, { "epoch": 23.660179640718564, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1324, "step": 15805 }, { "epoch": 23.66167664670659, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1364, "step": 15806 }, { "epoch": 23.66317365269461, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1282, "step": 15807 }, { "epoch": 23.664670658682635, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1288, "step": 15808 }, { "epoch": 23.66616766467066, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1341, "step": 15809 }, { "epoch": 23.66766467065868, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1344, "step": 15810 }, { "epoch": 23.669161676646706, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.133, "step": 15811 }, { "epoch": 23.67065868263473, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1335, "step": 15812 }, { "epoch": 23.672155688622755, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1297, "step": 15813 }, { "epoch": 23.67365269461078, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1305, "step": 15814 }, { "epoch": 23.6751497005988, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1261, "step": 15815 }, { "epoch": 23.676646706586826, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1229, "step": 15816 }, { "epoch": 23.67814371257485, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1305, "step": 15817 }, { "epoch": 23.679640718562876, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.135, "step": 15818 }, { "epoch": 23.681137724550897, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1292, "step": 15819 }, { "epoch": 23.682634730538922, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1328, "step": 15820 }, { "epoch": 23.684131736526947, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.135, "step": 15821 }, { "epoch": 23.68562874251497, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1337, "step": 15822 }, { "epoch": 23.687125748502993, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1286, "step": 15823 }, { "epoch": 23.688622754491018, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1308, "step": 15824 }, { "epoch": 23.690119760479043, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1314, "step": 15825 }, { "epoch": 23.691616766467067, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1246, "step": 15826 }, { "epoch": 23.69311377245509, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.13, "step": 15827 }, { "epoch": 23.694610778443113, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1305, "step": 15828 }, { "epoch": 23.69610778443114, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1343, "step": 15829 }, { "epoch": 23.697604790419163, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.13, "step": 15830 }, { "epoch": 23.699101796407184, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1256, "step": 15831 }, { "epoch": 23.70059880239521, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1304, "step": 15832 }, { "epoch": 23.702095808383234, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1308, "step": 15833 }, { "epoch": 23.70359281437126, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1326, "step": 15834 }, { "epoch": 23.70508982035928, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.129, "step": 15835 }, { "epoch": 23.706586826347305, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.13, "step": 15836 }, { "epoch": 23.70808383233533, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1276, "step": 15837 }, { "epoch": 23.709580838323355, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1329, "step": 15838 }, { "epoch": 23.711077844311376, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1364, "step": 15839 }, { "epoch": 23.7125748502994, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1303, "step": 15840 }, { "epoch": 23.714071856287426, "grad_norm": 0.05078125, "learning_rate": 0.0008, "loss": 1.1369, "step": 15841 }, { "epoch": 23.71556886227545, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1222, "step": 15842 }, { "epoch": 23.71706586826347, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1353, "step": 15843 }, { "epoch": 23.718562874251496, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1308, "step": 15844 }, { "epoch": 23.72005988023952, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1358, "step": 15845 }, { "epoch": 23.721556886227546, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.13, "step": 15846 }, { "epoch": 23.723053892215567, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1387, "step": 15847 }, { "epoch": 23.724550898203592, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1354, "step": 15848 }, { "epoch": 23.726047904191617, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1294, "step": 15849 }, { "epoch": 23.727544910179642, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.128, "step": 15850 }, { "epoch": 23.729041916167663, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1334, "step": 15851 }, { "epoch": 23.730538922155688, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1352, "step": 15852 }, { "epoch": 23.732035928143713, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1293, "step": 15853 }, { "epoch": 23.733532934131738, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1358, "step": 15854 }, { "epoch": 23.73502994011976, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1297, "step": 15855 }, { "epoch": 23.736526946107784, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1291, "step": 15856 }, { "epoch": 23.73802395209581, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1295, "step": 15857 }, { "epoch": 23.739520958083833, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1281, "step": 15858 }, { "epoch": 23.741017964071855, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1415, "step": 15859 }, { "epoch": 23.74251497005988, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1318, "step": 15860 }, { "epoch": 23.744011976047904, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.129, "step": 15861 }, { "epoch": 23.74550898203593, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1375, "step": 15862 }, { "epoch": 23.74700598802395, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.139, "step": 15863 }, { "epoch": 23.748502994011975, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1339, "step": 15864 }, { "epoch": 23.75, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1331, "step": 15865 }, { "epoch": 23.751497005988025, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1363, "step": 15866 }, { "epoch": 23.75299401197605, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1372, "step": 15867 }, { "epoch": 23.75449101796407, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1267, "step": 15868 }, { "epoch": 23.755988023952096, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1287, "step": 15869 }, { "epoch": 23.75748502994012, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1378, "step": 15870 }, { "epoch": 23.758982035928145, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1358, "step": 15871 }, { "epoch": 23.760479041916167, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1371, "step": 15872 }, { "epoch": 23.76197604790419, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1247, "step": 15873 }, { "epoch": 23.763473053892216, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1412, "step": 15874 }, { "epoch": 23.76497005988024, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.137, "step": 15875 }, { "epoch": 23.766467065868262, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1281, "step": 15876 }, { "epoch": 23.767964071856287, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1264, "step": 15877 }, { "epoch": 23.769461077844312, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1275, "step": 15878 }, { "epoch": 23.770958083832337, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1343, "step": 15879 }, { "epoch": 23.772455089820358, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1327, "step": 15880 }, { "epoch": 23.773952095808383, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1335, "step": 15881 }, { "epoch": 23.775449101796408, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1335, "step": 15882 }, { "epoch": 23.776946107784433, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1264, "step": 15883 }, { "epoch": 23.778443113772454, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1305, "step": 15884 }, { "epoch": 23.77994011976048, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.13, "step": 15885 }, { "epoch": 23.781437125748504, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1359, "step": 15886 }, { "epoch": 23.78293413173653, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1405, "step": 15887 }, { "epoch": 23.78443113772455, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1354, "step": 15888 }, { "epoch": 23.785928143712574, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1341, "step": 15889 }, { "epoch": 23.7874251497006, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1279, "step": 15890 }, { "epoch": 23.788922155688624, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1332, "step": 15891 }, { "epoch": 23.790419161676645, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1342, "step": 15892 }, { "epoch": 23.79191616766467, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1268, "step": 15893 }, { "epoch": 23.793413173652695, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1262, "step": 15894 }, { "epoch": 23.79491017964072, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1314, "step": 15895 }, { "epoch": 23.79640718562874, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1365, "step": 15896 }, { "epoch": 23.797904191616766, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1329, "step": 15897 }, { "epoch": 23.79940119760479, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1235, "step": 15898 }, { "epoch": 23.800898203592816, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1358, "step": 15899 }, { "epoch": 23.802395209580837, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1337, "step": 15900 }, { "epoch": 23.80389221556886, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1337, "step": 15901 }, { "epoch": 23.805389221556887, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1374, "step": 15902 }, { "epoch": 23.80688622754491, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1253, "step": 15903 }, { "epoch": 23.808383233532933, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1287, "step": 15904 }, { "epoch": 23.809880239520957, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1264, "step": 15905 }, { "epoch": 23.811377245508982, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1397, "step": 15906 }, { "epoch": 23.812874251497007, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1306, "step": 15907 }, { "epoch": 23.81437125748503, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1361, "step": 15908 }, { "epoch": 23.815868263473053, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.127, "step": 15909 }, { "epoch": 23.817365269461078, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1324, "step": 15910 }, { "epoch": 23.818862275449103, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.132, "step": 15911 }, { "epoch": 23.820359281437124, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1311, "step": 15912 }, { "epoch": 23.82185628742515, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.126, "step": 15913 }, { "epoch": 23.823353293413174, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1303, "step": 15914 }, { "epoch": 23.8248502994012, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1332, "step": 15915 }, { "epoch": 23.82634730538922, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1291, "step": 15916 }, { "epoch": 23.827844311377245, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1328, "step": 15917 }, { "epoch": 23.82934131736527, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1312, "step": 15918 }, { "epoch": 23.830838323353294, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.1347, "step": 15919 }, { "epoch": 23.83233532934132, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1354, "step": 15920 }, { "epoch": 23.83383233532934, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.133, "step": 15921 }, { "epoch": 23.835329341317365, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1303, "step": 15922 }, { "epoch": 23.83682634730539, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.133, "step": 15923 }, { "epoch": 23.83832335329341, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1334, "step": 15924 }, { "epoch": 23.839820359281436, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.131, "step": 15925 }, { "epoch": 23.84131736526946, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1379, "step": 15926 }, { "epoch": 23.842814371257486, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1344, "step": 15927 }, { "epoch": 23.84431137724551, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1349, "step": 15928 }, { "epoch": 23.845808383233532, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.135, "step": 15929 }, { "epoch": 23.847305389221557, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1366, "step": 15930 }, { "epoch": 23.84880239520958, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1304, "step": 15931 }, { "epoch": 23.850299401197606, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1381, "step": 15932 }, { "epoch": 23.851796407185628, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1262, "step": 15933 }, { "epoch": 23.853293413173652, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1305, "step": 15934 }, { "epoch": 23.854790419161677, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1273, "step": 15935 }, { "epoch": 23.856287425149702, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1268, "step": 15936 }, { "epoch": 23.857784431137723, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1262, "step": 15937 }, { "epoch": 23.85928143712575, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1326, "step": 15938 }, { "epoch": 23.860778443113773, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1332, "step": 15939 }, { "epoch": 23.862275449101798, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1373, "step": 15940 }, { "epoch": 23.86377245508982, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1237, "step": 15941 }, { "epoch": 23.865269461077844, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1375, "step": 15942 }, { "epoch": 23.86676646706587, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1289, "step": 15943 }, { "epoch": 23.868263473053894, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1279, "step": 15944 }, { "epoch": 23.869760479041915, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1294, "step": 15945 }, { "epoch": 23.87125748502994, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1322, "step": 15946 }, { "epoch": 23.872754491017965, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.137, "step": 15947 }, { "epoch": 23.87425149700599, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1259, "step": 15948 }, { "epoch": 23.87574850299401, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.123, "step": 15949 }, { "epoch": 23.877245508982035, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1319, "step": 15950 }, { "epoch": 23.87874251497006, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1405, "step": 15951 }, { "epoch": 23.880239520958085, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1396, "step": 15952 }, { "epoch": 23.881736526946106, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1295, "step": 15953 }, { "epoch": 23.88323353293413, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1323, "step": 15954 }, { "epoch": 23.884730538922156, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.132, "step": 15955 }, { "epoch": 23.88622754491018, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1276, "step": 15956 }, { "epoch": 23.887724550898202, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.131, "step": 15957 }, { "epoch": 23.889221556886227, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1299, "step": 15958 }, { "epoch": 23.89071856287425, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1371, "step": 15959 }, { "epoch": 23.892215568862277, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1347, "step": 15960 }, { "epoch": 23.893712574850298, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1295, "step": 15961 }, { "epoch": 23.895209580838323, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.132, "step": 15962 }, { "epoch": 23.896706586826348, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1299, "step": 15963 }, { "epoch": 23.898203592814372, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1363, "step": 15964 }, { "epoch": 23.899700598802394, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1309, "step": 15965 }, { "epoch": 23.90119760479042, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1333, "step": 15966 }, { "epoch": 23.902694610778443, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.137, "step": 15967 }, { "epoch": 23.904191616766468, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1368, "step": 15968 }, { "epoch": 23.90568862275449, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.127, "step": 15969 }, { "epoch": 23.907185628742514, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1339, "step": 15970 }, { "epoch": 23.90868263473054, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1361, "step": 15971 }, { "epoch": 23.910179640718564, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1294, "step": 15972 }, { "epoch": 23.91167664670659, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1412, "step": 15973 }, { "epoch": 23.91317365269461, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1256, "step": 15974 }, { "epoch": 23.914670658682635, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1276, "step": 15975 }, { "epoch": 23.91616766467066, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.13, "step": 15976 }, { "epoch": 23.91766467065868, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1327, "step": 15977 }, { "epoch": 23.919161676646706, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1349, "step": 15978 }, { "epoch": 23.92065868263473, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1307, "step": 15979 }, { "epoch": 23.922155688622755, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1305, "step": 15980 }, { "epoch": 23.92365269461078, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1373, "step": 15981 }, { "epoch": 23.9251497005988, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1283, "step": 15982 }, { "epoch": 23.926646706586826, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1263, "step": 15983 }, { "epoch": 23.92814371257485, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1296, "step": 15984 }, { "epoch": 23.929640718562876, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1308, "step": 15985 }, { "epoch": 23.931137724550897, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1348, "step": 15986 }, { "epoch": 23.932634730538922, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1282, "step": 15987 }, { "epoch": 23.934131736526947, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1235, "step": 15988 }, { "epoch": 23.93562874251497, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1257, "step": 15989 }, { "epoch": 23.937125748502993, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1328, "step": 15990 }, { "epoch": 23.938622754491018, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1317, "step": 15991 }, { "epoch": 23.940119760479043, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1329, "step": 15992 }, { "epoch": 23.941616766467067, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.139, "step": 15993 }, { "epoch": 23.94311377245509, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1277, "step": 15994 }, { "epoch": 23.944610778443113, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1347, "step": 15995 }, { "epoch": 23.94610778443114, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1335, "step": 15996 }, { "epoch": 23.947604790419163, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1299, "step": 15997 }, { "epoch": 23.949101796407184, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1386, "step": 15998 }, { "epoch": 23.95059880239521, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1352, "step": 15999 }, { "epoch": 23.952095808383234, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1335, "step": 16000 }, { "epoch": 23.95359281437126, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1323, "step": 16001 }, { "epoch": 23.95508982035928, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1381, "step": 16002 }, { "epoch": 23.956586826347305, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1353, "step": 16003 }, { "epoch": 23.95808383233533, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1283, "step": 16004 }, { "epoch": 23.959580838323355, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1313, "step": 16005 }, { "epoch": 23.961077844311376, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.127, "step": 16006 }, { "epoch": 23.9625748502994, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1309, "step": 16007 }, { "epoch": 23.964071856287426, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1395, "step": 16008 }, { "epoch": 23.96556886227545, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.133, "step": 16009 }, { "epoch": 23.96706586826347, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1318, "step": 16010 }, { "epoch": 23.968562874251496, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1364, "step": 16011 }, { "epoch": 23.97005988023952, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1349, "step": 16012 }, { "epoch": 23.971556886227546, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1332, "step": 16013 }, { "epoch": 23.973053892215567, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1319, "step": 16014 }, { "epoch": 23.974550898203592, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1332, "step": 16015 }, { "epoch": 23.976047904191617, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1323, "step": 16016 }, { "epoch": 23.977544910179642, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1303, "step": 16017 }, { "epoch": 23.979041916167663, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1364, "step": 16018 }, { "epoch": 23.980538922155688, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1335, "step": 16019 }, { "epoch": 23.982035928143713, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1291, "step": 16020 }, { "epoch": 23.983532934131738, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.135, "step": 16021 }, { "epoch": 23.98502994011976, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1331, "step": 16022 }, { "epoch": 23.986526946107784, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1294, "step": 16023 }, { "epoch": 23.98802395209581, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1256, "step": 16024 }, { "epoch": 23.989520958083833, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1397, "step": 16025 }, { "epoch": 23.991017964071855, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1309, "step": 16026 }, { "epoch": 23.99251497005988, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1322, "step": 16027 }, { "epoch": 23.994011976047904, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1376, "step": 16028 }, { "epoch": 23.99550898203593, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1297, "step": 16029 }, { "epoch": 23.99700598802395, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1345, "step": 16030 }, { "epoch": 23.998502994011975, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1332, "step": 16031 }, { "epoch": 24.0, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 16032 }, { "epoch": 24.001497005988025, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1311, "step": 16033 }, { "epoch": 24.00299401197605, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1313, "step": 16034 }, { "epoch": 24.00449101796407, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1307, "step": 16035 }, { "epoch": 24.005988023952096, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1313, "step": 16036 }, { "epoch": 24.00748502994012, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1334, "step": 16037 }, { "epoch": 24.008982035928145, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.133, "step": 16038 }, { "epoch": 24.010479041916167, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1292, "step": 16039 }, { "epoch": 24.01197604790419, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1321, "step": 16040 }, { "epoch": 24.013473053892216, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1363, "step": 16041 }, { "epoch": 24.01497005988024, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1277, "step": 16042 }, { "epoch": 24.016467065868262, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1386, "step": 16043 }, { "epoch": 24.017964071856287, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1264, "step": 16044 }, { "epoch": 24.019461077844312, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1337, "step": 16045 }, { "epoch": 24.020958083832337, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1296, "step": 16046 }, { "epoch": 24.022455089820358, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1293, "step": 16047 }, { "epoch": 24.023952095808383, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1292, "step": 16048 }, { "epoch": 24.025449101796408, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1327, "step": 16049 }, { "epoch": 24.026946107784433, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1361, "step": 16050 }, { "epoch": 24.028443113772454, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1337, "step": 16051 }, { "epoch": 24.02994011976048, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1347, "step": 16052 }, { "epoch": 24.031437125748504, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1338, "step": 16053 }, { "epoch": 24.03293413173653, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1266, "step": 16054 }, { "epoch": 24.03443113772455, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1275, "step": 16055 }, { "epoch": 24.035928143712574, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1211, "step": 16056 }, { "epoch": 24.0374251497006, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.128, "step": 16057 }, { "epoch": 24.038922155688624, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1314, "step": 16058 }, { "epoch": 24.040419161676645, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1321, "step": 16059 }, { "epoch": 24.04191616766467, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1302, "step": 16060 }, { "epoch": 24.043413173652695, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1289, "step": 16061 }, { "epoch": 24.04491017964072, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1315, "step": 16062 }, { "epoch": 24.04640718562874, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1304, "step": 16063 }, { "epoch": 24.047904191616766, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1308, "step": 16064 }, { "epoch": 24.04940119760479, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1292, "step": 16065 }, { "epoch": 24.050898203592816, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1267, "step": 16066 }, { "epoch": 24.052395209580837, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1324, "step": 16067 }, { "epoch": 24.05389221556886, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1284, "step": 16068 }, { "epoch": 24.055389221556887, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1272, "step": 16069 }, { "epoch": 24.05688622754491, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1301, "step": 16070 }, { "epoch": 24.058383233532933, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1317, "step": 16071 }, { "epoch": 24.059880239520957, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1287, "step": 16072 }, { "epoch": 24.061377245508982, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1329, "step": 16073 }, { "epoch": 24.062874251497007, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1284, "step": 16074 }, { "epoch": 24.06437125748503, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.138, "step": 16075 }, { "epoch": 24.065868263473053, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1264, "step": 16076 }, { "epoch": 24.067365269461078, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.133, "step": 16077 }, { "epoch": 24.068862275449103, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1319, "step": 16078 }, { "epoch": 24.070359281437124, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1284, "step": 16079 }, { "epoch": 24.07185628742515, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1311, "step": 16080 }, { "epoch": 24.073353293413174, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1364, "step": 16081 }, { "epoch": 24.0748502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1285, "step": 16082 }, { "epoch": 24.07634730538922, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1318, "step": 16083 }, { "epoch": 24.077844311377245, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1342, "step": 16084 }, { "epoch": 24.07934131736527, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1281, "step": 16085 }, { "epoch": 24.080838323353294, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1331, "step": 16086 }, { "epoch": 24.082335329341316, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1295, "step": 16087 }, { "epoch": 24.08383233532934, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1288, "step": 16088 }, { "epoch": 24.085329341317365, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1361, "step": 16089 }, { "epoch": 24.08682634730539, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1342, "step": 16090 }, { "epoch": 24.088323353293415, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.133, "step": 16091 }, { "epoch": 24.089820359281436, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1385, "step": 16092 }, { "epoch": 24.09131736526946, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1304, "step": 16093 }, { "epoch": 24.092814371257486, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1321, "step": 16094 }, { "epoch": 24.09431137724551, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1241, "step": 16095 }, { "epoch": 24.095808383233532, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1311, "step": 16096 }, { "epoch": 24.097305389221557, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1354, "step": 16097 }, { "epoch": 24.09880239520958, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1261, "step": 16098 }, { "epoch": 24.100299401197606, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1304, "step": 16099 }, { "epoch": 24.101796407185628, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1353, "step": 16100 }, { "epoch": 24.103293413173652, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1269, "step": 16101 }, { "epoch": 24.104790419161677, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.133, "step": 16102 }, { "epoch": 24.106287425149702, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1252, "step": 16103 }, { "epoch": 24.107784431137723, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1292, "step": 16104 }, { "epoch": 24.10928143712575, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1333, "step": 16105 }, { "epoch": 24.110778443113773, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1312, "step": 16106 }, { "epoch": 24.112275449101798, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1262, "step": 16107 }, { "epoch": 24.11377245508982, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1325, "step": 16108 }, { "epoch": 24.115269461077844, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1289, "step": 16109 }, { "epoch": 24.11676646706587, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1277, "step": 16110 }, { "epoch": 24.118263473053894, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1289, "step": 16111 }, { "epoch": 24.119760479041915, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1334, "step": 16112 }, { "epoch": 24.12125748502994, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1327, "step": 16113 }, { "epoch": 24.122754491017965, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1288, "step": 16114 }, { "epoch": 24.12425149700599, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1281, "step": 16115 }, { "epoch": 24.12574850299401, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1316, "step": 16116 }, { "epoch": 24.127245508982035, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1247, "step": 16117 }, { "epoch": 24.12874251497006, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1328, "step": 16118 }, { "epoch": 24.130239520958085, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1328, "step": 16119 }, { "epoch": 24.131736526946106, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1283, "step": 16120 }, { "epoch": 24.13323353293413, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1273, "step": 16121 }, { "epoch": 24.134730538922156, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1298, "step": 16122 }, { "epoch": 24.13622754491018, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1292, "step": 16123 }, { "epoch": 24.137724550898202, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1303, "step": 16124 }, { "epoch": 24.139221556886227, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1377, "step": 16125 }, { "epoch": 24.14071856287425, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1367, "step": 16126 }, { "epoch": 24.142215568862277, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.138, "step": 16127 }, { "epoch": 24.143712574850298, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1353, "step": 16128 }, { "epoch": 24.145209580838323, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1336, "step": 16129 }, { "epoch": 24.146706586826348, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1298, "step": 16130 }, { "epoch": 24.148203592814372, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1269, "step": 16131 }, { "epoch": 24.149700598802394, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1288, "step": 16132 }, { "epoch": 24.15119760479042, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1398, "step": 16133 }, { "epoch": 24.152694610778443, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1301, "step": 16134 }, { "epoch": 24.154191616766468, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1302, "step": 16135 }, { "epoch": 24.15568862275449, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1347, "step": 16136 }, { "epoch": 24.157185628742514, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1326, "step": 16137 }, { "epoch": 24.15868263473054, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1352, "step": 16138 }, { "epoch": 24.160179640718564, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1256, "step": 16139 }, { "epoch": 24.161676646706585, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1271, "step": 16140 }, { "epoch": 24.16317365269461, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1293, "step": 16141 }, { "epoch": 24.164670658682635, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1304, "step": 16142 }, { "epoch": 24.16616766467066, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1242, "step": 16143 }, { "epoch": 24.16766467065868, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1328, "step": 16144 }, { "epoch": 24.169161676646706, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1295, "step": 16145 }, { "epoch": 24.17065868263473, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1348, "step": 16146 }, { "epoch": 24.172155688622755, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1254, "step": 16147 }, { "epoch": 24.17365269461078, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 16148 }, { "epoch": 24.1751497005988, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1276, "step": 16149 }, { "epoch": 24.176646706586826, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1334, "step": 16150 }, { "epoch": 24.17814371257485, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1311, "step": 16151 }, { "epoch": 24.179640718562876, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.14, "step": 16152 }, { "epoch": 24.181137724550897, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1308, "step": 16153 }, { "epoch": 24.182634730538922, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1281, "step": 16154 }, { "epoch": 24.184131736526947, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1309, "step": 16155 }, { "epoch": 24.18562874251497, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1288, "step": 16156 }, { "epoch": 24.187125748502993, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1324, "step": 16157 }, { "epoch": 24.188622754491018, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1277, "step": 16158 }, { "epoch": 24.190119760479043, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1328, "step": 16159 }, { "epoch": 24.191616766467067, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1342, "step": 16160 }, { "epoch": 24.19311377245509, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1267, "step": 16161 }, { "epoch": 24.194610778443113, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1346, "step": 16162 }, { "epoch": 24.19610778443114, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.132, "step": 16163 }, { "epoch": 24.197604790419163, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1297, "step": 16164 }, { "epoch": 24.199101796407184, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1323, "step": 16165 }, { "epoch": 24.20059880239521, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1273, "step": 16166 }, { "epoch": 24.202095808383234, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.134, "step": 16167 }, { "epoch": 24.20359281437126, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1321, "step": 16168 }, { "epoch": 24.20508982035928, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.13, "step": 16169 }, { "epoch": 24.206586826347305, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1343, "step": 16170 }, { "epoch": 24.20808383233533, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.135, "step": 16171 }, { "epoch": 24.209580838323355, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1294, "step": 16172 }, { "epoch": 24.211077844311376, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1355, "step": 16173 }, { "epoch": 24.2125748502994, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1301, "step": 16174 }, { "epoch": 24.214071856287426, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1366, "step": 16175 }, { "epoch": 24.21556886227545, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.135, "step": 16176 }, { "epoch": 24.21706586826347, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1312, "step": 16177 }, { "epoch": 24.218562874251496, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1298, "step": 16178 }, { "epoch": 24.22005988023952, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.132, "step": 16179 }, { "epoch": 24.221556886227546, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1338, "step": 16180 }, { "epoch": 24.223053892215567, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1297, "step": 16181 }, { "epoch": 24.224550898203592, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.125, "step": 16182 }, { "epoch": 24.226047904191617, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1333, "step": 16183 }, { "epoch": 24.227544910179642, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1332, "step": 16184 }, { "epoch": 24.229041916167663, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1321, "step": 16185 }, { "epoch": 24.230538922155688, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1337, "step": 16186 }, { "epoch": 24.232035928143713, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.132, "step": 16187 }, { "epoch": 24.233532934131738, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1326, "step": 16188 }, { "epoch": 24.23502994011976, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1309, "step": 16189 }, { "epoch": 24.236526946107784, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1355, "step": 16190 }, { "epoch": 24.23802395209581, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1382, "step": 16191 }, { "epoch": 24.239520958083833, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1286, "step": 16192 }, { "epoch": 24.241017964071855, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1317, "step": 16193 }, { "epoch": 24.24251497005988, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1293, "step": 16194 }, { "epoch": 24.244011976047904, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1377, "step": 16195 }, { "epoch": 24.24550898203593, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1315, "step": 16196 }, { "epoch": 24.24700598802395, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1366, "step": 16197 }, { "epoch": 24.248502994011975, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1284, "step": 16198 }, { "epoch": 24.25, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1312, "step": 16199 }, { "epoch": 24.251497005988025, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1313, "step": 16200 }, { "epoch": 24.25299401197605, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1304, "step": 16201 }, { "epoch": 24.25449101796407, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1308, "step": 16202 }, { "epoch": 24.255988023952096, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1321, "step": 16203 }, { "epoch": 24.25748502994012, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1257, "step": 16204 }, { "epoch": 24.258982035928145, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1299, "step": 16205 }, { "epoch": 24.260479041916167, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1387, "step": 16206 }, { "epoch": 24.26197604790419, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1308, "step": 16207 }, { "epoch": 24.263473053892216, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1345, "step": 16208 }, { "epoch": 24.26497005988024, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.131, "step": 16209 }, { "epoch": 24.266467065868262, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1283, "step": 16210 }, { "epoch": 24.267964071856287, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1288, "step": 16211 }, { "epoch": 24.269461077844312, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1329, "step": 16212 }, { "epoch": 24.270958083832337, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1296, "step": 16213 }, { "epoch": 24.272455089820358, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1262, "step": 16214 }, { "epoch": 24.273952095808383, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1285, "step": 16215 }, { "epoch": 24.275449101796408, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1234, "step": 16216 }, { "epoch": 24.276946107784433, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1302, "step": 16217 }, { "epoch": 24.278443113772454, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1378, "step": 16218 }, { "epoch": 24.27994011976048, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1323, "step": 16219 }, { "epoch": 24.281437125748504, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.132, "step": 16220 }, { "epoch": 24.28293413173653, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1327, "step": 16221 }, { "epoch": 24.28443113772455, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1266, "step": 16222 }, { "epoch": 24.285928143712574, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1325, "step": 16223 }, { "epoch": 24.2874251497006, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1324, "step": 16224 }, { "epoch": 24.288922155688624, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1335, "step": 16225 }, { "epoch": 24.290419161676645, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1306, "step": 16226 }, { "epoch": 24.29191616766467, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1364, "step": 16227 }, { "epoch": 24.293413173652695, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1408, "step": 16228 }, { "epoch": 24.29491017964072, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1328, "step": 16229 }, { "epoch": 24.29640718562874, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1278, "step": 16230 }, { "epoch": 24.297904191616766, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1267, "step": 16231 }, { "epoch": 24.29940119760479, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1358, "step": 16232 }, { "epoch": 24.300898203592816, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1317, "step": 16233 }, { "epoch": 24.302395209580837, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1323, "step": 16234 }, { "epoch": 24.30389221556886, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1279, "step": 16235 }, { "epoch": 24.305389221556887, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1399, "step": 16236 }, { "epoch": 24.30688622754491, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1268, "step": 16237 }, { "epoch": 24.308383233532933, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1271, "step": 16238 }, { "epoch": 24.309880239520957, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1335, "step": 16239 }, { "epoch": 24.311377245508982, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1354, "step": 16240 }, { "epoch": 24.312874251497007, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1338, "step": 16241 }, { "epoch": 24.31437125748503, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.133, "step": 16242 }, { "epoch": 24.315868263473053, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1351, "step": 16243 }, { "epoch": 24.317365269461078, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1308, "step": 16244 }, { "epoch": 24.318862275449103, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1284, "step": 16245 }, { "epoch": 24.320359281437124, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.131, "step": 16246 }, { "epoch": 24.32185628742515, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1301, "step": 16247 }, { "epoch": 24.323353293413174, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1278, "step": 16248 }, { "epoch": 24.3248502994012, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1248, "step": 16249 }, { "epoch": 24.32634730538922, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1233, "step": 16250 }, { "epoch": 24.327844311377245, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.132, "step": 16251 }, { "epoch": 24.32934131736527, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1326, "step": 16252 }, { "epoch": 24.330838323353294, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1281, "step": 16253 }, { "epoch": 24.33233532934132, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.129, "step": 16254 }, { "epoch": 24.33383233532934, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.138, "step": 16255 }, { "epoch": 24.335329341317365, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1275, "step": 16256 }, { "epoch": 24.33682634730539, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1324, "step": 16257 }, { "epoch": 24.338323353293415, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1315, "step": 16258 }, { "epoch": 24.339820359281436, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1333, "step": 16259 }, { "epoch": 24.34131736526946, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1218, "step": 16260 }, { "epoch": 24.342814371257486, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1353, "step": 16261 }, { "epoch": 24.34431137724551, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1271, "step": 16262 }, { "epoch": 24.345808383233532, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1349, "step": 16263 }, { "epoch": 24.347305389221557, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.135, "step": 16264 }, { "epoch": 24.34880239520958, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1379, "step": 16265 }, { "epoch": 24.350299401197606, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1309, "step": 16266 }, { "epoch": 24.351796407185628, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1334, "step": 16267 }, { "epoch": 24.353293413173652, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1338, "step": 16268 }, { "epoch": 24.354790419161677, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1218, "step": 16269 }, { "epoch": 24.356287425149702, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1315, "step": 16270 }, { "epoch": 24.357784431137723, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1316, "step": 16271 }, { "epoch": 24.35928143712575, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1281, "step": 16272 }, { "epoch": 24.360778443113773, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1348, "step": 16273 }, { "epoch": 24.362275449101798, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1309, "step": 16274 }, { "epoch": 24.36377245508982, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1319, "step": 16275 }, { "epoch": 24.365269461077844, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1359, "step": 16276 }, { "epoch": 24.36676646706587, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.133, "step": 16277 }, { "epoch": 24.368263473053894, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1357, "step": 16278 }, { "epoch": 24.369760479041915, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1316, "step": 16279 }, { "epoch": 24.37125748502994, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1324, "step": 16280 }, { "epoch": 24.372754491017965, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.132, "step": 16281 }, { "epoch": 24.37425149700599, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1287, "step": 16282 }, { "epoch": 24.37574850299401, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1388, "step": 16283 }, { "epoch": 24.377245508982035, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1295, "step": 16284 }, { "epoch": 24.37874251497006, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1332, "step": 16285 }, { "epoch": 24.380239520958085, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1418, "step": 16286 }, { "epoch": 24.381736526946106, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1325, "step": 16287 }, { "epoch": 24.38323353293413, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1295, "step": 16288 }, { "epoch": 24.384730538922156, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1377, "step": 16289 }, { "epoch": 24.38622754491018, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1309, "step": 16290 }, { "epoch": 24.387724550898202, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1366, "step": 16291 }, { "epoch": 24.389221556886227, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1283, "step": 16292 }, { "epoch": 24.39071856287425, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1268, "step": 16293 }, { "epoch": 24.392215568862277, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1323, "step": 16294 }, { "epoch": 24.393712574850298, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1316, "step": 16295 }, { "epoch": 24.395209580838323, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1305, "step": 16296 }, { "epoch": 24.396706586826348, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1285, "step": 16297 }, { "epoch": 24.398203592814372, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1339, "step": 16298 }, { "epoch": 24.399700598802394, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1318, "step": 16299 }, { "epoch": 24.40119760479042, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1291, "step": 16300 }, { "epoch": 24.402694610778443, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1276, "step": 16301 }, { "epoch": 24.404191616766468, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1283, "step": 16302 }, { "epoch": 24.40568862275449, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.13, "step": 16303 }, { "epoch": 24.407185628742514, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1296, "step": 16304 }, { "epoch": 24.40868263473054, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1272, "step": 16305 }, { "epoch": 24.410179640718564, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1334, "step": 16306 }, { "epoch": 24.411676646706585, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1299, "step": 16307 }, { "epoch": 24.41317365269461, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1274, "step": 16308 }, { "epoch": 24.414670658682635, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.131, "step": 16309 }, { "epoch": 24.41616766467066, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1375, "step": 16310 }, { "epoch": 24.41766467065868, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1304, "step": 16311 }, { "epoch": 24.419161676646706, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1371, "step": 16312 }, { "epoch": 24.42065868263473, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1364, "step": 16313 }, { "epoch": 24.422155688622755, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1357, "step": 16314 }, { "epoch": 24.42365269461078, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1323, "step": 16315 }, { "epoch": 24.4251497005988, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1318, "step": 16316 }, { "epoch": 24.426646706586826, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1295, "step": 16317 }, { "epoch": 24.42814371257485, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1359, "step": 16318 }, { "epoch": 24.429640718562876, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1331, "step": 16319 }, { "epoch": 24.431137724550897, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1307, "step": 16320 }, { "epoch": 24.432634730538922, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1335, "step": 16321 }, { "epoch": 24.434131736526947, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.137, "step": 16322 }, { "epoch": 24.43562874251497, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1308, "step": 16323 }, { "epoch": 24.437125748502993, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1263, "step": 16324 }, { "epoch": 24.438622754491018, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1303, "step": 16325 }, { "epoch": 24.440119760479043, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1283, "step": 16326 }, { "epoch": 24.441616766467067, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1323, "step": 16327 }, { "epoch": 24.44311377245509, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1241, "step": 16328 }, { "epoch": 24.444610778443113, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1342, "step": 16329 }, { "epoch": 24.44610778443114, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1315, "step": 16330 }, { "epoch": 24.447604790419163, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1342, "step": 16331 }, { "epoch": 24.449101796407184, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1277, "step": 16332 }, { "epoch": 24.45059880239521, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1267, "step": 16333 }, { "epoch": 24.452095808383234, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1289, "step": 16334 }, { "epoch": 24.45359281437126, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.137, "step": 16335 }, { "epoch": 24.45508982035928, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1286, "step": 16336 }, { "epoch": 24.456586826347305, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1351, "step": 16337 }, { "epoch": 24.45808383233533, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1293, "step": 16338 }, { "epoch": 24.459580838323355, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.133, "step": 16339 }, { "epoch": 24.461077844311376, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.135, "step": 16340 }, { "epoch": 24.4625748502994, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1297, "step": 16341 }, { "epoch": 24.464071856287426, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1347, "step": 16342 }, { "epoch": 24.46556886227545, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.134, "step": 16343 }, { "epoch": 24.46706586826347, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1314, "step": 16344 }, { "epoch": 24.468562874251496, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1364, "step": 16345 }, { "epoch": 24.47005988023952, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1301, "step": 16346 }, { "epoch": 24.471556886227546, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1275, "step": 16347 }, { "epoch": 24.473053892215567, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1307, "step": 16348 }, { "epoch": 24.474550898203592, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1275, "step": 16349 }, { "epoch": 24.476047904191617, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1284, "step": 16350 }, { "epoch": 24.477544910179642, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1351, "step": 16351 }, { "epoch": 24.479041916167663, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1312, "step": 16352 }, { "epoch": 24.480538922155688, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1287, "step": 16353 }, { "epoch": 24.482035928143713, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1359, "step": 16354 }, { "epoch": 24.483532934131738, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1326, "step": 16355 }, { "epoch": 24.48502994011976, "grad_norm": 0.1474609375, "learning_rate": 0.0008, "loss": 1.1322, "step": 16356 }, { "epoch": 24.486526946107784, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1299, "step": 16357 }, { "epoch": 24.48802395209581, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1235, "step": 16358 }, { "epoch": 24.489520958083833, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1293, "step": 16359 }, { "epoch": 24.491017964071855, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1311, "step": 16360 }, { "epoch": 24.49251497005988, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.135, "step": 16361 }, { "epoch": 24.494011976047904, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1283, "step": 16362 }, { "epoch": 24.49550898203593, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1331, "step": 16363 }, { "epoch": 24.49700598802395, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1357, "step": 16364 }, { "epoch": 24.498502994011975, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1312, "step": 16365 }, { "epoch": 24.5, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.129, "step": 16366 }, { "epoch": 24.501497005988025, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.128, "step": 16367 }, { "epoch": 24.50299401197605, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1281, "step": 16368 }, { "epoch": 24.50449101796407, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1376, "step": 16369 }, { "epoch": 24.505988023952096, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1236, "step": 16370 }, { "epoch": 24.50748502994012, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1364, "step": 16371 }, { "epoch": 24.508982035928145, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1359, "step": 16372 }, { "epoch": 24.510479041916167, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.13, "step": 16373 }, { "epoch": 24.51197604790419, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1293, "step": 16374 }, { "epoch": 24.513473053892216, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1351, "step": 16375 }, { "epoch": 24.51497005988024, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1366, "step": 16376 }, { "epoch": 24.516467065868262, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1292, "step": 16377 }, { "epoch": 24.517964071856287, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1368, "step": 16378 }, { "epoch": 24.519461077844312, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1278, "step": 16379 }, { "epoch": 24.520958083832337, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1317, "step": 16380 }, { "epoch": 24.522455089820358, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1233, "step": 16381 }, { "epoch": 24.523952095808383, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1365, "step": 16382 }, { "epoch": 24.525449101796408, "grad_norm": 0.1220703125, "learning_rate": 0.0008, "loss": 1.1354, "step": 16383 }, { "epoch": 24.526946107784433, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1375, "step": 16384 }, { "epoch": 24.528443113772454, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1379, "step": 16385 }, { "epoch": 24.52994011976048, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1327, "step": 16386 }, { "epoch": 24.531437125748504, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.13, "step": 16387 }, { "epoch": 24.53293413173653, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1256, "step": 16388 }, { "epoch": 24.53443113772455, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1348, "step": 16389 }, { "epoch": 24.535928143712574, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1322, "step": 16390 }, { "epoch": 24.5374251497006, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1245, "step": 16391 }, { "epoch": 24.538922155688624, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.131, "step": 16392 }, { "epoch": 24.540419161676645, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1331, "step": 16393 }, { "epoch": 24.54191616766467, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1312, "step": 16394 }, { "epoch": 24.543413173652695, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1355, "step": 16395 }, { "epoch": 24.54491017964072, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1298, "step": 16396 }, { "epoch": 24.54640718562874, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1319, "step": 16397 }, { "epoch": 24.547904191616766, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.135, "step": 16398 }, { "epoch": 24.54940119760479, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1321, "step": 16399 }, { "epoch": 24.550898203592816, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1305, "step": 16400 }, { "epoch": 24.552395209580837, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1302, "step": 16401 }, { "epoch": 24.55389221556886, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1371, "step": 16402 }, { "epoch": 24.555389221556887, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1299, "step": 16403 }, { "epoch": 24.55688622754491, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1394, "step": 16404 }, { "epoch": 24.558383233532933, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1275, "step": 16405 }, { "epoch": 24.559880239520957, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1351, "step": 16406 }, { "epoch": 24.561377245508982, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1283, "step": 16407 }, { "epoch": 24.562874251497007, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1256, "step": 16408 }, { "epoch": 24.56437125748503, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1253, "step": 16409 }, { "epoch": 24.565868263473053, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1301, "step": 16410 }, { "epoch": 24.567365269461078, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1291, "step": 16411 }, { "epoch": 24.568862275449103, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.136, "step": 16412 }, { "epoch": 24.570359281437124, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1335, "step": 16413 }, { "epoch": 24.57185628742515, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1313, "step": 16414 }, { "epoch": 24.573353293413174, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1336, "step": 16415 }, { "epoch": 24.5748502994012, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1295, "step": 16416 }, { "epoch": 24.57634730538922, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1408, "step": 16417 }, { "epoch": 24.577844311377245, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1274, "step": 16418 }, { "epoch": 24.57934131736527, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1288, "step": 16419 }, { "epoch": 24.580838323353294, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1346, "step": 16420 }, { "epoch": 24.58233532934132, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1391, "step": 16421 }, { "epoch": 24.58383233532934, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1282, "step": 16422 }, { "epoch": 24.585329341317365, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1349, "step": 16423 }, { "epoch": 24.58682634730539, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1326, "step": 16424 }, { "epoch": 24.58832335329341, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1351, "step": 16425 }, { "epoch": 24.589820359281436, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1364, "step": 16426 }, { "epoch": 24.59131736526946, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.13, "step": 16427 }, { "epoch": 24.592814371257486, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1369, "step": 16428 }, { "epoch": 24.59431137724551, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.133, "step": 16429 }, { "epoch": 24.595808383233532, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1347, "step": 16430 }, { "epoch": 24.597305389221557, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1326, "step": 16431 }, { "epoch": 24.59880239520958, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 16432 }, { "epoch": 24.600299401197606, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1348, "step": 16433 }, { "epoch": 24.601796407185628, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1307, "step": 16434 }, { "epoch": 24.603293413173652, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1327, "step": 16435 }, { "epoch": 24.604790419161677, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1296, "step": 16436 }, { "epoch": 24.606287425149702, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1364, "step": 16437 }, { "epoch": 24.607784431137723, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1363, "step": 16438 }, { "epoch": 24.60928143712575, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1311, "step": 16439 }, { "epoch": 24.610778443113773, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1278, "step": 16440 }, { "epoch": 24.612275449101798, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1291, "step": 16441 }, { "epoch": 24.61377245508982, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1288, "step": 16442 }, { "epoch": 24.615269461077844, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1354, "step": 16443 }, { "epoch": 24.61676646706587, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1258, "step": 16444 }, { "epoch": 24.618263473053894, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1321, "step": 16445 }, { "epoch": 24.619760479041915, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1322, "step": 16446 }, { "epoch": 24.62125748502994, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1327, "step": 16447 }, { "epoch": 24.622754491017965, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1234, "step": 16448 }, { "epoch": 24.62425149700599, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1284, "step": 16449 }, { "epoch": 24.62574850299401, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1232, "step": 16450 }, { "epoch": 24.627245508982035, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1368, "step": 16451 }, { "epoch": 24.62874251497006, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1279, "step": 16452 }, { "epoch": 24.630239520958085, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1312, "step": 16453 }, { "epoch": 24.631736526946106, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1264, "step": 16454 }, { "epoch": 24.63323353293413, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1352, "step": 16455 }, { "epoch": 24.634730538922156, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1266, "step": 16456 }, { "epoch": 24.63622754491018, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.13, "step": 16457 }, { "epoch": 24.637724550898202, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1387, "step": 16458 }, { "epoch": 24.639221556886227, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1427, "step": 16459 }, { "epoch": 24.64071856287425, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1294, "step": 16460 }, { "epoch": 24.642215568862277, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1283, "step": 16461 }, { "epoch": 24.643712574850298, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.127, "step": 16462 }, { "epoch": 24.645209580838323, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1338, "step": 16463 }, { "epoch": 24.646706586826348, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1334, "step": 16464 }, { "epoch": 24.648203592814372, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1369, "step": 16465 }, { "epoch": 24.649700598802394, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1355, "step": 16466 }, { "epoch": 24.65119760479042, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.126, "step": 16467 }, { "epoch": 24.652694610778443, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1279, "step": 16468 }, { "epoch": 24.654191616766468, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.126, "step": 16469 }, { "epoch": 24.65568862275449, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1289, "step": 16470 }, { "epoch": 24.657185628742514, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1343, "step": 16471 }, { "epoch": 24.65868263473054, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1372, "step": 16472 }, { "epoch": 24.660179640718564, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1258, "step": 16473 }, { "epoch": 24.66167664670659, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1281, "step": 16474 }, { "epoch": 24.66317365269461, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1355, "step": 16475 }, { "epoch": 24.664670658682635, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1323, "step": 16476 }, { "epoch": 24.66616766467066, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.126, "step": 16477 }, { "epoch": 24.66766467065868, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1405, "step": 16478 }, { "epoch": 24.669161676646706, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1304, "step": 16479 }, { "epoch": 24.67065868263473, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1366, "step": 16480 }, { "epoch": 24.672155688622755, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1308, "step": 16481 }, { "epoch": 24.67365269461078, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1318, "step": 16482 }, { "epoch": 24.6751497005988, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1287, "step": 16483 }, { "epoch": 24.676646706586826, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1346, "step": 16484 }, { "epoch": 24.67814371257485, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1283, "step": 16485 }, { "epoch": 24.679640718562876, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1326, "step": 16486 }, { "epoch": 24.681137724550897, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1348, "step": 16487 }, { "epoch": 24.682634730538922, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1354, "step": 16488 }, { "epoch": 24.684131736526947, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.127, "step": 16489 }, { "epoch": 24.68562874251497, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1304, "step": 16490 }, { "epoch": 24.687125748502993, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.129, "step": 16491 }, { "epoch": 24.688622754491018, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1301, "step": 16492 }, { "epoch": 24.690119760479043, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1353, "step": 16493 }, { "epoch": 24.691616766467067, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1327, "step": 16494 }, { "epoch": 24.69311377245509, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1362, "step": 16495 }, { "epoch": 24.694610778443113, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1291, "step": 16496 }, { "epoch": 24.69610778443114, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1285, "step": 16497 }, { "epoch": 24.697604790419163, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1361, "step": 16498 }, { "epoch": 24.699101796407184, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1353, "step": 16499 }, { "epoch": 24.70059880239521, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1348, "step": 16500 }, { "epoch": 24.702095808383234, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1296, "step": 16501 }, { "epoch": 24.70359281437126, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1275, "step": 16502 }, { "epoch": 24.70508982035928, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1326, "step": 16503 }, { "epoch": 24.706586826347305, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1373, "step": 16504 }, { "epoch": 24.70808383233533, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1286, "step": 16505 }, { "epoch": 24.709580838323355, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1283, "step": 16506 }, { "epoch": 24.711077844311376, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1304, "step": 16507 }, { "epoch": 24.7125748502994, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1391, "step": 16508 }, { "epoch": 24.714071856287426, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1317, "step": 16509 }, { "epoch": 24.71556886227545, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1309, "step": 16510 }, { "epoch": 24.71706586826347, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1255, "step": 16511 }, { "epoch": 24.718562874251496, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1279, "step": 16512 }, { "epoch": 24.72005988023952, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1267, "step": 16513 }, { "epoch": 24.721556886227546, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1263, "step": 16514 }, { "epoch": 24.723053892215567, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1301, "step": 16515 }, { "epoch": 24.724550898203592, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.126, "step": 16516 }, { "epoch": 24.726047904191617, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1292, "step": 16517 }, { "epoch": 24.727544910179642, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1232, "step": 16518 }, { "epoch": 24.729041916167663, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.126, "step": 16519 }, { "epoch": 24.730538922155688, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1345, "step": 16520 }, { "epoch": 24.732035928143713, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1329, "step": 16521 }, { "epoch": 24.733532934131738, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1348, "step": 16522 }, { "epoch": 24.73502994011976, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1347, "step": 16523 }, { "epoch": 24.736526946107784, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1328, "step": 16524 }, { "epoch": 24.73802395209581, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1369, "step": 16525 }, { "epoch": 24.739520958083833, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1289, "step": 16526 }, { "epoch": 24.741017964071855, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1331, "step": 16527 }, { "epoch": 24.74251497005988, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1363, "step": 16528 }, { "epoch": 24.744011976047904, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1382, "step": 16529 }, { "epoch": 24.74550898203593, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.133, "step": 16530 }, { "epoch": 24.74700598802395, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1335, "step": 16531 }, { "epoch": 24.748502994011975, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1324, "step": 16532 }, { "epoch": 24.75, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1332, "step": 16533 }, { "epoch": 24.751497005988025, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1341, "step": 16534 }, { "epoch": 24.75299401197605, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1291, "step": 16535 }, { "epoch": 24.75449101796407, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1397, "step": 16536 }, { "epoch": 24.755988023952096, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.13, "step": 16537 }, { "epoch": 24.75748502994012, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1363, "step": 16538 }, { "epoch": 24.758982035928145, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1329, "step": 16539 }, { "epoch": 24.760479041916167, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1374, "step": 16540 }, { "epoch": 24.76197604790419, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1284, "step": 16541 }, { "epoch": 24.763473053892216, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1226, "step": 16542 }, { "epoch": 24.76497005988024, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1267, "step": 16543 }, { "epoch": 24.766467065868262, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1299, "step": 16544 }, { "epoch": 24.767964071856287, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1316, "step": 16545 }, { "epoch": 24.769461077844312, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1265, "step": 16546 }, { "epoch": 24.770958083832337, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1351, "step": 16547 }, { "epoch": 24.772455089820358, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.136, "step": 16548 }, { "epoch": 24.773952095808383, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1347, "step": 16549 }, { "epoch": 24.775449101796408, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.132, "step": 16550 }, { "epoch": 24.776946107784433, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1298, "step": 16551 }, { "epoch": 24.778443113772454, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1296, "step": 16552 }, { "epoch": 24.77994011976048, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1296, "step": 16553 }, { "epoch": 24.781437125748504, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1348, "step": 16554 }, { "epoch": 24.78293413173653, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1333, "step": 16555 }, { "epoch": 24.78443113772455, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1348, "step": 16556 }, { "epoch": 24.785928143712574, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1298, "step": 16557 }, { "epoch": 24.7874251497006, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1346, "step": 16558 }, { "epoch": 24.788922155688624, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1344, "step": 16559 }, { "epoch": 24.790419161676645, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1229, "step": 16560 }, { "epoch": 24.79191616766467, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1299, "step": 16561 }, { "epoch": 24.793413173652695, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1395, "step": 16562 }, { "epoch": 24.79491017964072, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1364, "step": 16563 }, { "epoch": 24.79640718562874, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1314, "step": 16564 }, { "epoch": 24.797904191616766, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.127, "step": 16565 }, { "epoch": 24.79940119760479, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1307, "step": 16566 }, { "epoch": 24.800898203592816, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1417, "step": 16567 }, { "epoch": 24.802395209580837, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1295, "step": 16568 }, { "epoch": 24.80389221556886, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.125, "step": 16569 }, { "epoch": 24.805389221556887, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1296, "step": 16570 }, { "epoch": 24.80688622754491, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1301, "step": 16571 }, { "epoch": 24.808383233532933, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1286, "step": 16572 }, { "epoch": 24.809880239520957, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1303, "step": 16573 }, { "epoch": 24.811377245508982, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1343, "step": 16574 }, { "epoch": 24.812874251497007, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1302, "step": 16575 }, { "epoch": 24.81437125748503, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1326, "step": 16576 }, { "epoch": 24.815868263473053, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1281, "step": 16577 }, { "epoch": 24.817365269461078, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1328, "step": 16578 }, { "epoch": 24.818862275449103, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1294, "step": 16579 }, { "epoch": 24.820359281437124, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1296, "step": 16580 }, { "epoch": 24.82185628742515, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1323, "step": 16581 }, { "epoch": 24.823353293413174, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1335, "step": 16582 }, { "epoch": 24.8248502994012, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.125, "step": 16583 }, { "epoch": 24.82634730538922, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1285, "step": 16584 }, { "epoch": 24.827844311377245, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1341, "step": 16585 }, { "epoch": 24.82934131736527, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1283, "step": 16586 }, { "epoch": 24.830838323353294, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1311, "step": 16587 }, { "epoch": 24.83233532934132, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1216, "step": 16588 }, { "epoch": 24.83383233532934, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.134, "step": 16589 }, { "epoch": 24.835329341317365, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.136, "step": 16590 }, { "epoch": 24.83682634730539, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1226, "step": 16591 }, { "epoch": 24.83832335329341, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1269, "step": 16592 }, { "epoch": 24.839820359281436, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1312, "step": 16593 }, { "epoch": 24.84131736526946, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1373, "step": 16594 }, { "epoch": 24.842814371257486, "grad_norm": 0.2158203125, "learning_rate": 0.0008, "loss": 1.1371, "step": 16595 }, { "epoch": 24.84431137724551, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1348, "step": 16596 }, { "epoch": 24.845808383233532, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1341, "step": 16597 }, { "epoch": 24.847305389221557, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.132, "step": 16598 }, { "epoch": 24.84880239520958, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1322, "step": 16599 }, { "epoch": 24.850299401197606, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1357, "step": 16600 }, { "epoch": 24.851796407185628, "grad_norm": 0.130859375, "learning_rate": 0.0008, "loss": 1.1265, "step": 16601 }, { "epoch": 24.853293413173652, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1291, "step": 16602 }, { "epoch": 24.854790419161677, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1311, "step": 16603 }, { "epoch": 24.856287425149702, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1322, "step": 16604 }, { "epoch": 24.857784431137723, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1277, "step": 16605 }, { "epoch": 24.85928143712575, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1359, "step": 16606 }, { "epoch": 24.860778443113773, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.127, "step": 16607 }, { "epoch": 24.862275449101798, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1371, "step": 16608 }, { "epoch": 24.86377245508982, "grad_norm": 0.11181640625, "learning_rate": 0.0008, "loss": 1.1261, "step": 16609 }, { "epoch": 24.865269461077844, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1325, "step": 16610 }, { "epoch": 24.86676646706587, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1262, "step": 16611 }, { "epoch": 24.868263473053894, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.124, "step": 16612 }, { "epoch": 24.869760479041915, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1318, "step": 16613 }, { "epoch": 24.87125748502994, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.131, "step": 16614 }, { "epoch": 24.872754491017965, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1275, "step": 16615 }, { "epoch": 24.87425149700599, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1371, "step": 16616 }, { "epoch": 24.87574850299401, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1221, "step": 16617 }, { "epoch": 24.877245508982035, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1313, "step": 16618 }, { "epoch": 24.87874251497006, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1273, "step": 16619 }, { "epoch": 24.880239520958085, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.128, "step": 16620 }, { "epoch": 24.881736526946106, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1309, "step": 16621 }, { "epoch": 24.88323353293413, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1293, "step": 16622 }, { "epoch": 24.884730538922156, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1328, "step": 16623 }, { "epoch": 24.88622754491018, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1279, "step": 16624 }, { "epoch": 24.887724550898202, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.125, "step": 16625 }, { "epoch": 24.889221556886227, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.122, "step": 16626 }, { "epoch": 24.89071856287425, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1247, "step": 16627 }, { "epoch": 24.892215568862277, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1303, "step": 16628 }, { "epoch": 24.893712574850298, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1268, "step": 16629 }, { "epoch": 24.895209580838323, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.137, "step": 16630 }, { "epoch": 24.896706586826348, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1307, "step": 16631 }, { "epoch": 24.898203592814372, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1309, "step": 16632 }, { "epoch": 24.899700598802394, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1271, "step": 16633 }, { "epoch": 24.90119760479042, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1254, "step": 16634 }, { "epoch": 24.902694610778443, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1335, "step": 16635 }, { "epoch": 24.904191616766468, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1305, "step": 16636 }, { "epoch": 24.90568862275449, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1292, "step": 16637 }, { "epoch": 24.907185628742514, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1285, "step": 16638 }, { "epoch": 24.90868263473054, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1309, "step": 16639 }, { "epoch": 24.910179640718564, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1343, "step": 16640 }, { "epoch": 24.91167664670659, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1342, "step": 16641 }, { "epoch": 24.91317365269461, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1319, "step": 16642 }, { "epoch": 24.914670658682635, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1341, "step": 16643 }, { "epoch": 24.91616766467066, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1365, "step": 16644 }, { "epoch": 24.91766467065868, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1391, "step": 16645 }, { "epoch": 24.919161676646706, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1333, "step": 16646 }, { "epoch": 24.92065868263473, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1291, "step": 16647 }, { "epoch": 24.922155688622755, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1334, "step": 16648 }, { "epoch": 24.92365269461078, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1292, "step": 16649 }, { "epoch": 24.9251497005988, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1296, "step": 16650 }, { "epoch": 24.926646706586826, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1284, "step": 16651 }, { "epoch": 24.92814371257485, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1298, "step": 16652 }, { "epoch": 24.929640718562876, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1277, "step": 16653 }, { "epoch": 24.931137724550897, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1311, "step": 16654 }, { "epoch": 24.932634730538922, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1248, "step": 16655 }, { "epoch": 24.934131736526947, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1242, "step": 16656 }, { "epoch": 24.93562874251497, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1311, "step": 16657 }, { "epoch": 24.937125748502993, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.13, "step": 16658 }, { "epoch": 24.938622754491018, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1315, "step": 16659 }, { "epoch": 24.940119760479043, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1357, "step": 16660 }, { "epoch": 24.941616766467067, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1276, "step": 16661 }, { "epoch": 24.94311377245509, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1314, "step": 16662 }, { "epoch": 24.944610778443113, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.133, "step": 16663 }, { "epoch": 24.94610778443114, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1302, "step": 16664 }, { "epoch": 24.947604790419163, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1415, "step": 16665 }, { "epoch": 24.949101796407184, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1353, "step": 16666 }, { "epoch": 24.95059880239521, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1289, "step": 16667 }, { "epoch": 24.952095808383234, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1312, "step": 16668 }, { "epoch": 24.95359281437126, "grad_norm": 0.11767578125, "learning_rate": 0.0008, "loss": 1.1244, "step": 16669 }, { "epoch": 24.95508982035928, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1316, "step": 16670 }, { "epoch": 24.956586826347305, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1308, "step": 16671 }, { "epoch": 24.95808383233533, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1296, "step": 16672 }, { "epoch": 24.959580838323355, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1353, "step": 16673 }, { "epoch": 24.961077844311376, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1314, "step": 16674 }, { "epoch": 24.9625748502994, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1415, "step": 16675 }, { "epoch": 24.964071856287426, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1347, "step": 16676 }, { "epoch": 24.96556886227545, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1303, "step": 16677 }, { "epoch": 24.96706586826347, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1248, "step": 16678 }, { "epoch": 24.968562874251496, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1353, "step": 16679 }, { "epoch": 24.97005988023952, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.133, "step": 16680 }, { "epoch": 24.971556886227546, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1297, "step": 16681 }, { "epoch": 24.973053892215567, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1283, "step": 16682 }, { "epoch": 24.974550898203592, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1239, "step": 16683 }, { "epoch": 24.976047904191617, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1342, "step": 16684 }, { "epoch": 24.977544910179642, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1347, "step": 16685 }, { "epoch": 24.979041916167663, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1323, "step": 16686 }, { "epoch": 24.980538922155688, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1302, "step": 16687 }, { "epoch": 24.982035928143713, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1232, "step": 16688 }, { "epoch": 24.983532934131738, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.136, "step": 16689 }, { "epoch": 24.98502994011976, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1225, "step": 16690 }, { "epoch": 24.986526946107784, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1394, "step": 16691 }, { "epoch": 24.98802395209581, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1288, "step": 16692 }, { "epoch": 24.989520958083833, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.133, "step": 16693 }, { "epoch": 24.991017964071855, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1256, "step": 16694 }, { "epoch": 24.99251497005988, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1298, "step": 16695 }, { "epoch": 24.994011976047904, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1326, "step": 16696 }, { "epoch": 24.99550898203593, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1337, "step": 16697 }, { "epoch": 24.99700598802395, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1274, "step": 16698 }, { "epoch": 24.998502994011975, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1336, "step": 16699 }, { "epoch": 25.0, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1276, "step": 16700 }, { "epoch": 25.001497005988025, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1316, "step": 16701 }, { "epoch": 25.00299401197605, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1349, "step": 16702 }, { "epoch": 25.00449101796407, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1254, "step": 16703 }, { "epoch": 25.005988023952096, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1375, "step": 16704 }, { "epoch": 25.00748502994012, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1238, "step": 16705 }, { "epoch": 25.008982035928145, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1344, "step": 16706 }, { "epoch": 25.010479041916167, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1347, "step": 16707 }, { "epoch": 25.01197604790419, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1298, "step": 16708 }, { "epoch": 25.013473053892216, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1357, "step": 16709 }, { "epoch": 25.01497005988024, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1278, "step": 16710 }, { "epoch": 25.016467065868262, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1295, "step": 16711 }, { "epoch": 25.017964071856287, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1348, "step": 16712 }, { "epoch": 25.019461077844312, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1286, "step": 16713 }, { "epoch": 25.020958083832337, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1335, "step": 16714 }, { "epoch": 25.022455089820358, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1318, "step": 16715 }, { "epoch": 25.023952095808383, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.1297, "step": 16716 }, { "epoch": 25.025449101796408, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.128, "step": 16717 }, { "epoch": 25.026946107784433, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.134, "step": 16718 }, { "epoch": 25.028443113772454, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1298, "step": 16719 }, { "epoch": 25.02994011976048, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1363, "step": 16720 }, { "epoch": 25.031437125748504, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1333, "step": 16721 }, { "epoch": 25.03293413173653, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1356, "step": 16722 }, { "epoch": 25.03443113772455, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1341, "step": 16723 }, { "epoch": 25.035928143712574, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1289, "step": 16724 }, { "epoch": 25.0374251497006, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1253, "step": 16725 }, { "epoch": 25.038922155688624, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1298, "step": 16726 }, { "epoch": 25.040419161676645, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1301, "step": 16727 }, { "epoch": 25.04191616766467, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1306, "step": 16728 }, { "epoch": 25.043413173652695, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1281, "step": 16729 }, { "epoch": 25.04491017964072, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1321, "step": 16730 }, { "epoch": 25.04640718562874, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1307, "step": 16731 }, { "epoch": 25.047904191616766, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1332, "step": 16732 }, { "epoch": 25.04940119760479, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1279, "step": 16733 }, { "epoch": 25.050898203592816, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1341, "step": 16734 }, { "epoch": 25.052395209580837, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1299, "step": 16735 }, { "epoch": 25.05389221556886, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1296, "step": 16736 }, { "epoch": 25.055389221556887, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1281, "step": 16737 }, { "epoch": 25.05688622754491, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1336, "step": 16738 }, { "epoch": 25.058383233532933, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1354, "step": 16739 }, { "epoch": 25.059880239520957, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1305, "step": 16740 }, { "epoch": 25.061377245508982, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.13, "step": 16741 }, { "epoch": 25.062874251497007, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1313, "step": 16742 }, { "epoch": 25.06437125748503, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1309, "step": 16743 }, { "epoch": 25.065868263473053, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1306, "step": 16744 }, { "epoch": 25.067365269461078, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1291, "step": 16745 }, { "epoch": 25.068862275449103, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1309, "step": 16746 }, { "epoch": 25.070359281437124, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1276, "step": 16747 }, { "epoch": 25.07185628742515, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1266, "step": 16748 }, { "epoch": 25.073353293413174, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1258, "step": 16749 }, { "epoch": 25.0748502994012, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1321, "step": 16750 }, { "epoch": 25.07634730538922, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1331, "step": 16751 }, { "epoch": 25.077844311377245, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1332, "step": 16752 }, { "epoch": 25.07934131736527, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1311, "step": 16753 }, { "epoch": 25.080838323353294, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1265, "step": 16754 }, { "epoch": 25.082335329341316, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1379, "step": 16755 }, { "epoch": 25.08383233532934, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.132, "step": 16756 }, { "epoch": 25.085329341317365, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1296, "step": 16757 }, { "epoch": 25.08682634730539, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1316, "step": 16758 }, { "epoch": 25.088323353293415, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1303, "step": 16759 }, { "epoch": 25.089820359281436, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1293, "step": 16760 }, { "epoch": 25.09131736526946, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1282, "step": 16761 }, { "epoch": 25.092814371257486, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1332, "step": 16762 }, { "epoch": 25.09431137724551, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1296, "step": 16763 }, { "epoch": 25.095808383233532, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1324, "step": 16764 }, { "epoch": 25.097305389221557, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1325, "step": 16765 }, { "epoch": 25.09880239520958, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1348, "step": 16766 }, { "epoch": 25.100299401197606, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1407, "step": 16767 }, { "epoch": 25.101796407185628, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1314, "step": 16768 }, { "epoch": 25.103293413173652, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1264, "step": 16769 }, { "epoch": 25.104790419161677, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1301, "step": 16770 }, { "epoch": 25.106287425149702, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1318, "step": 16771 }, { "epoch": 25.107784431137723, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1376, "step": 16772 }, { "epoch": 25.10928143712575, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1371, "step": 16773 }, { "epoch": 25.110778443113773, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1307, "step": 16774 }, { "epoch": 25.112275449101798, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.126, "step": 16775 }, { "epoch": 25.11377245508982, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1388, "step": 16776 }, { "epoch": 25.115269461077844, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1288, "step": 16777 }, { "epoch": 25.11676646706587, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1281, "step": 16778 }, { "epoch": 25.118263473053894, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1348, "step": 16779 }, { "epoch": 25.119760479041915, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1292, "step": 16780 }, { "epoch": 25.12125748502994, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1281, "step": 16781 }, { "epoch": 25.122754491017965, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1309, "step": 16782 }, { "epoch": 25.12425149700599, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 16783 }, { "epoch": 25.12574850299401, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1362, "step": 16784 }, { "epoch": 25.127245508982035, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1281, "step": 16785 }, { "epoch": 25.12874251497006, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1348, "step": 16786 }, { "epoch": 25.130239520958085, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1308, "step": 16787 }, { "epoch": 25.131736526946106, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1259, "step": 16788 }, { "epoch": 25.13323353293413, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1295, "step": 16789 }, { "epoch": 25.134730538922156, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1271, "step": 16790 }, { "epoch": 25.13622754491018, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.13, "step": 16791 }, { "epoch": 25.137724550898202, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1292, "step": 16792 }, { "epoch": 25.139221556886227, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1308, "step": 16793 }, { "epoch": 25.14071856287425, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1315, "step": 16794 }, { "epoch": 25.142215568862277, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1321, "step": 16795 }, { "epoch": 25.143712574850298, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1298, "step": 16796 }, { "epoch": 25.145209580838323, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1263, "step": 16797 }, { "epoch": 25.146706586826348, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1275, "step": 16798 }, { "epoch": 25.148203592814372, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1372, "step": 16799 }, { "epoch": 25.149700598802394, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1287, "step": 16800 }, { "epoch": 25.15119760479042, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1306, "step": 16801 }, { "epoch": 25.152694610778443, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1288, "step": 16802 }, { "epoch": 25.154191616766468, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1328, "step": 16803 }, { "epoch": 25.15568862275449, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1328, "step": 16804 }, { "epoch": 25.157185628742514, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1263, "step": 16805 }, { "epoch": 25.15868263473054, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1287, "step": 16806 }, { "epoch": 25.160179640718564, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1336, "step": 16807 }, { "epoch": 25.161676646706585, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1313, "step": 16808 }, { "epoch": 25.16317365269461, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1325, "step": 16809 }, { "epoch": 25.164670658682635, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.125, "step": 16810 }, { "epoch": 25.16616766467066, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1311, "step": 16811 }, { "epoch": 25.16766467065868, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.132, "step": 16812 }, { "epoch": 25.169161676646706, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1323, "step": 16813 }, { "epoch": 25.17065868263473, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1283, "step": 16814 }, { "epoch": 25.172155688622755, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1264, "step": 16815 }, { "epoch": 25.17365269461078, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1295, "step": 16816 }, { "epoch": 25.1751497005988, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1292, "step": 16817 }, { "epoch": 25.176646706586826, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1383, "step": 16818 }, { "epoch": 25.17814371257485, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.131, "step": 16819 }, { "epoch": 25.179640718562876, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1325, "step": 16820 }, { "epoch": 25.181137724550897, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1285, "step": 16821 }, { "epoch": 25.182634730538922, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1315, "step": 16822 }, { "epoch": 25.184131736526947, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1343, "step": 16823 }, { "epoch": 25.18562874251497, "grad_norm": 0.05419921875, "learning_rate": 0.0008, "loss": 1.135, "step": 16824 }, { "epoch": 25.187125748502993, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1235, "step": 16825 }, { "epoch": 25.188622754491018, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1372, "step": 16826 }, { "epoch": 25.190119760479043, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1267, "step": 16827 }, { "epoch": 25.191616766467067, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1294, "step": 16828 }, { "epoch": 25.19311377245509, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1292, "step": 16829 }, { "epoch": 25.194610778443113, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1326, "step": 16830 }, { "epoch": 25.19610778443114, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1241, "step": 16831 }, { "epoch": 25.197604790419163, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1293, "step": 16832 }, { "epoch": 25.199101796407184, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1345, "step": 16833 }, { "epoch": 25.20059880239521, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1348, "step": 16834 }, { "epoch": 25.202095808383234, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1323, "step": 16835 }, { "epoch": 25.20359281437126, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1359, "step": 16836 }, { "epoch": 25.20508982035928, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1271, "step": 16837 }, { "epoch": 25.206586826347305, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1319, "step": 16838 }, { "epoch": 25.20808383233533, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1293, "step": 16839 }, { "epoch": 25.209580838323355, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1265, "step": 16840 }, { "epoch": 25.211077844311376, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1386, "step": 16841 }, { "epoch": 25.2125748502994, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1331, "step": 16842 }, { "epoch": 25.214071856287426, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1351, "step": 16843 }, { "epoch": 25.21556886227545, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1295, "step": 16844 }, { "epoch": 25.21706586826347, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1346, "step": 16845 }, { "epoch": 25.218562874251496, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1382, "step": 16846 }, { "epoch": 25.22005988023952, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1249, "step": 16847 }, { "epoch": 25.221556886227546, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.135, "step": 16848 }, { "epoch": 25.223053892215567, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1346, "step": 16849 }, { "epoch": 25.224550898203592, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.135, "step": 16850 }, { "epoch": 25.226047904191617, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1269, "step": 16851 }, { "epoch": 25.227544910179642, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1318, "step": 16852 }, { "epoch": 25.229041916167663, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1294, "step": 16853 }, { "epoch": 25.230538922155688, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1248, "step": 16854 }, { "epoch": 25.232035928143713, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1367, "step": 16855 }, { "epoch": 25.233532934131738, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1358, "step": 16856 }, { "epoch": 25.23502994011976, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.123, "step": 16857 }, { "epoch": 25.236526946107784, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1298, "step": 16858 }, { "epoch": 25.23802395209581, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1261, "step": 16859 }, { "epoch": 25.239520958083833, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1335, "step": 16860 }, { "epoch": 25.241017964071855, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1282, "step": 16861 }, { "epoch": 25.24251497005988, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 16862 }, { "epoch": 25.244011976047904, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1352, "step": 16863 }, { "epoch": 25.24550898203593, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1285, "step": 16864 }, { "epoch": 25.24700598802395, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1258, "step": 16865 }, { "epoch": 25.248502994011975, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1348, "step": 16866 }, { "epoch": 25.25, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1288, "step": 16867 }, { "epoch": 25.251497005988025, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1316, "step": 16868 }, { "epoch": 25.25299401197605, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1359, "step": 16869 }, { "epoch": 25.25449101796407, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1259, "step": 16870 }, { "epoch": 25.255988023952096, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1273, "step": 16871 }, { "epoch": 25.25748502994012, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1357, "step": 16872 }, { "epoch": 25.258982035928145, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1213, "step": 16873 }, { "epoch": 25.260479041916167, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1328, "step": 16874 }, { "epoch": 25.26197604790419, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1249, "step": 16875 }, { "epoch": 25.263473053892216, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1282, "step": 16876 }, { "epoch": 25.26497005988024, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1326, "step": 16877 }, { "epoch": 25.266467065868262, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1355, "step": 16878 }, { "epoch": 25.267964071856287, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1308, "step": 16879 }, { "epoch": 25.269461077844312, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1336, "step": 16880 }, { "epoch": 25.270958083832337, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.132, "step": 16881 }, { "epoch": 25.272455089820358, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1303, "step": 16882 }, { "epoch": 25.273952095808383, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1322, "step": 16883 }, { "epoch": 25.275449101796408, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1285, "step": 16884 }, { "epoch": 25.276946107784433, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1277, "step": 16885 }, { "epoch": 25.278443113772454, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1329, "step": 16886 }, { "epoch": 25.27994011976048, "grad_norm": 0.12451171875, "learning_rate": 0.0008, "loss": 1.1236, "step": 16887 }, { "epoch": 25.281437125748504, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1322, "step": 16888 }, { "epoch": 25.28293413173653, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1283, "step": 16889 }, { "epoch": 25.28443113772455, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1338, "step": 16890 }, { "epoch": 25.285928143712574, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1252, "step": 16891 }, { "epoch": 25.2874251497006, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1356, "step": 16892 }, { "epoch": 25.288922155688624, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1237, "step": 16893 }, { "epoch": 25.290419161676645, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1385, "step": 16894 }, { "epoch": 25.29191616766467, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1259, "step": 16895 }, { "epoch": 25.293413173652695, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1321, "step": 16896 }, { "epoch": 25.29491017964072, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1317, "step": 16897 }, { "epoch": 25.29640718562874, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1278, "step": 16898 }, { "epoch": 25.297904191616766, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1316, "step": 16899 }, { "epoch": 25.29940119760479, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1298, "step": 16900 }, { "epoch": 25.300898203592816, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1268, "step": 16901 }, { "epoch": 25.302395209580837, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1319, "step": 16902 }, { "epoch": 25.30389221556886, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1323, "step": 16903 }, { "epoch": 25.305389221556887, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1363, "step": 16904 }, { "epoch": 25.30688622754491, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1303, "step": 16905 }, { "epoch": 25.308383233532933, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1255, "step": 16906 }, { "epoch": 25.309880239520957, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1249, "step": 16907 }, { "epoch": 25.311377245508982, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1257, "step": 16908 }, { "epoch": 25.312874251497007, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1265, "step": 16909 }, { "epoch": 25.31437125748503, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1255, "step": 16910 }, { "epoch": 25.315868263473053, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1327, "step": 16911 }, { "epoch": 25.317365269461078, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1253, "step": 16912 }, { "epoch": 25.318862275449103, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1309, "step": 16913 }, { "epoch": 25.320359281437124, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1309, "step": 16914 }, { "epoch": 25.32185628742515, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1326, "step": 16915 }, { "epoch": 25.323353293413174, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1315, "step": 16916 }, { "epoch": 25.3248502994012, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1278, "step": 16917 }, { "epoch": 25.32634730538922, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.129, "step": 16918 }, { "epoch": 25.327844311377245, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1281, "step": 16919 }, { "epoch": 25.32934131736527, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1361, "step": 16920 }, { "epoch": 25.330838323353294, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.134, "step": 16921 }, { "epoch": 25.33233532934132, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1351, "step": 16922 }, { "epoch": 25.33383233532934, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1295, "step": 16923 }, { "epoch": 25.335329341317365, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1301, "step": 16924 }, { "epoch": 25.33682634730539, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1295, "step": 16925 }, { "epoch": 25.338323353293415, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1343, "step": 16926 }, { "epoch": 25.339820359281436, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1388, "step": 16927 }, { "epoch": 25.34131736526946, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1315, "step": 16928 }, { "epoch": 25.342814371257486, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1326, "step": 16929 }, { "epoch": 25.34431137724551, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1296, "step": 16930 }, { "epoch": 25.345808383233532, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1277, "step": 16931 }, { "epoch": 25.347305389221557, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1288, "step": 16932 }, { "epoch": 25.34880239520958, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1344, "step": 16933 }, { "epoch": 25.350299401197606, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.132, "step": 16934 }, { "epoch": 25.351796407185628, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1297, "step": 16935 }, { "epoch": 25.353293413173652, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1327, "step": 16936 }, { "epoch": 25.354790419161677, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1306, "step": 16937 }, { "epoch": 25.356287425149702, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1175, "step": 16938 }, { "epoch": 25.357784431137723, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1326, "step": 16939 }, { "epoch": 25.35928143712575, "grad_norm": 0.05126953125, "learning_rate": 0.0008, "loss": 1.1332, "step": 16940 }, { "epoch": 25.360778443113773, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.124, "step": 16941 }, { "epoch": 25.362275449101798, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1351, "step": 16942 }, { "epoch": 25.36377245508982, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.131, "step": 16943 }, { "epoch": 25.365269461077844, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1394, "step": 16944 }, { "epoch": 25.36676646706587, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1302, "step": 16945 }, { "epoch": 25.368263473053894, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1313, "step": 16946 }, { "epoch": 25.369760479041915, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1378, "step": 16947 }, { "epoch": 25.37125748502994, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1338, "step": 16948 }, { "epoch": 25.372754491017965, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1336, "step": 16949 }, { "epoch": 25.37425149700599, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1309, "step": 16950 }, { "epoch": 25.37574850299401, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1314, "step": 16951 }, { "epoch": 25.377245508982035, "grad_norm": 0.052734375, "learning_rate": 0.0008, "loss": 1.1418, "step": 16952 }, { "epoch": 25.37874251497006, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.133, "step": 16953 }, { "epoch": 25.380239520958085, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1282, "step": 16954 }, { "epoch": 25.381736526946106, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1269, "step": 16955 }, { "epoch": 25.38323353293413, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1308, "step": 16956 }, { "epoch": 25.384730538922156, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1264, "step": 16957 }, { "epoch": 25.38622754491018, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1279, "step": 16958 }, { "epoch": 25.387724550898202, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1354, "step": 16959 }, { "epoch": 25.389221556886227, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1263, "step": 16960 }, { "epoch": 25.39071856287425, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1293, "step": 16961 }, { "epoch": 25.392215568862277, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1316, "step": 16962 }, { "epoch": 25.393712574850298, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1294, "step": 16963 }, { "epoch": 25.395209580838323, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1331, "step": 16964 }, { "epoch": 25.396706586826348, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1296, "step": 16965 }, { "epoch": 25.398203592814372, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1264, "step": 16966 }, { "epoch": 25.399700598802394, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.13, "step": 16967 }, { "epoch": 25.40119760479042, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1349, "step": 16968 }, { "epoch": 25.402694610778443, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1301, "step": 16969 }, { "epoch": 25.404191616766468, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.132, "step": 16970 }, { "epoch": 25.40568862275449, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1342, "step": 16971 }, { "epoch": 25.407185628742514, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1369, "step": 16972 }, { "epoch": 25.40868263473054, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1303, "step": 16973 }, { "epoch": 25.410179640718564, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1286, "step": 16974 }, { "epoch": 25.411676646706585, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1245, "step": 16975 }, { "epoch": 25.41317365269461, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1266, "step": 16976 }, { "epoch": 25.414670658682635, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1234, "step": 16977 }, { "epoch": 25.41616766467066, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1348, "step": 16978 }, { "epoch": 25.41766467065868, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1322, "step": 16979 }, { "epoch": 25.419161676646706, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1342, "step": 16980 }, { "epoch": 25.42065868263473, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1244, "step": 16981 }, { "epoch": 25.422155688622755, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1288, "step": 16982 }, { "epoch": 25.42365269461078, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1317, "step": 16983 }, { "epoch": 25.4251497005988, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1269, "step": 16984 }, { "epoch": 25.426646706586826, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1374, "step": 16985 }, { "epoch": 25.42814371257485, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.132, "step": 16986 }, { "epoch": 25.429640718562876, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1277, "step": 16987 }, { "epoch": 25.431137724550897, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.131, "step": 16988 }, { "epoch": 25.432634730538922, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1318, "step": 16989 }, { "epoch": 25.434131736526947, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1222, "step": 16990 }, { "epoch": 25.43562874251497, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1278, "step": 16991 }, { "epoch": 25.437125748502993, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1283, "step": 16992 }, { "epoch": 25.438622754491018, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1362, "step": 16993 }, { "epoch": 25.440119760479043, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1287, "step": 16994 }, { "epoch": 25.441616766467067, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1265, "step": 16995 }, { "epoch": 25.44311377245509, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1327, "step": 16996 }, { "epoch": 25.444610778443113, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1338, "step": 16997 }, { "epoch": 25.44610778443114, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1346, "step": 16998 }, { "epoch": 25.447604790419163, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1363, "step": 16999 }, { "epoch": 25.449101796407184, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1299, "step": 17000 }, { "epoch": 25.45059880239521, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1269, "step": 17001 }, { "epoch": 25.452095808383234, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.124, "step": 17002 }, { "epoch": 25.45359281437126, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1279, "step": 17003 }, { "epoch": 25.45508982035928, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1257, "step": 17004 }, { "epoch": 25.456586826347305, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1334, "step": 17005 }, { "epoch": 25.45808383233533, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1327, "step": 17006 }, { "epoch": 25.459580838323355, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1307, "step": 17007 }, { "epoch": 25.461077844311376, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1383, "step": 17008 }, { "epoch": 25.4625748502994, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1313, "step": 17009 }, { "epoch": 25.464071856287426, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1297, "step": 17010 }, { "epoch": 25.46556886227545, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1295, "step": 17011 }, { "epoch": 25.46706586826347, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1306, "step": 17012 }, { "epoch": 25.468562874251496, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.132, "step": 17013 }, { "epoch": 25.47005988023952, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.129, "step": 17014 }, { "epoch": 25.471556886227546, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1363, "step": 17015 }, { "epoch": 25.473053892215567, "grad_norm": 0.05126953125, "learning_rate": 0.0008, "loss": 1.1345, "step": 17016 }, { "epoch": 25.474550898203592, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1318, "step": 17017 }, { "epoch": 25.476047904191617, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1271, "step": 17018 }, { "epoch": 25.477544910179642, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1344, "step": 17019 }, { "epoch": 25.479041916167663, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1336, "step": 17020 }, { "epoch": 25.480538922155688, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1295, "step": 17021 }, { "epoch": 25.482035928143713, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1327, "step": 17022 }, { "epoch": 25.483532934131738, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1332, "step": 17023 }, { "epoch": 25.48502994011976, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1297, "step": 17024 }, { "epoch": 25.486526946107784, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.132, "step": 17025 }, { "epoch": 25.48802395209581, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1253, "step": 17026 }, { "epoch": 25.489520958083833, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1257, "step": 17027 }, { "epoch": 25.491017964071855, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1348, "step": 17028 }, { "epoch": 25.49251497005988, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1345, "step": 17029 }, { "epoch": 25.494011976047904, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1273, "step": 17030 }, { "epoch": 25.49550898203593, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1312, "step": 17031 }, { "epoch": 25.49700598802395, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1347, "step": 17032 }, { "epoch": 25.498502994011975, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1325, "step": 17033 }, { "epoch": 25.5, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1303, "step": 17034 }, { "epoch": 25.501497005988025, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1276, "step": 17035 }, { "epoch": 25.50299401197605, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.138, "step": 17036 }, { "epoch": 25.50449101796407, "grad_norm": 0.1181640625, "learning_rate": 0.0008, "loss": 1.1296, "step": 17037 }, { "epoch": 25.505988023952096, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1333, "step": 17038 }, { "epoch": 25.50748502994012, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1399, "step": 17039 }, { "epoch": 25.508982035928145, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.127, "step": 17040 }, { "epoch": 25.510479041916167, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1374, "step": 17041 }, { "epoch": 25.51197604790419, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1287, "step": 17042 }, { "epoch": 25.513473053892216, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1355, "step": 17043 }, { "epoch": 25.51497005988024, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1305, "step": 17044 }, { "epoch": 25.516467065868262, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1302, "step": 17045 }, { "epoch": 25.517964071856287, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1263, "step": 17046 }, { "epoch": 25.519461077844312, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1361, "step": 17047 }, { "epoch": 25.520958083832337, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1307, "step": 17048 }, { "epoch": 25.522455089820358, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1246, "step": 17049 }, { "epoch": 25.523952095808383, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1294, "step": 17050 }, { "epoch": 25.525449101796408, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1338, "step": 17051 }, { "epoch": 25.526946107784433, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1245, "step": 17052 }, { "epoch": 25.528443113772454, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1306, "step": 17053 }, { "epoch": 25.52994011976048, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1285, "step": 17054 }, { "epoch": 25.531437125748504, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1239, "step": 17055 }, { "epoch": 25.53293413173653, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1336, "step": 17056 }, { "epoch": 25.53443113772455, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1316, "step": 17057 }, { "epoch": 25.535928143712574, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1427, "step": 17058 }, { "epoch": 25.5374251497006, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1287, "step": 17059 }, { "epoch": 25.538922155688624, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1327, "step": 17060 }, { "epoch": 25.540419161676645, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1254, "step": 17061 }, { "epoch": 25.54191616766467, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1262, "step": 17062 }, { "epoch": 25.543413173652695, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1341, "step": 17063 }, { "epoch": 25.54491017964072, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1315, "step": 17064 }, { "epoch": 25.54640718562874, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1342, "step": 17065 }, { "epoch": 25.547904191616766, "grad_norm": 0.103515625, "learning_rate": 0.0008, "loss": 1.1312, "step": 17066 }, { "epoch": 25.54940119760479, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1289, "step": 17067 }, { "epoch": 25.550898203592816, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1298, "step": 17068 }, { "epoch": 25.552395209580837, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1332, "step": 17069 }, { "epoch": 25.55389221556886, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.132, "step": 17070 }, { "epoch": 25.555389221556887, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1266, "step": 17071 }, { "epoch": 25.55688622754491, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1333, "step": 17072 }, { "epoch": 25.558383233532933, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1335, "step": 17073 }, { "epoch": 25.559880239520957, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1255, "step": 17074 }, { "epoch": 25.561377245508982, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1279, "step": 17075 }, { "epoch": 25.562874251497007, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1346, "step": 17076 }, { "epoch": 25.56437125748503, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1269, "step": 17077 }, { "epoch": 25.565868263473053, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 17078 }, { "epoch": 25.567365269461078, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.127, "step": 17079 }, { "epoch": 25.568862275449103, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.129, "step": 17080 }, { "epoch": 25.570359281437124, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1321, "step": 17081 }, { "epoch": 25.57185628742515, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1324, "step": 17082 }, { "epoch": 25.573353293413174, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1281, "step": 17083 }, { "epoch": 25.5748502994012, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1283, "step": 17084 }, { "epoch": 25.57634730538922, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1255, "step": 17085 }, { "epoch": 25.577844311377245, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1438, "step": 17086 }, { "epoch": 25.57934131736527, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.129, "step": 17087 }, { "epoch": 25.580838323353294, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1304, "step": 17088 }, { "epoch": 25.58233532934132, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1317, "step": 17089 }, { "epoch": 25.58383233532934, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1254, "step": 17090 }, { "epoch": 25.585329341317365, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1252, "step": 17091 }, { "epoch": 25.58682634730539, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1224, "step": 17092 }, { "epoch": 25.58832335329341, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1329, "step": 17093 }, { "epoch": 25.589820359281436, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1378, "step": 17094 }, { "epoch": 25.59131736526946, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1312, "step": 17095 }, { "epoch": 25.592814371257486, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1242, "step": 17096 }, { "epoch": 25.59431137724551, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1266, "step": 17097 }, { "epoch": 25.595808383233532, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1308, "step": 17098 }, { "epoch": 25.597305389221557, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1288, "step": 17099 }, { "epoch": 25.59880239520958, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1278, "step": 17100 }, { "epoch": 25.600299401197606, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1256, "step": 17101 }, { "epoch": 25.601796407185628, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1284, "step": 17102 }, { "epoch": 25.603293413173652, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1312, "step": 17103 }, { "epoch": 25.604790419161677, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1262, "step": 17104 }, { "epoch": 25.606287425149702, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.13, "step": 17105 }, { "epoch": 25.607784431137723, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1312, "step": 17106 }, { "epoch": 25.60928143712575, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1374, "step": 17107 }, { "epoch": 25.610778443113773, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1285, "step": 17108 }, { "epoch": 25.612275449101798, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1239, "step": 17109 }, { "epoch": 25.61377245508982, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1271, "step": 17110 }, { "epoch": 25.615269461077844, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1359, "step": 17111 }, { "epoch": 25.61676646706587, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1234, "step": 17112 }, { "epoch": 25.618263473053894, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1274, "step": 17113 }, { "epoch": 25.619760479041915, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1349, "step": 17114 }, { "epoch": 25.62125748502994, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1261, "step": 17115 }, { "epoch": 25.622754491017965, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1335, "step": 17116 }, { "epoch": 25.62425149700599, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1299, "step": 17117 }, { "epoch": 25.62574850299401, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1334, "step": 17118 }, { "epoch": 25.627245508982035, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1244, "step": 17119 }, { "epoch": 25.62874251497006, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1277, "step": 17120 }, { "epoch": 25.630239520958085, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1315, "step": 17121 }, { "epoch": 25.631736526946106, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1288, "step": 17122 }, { "epoch": 25.63323353293413, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1299, "step": 17123 }, { "epoch": 25.634730538922156, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1309, "step": 17124 }, { "epoch": 25.63622754491018, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1281, "step": 17125 }, { "epoch": 25.637724550898202, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1359, "step": 17126 }, { "epoch": 25.639221556886227, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1288, "step": 17127 }, { "epoch": 25.64071856287425, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1336, "step": 17128 }, { "epoch": 25.642215568862277, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1261, "step": 17129 }, { "epoch": 25.643712574850298, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1178, "step": 17130 }, { "epoch": 25.645209580838323, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1297, "step": 17131 }, { "epoch": 25.646706586826348, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1338, "step": 17132 }, { "epoch": 25.648203592814372, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1305, "step": 17133 }, { "epoch": 25.649700598802394, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.128, "step": 17134 }, { "epoch": 25.65119760479042, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1328, "step": 17135 }, { "epoch": 25.652694610778443, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1284, "step": 17136 }, { "epoch": 25.654191616766468, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1288, "step": 17137 }, { "epoch": 25.65568862275449, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1291, "step": 17138 }, { "epoch": 25.657185628742514, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1227, "step": 17139 }, { "epoch": 25.65868263473054, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.127, "step": 17140 }, { "epoch": 25.660179640718564, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1266, "step": 17141 }, { "epoch": 25.66167664670659, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1261, "step": 17142 }, { "epoch": 25.66317365269461, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1325, "step": 17143 }, { "epoch": 25.664670658682635, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1357, "step": 17144 }, { "epoch": 25.66616766467066, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1329, "step": 17145 }, { "epoch": 25.66766467065868, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1334, "step": 17146 }, { "epoch": 25.669161676646706, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1265, "step": 17147 }, { "epoch": 25.67065868263473, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1328, "step": 17148 }, { "epoch": 25.672155688622755, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1293, "step": 17149 }, { "epoch": 25.67365269461078, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1258, "step": 17150 }, { "epoch": 25.6751497005988, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1304, "step": 17151 }, { "epoch": 25.676646706586826, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1286, "step": 17152 }, { "epoch": 25.67814371257485, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1258, "step": 17153 }, { "epoch": 25.679640718562876, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1297, "step": 17154 }, { "epoch": 25.681137724550897, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1302, "step": 17155 }, { "epoch": 25.682634730538922, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1313, "step": 17156 }, { "epoch": 25.684131736526947, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1255, "step": 17157 }, { "epoch": 25.68562874251497, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1256, "step": 17158 }, { "epoch": 25.687125748502993, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1368, "step": 17159 }, { "epoch": 25.688622754491018, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1308, "step": 17160 }, { "epoch": 25.690119760479043, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1356, "step": 17161 }, { "epoch": 25.691616766467067, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1264, "step": 17162 }, { "epoch": 25.69311377245509, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1281, "step": 17163 }, { "epoch": 25.694610778443113, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1319, "step": 17164 }, { "epoch": 25.69610778443114, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1251, "step": 17165 }, { "epoch": 25.697604790419163, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1293, "step": 17166 }, { "epoch": 25.699101796407184, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1339, "step": 17167 }, { "epoch": 25.70059880239521, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1321, "step": 17168 }, { "epoch": 25.702095808383234, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1324, "step": 17169 }, { "epoch": 25.70359281437126, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1273, "step": 17170 }, { "epoch": 25.70508982035928, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1276, "step": 17171 }, { "epoch": 25.706586826347305, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1249, "step": 17172 }, { "epoch": 25.70808383233533, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1267, "step": 17173 }, { "epoch": 25.709580838323355, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1356, "step": 17174 }, { "epoch": 25.711077844311376, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1249, "step": 17175 }, { "epoch": 25.7125748502994, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1291, "step": 17176 }, { "epoch": 25.714071856287426, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1345, "step": 17177 }, { "epoch": 25.71556886227545, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1309, "step": 17178 }, { "epoch": 25.71706586826347, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1315, "step": 17179 }, { "epoch": 25.718562874251496, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.132, "step": 17180 }, { "epoch": 25.72005988023952, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1309, "step": 17181 }, { "epoch": 25.721556886227546, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1301, "step": 17182 }, { "epoch": 25.723053892215567, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.13, "step": 17183 }, { "epoch": 25.724550898203592, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1308, "step": 17184 }, { "epoch": 25.726047904191617, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1383, "step": 17185 }, { "epoch": 25.727544910179642, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1267, "step": 17186 }, { "epoch": 25.729041916167663, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1337, "step": 17187 }, { "epoch": 25.730538922155688, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.133, "step": 17188 }, { "epoch": 25.732035928143713, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1354, "step": 17189 }, { "epoch": 25.733532934131738, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1333, "step": 17190 }, { "epoch": 25.73502994011976, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1286, "step": 17191 }, { "epoch": 25.736526946107784, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1326, "step": 17192 }, { "epoch": 25.73802395209581, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1302, "step": 17193 }, { "epoch": 25.739520958083833, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1351, "step": 17194 }, { "epoch": 25.741017964071855, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1385, "step": 17195 }, { "epoch": 25.74251497005988, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1279, "step": 17196 }, { "epoch": 25.744011976047904, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1302, "step": 17197 }, { "epoch": 25.74550898203593, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1237, "step": 17198 }, { "epoch": 25.74700598802395, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1324, "step": 17199 }, { "epoch": 25.748502994011975, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1289, "step": 17200 }, { "epoch": 25.75, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1268, "step": 17201 }, { "epoch": 25.751497005988025, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1305, "step": 17202 }, { "epoch": 25.75299401197605, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1344, "step": 17203 }, { "epoch": 25.75449101796407, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1356, "step": 17204 }, { "epoch": 25.755988023952096, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1378, "step": 17205 }, { "epoch": 25.75748502994012, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1288, "step": 17206 }, { "epoch": 25.758982035928145, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1314, "step": 17207 }, { "epoch": 25.760479041916167, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1283, "step": 17208 }, { "epoch": 25.76197604790419, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1293, "step": 17209 }, { "epoch": 25.763473053892216, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1254, "step": 17210 }, { "epoch": 25.76497005988024, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1252, "step": 17211 }, { "epoch": 25.766467065868262, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1305, "step": 17212 }, { "epoch": 25.767964071856287, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1236, "step": 17213 }, { "epoch": 25.769461077844312, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1296, "step": 17214 }, { "epoch": 25.770958083832337, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1343, "step": 17215 }, { "epoch": 25.772455089820358, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1339, "step": 17216 }, { "epoch": 25.773952095808383, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1323, "step": 17217 }, { "epoch": 25.775449101796408, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1335, "step": 17218 }, { "epoch": 25.776946107784433, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.125, "step": 17219 }, { "epoch": 25.778443113772454, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1355, "step": 17220 }, { "epoch": 25.77994011976048, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1307, "step": 17221 }, { "epoch": 25.781437125748504, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.132, "step": 17222 }, { "epoch": 25.78293413173653, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1309, "step": 17223 }, { "epoch": 25.78443113772455, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.128, "step": 17224 }, { "epoch": 25.785928143712574, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1295, "step": 17225 }, { "epoch": 25.7874251497006, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1303, "step": 17226 }, { "epoch": 25.788922155688624, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1317, "step": 17227 }, { "epoch": 25.790419161676645, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1306, "step": 17228 }, { "epoch": 25.79191616766467, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1231, "step": 17229 }, { "epoch": 25.793413173652695, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.126, "step": 17230 }, { "epoch": 25.79491017964072, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1301, "step": 17231 }, { "epoch": 25.79640718562874, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1267, "step": 17232 }, { "epoch": 25.797904191616766, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1301, "step": 17233 }, { "epoch": 25.79940119760479, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1326, "step": 17234 }, { "epoch": 25.800898203592816, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1353, "step": 17235 }, { "epoch": 25.802395209580837, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1231, "step": 17236 }, { "epoch": 25.80389221556886, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1385, "step": 17237 }, { "epoch": 25.805389221556887, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1206, "step": 17238 }, { "epoch": 25.80688622754491, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1308, "step": 17239 }, { "epoch": 25.808383233532933, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1258, "step": 17240 }, { "epoch": 25.809880239520957, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1284, "step": 17241 }, { "epoch": 25.811377245508982, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1309, "step": 17242 }, { "epoch": 25.812874251497007, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1251, "step": 17243 }, { "epoch": 25.81437125748503, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1252, "step": 17244 }, { "epoch": 25.815868263473053, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1307, "step": 17245 }, { "epoch": 25.817365269461078, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1345, "step": 17246 }, { "epoch": 25.818862275449103, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1264, "step": 17247 }, { "epoch": 25.820359281437124, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.136, "step": 17248 }, { "epoch": 25.82185628742515, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1264, "step": 17249 }, { "epoch": 25.823353293413174, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1258, "step": 17250 }, { "epoch": 25.8248502994012, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1274, "step": 17251 }, { "epoch": 25.82634730538922, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1327, "step": 17252 }, { "epoch": 25.827844311377245, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1272, "step": 17253 }, { "epoch": 25.82934131736527, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1324, "step": 17254 }, { "epoch": 25.830838323353294, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1285, "step": 17255 }, { "epoch": 25.83233532934132, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1336, "step": 17256 }, { "epoch": 25.83383233532934, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1296, "step": 17257 }, { "epoch": 25.835329341317365, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.125, "step": 17258 }, { "epoch": 25.83682634730539, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1331, "step": 17259 }, { "epoch": 25.83832335329341, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1283, "step": 17260 }, { "epoch": 25.839820359281436, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1315, "step": 17261 }, { "epoch": 25.84131736526946, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1308, "step": 17262 }, { "epoch": 25.842814371257486, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1333, "step": 17263 }, { "epoch": 25.84431137724551, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1267, "step": 17264 }, { "epoch": 25.845808383233532, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1256, "step": 17265 }, { "epoch": 25.847305389221557, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.129, "step": 17266 }, { "epoch": 25.84880239520958, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1306, "step": 17267 }, { "epoch": 25.850299401197606, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1335, "step": 17268 }, { "epoch": 25.851796407185628, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.128, "step": 17269 }, { "epoch": 25.853293413173652, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1299, "step": 17270 }, { "epoch": 25.854790419161677, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1313, "step": 17271 }, { "epoch": 25.856287425149702, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1302, "step": 17272 }, { "epoch": 25.857784431137723, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1257, "step": 17273 }, { "epoch": 25.85928143712575, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1348, "step": 17274 }, { "epoch": 25.860778443113773, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1249, "step": 17275 }, { "epoch": 25.862275449101798, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1297, "step": 17276 }, { "epoch": 25.86377245508982, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1283, "step": 17277 }, { "epoch": 25.865269461077844, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.135, "step": 17278 }, { "epoch": 25.86676646706587, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1297, "step": 17279 }, { "epoch": 25.868263473053894, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1291, "step": 17280 }, { "epoch": 25.869760479041915, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.123, "step": 17281 }, { "epoch": 25.87125748502994, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1332, "step": 17282 }, { "epoch": 25.872754491017965, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1292, "step": 17283 }, { "epoch": 25.87425149700599, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1335, "step": 17284 }, { "epoch": 25.87574850299401, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1314, "step": 17285 }, { "epoch": 25.877245508982035, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1348, "step": 17286 }, { "epoch": 25.87874251497006, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1315, "step": 17287 }, { "epoch": 25.880239520958085, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1255, "step": 17288 }, { "epoch": 25.881736526946106, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1314, "step": 17289 }, { "epoch": 25.88323353293413, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1296, "step": 17290 }, { "epoch": 25.884730538922156, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1356, "step": 17291 }, { "epoch": 25.88622754491018, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1278, "step": 17292 }, { "epoch": 25.887724550898202, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1316, "step": 17293 }, { "epoch": 25.889221556886227, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1271, "step": 17294 }, { "epoch": 25.89071856287425, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1361, "step": 17295 }, { "epoch": 25.892215568862277, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1303, "step": 17296 }, { "epoch": 25.893712574850298, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1289, "step": 17297 }, { "epoch": 25.895209580838323, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1288, "step": 17298 }, { "epoch": 25.896706586826348, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1302, "step": 17299 }, { "epoch": 25.898203592814372, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1304, "step": 17300 }, { "epoch": 25.899700598802394, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1271, "step": 17301 }, { "epoch": 25.90119760479042, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1331, "step": 17302 }, { "epoch": 25.902694610778443, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1258, "step": 17303 }, { "epoch": 25.904191616766468, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1298, "step": 17304 }, { "epoch": 25.90568862275449, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1342, "step": 17305 }, { "epoch": 25.907185628742514, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1322, "step": 17306 }, { "epoch": 25.90868263473054, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1282, "step": 17307 }, { "epoch": 25.910179640718564, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.126, "step": 17308 }, { "epoch": 25.91167664670659, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1308, "step": 17309 }, { "epoch": 25.91317365269461, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1336, "step": 17310 }, { "epoch": 25.914670658682635, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1238, "step": 17311 }, { "epoch": 25.91616766467066, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1314, "step": 17312 }, { "epoch": 25.91766467065868, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1307, "step": 17313 }, { "epoch": 25.919161676646706, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1341, "step": 17314 }, { "epoch": 25.92065868263473, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1329, "step": 17315 }, { "epoch": 25.922155688622755, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1368, "step": 17316 }, { "epoch": 25.92365269461078, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1334, "step": 17317 }, { "epoch": 25.9251497005988, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1306, "step": 17318 }, { "epoch": 25.926646706586826, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1323, "step": 17319 }, { "epoch": 25.92814371257485, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1288, "step": 17320 }, { "epoch": 25.929640718562876, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1331, "step": 17321 }, { "epoch": 25.931137724550897, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1336, "step": 17322 }, { "epoch": 25.932634730538922, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1354, "step": 17323 }, { "epoch": 25.934131736526947, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1322, "step": 17324 }, { "epoch": 25.93562874251497, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1333, "step": 17325 }, { "epoch": 25.937125748502993, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1231, "step": 17326 }, { "epoch": 25.938622754491018, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1277, "step": 17327 }, { "epoch": 25.940119760479043, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1269, "step": 17328 }, { "epoch": 25.941616766467067, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1327, "step": 17329 }, { "epoch": 25.94311377245509, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1331, "step": 17330 }, { "epoch": 25.944610778443113, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1261, "step": 17331 }, { "epoch": 25.94610778443114, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1326, "step": 17332 }, { "epoch": 25.947604790419163, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1319, "step": 17333 }, { "epoch": 25.949101796407184, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1307, "step": 17334 }, { "epoch": 25.95059880239521, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1307, "step": 17335 }, { "epoch": 25.952095808383234, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1318, "step": 17336 }, { "epoch": 25.95359281437126, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1264, "step": 17337 }, { "epoch": 25.95508982035928, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1214, "step": 17338 }, { "epoch": 25.956586826347305, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1362, "step": 17339 }, { "epoch": 25.95808383233533, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1256, "step": 17340 }, { "epoch": 25.959580838323355, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1297, "step": 17341 }, { "epoch": 25.961077844311376, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1283, "step": 17342 }, { "epoch": 25.9625748502994, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1301, "step": 17343 }, { "epoch": 25.964071856287426, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1281, "step": 17344 }, { "epoch": 25.96556886227545, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1353, "step": 17345 }, { "epoch": 25.96706586826347, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.132, "step": 17346 }, { "epoch": 25.968562874251496, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1264, "step": 17347 }, { "epoch": 25.97005988023952, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1249, "step": 17348 }, { "epoch": 25.971556886227546, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1272, "step": 17349 }, { "epoch": 25.973053892215567, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1314, "step": 17350 }, { "epoch": 25.974550898203592, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1375, "step": 17351 }, { "epoch": 25.976047904191617, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1305, "step": 17352 }, { "epoch": 25.977544910179642, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1278, "step": 17353 }, { "epoch": 25.979041916167663, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1323, "step": 17354 }, { "epoch": 25.980538922155688, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1306, "step": 17355 }, { "epoch": 25.982035928143713, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1288, "step": 17356 }, { "epoch": 25.983532934131738, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.133, "step": 17357 }, { "epoch": 25.98502994011976, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1274, "step": 17358 }, { "epoch": 25.986526946107784, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1324, "step": 17359 }, { "epoch": 25.98802395209581, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1272, "step": 17360 }, { "epoch": 25.989520958083833, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1338, "step": 17361 }, { "epoch": 25.991017964071855, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1399, "step": 17362 }, { "epoch": 25.99251497005988, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1264, "step": 17363 }, { "epoch": 25.994011976047904, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1302, "step": 17364 }, { "epoch": 25.99550898203593, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1327, "step": 17365 }, { "epoch": 25.99700598802395, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1256, "step": 17366 }, { "epoch": 25.998502994011975, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1322, "step": 17367 }, { "epoch": 26.0, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1285, "step": 17368 }, { "epoch": 26.001497005988025, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.131, "step": 17369 }, { "epoch": 26.00299401197605, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1294, "step": 17370 }, { "epoch": 26.00449101796407, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1265, "step": 17371 }, { "epoch": 26.005988023952096, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1318, "step": 17372 }, { "epoch": 26.00748502994012, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1298, "step": 17373 }, { "epoch": 26.008982035928145, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1292, "step": 17374 }, { "epoch": 26.010479041916167, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1325, "step": 17375 }, { "epoch": 26.01197604790419, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1347, "step": 17376 }, { "epoch": 26.013473053892216, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1302, "step": 17377 }, { "epoch": 26.01497005988024, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1299, "step": 17378 }, { "epoch": 26.016467065868262, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1286, "step": 17379 }, { "epoch": 26.017964071856287, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.135, "step": 17380 }, { "epoch": 26.019461077844312, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1266, "step": 17381 }, { "epoch": 26.020958083832337, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1285, "step": 17382 }, { "epoch": 26.022455089820358, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1353, "step": 17383 }, { "epoch": 26.023952095808383, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1243, "step": 17384 }, { "epoch": 26.025449101796408, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1319, "step": 17385 }, { "epoch": 26.026946107784433, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1314, "step": 17386 }, { "epoch": 26.028443113772454, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1323, "step": 17387 }, { "epoch": 26.02994011976048, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1289, "step": 17388 }, { "epoch": 26.031437125748504, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1308, "step": 17389 }, { "epoch": 26.03293413173653, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.132, "step": 17390 }, { "epoch": 26.03443113772455, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1298, "step": 17391 }, { "epoch": 26.035928143712574, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1299, "step": 17392 }, { "epoch": 26.0374251497006, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1356, "step": 17393 }, { "epoch": 26.038922155688624, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1298, "step": 17394 }, { "epoch": 26.040419161676645, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1346, "step": 17395 }, { "epoch": 26.04191616766467, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1294, "step": 17396 }, { "epoch": 26.043413173652695, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1274, "step": 17397 }, { "epoch": 26.04491017964072, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.126, "step": 17398 }, { "epoch": 26.04640718562874, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1264, "step": 17399 }, { "epoch": 26.047904191616766, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1303, "step": 17400 }, { "epoch": 26.04940119760479, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1326, "step": 17401 }, { "epoch": 26.050898203592816, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1297, "step": 17402 }, { "epoch": 26.052395209580837, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1317, "step": 17403 }, { "epoch": 26.05389221556886, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1284, "step": 17404 }, { "epoch": 26.055389221556887, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1265, "step": 17405 }, { "epoch": 26.05688622754491, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1263, "step": 17406 }, { "epoch": 26.058383233532933, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.129, "step": 17407 }, { "epoch": 26.059880239520957, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1268, "step": 17408 }, { "epoch": 26.061377245508982, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1406, "step": 17409 }, { "epoch": 26.062874251497007, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1373, "step": 17410 }, { "epoch": 26.06437125748503, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1281, "step": 17411 }, { "epoch": 26.065868263473053, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1355, "step": 17412 }, { "epoch": 26.067365269461078, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.138, "step": 17413 }, { "epoch": 26.068862275449103, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1351, "step": 17414 }, { "epoch": 26.070359281437124, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1329, "step": 17415 }, { "epoch": 26.07185628742515, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1254, "step": 17416 }, { "epoch": 26.073353293413174, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1266, "step": 17417 }, { "epoch": 26.0748502994012, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1336, "step": 17418 }, { "epoch": 26.07634730538922, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1233, "step": 17419 }, { "epoch": 26.077844311377245, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1267, "step": 17420 }, { "epoch": 26.07934131736527, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1301, "step": 17421 }, { "epoch": 26.080838323353294, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1355, "step": 17422 }, { "epoch": 26.082335329341316, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1284, "step": 17423 }, { "epoch": 26.08383233532934, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1256, "step": 17424 }, { "epoch": 26.085329341317365, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1298, "step": 17425 }, { "epoch": 26.08682634730539, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1348, "step": 17426 }, { "epoch": 26.088323353293415, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1282, "step": 17427 }, { "epoch": 26.089820359281436, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1325, "step": 17428 }, { "epoch": 26.09131736526946, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1326, "step": 17429 }, { "epoch": 26.092814371257486, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1309, "step": 17430 }, { "epoch": 26.09431137724551, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1271, "step": 17431 }, { "epoch": 26.095808383233532, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1339, "step": 17432 }, { "epoch": 26.097305389221557, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.13, "step": 17433 }, { "epoch": 26.09880239520958, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.133, "step": 17434 }, { "epoch": 26.100299401197606, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1243, "step": 17435 }, { "epoch": 26.101796407185628, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1248, "step": 17436 }, { "epoch": 26.103293413173652, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1259, "step": 17437 }, { "epoch": 26.104790419161677, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1285, "step": 17438 }, { "epoch": 26.106287425149702, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1307, "step": 17439 }, { "epoch": 26.107784431137723, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.136, "step": 17440 }, { "epoch": 26.10928143712575, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1274, "step": 17441 }, { "epoch": 26.110778443113773, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1281, "step": 17442 }, { "epoch": 26.112275449101798, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1314, "step": 17443 }, { "epoch": 26.11377245508982, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1258, "step": 17444 }, { "epoch": 26.115269461077844, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1364, "step": 17445 }, { "epoch": 26.11676646706587, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.129, "step": 17446 }, { "epoch": 26.118263473053894, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1265, "step": 17447 }, { "epoch": 26.119760479041915, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1282, "step": 17448 }, { "epoch": 26.12125748502994, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1282, "step": 17449 }, { "epoch": 26.122754491017965, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1324, "step": 17450 }, { "epoch": 26.12425149700599, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1235, "step": 17451 }, { "epoch": 26.12574850299401, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1277, "step": 17452 }, { "epoch": 26.127245508982035, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1262, "step": 17453 }, { "epoch": 26.12874251497006, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1358, "step": 17454 }, { "epoch": 26.130239520958085, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.123, "step": 17455 }, { "epoch": 26.131736526946106, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1269, "step": 17456 }, { "epoch": 26.13323353293413, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1241, "step": 17457 }, { "epoch": 26.134730538922156, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1274, "step": 17458 }, { "epoch": 26.13622754491018, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1325, "step": 17459 }, { "epoch": 26.137724550898202, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1222, "step": 17460 }, { "epoch": 26.139221556886227, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1313, "step": 17461 }, { "epoch": 26.14071856287425, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1288, "step": 17462 }, { "epoch": 26.142215568862277, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1215, "step": 17463 }, { "epoch": 26.143712574850298, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1267, "step": 17464 }, { "epoch": 26.145209580838323, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1283, "step": 17465 }, { "epoch": 26.146706586826348, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1235, "step": 17466 }, { "epoch": 26.148203592814372, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1324, "step": 17467 }, { "epoch": 26.149700598802394, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1296, "step": 17468 }, { "epoch": 26.15119760479042, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1316, "step": 17469 }, { "epoch": 26.152694610778443, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1306, "step": 17470 }, { "epoch": 26.154191616766468, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1312, "step": 17471 }, { "epoch": 26.15568862275449, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1306, "step": 17472 }, { "epoch": 26.157185628742514, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1287, "step": 17473 }, { "epoch": 26.15868263473054, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1323, "step": 17474 }, { "epoch": 26.160179640718564, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1283, "step": 17475 }, { "epoch": 26.161676646706585, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1227, "step": 17476 }, { "epoch": 26.16317365269461, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1292, "step": 17477 }, { "epoch": 26.164670658682635, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1377, "step": 17478 }, { "epoch": 26.16616766467066, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1251, "step": 17479 }, { "epoch": 26.16766467065868, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1277, "step": 17480 }, { "epoch": 26.169161676646706, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1284, "step": 17481 }, { "epoch": 26.17065868263473, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1331, "step": 17482 }, { "epoch": 26.172155688622755, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.131, "step": 17483 }, { "epoch": 26.17365269461078, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1301, "step": 17484 }, { "epoch": 26.1751497005988, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1392, "step": 17485 }, { "epoch": 26.176646706586826, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.125, "step": 17486 }, { "epoch": 26.17814371257485, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1278, "step": 17487 }, { "epoch": 26.179640718562876, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1324, "step": 17488 }, { "epoch": 26.181137724550897, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.128, "step": 17489 }, { "epoch": 26.182634730538922, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1308, "step": 17490 }, { "epoch": 26.184131736526947, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1298, "step": 17491 }, { "epoch": 26.18562874251497, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1349, "step": 17492 }, { "epoch": 26.187125748502993, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1331, "step": 17493 }, { "epoch": 26.188622754491018, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1219, "step": 17494 }, { "epoch": 26.190119760479043, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1342, "step": 17495 }, { "epoch": 26.191616766467067, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1353, "step": 17496 }, { "epoch": 26.19311377245509, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1326, "step": 17497 }, { "epoch": 26.194610778443113, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1342, "step": 17498 }, { "epoch": 26.19610778443114, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.129, "step": 17499 }, { "epoch": 26.197604790419163, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1315, "step": 17500 }, { "epoch": 26.199101796407184, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1249, "step": 17501 }, { "epoch": 26.20059880239521, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1357, "step": 17502 }, { "epoch": 26.202095808383234, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1232, "step": 17503 }, { "epoch": 26.20359281437126, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1295, "step": 17504 }, { "epoch": 26.20508982035928, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.131, "step": 17505 }, { "epoch": 26.206586826347305, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1188, "step": 17506 }, { "epoch": 26.20808383233533, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1195, "step": 17507 }, { "epoch": 26.209580838323355, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1269, "step": 17508 }, { "epoch": 26.211077844311376, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1328, "step": 17509 }, { "epoch": 26.2125748502994, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1295, "step": 17510 }, { "epoch": 26.214071856287426, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1272, "step": 17511 }, { "epoch": 26.21556886227545, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1367, "step": 17512 }, { "epoch": 26.21706586826347, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1297, "step": 17513 }, { "epoch": 26.218562874251496, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1343, "step": 17514 }, { "epoch": 26.22005988023952, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1269, "step": 17515 }, { "epoch": 26.221556886227546, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1324, "step": 17516 }, { "epoch": 26.223053892215567, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1324, "step": 17517 }, { "epoch": 26.224550898203592, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1315, "step": 17518 }, { "epoch": 26.226047904191617, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1238, "step": 17519 }, { "epoch": 26.227544910179642, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1407, "step": 17520 }, { "epoch": 26.229041916167663, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1289, "step": 17521 }, { "epoch": 26.230538922155688, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1244, "step": 17522 }, { "epoch": 26.232035928143713, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1334, "step": 17523 }, { "epoch": 26.233532934131738, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.133, "step": 17524 }, { "epoch": 26.23502994011976, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1325, "step": 17525 }, { "epoch": 26.236526946107784, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1354, "step": 17526 }, { "epoch": 26.23802395209581, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.125, "step": 17527 }, { "epoch": 26.239520958083833, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1258, "step": 17528 }, { "epoch": 26.241017964071855, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.133, "step": 17529 }, { "epoch": 26.24251497005988, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1335, "step": 17530 }, { "epoch": 26.244011976047904, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1321, "step": 17531 }, { "epoch": 26.24550898203593, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1305, "step": 17532 }, { "epoch": 26.24700598802395, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1337, "step": 17533 }, { "epoch": 26.248502994011975, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1356, "step": 17534 }, { "epoch": 26.25, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1355, "step": 17535 }, { "epoch": 26.251497005988025, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.125, "step": 17536 }, { "epoch": 26.25299401197605, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1312, "step": 17537 }, { "epoch": 26.25449101796407, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1333, "step": 17538 }, { "epoch": 26.255988023952096, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1271, "step": 17539 }, { "epoch": 26.25748502994012, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1267, "step": 17540 }, { "epoch": 26.258982035928145, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1315, "step": 17541 }, { "epoch": 26.260479041916167, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1375, "step": 17542 }, { "epoch": 26.26197604790419, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1265, "step": 17543 }, { "epoch": 26.263473053892216, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.128, "step": 17544 }, { "epoch": 26.26497005988024, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1296, "step": 17545 }, { "epoch": 26.266467065868262, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1329, "step": 17546 }, { "epoch": 26.267964071856287, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1325, "step": 17547 }, { "epoch": 26.269461077844312, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1316, "step": 17548 }, { "epoch": 26.270958083832337, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1273, "step": 17549 }, { "epoch": 26.272455089820358, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1246, "step": 17550 }, { "epoch": 26.273952095808383, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1297, "step": 17551 }, { "epoch": 26.275449101796408, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1289, "step": 17552 }, { "epoch": 26.276946107784433, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1316, "step": 17553 }, { "epoch": 26.278443113772454, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1214, "step": 17554 }, { "epoch": 26.27994011976048, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1295, "step": 17555 }, { "epoch": 26.281437125748504, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1275, "step": 17556 }, { "epoch": 26.28293413173653, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1323, "step": 17557 }, { "epoch": 26.28443113772455, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1324, "step": 17558 }, { "epoch": 26.285928143712574, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1363, "step": 17559 }, { "epoch": 26.2874251497006, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1294, "step": 17560 }, { "epoch": 26.288922155688624, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1266, "step": 17561 }, { "epoch": 26.290419161676645, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1277, "step": 17562 }, { "epoch": 26.29191616766467, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1292, "step": 17563 }, { "epoch": 26.293413173652695, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.126, "step": 17564 }, { "epoch": 26.29491017964072, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1291, "step": 17565 }, { "epoch": 26.29640718562874, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1278, "step": 17566 }, { "epoch": 26.297904191616766, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1287, "step": 17567 }, { "epoch": 26.29940119760479, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.134, "step": 17568 }, { "epoch": 26.300898203592816, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.129, "step": 17569 }, { "epoch": 26.302395209580837, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1264, "step": 17570 }, { "epoch": 26.30389221556886, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1282, "step": 17571 }, { "epoch": 26.305389221556887, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1314, "step": 17572 }, { "epoch": 26.30688622754491, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1308, "step": 17573 }, { "epoch": 26.308383233532933, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1294, "step": 17574 }, { "epoch": 26.309880239520957, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1379, "step": 17575 }, { "epoch": 26.311377245508982, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1364, "step": 17576 }, { "epoch": 26.312874251497007, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.132, "step": 17577 }, { "epoch": 26.31437125748503, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1261, "step": 17578 }, { "epoch": 26.315868263473053, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1309, "step": 17579 }, { "epoch": 26.317365269461078, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1276, "step": 17580 }, { "epoch": 26.318862275449103, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1312, "step": 17581 }, { "epoch": 26.320359281437124, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1288, "step": 17582 }, { "epoch": 26.32185628742515, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1337, "step": 17583 }, { "epoch": 26.323353293413174, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1355, "step": 17584 }, { "epoch": 26.3248502994012, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1306, "step": 17585 }, { "epoch": 26.32634730538922, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1289, "step": 17586 }, { "epoch": 26.327844311377245, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1229, "step": 17587 }, { "epoch": 26.32934131736527, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1372, "step": 17588 }, { "epoch": 26.330838323353294, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1275, "step": 17589 }, { "epoch": 26.33233532934132, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1227, "step": 17590 }, { "epoch": 26.33383233532934, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1276, "step": 17591 }, { "epoch": 26.335329341317365, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1309, "step": 17592 }, { "epoch": 26.33682634730539, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1253, "step": 17593 }, { "epoch": 26.338323353293415, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1354, "step": 17594 }, { "epoch": 26.339820359281436, "grad_norm": 0.04931640625, "learning_rate": 0.0008, "loss": 1.1258, "step": 17595 }, { "epoch": 26.34131736526946, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1299, "step": 17596 }, { "epoch": 26.342814371257486, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1241, "step": 17597 }, { "epoch": 26.34431137724551, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1345, "step": 17598 }, { "epoch": 26.345808383233532, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1357, "step": 17599 }, { "epoch": 26.347305389221557, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1284, "step": 17600 }, { "epoch": 26.34880239520958, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1323, "step": 17601 }, { "epoch": 26.350299401197606, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1301, "step": 17602 }, { "epoch": 26.351796407185628, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1338, "step": 17603 }, { "epoch": 26.353293413173652, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1234, "step": 17604 }, { "epoch": 26.354790419161677, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1318, "step": 17605 }, { "epoch": 26.356287425149702, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1313, "step": 17606 }, { "epoch": 26.357784431137723, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1272, "step": 17607 }, { "epoch": 26.35928143712575, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1301, "step": 17608 }, { "epoch": 26.360778443113773, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1342, "step": 17609 }, { "epoch": 26.362275449101798, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1388, "step": 17610 }, { "epoch": 26.36377245508982, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1306, "step": 17611 }, { "epoch": 26.365269461077844, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1288, "step": 17612 }, { "epoch": 26.36676646706587, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1275, "step": 17613 }, { "epoch": 26.368263473053894, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.127, "step": 17614 }, { "epoch": 26.369760479041915, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1312, "step": 17615 }, { "epoch": 26.37125748502994, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1304, "step": 17616 }, { "epoch": 26.372754491017965, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1298, "step": 17617 }, { "epoch": 26.37425149700599, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1327, "step": 17618 }, { "epoch": 26.37574850299401, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1284, "step": 17619 }, { "epoch": 26.377245508982035, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1322, "step": 17620 }, { "epoch": 26.37874251497006, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.125, "step": 17621 }, { "epoch": 26.380239520958085, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1326, "step": 17622 }, { "epoch": 26.381736526946106, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1298, "step": 17623 }, { "epoch": 26.38323353293413, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1284, "step": 17624 }, { "epoch": 26.384730538922156, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1341, "step": 17625 }, { "epoch": 26.38622754491018, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1325, "step": 17626 }, { "epoch": 26.387724550898202, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1297, "step": 17627 }, { "epoch": 26.389221556886227, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1197, "step": 17628 }, { "epoch": 26.39071856287425, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1303, "step": 17629 }, { "epoch": 26.392215568862277, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1266, "step": 17630 }, { "epoch": 26.393712574850298, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1276, "step": 17631 }, { "epoch": 26.395209580838323, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1293, "step": 17632 }, { "epoch": 26.396706586826348, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1238, "step": 17633 }, { "epoch": 26.398203592814372, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1274, "step": 17634 }, { "epoch": 26.399700598802394, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1341, "step": 17635 }, { "epoch": 26.40119760479042, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1381, "step": 17636 }, { "epoch": 26.402694610778443, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1357, "step": 17637 }, { "epoch": 26.404191616766468, "grad_norm": 0.11083984375, "learning_rate": 0.0008, "loss": 1.1306, "step": 17638 }, { "epoch": 26.40568862275449, "grad_norm": 0.115234375, "learning_rate": 0.0008, "loss": 1.129, "step": 17639 }, { "epoch": 26.407185628742514, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1305, "step": 17640 }, { "epoch": 26.40868263473054, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1225, "step": 17641 }, { "epoch": 26.410179640718564, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1279, "step": 17642 }, { "epoch": 26.411676646706585, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1278, "step": 17643 }, { "epoch": 26.41317365269461, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1346, "step": 17644 }, { "epoch": 26.414670658682635, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1244, "step": 17645 }, { "epoch": 26.41616766467066, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1283, "step": 17646 }, { "epoch": 26.41766467065868, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1292, "step": 17647 }, { "epoch": 26.419161676646706, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1304, "step": 17648 }, { "epoch": 26.42065868263473, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1291, "step": 17649 }, { "epoch": 26.422155688622755, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1293, "step": 17650 }, { "epoch": 26.42365269461078, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.136, "step": 17651 }, { "epoch": 26.4251497005988, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1308, "step": 17652 }, { "epoch": 26.426646706586826, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.131, "step": 17653 }, { "epoch": 26.42814371257485, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1271, "step": 17654 }, { "epoch": 26.429640718562876, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1285, "step": 17655 }, { "epoch": 26.431137724550897, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1294, "step": 17656 }, { "epoch": 26.432634730538922, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1309, "step": 17657 }, { "epoch": 26.434131736526947, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1309, "step": 17658 }, { "epoch": 26.43562874251497, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1293, "step": 17659 }, { "epoch": 26.437125748502993, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1292, "step": 17660 }, { "epoch": 26.438622754491018, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1313, "step": 17661 }, { "epoch": 26.440119760479043, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1285, "step": 17662 }, { "epoch": 26.441616766467067, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1247, "step": 17663 }, { "epoch": 26.44311377245509, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1249, "step": 17664 }, { "epoch": 26.444610778443113, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1302, "step": 17665 }, { "epoch": 26.44610778443114, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1269, "step": 17666 }, { "epoch": 26.447604790419163, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1333, "step": 17667 }, { "epoch": 26.449101796407184, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.127, "step": 17668 }, { "epoch": 26.45059880239521, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1399, "step": 17669 }, { "epoch": 26.452095808383234, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1307, "step": 17670 }, { "epoch": 26.45359281437126, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1192, "step": 17671 }, { "epoch": 26.45508982035928, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1243, "step": 17672 }, { "epoch": 26.456586826347305, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1309, "step": 17673 }, { "epoch": 26.45808383233533, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1252, "step": 17674 }, { "epoch": 26.459580838323355, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1321, "step": 17675 }, { "epoch": 26.461077844311376, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1336, "step": 17676 }, { "epoch": 26.4625748502994, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1293, "step": 17677 }, { "epoch": 26.464071856287426, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1374, "step": 17678 }, { "epoch": 26.46556886227545, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1234, "step": 17679 }, { "epoch": 26.46706586826347, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1284, "step": 17680 }, { "epoch": 26.468562874251496, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1287, "step": 17681 }, { "epoch": 26.47005988023952, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1292, "step": 17682 }, { "epoch": 26.471556886227546, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1375, "step": 17683 }, { "epoch": 26.473053892215567, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1289, "step": 17684 }, { "epoch": 26.474550898203592, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.13, "step": 17685 }, { "epoch": 26.476047904191617, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1363, "step": 17686 }, { "epoch": 26.477544910179642, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1362, "step": 17687 }, { "epoch": 26.479041916167663, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1348, "step": 17688 }, { "epoch": 26.480538922155688, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1309, "step": 17689 }, { "epoch": 26.482035928143713, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1324, "step": 17690 }, { "epoch": 26.483532934131738, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1244, "step": 17691 }, { "epoch": 26.48502994011976, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.137, "step": 17692 }, { "epoch": 26.486526946107784, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1316, "step": 17693 }, { "epoch": 26.48802395209581, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1304, "step": 17694 }, { "epoch": 26.489520958083833, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.132, "step": 17695 }, { "epoch": 26.491017964071855, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1239, "step": 17696 }, { "epoch": 26.49251497005988, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1267, "step": 17697 }, { "epoch": 26.494011976047904, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1301, "step": 17698 }, { "epoch": 26.49550898203593, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1254, "step": 17699 }, { "epoch": 26.49700598802395, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1311, "step": 17700 }, { "epoch": 26.498502994011975, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1229, "step": 17701 }, { "epoch": 26.5, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1263, "step": 17702 }, { "epoch": 26.501497005988025, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1323, "step": 17703 }, { "epoch": 26.50299401197605, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1293, "step": 17704 }, { "epoch": 26.50449101796407, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1305, "step": 17705 }, { "epoch": 26.505988023952096, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1266, "step": 17706 }, { "epoch": 26.50748502994012, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1334, "step": 17707 }, { "epoch": 26.508982035928145, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1311, "step": 17708 }, { "epoch": 26.510479041916167, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1297, "step": 17709 }, { "epoch": 26.51197604790419, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1296, "step": 17710 }, { "epoch": 26.513473053892216, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1307, "step": 17711 }, { "epoch": 26.51497005988024, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1209, "step": 17712 }, { "epoch": 26.516467065868262, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1293, "step": 17713 }, { "epoch": 26.517964071856287, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1261, "step": 17714 }, { "epoch": 26.519461077844312, "grad_norm": 0.048095703125, "learning_rate": 0.0008, "loss": 1.1302, "step": 17715 }, { "epoch": 26.520958083832337, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.135, "step": 17716 }, { "epoch": 26.522455089820358, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1345, "step": 17717 }, { "epoch": 26.523952095808383, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1277, "step": 17718 }, { "epoch": 26.525449101796408, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1238, "step": 17719 }, { "epoch": 26.526946107784433, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1318, "step": 17720 }, { "epoch": 26.528443113772454, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1299, "step": 17721 }, { "epoch": 26.52994011976048, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1267, "step": 17722 }, { "epoch": 26.531437125748504, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1307, "step": 17723 }, { "epoch": 26.53293413173653, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1271, "step": 17724 }, { "epoch": 26.53443113772455, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1272, "step": 17725 }, { "epoch": 26.535928143712574, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1336, "step": 17726 }, { "epoch": 26.5374251497006, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1257, "step": 17727 }, { "epoch": 26.538922155688624, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1268, "step": 17728 }, { "epoch": 26.540419161676645, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1366, "step": 17729 }, { "epoch": 26.54191616766467, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.131, "step": 17730 }, { "epoch": 26.543413173652695, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1304, "step": 17731 }, { "epoch": 26.54491017964072, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1324, "step": 17732 }, { "epoch": 26.54640718562874, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1364, "step": 17733 }, { "epoch": 26.547904191616766, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1292, "step": 17734 }, { "epoch": 26.54940119760479, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1225, "step": 17735 }, { "epoch": 26.550898203592816, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1306, "step": 17736 }, { "epoch": 26.552395209580837, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1287, "step": 17737 }, { "epoch": 26.55389221556886, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1243, "step": 17738 }, { "epoch": 26.555389221556887, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1272, "step": 17739 }, { "epoch": 26.55688622754491, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1323, "step": 17740 }, { "epoch": 26.558383233532933, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1317, "step": 17741 }, { "epoch": 26.559880239520957, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1235, "step": 17742 }, { "epoch": 26.561377245508982, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1374, "step": 17743 }, { "epoch": 26.562874251497007, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1302, "step": 17744 }, { "epoch": 26.56437125748503, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1323, "step": 17745 }, { "epoch": 26.565868263473053, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1324, "step": 17746 }, { "epoch": 26.567365269461078, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1275, "step": 17747 }, { "epoch": 26.568862275449103, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1246, "step": 17748 }, { "epoch": 26.570359281437124, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1344, "step": 17749 }, { "epoch": 26.57185628742515, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1286, "step": 17750 }, { "epoch": 26.573353293413174, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1366, "step": 17751 }, { "epoch": 26.5748502994012, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1367, "step": 17752 }, { "epoch": 26.57634730538922, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1307, "step": 17753 }, { "epoch": 26.577844311377245, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1316, "step": 17754 }, { "epoch": 26.57934131736527, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1312, "step": 17755 }, { "epoch": 26.580838323353294, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1276, "step": 17756 }, { "epoch": 26.58233532934132, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1281, "step": 17757 }, { "epoch": 26.58383233532934, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1274, "step": 17758 }, { "epoch": 26.585329341317365, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.127, "step": 17759 }, { "epoch": 26.58682634730539, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1192, "step": 17760 }, { "epoch": 26.58832335329341, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1253, "step": 17761 }, { "epoch": 26.589820359281436, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1252, "step": 17762 }, { "epoch": 26.59131736526946, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1328, "step": 17763 }, { "epoch": 26.592814371257486, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1259, "step": 17764 }, { "epoch": 26.59431137724551, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1304, "step": 17765 }, { "epoch": 26.595808383233532, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1341, "step": 17766 }, { "epoch": 26.597305389221557, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1287, "step": 17767 }, { "epoch": 26.59880239520958, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1319, "step": 17768 }, { "epoch": 26.600299401197606, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1285, "step": 17769 }, { "epoch": 26.601796407185628, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1278, "step": 17770 }, { "epoch": 26.603293413173652, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1231, "step": 17771 }, { "epoch": 26.604790419161677, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1283, "step": 17772 }, { "epoch": 26.606287425149702, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1292, "step": 17773 }, { "epoch": 26.607784431137723, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1323, "step": 17774 }, { "epoch": 26.60928143712575, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1312, "step": 17775 }, { "epoch": 26.610778443113773, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.128, "step": 17776 }, { "epoch": 26.612275449101798, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1403, "step": 17777 }, { "epoch": 26.61377245508982, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1337, "step": 17778 }, { "epoch": 26.615269461077844, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1372, "step": 17779 }, { "epoch": 26.61676646706587, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1323, "step": 17780 }, { "epoch": 26.618263473053894, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1346, "step": 17781 }, { "epoch": 26.619760479041915, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1314, "step": 17782 }, { "epoch": 26.62125748502994, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1309, "step": 17783 }, { "epoch": 26.622754491017965, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1258, "step": 17784 }, { "epoch": 26.62425149700599, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1309, "step": 17785 }, { "epoch": 26.62574850299401, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1286, "step": 17786 }, { "epoch": 26.627245508982035, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1359, "step": 17787 }, { "epoch": 26.62874251497006, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1277, "step": 17788 }, { "epoch": 26.630239520958085, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1256, "step": 17789 }, { "epoch": 26.631736526946106, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1367, "step": 17790 }, { "epoch": 26.63323353293413, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1343, "step": 17791 }, { "epoch": 26.634730538922156, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1364, "step": 17792 }, { "epoch": 26.63622754491018, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1296, "step": 17793 }, { "epoch": 26.637724550898202, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1281, "step": 17794 }, { "epoch": 26.639221556886227, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1303, "step": 17795 }, { "epoch": 26.64071856287425, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1285, "step": 17796 }, { "epoch": 26.642215568862277, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1318, "step": 17797 }, { "epoch": 26.643712574850298, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1249, "step": 17798 }, { "epoch": 26.645209580838323, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1283, "step": 17799 }, { "epoch": 26.646706586826348, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1314, "step": 17800 }, { "epoch": 26.648203592814372, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1275, "step": 17801 }, { "epoch": 26.649700598802394, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1276, "step": 17802 }, { "epoch": 26.65119760479042, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1294, "step": 17803 }, { "epoch": 26.652694610778443, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.133, "step": 17804 }, { "epoch": 26.654191616766468, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1262, "step": 17805 }, { "epoch": 26.65568862275449, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1276, "step": 17806 }, { "epoch": 26.657185628742514, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1247, "step": 17807 }, { "epoch": 26.65868263473054, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1308, "step": 17808 }, { "epoch": 26.660179640718564, "grad_norm": 0.052734375, "learning_rate": 0.0008, "loss": 1.1324, "step": 17809 }, { "epoch": 26.66167664670659, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1361, "step": 17810 }, { "epoch": 26.66317365269461, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1352, "step": 17811 }, { "epoch": 26.664670658682635, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1299, "step": 17812 }, { "epoch": 26.66616766467066, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1315, "step": 17813 }, { "epoch": 26.66766467065868, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.127, "step": 17814 }, { "epoch": 26.669161676646706, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.129, "step": 17815 }, { "epoch": 26.67065868263473, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1285, "step": 17816 }, { "epoch": 26.672155688622755, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1255, "step": 17817 }, { "epoch": 26.67365269461078, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1288, "step": 17818 }, { "epoch": 26.6751497005988, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1275, "step": 17819 }, { "epoch": 26.676646706586826, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1286, "step": 17820 }, { "epoch": 26.67814371257485, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1281, "step": 17821 }, { "epoch": 26.679640718562876, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1309, "step": 17822 }, { "epoch": 26.681137724550897, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1267, "step": 17823 }, { "epoch": 26.682634730538922, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1323, "step": 17824 }, { "epoch": 26.684131736526947, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.133, "step": 17825 }, { "epoch": 26.68562874251497, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1261, "step": 17826 }, { "epoch": 26.687125748502993, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1314, "step": 17827 }, { "epoch": 26.688622754491018, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1301, "step": 17828 }, { "epoch": 26.690119760479043, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.128, "step": 17829 }, { "epoch": 26.691616766467067, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1309, "step": 17830 }, { "epoch": 26.69311377245509, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1309, "step": 17831 }, { "epoch": 26.694610778443113, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1294, "step": 17832 }, { "epoch": 26.69610778443114, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1246, "step": 17833 }, { "epoch": 26.697604790419163, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1257, "step": 17834 }, { "epoch": 26.699101796407184, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1292, "step": 17835 }, { "epoch": 26.70059880239521, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1315, "step": 17836 }, { "epoch": 26.702095808383234, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1318, "step": 17837 }, { "epoch": 26.70359281437126, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.13, "step": 17838 }, { "epoch": 26.70508982035928, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1298, "step": 17839 }, { "epoch": 26.706586826347305, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1207, "step": 17840 }, { "epoch": 26.70808383233533, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1316, "step": 17841 }, { "epoch": 26.709580838323355, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1348, "step": 17842 }, { "epoch": 26.711077844311376, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1301, "step": 17843 }, { "epoch": 26.7125748502994, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1289, "step": 17844 }, { "epoch": 26.714071856287426, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1289, "step": 17845 }, { "epoch": 26.71556886227545, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1335, "step": 17846 }, { "epoch": 26.71706586826347, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1234, "step": 17847 }, { "epoch": 26.718562874251496, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1265, "step": 17848 }, { "epoch": 26.72005988023952, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1284, "step": 17849 }, { "epoch": 26.721556886227546, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1299, "step": 17850 }, { "epoch": 26.723053892215567, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1302, "step": 17851 }, { "epoch": 26.724550898203592, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1303, "step": 17852 }, { "epoch": 26.726047904191617, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1334, "step": 17853 }, { "epoch": 26.727544910179642, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1218, "step": 17854 }, { "epoch": 26.729041916167663, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1332, "step": 17855 }, { "epoch": 26.730538922155688, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1365, "step": 17856 }, { "epoch": 26.732035928143713, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1268, "step": 17857 }, { "epoch": 26.733532934131738, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1346, "step": 17858 }, { "epoch": 26.73502994011976, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1267, "step": 17859 }, { "epoch": 26.736526946107784, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 17860 }, { "epoch": 26.73802395209581, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1296, "step": 17861 }, { "epoch": 26.739520958083833, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1349, "step": 17862 }, { "epoch": 26.741017964071855, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.131, "step": 17863 }, { "epoch": 26.74251497005988, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1275, "step": 17864 }, { "epoch": 26.744011976047904, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1274, "step": 17865 }, { "epoch": 26.74550898203593, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1314, "step": 17866 }, { "epoch": 26.74700598802395, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1288, "step": 17867 }, { "epoch": 26.748502994011975, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1281, "step": 17868 }, { "epoch": 26.75, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1285, "step": 17869 }, { "epoch": 26.751497005988025, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1322, "step": 17870 }, { "epoch": 26.75299401197605, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1263, "step": 17871 }, { "epoch": 26.75449101796407, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1254, "step": 17872 }, { "epoch": 26.755988023952096, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1273, "step": 17873 }, { "epoch": 26.75748502994012, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1254, "step": 17874 }, { "epoch": 26.758982035928145, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1311, "step": 17875 }, { "epoch": 26.760479041916167, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.129, "step": 17876 }, { "epoch": 26.76197604790419, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1242, "step": 17877 }, { "epoch": 26.763473053892216, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1249, "step": 17878 }, { "epoch": 26.76497005988024, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1251, "step": 17879 }, { "epoch": 26.766467065868262, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1287, "step": 17880 }, { "epoch": 26.767964071856287, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1273, "step": 17881 }, { "epoch": 26.769461077844312, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1372, "step": 17882 }, { "epoch": 26.770958083832337, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1298, "step": 17883 }, { "epoch": 26.772455089820358, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1277, "step": 17884 }, { "epoch": 26.773952095808383, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1341, "step": 17885 }, { "epoch": 26.775449101796408, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1331, "step": 17886 }, { "epoch": 26.776946107784433, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1191, "step": 17887 }, { "epoch": 26.778443113772454, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1312, "step": 17888 }, { "epoch": 26.77994011976048, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.13, "step": 17889 }, { "epoch": 26.781437125748504, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1288, "step": 17890 }, { "epoch": 26.78293413173653, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1288, "step": 17891 }, { "epoch": 26.78443113772455, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1274, "step": 17892 }, { "epoch": 26.785928143712574, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1262, "step": 17893 }, { "epoch": 26.7874251497006, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1267, "step": 17894 }, { "epoch": 26.788922155688624, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1296, "step": 17895 }, { "epoch": 26.790419161676645, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1311, "step": 17896 }, { "epoch": 26.79191616766467, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1348, "step": 17897 }, { "epoch": 26.793413173652695, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1342, "step": 17898 }, { "epoch": 26.79491017964072, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1262, "step": 17899 }, { "epoch": 26.79640718562874, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1248, "step": 17900 }, { "epoch": 26.797904191616766, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.134, "step": 17901 }, { "epoch": 26.79940119760479, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1293, "step": 17902 }, { "epoch": 26.800898203592816, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 17903 }, { "epoch": 26.802395209580837, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1254, "step": 17904 }, { "epoch": 26.80389221556886, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1317, "step": 17905 }, { "epoch": 26.805389221556887, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1313, "step": 17906 }, { "epoch": 26.80688622754491, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.122, "step": 17907 }, { "epoch": 26.808383233532933, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1267, "step": 17908 }, { "epoch": 26.809880239520957, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1265, "step": 17909 }, { "epoch": 26.811377245508982, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1341, "step": 17910 }, { "epoch": 26.812874251497007, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.125, "step": 17911 }, { "epoch": 26.81437125748503, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1298, "step": 17912 }, { "epoch": 26.815868263473053, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1304, "step": 17913 }, { "epoch": 26.817365269461078, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1277, "step": 17914 }, { "epoch": 26.818862275449103, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.127, "step": 17915 }, { "epoch": 26.820359281437124, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 17916 }, { "epoch": 26.82185628742515, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1282, "step": 17917 }, { "epoch": 26.823353293413174, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1282, "step": 17918 }, { "epoch": 26.8248502994012, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1272, "step": 17919 }, { "epoch": 26.82634730538922, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1263, "step": 17920 }, { "epoch": 26.827844311377245, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.129, "step": 17921 }, { "epoch": 26.82934131736527, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1329, "step": 17922 }, { "epoch": 26.830838323353294, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1306, "step": 17923 }, { "epoch": 26.83233532934132, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1289, "step": 17924 }, { "epoch": 26.83383233532934, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1291, "step": 17925 }, { "epoch": 26.835329341317365, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1281, "step": 17926 }, { "epoch": 26.83682634730539, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1325, "step": 17927 }, { "epoch": 26.83832335329341, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.13, "step": 17928 }, { "epoch": 26.839820359281436, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1294, "step": 17929 }, { "epoch": 26.84131736526946, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1232, "step": 17930 }, { "epoch": 26.842814371257486, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1348, "step": 17931 }, { "epoch": 26.84431137724551, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1358, "step": 17932 }, { "epoch": 26.845808383233532, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1298, "step": 17933 }, { "epoch": 26.847305389221557, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1327, "step": 17934 }, { "epoch": 26.84880239520958, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1321, "step": 17935 }, { "epoch": 26.850299401197606, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1317, "step": 17936 }, { "epoch": 26.851796407185628, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1336, "step": 17937 }, { "epoch": 26.853293413173652, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1253, "step": 17938 }, { "epoch": 26.854790419161677, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1301, "step": 17939 }, { "epoch": 26.856287425149702, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1375, "step": 17940 }, { "epoch": 26.857784431137723, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1306, "step": 17941 }, { "epoch": 26.85928143712575, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1288, "step": 17942 }, { "epoch": 26.860778443113773, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1278, "step": 17943 }, { "epoch": 26.862275449101798, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1329, "step": 17944 }, { "epoch": 26.86377245508982, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1325, "step": 17945 }, { "epoch": 26.865269461077844, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1321, "step": 17946 }, { "epoch": 26.86676646706587, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1271, "step": 17947 }, { "epoch": 26.868263473053894, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1279, "step": 17948 }, { "epoch": 26.869760479041915, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1284, "step": 17949 }, { "epoch": 26.87125748502994, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1283, "step": 17950 }, { "epoch": 26.872754491017965, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1253, "step": 17951 }, { "epoch": 26.87425149700599, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1284, "step": 17952 }, { "epoch": 26.87574850299401, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1279, "step": 17953 }, { "epoch": 26.877245508982035, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1291, "step": 17954 }, { "epoch": 26.87874251497006, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1281, "step": 17955 }, { "epoch": 26.880239520958085, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1318, "step": 17956 }, { "epoch": 26.881736526946106, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.129, "step": 17957 }, { "epoch": 26.88323353293413, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1277, "step": 17958 }, { "epoch": 26.884730538922156, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1257, "step": 17959 }, { "epoch": 26.88622754491018, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1314, "step": 17960 }, { "epoch": 26.887724550898202, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.131, "step": 17961 }, { "epoch": 26.889221556886227, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1274, "step": 17962 }, { "epoch": 26.89071856287425, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.131, "step": 17963 }, { "epoch": 26.892215568862277, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1294, "step": 17964 }, { "epoch": 26.893712574850298, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1323, "step": 17965 }, { "epoch": 26.895209580838323, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1368, "step": 17966 }, { "epoch": 26.896706586826348, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1356, "step": 17967 }, { "epoch": 26.898203592814372, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1316, "step": 17968 }, { "epoch": 26.899700598802394, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1296, "step": 17969 }, { "epoch": 26.90119760479042, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.131, "step": 17970 }, { "epoch": 26.902694610778443, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1335, "step": 17971 }, { "epoch": 26.904191616766468, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1383, "step": 17972 }, { "epoch": 26.90568862275449, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1301, "step": 17973 }, { "epoch": 26.907185628742514, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1295, "step": 17974 }, { "epoch": 26.90868263473054, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1299, "step": 17975 }, { "epoch": 26.910179640718564, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1281, "step": 17976 }, { "epoch": 26.91167664670659, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.123, "step": 17977 }, { "epoch": 26.91317365269461, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1354, "step": 17978 }, { "epoch": 26.914670658682635, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1339, "step": 17979 }, { "epoch": 26.91616766467066, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1314, "step": 17980 }, { "epoch": 26.91766467065868, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.127, "step": 17981 }, { "epoch": 26.919161676646706, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1312, "step": 17982 }, { "epoch": 26.92065868263473, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1256, "step": 17983 }, { "epoch": 26.922155688622755, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.128, "step": 17984 }, { "epoch": 26.92365269461078, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1317, "step": 17985 }, { "epoch": 26.9251497005988, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1238, "step": 17986 }, { "epoch": 26.926646706586826, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1298, "step": 17987 }, { "epoch": 26.92814371257485, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1315, "step": 17988 }, { "epoch": 26.929640718562876, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1298, "step": 17989 }, { "epoch": 26.931137724550897, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1283, "step": 17990 }, { "epoch": 26.932634730538922, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1299, "step": 17991 }, { "epoch": 26.934131736526947, "grad_norm": 0.12060546875, "learning_rate": 0.0008, "loss": 1.1281, "step": 17992 }, { "epoch": 26.93562874251497, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.125, "step": 17993 }, { "epoch": 26.937125748502993, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1279, "step": 17994 }, { "epoch": 26.938622754491018, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1262, "step": 17995 }, { "epoch": 26.940119760479043, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1315, "step": 17996 }, { "epoch": 26.941616766467067, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1318, "step": 17997 }, { "epoch": 26.94311377245509, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1314, "step": 17998 }, { "epoch": 26.944610778443113, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1294, "step": 17999 }, { "epoch": 26.94610778443114, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1291, "step": 18000 }, { "epoch": 26.947604790419163, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1255, "step": 18001 }, { "epoch": 26.949101796407184, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1312, "step": 18002 }, { "epoch": 26.95059880239521, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.133, "step": 18003 }, { "epoch": 26.952095808383234, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1319, "step": 18004 }, { "epoch": 26.95359281437126, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1332, "step": 18005 }, { "epoch": 26.95508982035928, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1308, "step": 18006 }, { "epoch": 26.956586826347305, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.128, "step": 18007 }, { "epoch": 26.95808383233533, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1252, "step": 18008 }, { "epoch": 26.959580838323355, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1277, "step": 18009 }, { "epoch": 26.961077844311376, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1312, "step": 18010 }, { "epoch": 26.9625748502994, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.135, "step": 18011 }, { "epoch": 26.964071856287426, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1251, "step": 18012 }, { "epoch": 26.96556886227545, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1277, "step": 18013 }, { "epoch": 26.96706586826347, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1261, "step": 18014 }, { "epoch": 26.968562874251496, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1303, "step": 18015 }, { "epoch": 26.97005988023952, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1309, "step": 18016 }, { "epoch": 26.971556886227546, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.126, "step": 18017 }, { "epoch": 26.973053892215567, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1263, "step": 18018 }, { "epoch": 26.974550898203592, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1308, "step": 18019 }, { "epoch": 26.976047904191617, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1332, "step": 18020 }, { "epoch": 26.977544910179642, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1308, "step": 18021 }, { "epoch": 26.979041916167663, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.13, "step": 18022 }, { "epoch": 26.980538922155688, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1201, "step": 18023 }, { "epoch": 26.982035928143713, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1316, "step": 18024 }, { "epoch": 26.983532934131738, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1303, "step": 18025 }, { "epoch": 26.98502994011976, "grad_norm": 0.1298828125, "learning_rate": 0.0008, "loss": 1.131, "step": 18026 }, { "epoch": 26.986526946107784, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1313, "step": 18027 }, { "epoch": 26.98802395209581, "grad_norm": 0.1689453125, "learning_rate": 0.0008, "loss": 1.1338, "step": 18028 }, { "epoch": 26.989520958083833, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1384, "step": 18029 }, { "epoch": 26.991017964071855, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1309, "step": 18030 }, { "epoch": 26.99251497005988, "grad_norm": 0.142578125, "learning_rate": 0.0008, "loss": 1.1374, "step": 18031 }, { "epoch": 26.994011976047904, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1345, "step": 18032 }, { "epoch": 26.99550898203593, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.129, "step": 18033 }, { "epoch": 26.99700598802395, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1236, "step": 18034 }, { "epoch": 26.998502994011975, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1287, "step": 18035 }, { "epoch": 27.0, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1331, "step": 18036 }, { "epoch": 27.001497005988025, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1258, "step": 18037 }, { "epoch": 27.00299401197605, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.128, "step": 18038 }, { "epoch": 27.00449101796407, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1274, "step": 18039 }, { "epoch": 27.005988023952096, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1239, "step": 18040 }, { "epoch": 27.00748502994012, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1299, "step": 18041 }, { "epoch": 27.008982035928145, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1332, "step": 18042 }, { "epoch": 27.010479041916167, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1286, "step": 18043 }, { "epoch": 27.01197604790419, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1332, "step": 18044 }, { "epoch": 27.013473053892216, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1307, "step": 18045 }, { "epoch": 27.01497005988024, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1291, "step": 18046 }, { "epoch": 27.016467065868262, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1269, "step": 18047 }, { "epoch": 27.017964071856287, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1231, "step": 18048 }, { "epoch": 27.019461077844312, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1301, "step": 18049 }, { "epoch": 27.020958083832337, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1255, "step": 18050 }, { "epoch": 27.022455089820358, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1317, "step": 18051 }, { "epoch": 27.023952095808383, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1326, "step": 18052 }, { "epoch": 27.025449101796408, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1266, "step": 18053 }, { "epoch": 27.026946107784433, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1216, "step": 18054 }, { "epoch": 27.028443113772454, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1325, "step": 18055 }, { "epoch": 27.02994011976048, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1274, "step": 18056 }, { "epoch": 27.031437125748504, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1342, "step": 18057 }, { "epoch": 27.03293413173653, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1233, "step": 18058 }, { "epoch": 27.03443113772455, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1326, "step": 18059 }, { "epoch": 27.035928143712574, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1315, "step": 18060 }, { "epoch": 27.0374251497006, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1307, "step": 18061 }, { "epoch": 27.038922155688624, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1286, "step": 18062 }, { "epoch": 27.040419161676645, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1327, "step": 18063 }, { "epoch": 27.04191616766467, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1302, "step": 18064 }, { "epoch": 27.043413173652695, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1278, "step": 18065 }, { "epoch": 27.04491017964072, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1279, "step": 18066 }, { "epoch": 27.04640718562874, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1293, "step": 18067 }, { "epoch": 27.047904191616766, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1297, "step": 18068 }, { "epoch": 27.04940119760479, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1291, "step": 18069 }, { "epoch": 27.050898203592816, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1242, "step": 18070 }, { "epoch": 27.052395209580837, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1305, "step": 18071 }, { "epoch": 27.05389221556886, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1327, "step": 18072 }, { "epoch": 27.055389221556887, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1207, "step": 18073 }, { "epoch": 27.05688622754491, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1328, "step": 18074 }, { "epoch": 27.058383233532933, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.13, "step": 18075 }, { "epoch": 27.059880239520957, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1281, "step": 18076 }, { "epoch": 27.061377245508982, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1277, "step": 18077 }, { "epoch": 27.062874251497007, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1242, "step": 18078 }, { "epoch": 27.06437125748503, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1308, "step": 18079 }, { "epoch": 27.065868263473053, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1243, "step": 18080 }, { "epoch": 27.067365269461078, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1336, "step": 18081 }, { "epoch": 27.068862275449103, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1327, "step": 18082 }, { "epoch": 27.070359281437124, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1378, "step": 18083 }, { "epoch": 27.07185628742515, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.127, "step": 18084 }, { "epoch": 27.073353293413174, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1301, "step": 18085 }, { "epoch": 27.0748502994012, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1294, "step": 18086 }, { "epoch": 27.07634730538922, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1299, "step": 18087 }, { "epoch": 27.077844311377245, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1333, "step": 18088 }, { "epoch": 27.07934131736527, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1283, "step": 18089 }, { "epoch": 27.080838323353294, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1257, "step": 18090 }, { "epoch": 27.082335329341316, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1255, "step": 18091 }, { "epoch": 27.08383233532934, "grad_norm": 0.048828125, "learning_rate": 0.0008, "loss": 1.1289, "step": 18092 }, { "epoch": 27.085329341317365, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1283, "step": 18093 }, { "epoch": 27.08682634730539, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1187, "step": 18094 }, { "epoch": 27.088323353293415, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1301, "step": 18095 }, { "epoch": 27.089820359281436, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1237, "step": 18096 }, { "epoch": 27.09131736526946, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1258, "step": 18097 }, { "epoch": 27.092814371257486, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1368, "step": 18098 }, { "epoch": 27.09431137724551, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1261, "step": 18099 }, { "epoch": 27.095808383233532, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1287, "step": 18100 }, { "epoch": 27.097305389221557, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1293, "step": 18101 }, { "epoch": 27.09880239520958, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1313, "step": 18102 }, { "epoch": 27.100299401197606, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1221, "step": 18103 }, { "epoch": 27.101796407185628, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1338, "step": 18104 }, { "epoch": 27.103293413173652, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1314, "step": 18105 }, { "epoch": 27.104790419161677, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1314, "step": 18106 }, { "epoch": 27.106287425149702, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.13, "step": 18107 }, { "epoch": 27.107784431137723, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1324, "step": 18108 }, { "epoch": 27.10928143712575, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1329, "step": 18109 }, { "epoch": 27.110778443113773, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1309, "step": 18110 }, { "epoch": 27.112275449101798, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1301, "step": 18111 }, { "epoch": 27.11377245508982, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1267, "step": 18112 }, { "epoch": 27.115269461077844, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1268, "step": 18113 }, { "epoch": 27.11676646706587, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1258, "step": 18114 }, { "epoch": 27.118263473053894, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1247, "step": 18115 }, { "epoch": 27.119760479041915, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1282, "step": 18116 }, { "epoch": 27.12125748502994, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1339, "step": 18117 }, { "epoch": 27.122754491017965, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1284, "step": 18118 }, { "epoch": 27.12425149700599, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1295, "step": 18119 }, { "epoch": 27.12574850299401, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1274, "step": 18120 }, { "epoch": 27.127245508982035, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1183, "step": 18121 }, { "epoch": 27.12874251497006, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1308, "step": 18122 }, { "epoch": 27.130239520958085, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1334, "step": 18123 }, { "epoch": 27.131736526946106, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 18124 }, { "epoch": 27.13323353293413, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1295, "step": 18125 }, { "epoch": 27.134730538922156, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1275, "step": 18126 }, { "epoch": 27.13622754491018, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1258, "step": 18127 }, { "epoch": 27.137724550898202, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1257, "step": 18128 }, { "epoch": 27.139221556886227, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1309, "step": 18129 }, { "epoch": 27.14071856287425, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1246, "step": 18130 }, { "epoch": 27.142215568862277, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1315, "step": 18131 }, { "epoch": 27.143712574850298, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1227, "step": 18132 }, { "epoch": 27.145209580838323, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1281, "step": 18133 }, { "epoch": 27.146706586826348, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1327, "step": 18134 }, { "epoch": 27.148203592814372, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1332, "step": 18135 }, { "epoch": 27.149700598802394, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1327, "step": 18136 }, { "epoch": 27.15119760479042, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1283, "step": 18137 }, { "epoch": 27.152694610778443, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1356, "step": 18138 }, { "epoch": 27.154191616766468, "grad_norm": 0.12109375, "learning_rate": 0.0008, "loss": 1.1286, "step": 18139 }, { "epoch": 27.15568862275449, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1348, "step": 18140 }, { "epoch": 27.157185628742514, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1337, "step": 18141 }, { "epoch": 27.15868263473054, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.128, "step": 18142 }, { "epoch": 27.160179640718564, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1307, "step": 18143 }, { "epoch": 27.161676646706585, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1224, "step": 18144 }, { "epoch": 27.16317365269461, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1267, "step": 18145 }, { "epoch": 27.164670658682635, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1228, "step": 18146 }, { "epoch": 27.16616766467066, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1296, "step": 18147 }, { "epoch": 27.16766467065868, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1305, "step": 18148 }, { "epoch": 27.169161676646706, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1269, "step": 18149 }, { "epoch": 27.17065868263473, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1241, "step": 18150 }, { "epoch": 27.172155688622755, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1311, "step": 18151 }, { "epoch": 27.17365269461078, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1293, "step": 18152 }, { "epoch": 27.1751497005988, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.128, "step": 18153 }, { "epoch": 27.176646706586826, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1297, "step": 18154 }, { "epoch": 27.17814371257485, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1276, "step": 18155 }, { "epoch": 27.179640718562876, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1244, "step": 18156 }, { "epoch": 27.181137724550897, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1288, "step": 18157 }, { "epoch": 27.182634730538922, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1312, "step": 18158 }, { "epoch": 27.184131736526947, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1256, "step": 18159 }, { "epoch": 27.18562874251497, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1295, "step": 18160 }, { "epoch": 27.187125748502993, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1281, "step": 18161 }, { "epoch": 27.188622754491018, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1257, "step": 18162 }, { "epoch": 27.190119760479043, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1282, "step": 18163 }, { "epoch": 27.191616766467067, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1304, "step": 18164 }, { "epoch": 27.19311377245509, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1285, "step": 18165 }, { "epoch": 27.194610778443113, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1263, "step": 18166 }, { "epoch": 27.19610778443114, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1301, "step": 18167 }, { "epoch": 27.197604790419163, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1304, "step": 18168 }, { "epoch": 27.199101796407184, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1243, "step": 18169 }, { "epoch": 27.20059880239521, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1211, "step": 18170 }, { "epoch": 27.202095808383234, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1283, "step": 18171 }, { "epoch": 27.20359281437126, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1209, "step": 18172 }, { "epoch": 27.20508982035928, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1254, "step": 18173 }, { "epoch": 27.206586826347305, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1302, "step": 18174 }, { "epoch": 27.20808383233533, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1318, "step": 18175 }, { "epoch": 27.209580838323355, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1336, "step": 18176 }, { "epoch": 27.211077844311376, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1273, "step": 18177 }, { "epoch": 27.2125748502994, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1297, "step": 18178 }, { "epoch": 27.214071856287426, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1298, "step": 18179 }, { "epoch": 27.21556886227545, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1284, "step": 18180 }, { "epoch": 27.21706586826347, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1357, "step": 18181 }, { "epoch": 27.218562874251496, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.125, "step": 18182 }, { "epoch": 27.22005988023952, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1301, "step": 18183 }, { "epoch": 27.221556886227546, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1248, "step": 18184 }, { "epoch": 27.223053892215567, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1286, "step": 18185 }, { "epoch": 27.224550898203592, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1322, "step": 18186 }, { "epoch": 27.226047904191617, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1297, "step": 18187 }, { "epoch": 27.227544910179642, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1308, "step": 18188 }, { "epoch": 27.229041916167663, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1295, "step": 18189 }, { "epoch": 27.230538922155688, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1259, "step": 18190 }, { "epoch": 27.232035928143713, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1299, "step": 18191 }, { "epoch": 27.233532934131738, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1256, "step": 18192 }, { "epoch": 27.23502994011976, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1302, "step": 18193 }, { "epoch": 27.236526946107784, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1364, "step": 18194 }, { "epoch": 27.23802395209581, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1307, "step": 18195 }, { "epoch": 27.239520958083833, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1276, "step": 18196 }, { "epoch": 27.241017964071855, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.125, "step": 18197 }, { "epoch": 27.24251497005988, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1289, "step": 18198 }, { "epoch": 27.244011976047904, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1291, "step": 18199 }, { "epoch": 27.24550898203593, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1293, "step": 18200 }, { "epoch": 27.24700598802395, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1314, "step": 18201 }, { "epoch": 27.248502994011975, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1356, "step": 18202 }, { "epoch": 27.25, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1352, "step": 18203 }, { "epoch": 27.251497005988025, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1242, "step": 18204 }, { "epoch": 27.25299401197605, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1349, "step": 18205 }, { "epoch": 27.25449101796407, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1298, "step": 18206 }, { "epoch": 27.255988023952096, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1312, "step": 18207 }, { "epoch": 27.25748502994012, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1367, "step": 18208 }, { "epoch": 27.258982035928145, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.13, "step": 18209 }, { "epoch": 27.260479041916167, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1271, "step": 18210 }, { "epoch": 27.26197604790419, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1321, "step": 18211 }, { "epoch": 27.263473053892216, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.128, "step": 18212 }, { "epoch": 27.26497005988024, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1303, "step": 18213 }, { "epoch": 27.266467065868262, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1297, "step": 18214 }, { "epoch": 27.267964071856287, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1249, "step": 18215 }, { "epoch": 27.269461077844312, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1325, "step": 18216 }, { "epoch": 27.270958083832337, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1271, "step": 18217 }, { "epoch": 27.272455089820358, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1273, "step": 18218 }, { "epoch": 27.273952095808383, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1327, "step": 18219 }, { "epoch": 27.275449101796408, "grad_norm": 0.11962890625, "learning_rate": 0.0008, "loss": 1.1268, "step": 18220 }, { "epoch": 27.276946107784433, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1355, "step": 18221 }, { "epoch": 27.278443113772454, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1296, "step": 18222 }, { "epoch": 27.27994011976048, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1248, "step": 18223 }, { "epoch": 27.281437125748504, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1295, "step": 18224 }, { "epoch": 27.28293413173653, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1338, "step": 18225 }, { "epoch": 27.28443113772455, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1269, "step": 18226 }, { "epoch": 27.285928143712574, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1345, "step": 18227 }, { "epoch": 27.2874251497006, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1305, "step": 18228 }, { "epoch": 27.288922155688624, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1247, "step": 18229 }, { "epoch": 27.290419161676645, "grad_norm": 0.05224609375, "learning_rate": 0.0008, "loss": 1.13, "step": 18230 }, { "epoch": 27.29191616766467, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1324, "step": 18231 }, { "epoch": 27.293413173652695, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1276, "step": 18232 }, { "epoch": 27.29491017964072, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1294, "step": 18233 }, { "epoch": 27.29640718562874, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1307, "step": 18234 }, { "epoch": 27.297904191616766, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1258, "step": 18235 }, { "epoch": 27.29940119760479, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1263, "step": 18236 }, { "epoch": 27.300898203592816, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1314, "step": 18237 }, { "epoch": 27.302395209580837, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1306, "step": 18238 }, { "epoch": 27.30389221556886, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1226, "step": 18239 }, { "epoch": 27.305389221556887, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1328, "step": 18240 }, { "epoch": 27.30688622754491, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1196, "step": 18241 }, { "epoch": 27.308383233532933, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1331, "step": 18242 }, { "epoch": 27.309880239520957, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1331, "step": 18243 }, { "epoch": 27.311377245508982, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1221, "step": 18244 }, { "epoch": 27.312874251497007, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1243, "step": 18245 }, { "epoch": 27.31437125748503, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1301, "step": 18246 }, { "epoch": 27.315868263473053, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1334, "step": 18247 }, { "epoch": 27.317365269461078, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1271, "step": 18248 }, { "epoch": 27.318862275449103, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1301, "step": 18249 }, { "epoch": 27.320359281437124, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1309, "step": 18250 }, { "epoch": 27.32185628742515, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1315, "step": 18251 }, { "epoch": 27.323353293413174, "grad_norm": 0.0517578125, "learning_rate": 0.0008, "loss": 1.1316, "step": 18252 }, { "epoch": 27.3248502994012, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1269, "step": 18253 }, { "epoch": 27.32634730538922, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1237, "step": 18254 }, { "epoch": 27.327844311377245, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1295, "step": 18255 }, { "epoch": 27.32934131736527, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1342, "step": 18256 }, { "epoch": 27.330838323353294, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1233, "step": 18257 }, { "epoch": 27.33233532934132, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1283, "step": 18258 }, { "epoch": 27.33383233532934, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1255, "step": 18259 }, { "epoch": 27.335329341317365, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.129, "step": 18260 }, { "epoch": 27.33682634730539, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1328, "step": 18261 }, { "epoch": 27.338323353293415, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1291, "step": 18262 }, { "epoch": 27.339820359281436, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1272, "step": 18263 }, { "epoch": 27.34131736526946, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1321, "step": 18264 }, { "epoch": 27.342814371257486, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1302, "step": 18265 }, { "epoch": 27.34431137724551, "grad_norm": 0.053466796875, "learning_rate": 0.0008, "loss": 1.1281, "step": 18266 }, { "epoch": 27.345808383233532, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1248, "step": 18267 }, { "epoch": 27.347305389221557, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1326, "step": 18268 }, { "epoch": 27.34880239520958, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1274, "step": 18269 }, { "epoch": 27.350299401197606, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1331, "step": 18270 }, { "epoch": 27.351796407185628, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1309, "step": 18271 }, { "epoch": 27.353293413173652, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.13, "step": 18272 }, { "epoch": 27.354790419161677, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1293, "step": 18273 }, { "epoch": 27.356287425149702, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1184, "step": 18274 }, { "epoch": 27.357784431137723, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1271, "step": 18275 }, { "epoch": 27.35928143712575, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1319, "step": 18276 }, { "epoch": 27.360778443113773, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1291, "step": 18277 }, { "epoch": 27.362275449101798, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.127, "step": 18278 }, { "epoch": 27.36377245508982, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1324, "step": 18279 }, { "epoch": 27.365269461077844, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1316, "step": 18280 }, { "epoch": 27.36676646706587, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.129, "step": 18281 }, { "epoch": 27.368263473053894, "grad_norm": 0.10546875, "learning_rate": 0.0008, "loss": 1.1309, "step": 18282 }, { "epoch": 27.369760479041915, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1301, "step": 18283 }, { "epoch": 27.37125748502994, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1318, "step": 18284 }, { "epoch": 27.372754491017965, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1297, "step": 18285 }, { "epoch": 27.37425149700599, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.128, "step": 18286 }, { "epoch": 27.37574850299401, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1334, "step": 18287 }, { "epoch": 27.377245508982035, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1265, "step": 18288 }, { "epoch": 27.37874251497006, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1223, "step": 18289 }, { "epoch": 27.380239520958085, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1327, "step": 18290 }, { "epoch": 27.381736526946106, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1306, "step": 18291 }, { "epoch": 27.38323353293413, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.129, "step": 18292 }, { "epoch": 27.384730538922156, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1255, "step": 18293 }, { "epoch": 27.38622754491018, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1291, "step": 18294 }, { "epoch": 27.387724550898202, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1267, "step": 18295 }, { "epoch": 27.389221556886227, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1281, "step": 18296 }, { "epoch": 27.39071856287425, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1252, "step": 18297 }, { "epoch": 27.392215568862277, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1264, "step": 18298 }, { "epoch": 27.393712574850298, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1267, "step": 18299 }, { "epoch": 27.395209580838323, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1333, "step": 18300 }, { "epoch": 27.396706586826348, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.129, "step": 18301 }, { "epoch": 27.398203592814372, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1274, "step": 18302 }, { "epoch": 27.399700598802394, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1344, "step": 18303 }, { "epoch": 27.40119760479042, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1292, "step": 18304 }, { "epoch": 27.402694610778443, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1314, "step": 18305 }, { "epoch": 27.404191616766468, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1222, "step": 18306 }, { "epoch": 27.40568862275449, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1314, "step": 18307 }, { "epoch": 27.407185628742514, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1303, "step": 18308 }, { "epoch": 27.40868263473054, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1326, "step": 18309 }, { "epoch": 27.410179640718564, "grad_norm": 0.054931640625, "learning_rate": 0.0008, "loss": 1.1268, "step": 18310 }, { "epoch": 27.411676646706585, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1262, "step": 18311 }, { "epoch": 27.41317365269461, "grad_norm": 0.052001953125, "learning_rate": 0.0008, "loss": 1.1306, "step": 18312 }, { "epoch": 27.414670658682635, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1384, "step": 18313 }, { "epoch": 27.41616766467066, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1254, "step": 18314 }, { "epoch": 27.41766467065868, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1338, "step": 18315 }, { "epoch": 27.419161676646706, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1267, "step": 18316 }, { "epoch": 27.42065868263473, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1272, "step": 18317 }, { "epoch": 27.422155688622755, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1274, "step": 18318 }, { "epoch": 27.42365269461078, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1307, "step": 18319 }, { "epoch": 27.4251497005988, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.136, "step": 18320 }, { "epoch": 27.426646706586826, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1287, "step": 18321 }, { "epoch": 27.42814371257485, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1355, "step": 18322 }, { "epoch": 27.429640718562876, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1338, "step": 18323 }, { "epoch": 27.431137724550897, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1255, "step": 18324 }, { "epoch": 27.432634730538922, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1246, "step": 18325 }, { "epoch": 27.434131736526947, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1294, "step": 18326 }, { "epoch": 27.43562874251497, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1283, "step": 18327 }, { "epoch": 27.437125748502993, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.131, "step": 18328 }, { "epoch": 27.438622754491018, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1259, "step": 18329 }, { "epoch": 27.440119760479043, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1302, "step": 18330 }, { "epoch": 27.441616766467067, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1316, "step": 18331 }, { "epoch": 27.44311377245509, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1362, "step": 18332 }, { "epoch": 27.444610778443113, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1304, "step": 18333 }, { "epoch": 27.44610778443114, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1267, "step": 18334 }, { "epoch": 27.447604790419163, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.125, "step": 18335 }, { "epoch": 27.449101796407184, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1277, "step": 18336 }, { "epoch": 27.45059880239521, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1208, "step": 18337 }, { "epoch": 27.452095808383234, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1319, "step": 18338 }, { "epoch": 27.45359281437126, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.1258, "step": 18339 }, { "epoch": 27.45508982035928, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.134, "step": 18340 }, { "epoch": 27.456586826347305, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.127, "step": 18341 }, { "epoch": 27.45808383233533, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1311, "step": 18342 }, { "epoch": 27.459580838323355, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1251, "step": 18343 }, { "epoch": 27.461077844311376, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1248, "step": 18344 }, { "epoch": 27.4625748502994, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1275, "step": 18345 }, { "epoch": 27.464071856287426, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.131, "step": 18346 }, { "epoch": 27.46556886227545, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1288, "step": 18347 }, { "epoch": 27.46706586826347, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1258, "step": 18348 }, { "epoch": 27.468562874251496, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1238, "step": 18349 }, { "epoch": 27.47005988023952, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1332, "step": 18350 }, { "epoch": 27.471556886227546, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1299, "step": 18351 }, { "epoch": 27.473053892215567, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1316, "step": 18352 }, { "epoch": 27.474550898203592, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1317, "step": 18353 }, { "epoch": 27.476047904191617, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1186, "step": 18354 }, { "epoch": 27.477544910179642, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1276, "step": 18355 }, { "epoch": 27.479041916167663, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1196, "step": 18356 }, { "epoch": 27.480538922155688, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1328, "step": 18357 }, { "epoch": 27.482035928143713, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1323, "step": 18358 }, { "epoch": 27.483532934131738, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.135, "step": 18359 }, { "epoch": 27.48502994011976, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1254, "step": 18360 }, { "epoch": 27.486526946107784, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1297, "step": 18361 }, { "epoch": 27.48802395209581, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1281, "step": 18362 }, { "epoch": 27.489520958083833, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1302, "step": 18363 }, { "epoch": 27.491017964071855, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1256, "step": 18364 }, { "epoch": 27.49251497005988, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.127, "step": 18365 }, { "epoch": 27.494011976047904, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1267, "step": 18366 }, { "epoch": 27.49550898203593, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1226, "step": 18367 }, { "epoch": 27.49700598802395, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1315, "step": 18368 }, { "epoch": 27.498502994011975, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1316, "step": 18369 }, { "epoch": 27.5, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.133, "step": 18370 }, { "epoch": 27.501497005988025, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1281, "step": 18371 }, { "epoch": 27.50299401197605, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.134, "step": 18372 }, { "epoch": 27.50449101796407, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1329, "step": 18373 }, { "epoch": 27.505988023952096, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1305, "step": 18374 }, { "epoch": 27.50748502994012, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1306, "step": 18375 }, { "epoch": 27.508982035928145, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1245, "step": 18376 }, { "epoch": 27.510479041916167, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1336, "step": 18377 }, { "epoch": 27.51197604790419, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1331, "step": 18378 }, { "epoch": 27.513473053892216, "grad_norm": 0.0517578125, "learning_rate": 0.0008, "loss": 1.1286, "step": 18379 }, { "epoch": 27.51497005988024, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1301, "step": 18380 }, { "epoch": 27.516467065868262, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1302, "step": 18381 }, { "epoch": 27.517964071856287, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1281, "step": 18382 }, { "epoch": 27.519461077844312, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1292, "step": 18383 }, { "epoch": 27.520958083832337, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.13, "step": 18384 }, { "epoch": 27.522455089820358, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1241, "step": 18385 }, { "epoch": 27.523952095808383, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.125, "step": 18386 }, { "epoch": 27.525449101796408, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1269, "step": 18387 }, { "epoch": 27.526946107784433, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1299, "step": 18388 }, { "epoch": 27.528443113772454, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1341, "step": 18389 }, { "epoch": 27.52994011976048, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1303, "step": 18390 }, { "epoch": 27.531437125748504, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.129, "step": 18391 }, { "epoch": 27.53293413173653, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1328, "step": 18392 }, { "epoch": 27.53443113772455, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1288, "step": 18393 }, { "epoch": 27.535928143712574, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1275, "step": 18394 }, { "epoch": 27.5374251497006, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1306, "step": 18395 }, { "epoch": 27.538922155688624, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1242, "step": 18396 }, { "epoch": 27.540419161676645, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1385, "step": 18397 }, { "epoch": 27.54191616766467, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.126, "step": 18398 }, { "epoch": 27.543413173652695, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1244, "step": 18399 }, { "epoch": 27.54491017964072, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1226, "step": 18400 }, { "epoch": 27.54640718562874, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1322, "step": 18401 }, { "epoch": 27.547904191616766, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1312, "step": 18402 }, { "epoch": 27.54940119760479, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1253, "step": 18403 }, { "epoch": 27.550898203592816, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1355, "step": 18404 }, { "epoch": 27.552395209580837, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1296, "step": 18405 }, { "epoch": 27.55389221556886, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1326, "step": 18406 }, { "epoch": 27.555389221556887, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1283, "step": 18407 }, { "epoch": 27.55688622754491, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1284, "step": 18408 }, { "epoch": 27.558383233532933, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.135, "step": 18409 }, { "epoch": 27.559880239520957, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1254, "step": 18410 }, { "epoch": 27.561377245508982, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1321, "step": 18411 }, { "epoch": 27.562874251497007, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1321, "step": 18412 }, { "epoch": 27.56437125748503, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1359, "step": 18413 }, { "epoch": 27.565868263473053, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.13, "step": 18414 }, { "epoch": 27.567365269461078, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1306, "step": 18415 }, { "epoch": 27.568862275449103, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.135, "step": 18416 }, { "epoch": 27.570359281437124, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1268, "step": 18417 }, { "epoch": 27.57185628742515, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1296, "step": 18418 }, { "epoch": 27.573353293413174, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1277, "step": 18419 }, { "epoch": 27.5748502994012, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.126, "step": 18420 }, { "epoch": 27.57634730538922, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1269, "step": 18421 }, { "epoch": 27.577844311377245, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1265, "step": 18422 }, { "epoch": 27.57934131736527, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1228, "step": 18423 }, { "epoch": 27.580838323353294, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1394, "step": 18424 }, { "epoch": 27.58233532934132, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1254, "step": 18425 }, { "epoch": 27.58383233532934, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1237, "step": 18426 }, { "epoch": 27.585329341317365, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1269, "step": 18427 }, { "epoch": 27.58682634730539, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1336, "step": 18428 }, { "epoch": 27.58832335329341, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1392, "step": 18429 }, { "epoch": 27.589820359281436, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1309, "step": 18430 }, { "epoch": 27.59131736526946, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1299, "step": 18431 }, { "epoch": 27.592814371257486, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1274, "step": 18432 }, { "epoch": 27.59431137724551, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1342, "step": 18433 }, { "epoch": 27.595808383233532, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1295, "step": 18434 }, { "epoch": 27.597305389221557, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1318, "step": 18435 }, { "epoch": 27.59880239520958, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1262, "step": 18436 }, { "epoch": 27.600299401197606, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1268, "step": 18437 }, { "epoch": 27.601796407185628, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1305, "step": 18438 }, { "epoch": 27.603293413173652, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1329, "step": 18439 }, { "epoch": 27.604790419161677, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1297, "step": 18440 }, { "epoch": 27.606287425149702, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1314, "step": 18441 }, { "epoch": 27.607784431137723, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.132, "step": 18442 }, { "epoch": 27.60928143712575, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1252, "step": 18443 }, { "epoch": 27.610778443113773, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1315, "step": 18444 }, { "epoch": 27.612275449101798, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1313, "step": 18445 }, { "epoch": 27.61377245508982, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1253, "step": 18446 }, { "epoch": 27.615269461077844, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1211, "step": 18447 }, { "epoch": 27.61676646706587, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.132, "step": 18448 }, { "epoch": 27.618263473053894, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1252, "step": 18449 }, { "epoch": 27.619760479041915, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1323, "step": 18450 }, { "epoch": 27.62125748502994, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1339, "step": 18451 }, { "epoch": 27.622754491017965, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1226, "step": 18452 }, { "epoch": 27.62425149700599, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1258, "step": 18453 }, { "epoch": 27.62574850299401, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1263, "step": 18454 }, { "epoch": 27.627245508982035, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1345, "step": 18455 }, { "epoch": 27.62874251497006, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1315, "step": 18456 }, { "epoch": 27.630239520958085, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1307, "step": 18457 }, { "epoch": 27.631736526946106, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1285, "step": 18458 }, { "epoch": 27.63323353293413, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1251, "step": 18459 }, { "epoch": 27.634730538922156, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1373, "step": 18460 }, { "epoch": 27.63622754491018, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1323, "step": 18461 }, { "epoch": 27.637724550898202, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1307, "step": 18462 }, { "epoch": 27.639221556886227, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1303, "step": 18463 }, { "epoch": 27.64071856287425, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1293, "step": 18464 }, { "epoch": 27.642215568862277, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1322, "step": 18465 }, { "epoch": 27.643712574850298, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1345, "step": 18466 }, { "epoch": 27.645209580838323, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1273, "step": 18467 }, { "epoch": 27.646706586826348, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1274, "step": 18468 }, { "epoch": 27.648203592814372, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1286, "step": 18469 }, { "epoch": 27.649700598802394, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.123, "step": 18470 }, { "epoch": 27.65119760479042, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1325, "step": 18471 }, { "epoch": 27.652694610778443, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1353, "step": 18472 }, { "epoch": 27.654191616766468, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1298, "step": 18473 }, { "epoch": 27.65568862275449, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.125, "step": 18474 }, { "epoch": 27.657185628742514, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1362, "step": 18475 }, { "epoch": 27.65868263473054, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1281, "step": 18476 }, { "epoch": 27.660179640718564, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1293, "step": 18477 }, { "epoch": 27.66167664670659, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1262, "step": 18478 }, { "epoch": 27.66317365269461, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1281, "step": 18479 }, { "epoch": 27.664670658682635, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1308, "step": 18480 }, { "epoch": 27.66616766467066, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1283, "step": 18481 }, { "epoch": 27.66766467065868, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1321, "step": 18482 }, { "epoch": 27.669161676646706, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1276, "step": 18483 }, { "epoch": 27.67065868263473, "grad_norm": 0.123046875, "learning_rate": 0.0008, "loss": 1.1378, "step": 18484 }, { "epoch": 27.672155688622755, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1366, "step": 18485 }, { "epoch": 27.67365269461078, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1271, "step": 18486 }, { "epoch": 27.6751497005988, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1317, "step": 18487 }, { "epoch": 27.676646706586826, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1288, "step": 18488 }, { "epoch": 27.67814371257485, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1245, "step": 18489 }, { "epoch": 27.679640718562876, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1297, "step": 18490 }, { "epoch": 27.681137724550897, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1308, "step": 18491 }, { "epoch": 27.682634730538922, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1255, "step": 18492 }, { "epoch": 27.684131736526947, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1221, "step": 18493 }, { "epoch": 27.68562874251497, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1277, "step": 18494 }, { "epoch": 27.687125748502993, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1346, "step": 18495 }, { "epoch": 27.688622754491018, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.129, "step": 18496 }, { "epoch": 27.690119760479043, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1293, "step": 18497 }, { "epoch": 27.691616766467067, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1285, "step": 18498 }, { "epoch": 27.69311377245509, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1311, "step": 18499 }, { "epoch": 27.694610778443113, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.133, "step": 18500 }, { "epoch": 27.69610778443114, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1356, "step": 18501 }, { "epoch": 27.697604790419163, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.131, "step": 18502 }, { "epoch": 27.699101796407184, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1331, "step": 18503 }, { "epoch": 27.70059880239521, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1288, "step": 18504 }, { "epoch": 27.702095808383234, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1277, "step": 18505 }, { "epoch": 27.70359281437126, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1262, "step": 18506 }, { "epoch": 27.70508982035928, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1265, "step": 18507 }, { "epoch": 27.706586826347305, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1272, "step": 18508 }, { "epoch": 27.70808383233533, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1274, "step": 18509 }, { "epoch": 27.709580838323355, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1316, "step": 18510 }, { "epoch": 27.711077844311376, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1253, "step": 18511 }, { "epoch": 27.7125748502994, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1327, "step": 18512 }, { "epoch": 27.714071856287426, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1309, "step": 18513 }, { "epoch": 27.71556886227545, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1324, "step": 18514 }, { "epoch": 27.71706586826347, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1266, "step": 18515 }, { "epoch": 27.718562874251496, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1303, "step": 18516 }, { "epoch": 27.72005988023952, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1304, "step": 18517 }, { "epoch": 27.721556886227546, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1329, "step": 18518 }, { "epoch": 27.723053892215567, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1222, "step": 18519 }, { "epoch": 27.724550898203592, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.132, "step": 18520 }, { "epoch": 27.726047904191617, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1278, "step": 18521 }, { "epoch": 27.727544910179642, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1316, "step": 18522 }, { "epoch": 27.729041916167663, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1334, "step": 18523 }, { "epoch": 27.730538922155688, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1252, "step": 18524 }, { "epoch": 27.732035928143713, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1302, "step": 18525 }, { "epoch": 27.733532934131738, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1301, "step": 18526 }, { "epoch": 27.73502994011976, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1228, "step": 18527 }, { "epoch": 27.736526946107784, "grad_norm": 0.05078125, "learning_rate": 0.0008, "loss": 1.1283, "step": 18528 }, { "epoch": 27.73802395209581, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1288, "step": 18529 }, { "epoch": 27.739520958083833, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1231, "step": 18530 }, { "epoch": 27.741017964071855, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1242, "step": 18531 }, { "epoch": 27.74251497005988, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1314, "step": 18532 }, { "epoch": 27.744011976047904, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1288, "step": 18533 }, { "epoch": 27.74550898203593, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1294, "step": 18534 }, { "epoch": 27.74700598802395, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1274, "step": 18535 }, { "epoch": 27.748502994011975, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1309, "step": 18536 }, { "epoch": 27.75, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1321, "step": 18537 }, { "epoch": 27.751497005988025, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1288, "step": 18538 }, { "epoch": 27.75299401197605, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.125, "step": 18539 }, { "epoch": 27.75449101796407, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1278, "step": 18540 }, { "epoch": 27.755988023952096, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1248, "step": 18541 }, { "epoch": 27.75748502994012, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1333, "step": 18542 }, { "epoch": 27.758982035928145, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1265, "step": 18543 }, { "epoch": 27.760479041916167, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1328, "step": 18544 }, { "epoch": 27.76197604790419, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1271, "step": 18545 }, { "epoch": 27.763473053892216, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1274, "step": 18546 }, { "epoch": 27.76497005988024, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1257, "step": 18547 }, { "epoch": 27.766467065868262, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1282, "step": 18548 }, { "epoch": 27.767964071856287, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1236, "step": 18549 }, { "epoch": 27.769461077844312, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1262, "step": 18550 }, { "epoch": 27.770958083832337, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1307, "step": 18551 }, { "epoch": 27.772455089820358, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1283, "step": 18552 }, { "epoch": 27.773952095808383, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1256, "step": 18553 }, { "epoch": 27.775449101796408, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1294, "step": 18554 }, { "epoch": 27.776946107784433, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1229, "step": 18555 }, { "epoch": 27.778443113772454, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.1274, "step": 18556 }, { "epoch": 27.77994011976048, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1269, "step": 18557 }, { "epoch": 27.781437125748504, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1245, "step": 18558 }, { "epoch": 27.78293413173653, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1306, "step": 18559 }, { "epoch": 27.78443113772455, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1234, "step": 18560 }, { "epoch": 27.785928143712574, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1257, "step": 18561 }, { "epoch": 27.7874251497006, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1258, "step": 18562 }, { "epoch": 27.788922155688624, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1275, "step": 18563 }, { "epoch": 27.790419161676645, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1285, "step": 18564 }, { "epoch": 27.79191616766467, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1358, "step": 18565 }, { "epoch": 27.793413173652695, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1273, "step": 18566 }, { "epoch": 27.79491017964072, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1337, "step": 18567 }, { "epoch": 27.79640718562874, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1339, "step": 18568 }, { "epoch": 27.797904191616766, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1312, "step": 18569 }, { "epoch": 27.79940119760479, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1283, "step": 18570 }, { "epoch": 27.800898203592816, "grad_norm": 0.0537109375, "learning_rate": 0.0008, "loss": 1.1285, "step": 18571 }, { "epoch": 27.802395209580837, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.14, "step": 18572 }, { "epoch": 27.80389221556886, "grad_norm": 0.050537109375, "learning_rate": 0.0008, "loss": 1.1277, "step": 18573 }, { "epoch": 27.805389221556887, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1319, "step": 18574 }, { "epoch": 27.80688622754491, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1351, "step": 18575 }, { "epoch": 27.808383233532933, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1266, "step": 18576 }, { "epoch": 27.809880239520957, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.127, "step": 18577 }, { "epoch": 27.811377245508982, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1289, "step": 18578 }, { "epoch": 27.812874251497007, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1252, "step": 18579 }, { "epoch": 27.81437125748503, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1323, "step": 18580 }, { "epoch": 27.815868263473053, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1247, "step": 18581 }, { "epoch": 27.817365269461078, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1267, "step": 18582 }, { "epoch": 27.818862275449103, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1343, "step": 18583 }, { "epoch": 27.820359281437124, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1274, "step": 18584 }, { "epoch": 27.82185628742515, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.125, "step": 18585 }, { "epoch": 27.823353293413174, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1323, "step": 18586 }, { "epoch": 27.8248502994012, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1316, "step": 18587 }, { "epoch": 27.82634730538922, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1274, "step": 18588 }, { "epoch": 27.827844311377245, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1337, "step": 18589 }, { "epoch": 27.82934131736527, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1221, "step": 18590 }, { "epoch": 27.830838323353294, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1254, "step": 18591 }, { "epoch": 27.83233532934132, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1284, "step": 18592 }, { "epoch": 27.83383233532934, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1242, "step": 18593 }, { "epoch": 27.835329341317365, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1319, "step": 18594 }, { "epoch": 27.83682634730539, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1292, "step": 18595 }, { "epoch": 27.83832335329341, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1305, "step": 18596 }, { "epoch": 27.839820359281436, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1263, "step": 18597 }, { "epoch": 27.84131736526946, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1301, "step": 18598 }, { "epoch": 27.842814371257486, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1308, "step": 18599 }, { "epoch": 27.84431137724551, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1342, "step": 18600 }, { "epoch": 27.845808383233532, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1293, "step": 18601 }, { "epoch": 27.847305389221557, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1276, "step": 18602 }, { "epoch": 27.84880239520958, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.124, "step": 18603 }, { "epoch": 27.850299401197606, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1263, "step": 18604 }, { "epoch": 27.851796407185628, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1282, "step": 18605 }, { "epoch": 27.853293413173652, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1302, "step": 18606 }, { "epoch": 27.854790419161677, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1253, "step": 18607 }, { "epoch": 27.856287425149702, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.131, "step": 18608 }, { "epoch": 27.857784431137723, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1324, "step": 18609 }, { "epoch": 27.85928143712575, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1214, "step": 18610 }, { "epoch": 27.860778443113773, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1306, "step": 18611 }, { "epoch": 27.862275449101798, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1255, "step": 18612 }, { "epoch": 27.86377245508982, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1298, "step": 18613 }, { "epoch": 27.865269461077844, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1312, "step": 18614 }, { "epoch": 27.86676646706587, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1291, "step": 18615 }, { "epoch": 27.868263473053894, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1351, "step": 18616 }, { "epoch": 27.869760479041915, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1333, "step": 18617 }, { "epoch": 27.87125748502994, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1296, "step": 18618 }, { "epoch": 27.872754491017965, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1278, "step": 18619 }, { "epoch": 27.87425149700599, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1326, "step": 18620 }, { "epoch": 27.87574850299401, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1272, "step": 18621 }, { "epoch": 27.877245508982035, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1306, "step": 18622 }, { "epoch": 27.87874251497006, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1288, "step": 18623 }, { "epoch": 27.880239520958085, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1268, "step": 18624 }, { "epoch": 27.881736526946106, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1239, "step": 18625 }, { "epoch": 27.88323353293413, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.125, "step": 18626 }, { "epoch": 27.884730538922156, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1237, "step": 18627 }, { "epoch": 27.88622754491018, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1325, "step": 18628 }, { "epoch": 27.887724550898202, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1313, "step": 18629 }, { "epoch": 27.889221556886227, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1302, "step": 18630 }, { "epoch": 27.89071856287425, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1232, "step": 18631 }, { "epoch": 27.892215568862277, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1199, "step": 18632 }, { "epoch": 27.893712574850298, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1313, "step": 18633 }, { "epoch": 27.895209580838323, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1343, "step": 18634 }, { "epoch": 27.896706586826348, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1289, "step": 18635 }, { "epoch": 27.898203592814372, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1263, "step": 18636 }, { "epoch": 27.899700598802394, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1285, "step": 18637 }, { "epoch": 27.90119760479042, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1335, "step": 18638 }, { "epoch": 27.902694610778443, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.123, "step": 18639 }, { "epoch": 27.904191616766468, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1328, "step": 18640 }, { "epoch": 27.90568862275449, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1238, "step": 18641 }, { "epoch": 27.907185628742514, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1321, "step": 18642 }, { "epoch": 27.90868263473054, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1242, "step": 18643 }, { "epoch": 27.910179640718564, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1277, "step": 18644 }, { "epoch": 27.91167664670659, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.127, "step": 18645 }, { "epoch": 27.91317365269461, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1375, "step": 18646 }, { "epoch": 27.914670658682635, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1342, "step": 18647 }, { "epoch": 27.91616766467066, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1335, "step": 18648 }, { "epoch": 27.91766467065868, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1308, "step": 18649 }, { "epoch": 27.919161676646706, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1268, "step": 18650 }, { "epoch": 27.92065868263473, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1246, "step": 18651 }, { "epoch": 27.922155688622755, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1289, "step": 18652 }, { "epoch": 27.92365269461078, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1253, "step": 18653 }, { "epoch": 27.9251497005988, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1266, "step": 18654 }, { "epoch": 27.926646706586826, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1305, "step": 18655 }, { "epoch": 27.92814371257485, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1274, "step": 18656 }, { "epoch": 27.929640718562876, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1192, "step": 18657 }, { "epoch": 27.931137724550897, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1272, "step": 18658 }, { "epoch": 27.932634730538922, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1365, "step": 18659 }, { "epoch": 27.934131736526947, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1344, "step": 18660 }, { "epoch": 27.93562874251497, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1273, "step": 18661 }, { "epoch": 27.937125748502993, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1296, "step": 18662 }, { "epoch": 27.938622754491018, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1235, "step": 18663 }, { "epoch": 27.940119760479043, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1279, "step": 18664 }, { "epoch": 27.941616766467067, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.13, "step": 18665 }, { "epoch": 27.94311377245509, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1272, "step": 18666 }, { "epoch": 27.944610778443113, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1348, "step": 18667 }, { "epoch": 27.94610778443114, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1242, "step": 18668 }, { "epoch": 27.947604790419163, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1301, "step": 18669 }, { "epoch": 27.949101796407184, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.1246, "step": 18670 }, { "epoch": 27.95059880239521, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1242, "step": 18671 }, { "epoch": 27.952095808383234, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1219, "step": 18672 }, { "epoch": 27.95359281437126, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1329, "step": 18673 }, { "epoch": 27.95508982035928, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1349, "step": 18674 }, { "epoch": 27.956586826347305, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.133, "step": 18675 }, { "epoch": 27.95808383233533, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.133, "step": 18676 }, { "epoch": 27.959580838323355, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1347, "step": 18677 }, { "epoch": 27.961077844311376, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1276, "step": 18678 }, { "epoch": 27.9625748502994, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1265, "step": 18679 }, { "epoch": 27.964071856287426, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.1331, "step": 18680 }, { "epoch": 27.96556886227545, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1354, "step": 18681 }, { "epoch": 27.96706586826347, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1251, "step": 18682 }, { "epoch": 27.968562874251496, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1251, "step": 18683 }, { "epoch": 27.97005988023952, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1327, "step": 18684 }, { "epoch": 27.971556886227546, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1237, "step": 18685 }, { "epoch": 27.973053892215567, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1319, "step": 18686 }, { "epoch": 27.974550898203592, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.128, "step": 18687 }, { "epoch": 27.976047904191617, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1269, "step": 18688 }, { "epoch": 27.977544910179642, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.13, "step": 18689 }, { "epoch": 27.979041916167663, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1322, "step": 18690 }, { "epoch": 27.980538922155688, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1303, "step": 18691 }, { "epoch": 27.982035928143713, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1279, "step": 18692 }, { "epoch": 27.983532934131738, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1323, "step": 18693 }, { "epoch": 27.98502994011976, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1314, "step": 18694 }, { "epoch": 27.986526946107784, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.13, "step": 18695 }, { "epoch": 27.98802395209581, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1261, "step": 18696 }, { "epoch": 27.989520958083833, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1296, "step": 18697 }, { "epoch": 27.991017964071855, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1282, "step": 18698 }, { "epoch": 27.99251497005988, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.128, "step": 18699 }, { "epoch": 27.994011976047904, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1204, "step": 18700 }, { "epoch": 27.99550898203593, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1231, "step": 18701 }, { "epoch": 27.99700598802395, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1282, "step": 18702 }, { "epoch": 27.998502994011975, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1304, "step": 18703 }, { "epoch": 28.0, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1279, "step": 18704 }, { "epoch": 28.001497005988025, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1232, "step": 18705 }, { "epoch": 28.00299401197605, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1323, "step": 18706 }, { "epoch": 28.00449101796407, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1301, "step": 18707 }, { "epoch": 28.005988023952096, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1273, "step": 18708 }, { "epoch": 28.00748502994012, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1315, "step": 18709 }, { "epoch": 28.008982035928145, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1331, "step": 18710 }, { "epoch": 28.010479041916167, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1206, "step": 18711 }, { "epoch": 28.01197604790419, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1312, "step": 18712 }, { "epoch": 28.013473053892216, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1221, "step": 18713 }, { "epoch": 28.01497005988024, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.1282, "step": 18714 }, { "epoch": 28.016467065868262, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1267, "step": 18715 }, { "epoch": 28.017964071856287, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1291, "step": 18716 }, { "epoch": 28.019461077844312, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1277, "step": 18717 }, { "epoch": 28.020958083832337, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1318, "step": 18718 }, { "epoch": 28.022455089820358, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1326, "step": 18719 }, { "epoch": 28.023952095808383, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1267, "step": 18720 }, { "epoch": 28.025449101796408, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1298, "step": 18721 }, { "epoch": 28.026946107784433, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1218, "step": 18722 }, { "epoch": 28.028443113772454, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1254, "step": 18723 }, { "epoch": 28.02994011976048, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1258, "step": 18724 }, { "epoch": 28.031437125748504, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.127, "step": 18725 }, { "epoch": 28.03293413173653, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1316, "step": 18726 }, { "epoch": 28.03443113772455, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1261, "step": 18727 }, { "epoch": 28.035928143712574, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1338, "step": 18728 }, { "epoch": 28.0374251497006, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1299, "step": 18729 }, { "epoch": 28.038922155688624, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1316, "step": 18730 }, { "epoch": 28.040419161676645, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1199, "step": 18731 }, { "epoch": 28.04191616766467, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1233, "step": 18732 }, { "epoch": 28.043413173652695, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1278, "step": 18733 }, { "epoch": 28.04491017964072, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1375, "step": 18734 }, { "epoch": 28.04640718562874, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1338, "step": 18735 }, { "epoch": 28.047904191616766, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1233, "step": 18736 }, { "epoch": 28.04940119760479, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1293, "step": 18737 }, { "epoch": 28.050898203592816, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1309, "step": 18738 }, { "epoch": 28.052395209580837, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1214, "step": 18739 }, { "epoch": 28.05389221556886, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1274, "step": 18740 }, { "epoch": 28.055389221556887, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1342, "step": 18741 }, { "epoch": 28.05688622754491, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1247, "step": 18742 }, { "epoch": 28.058383233532933, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.126, "step": 18743 }, { "epoch": 28.059880239520957, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1217, "step": 18744 }, { "epoch": 28.061377245508982, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1302, "step": 18745 }, { "epoch": 28.062874251497007, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1286, "step": 18746 }, { "epoch": 28.06437125748503, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1309, "step": 18747 }, { "epoch": 28.065868263473053, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1242, "step": 18748 }, { "epoch": 28.067365269461078, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1275, "step": 18749 }, { "epoch": 28.068862275449103, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.128, "step": 18750 }, { "epoch": 28.070359281437124, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1219, "step": 18751 }, { "epoch": 28.07185628742515, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1332, "step": 18752 }, { "epoch": 28.073353293413174, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1344, "step": 18753 }, { "epoch": 28.0748502994012, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1254, "step": 18754 }, { "epoch": 28.07634730538922, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1287, "step": 18755 }, { "epoch": 28.077844311377245, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1272, "step": 18756 }, { "epoch": 28.07934131736527, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1241, "step": 18757 }, { "epoch": 28.080838323353294, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1296, "step": 18758 }, { "epoch": 28.082335329341316, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1252, "step": 18759 }, { "epoch": 28.08383233532934, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1247, "step": 18760 }, { "epoch": 28.085329341317365, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.131, "step": 18761 }, { "epoch": 28.08682634730539, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1295, "step": 18762 }, { "epoch": 28.088323353293415, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1261, "step": 18763 }, { "epoch": 28.089820359281436, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1226, "step": 18764 }, { "epoch": 28.09131736526946, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1247, "step": 18765 }, { "epoch": 28.092814371257486, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.129, "step": 18766 }, { "epoch": 28.09431137724551, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1256, "step": 18767 }, { "epoch": 28.095808383233532, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1253, "step": 18768 }, { "epoch": 28.097305389221557, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1267, "step": 18769 }, { "epoch": 28.09880239520958, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1276, "step": 18770 }, { "epoch": 28.100299401197606, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1325, "step": 18771 }, { "epoch": 28.101796407185628, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1315, "step": 18772 }, { "epoch": 28.103293413173652, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1252, "step": 18773 }, { "epoch": 28.104790419161677, "grad_norm": 0.0546875, "learning_rate": 0.0008, "loss": 1.1302, "step": 18774 }, { "epoch": 28.106287425149702, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1274, "step": 18775 }, { "epoch": 28.107784431137723, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1281, "step": 18776 }, { "epoch": 28.10928143712575, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1363, "step": 18777 }, { "epoch": 28.110778443113773, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1291, "step": 18778 }, { "epoch": 28.112275449101798, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1285, "step": 18779 }, { "epoch": 28.11377245508982, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.129, "step": 18780 }, { "epoch": 28.115269461077844, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1312, "step": 18781 }, { "epoch": 28.11676646706587, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1284, "step": 18782 }, { "epoch": 28.118263473053894, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1257, "step": 18783 }, { "epoch": 28.119760479041915, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1321, "step": 18784 }, { "epoch": 28.12125748502994, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1288, "step": 18785 }, { "epoch": 28.122754491017965, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1318, "step": 18786 }, { "epoch": 28.12425149700599, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1298, "step": 18787 }, { "epoch": 28.12574850299401, "grad_norm": 0.10986328125, "learning_rate": 0.0008, "loss": 1.1277, "step": 18788 }, { "epoch": 28.127245508982035, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.129, "step": 18789 }, { "epoch": 28.12874251497006, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1262, "step": 18790 }, { "epoch": 28.130239520958085, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1332, "step": 18791 }, { "epoch": 28.131736526946106, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1283, "step": 18792 }, { "epoch": 28.13323353293413, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1237, "step": 18793 }, { "epoch": 28.134730538922156, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1268, "step": 18794 }, { "epoch": 28.13622754491018, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1279, "step": 18795 }, { "epoch": 28.137724550898202, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1347, "step": 18796 }, { "epoch": 28.139221556886227, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1298, "step": 18797 }, { "epoch": 28.14071856287425, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1267, "step": 18798 }, { "epoch": 28.142215568862277, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.129, "step": 18799 }, { "epoch": 28.143712574850298, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1276, "step": 18800 }, { "epoch": 28.145209580838323, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.126, "step": 18801 }, { "epoch": 28.146706586826348, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1293, "step": 18802 }, { "epoch": 28.148203592814372, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1283, "step": 18803 }, { "epoch": 28.149700598802394, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1273, "step": 18804 }, { "epoch": 28.15119760479042, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1278, "step": 18805 }, { "epoch": 28.152694610778443, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1253, "step": 18806 }, { "epoch": 28.154191616766468, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1281, "step": 18807 }, { "epoch": 28.15568862275449, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1358, "step": 18808 }, { "epoch": 28.157185628742514, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1254, "step": 18809 }, { "epoch": 28.15868263473054, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1226, "step": 18810 }, { "epoch": 28.160179640718564, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1239, "step": 18811 }, { "epoch": 28.161676646706585, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1311, "step": 18812 }, { "epoch": 28.16317365269461, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.124, "step": 18813 }, { "epoch": 28.164670658682635, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1283, "step": 18814 }, { "epoch": 28.16616766467066, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1258, "step": 18815 }, { "epoch": 28.16766467065868, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1205, "step": 18816 }, { "epoch": 28.169161676646706, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1294, "step": 18817 }, { "epoch": 28.17065868263473, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1311, "step": 18818 }, { "epoch": 28.172155688622755, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1322, "step": 18819 }, { "epoch": 28.17365269461078, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1274, "step": 18820 }, { "epoch": 28.1751497005988, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1234, "step": 18821 }, { "epoch": 28.176646706586826, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1275, "step": 18822 }, { "epoch": 28.17814371257485, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1262, "step": 18823 }, { "epoch": 28.179640718562876, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1298, "step": 18824 }, { "epoch": 28.181137724550897, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1299, "step": 18825 }, { "epoch": 28.182634730538922, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.127, "step": 18826 }, { "epoch": 28.184131736526947, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1297, "step": 18827 }, { "epoch": 28.18562874251497, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.127, "step": 18828 }, { "epoch": 28.187125748502993, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1287, "step": 18829 }, { "epoch": 28.188622754491018, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1253, "step": 18830 }, { "epoch": 28.190119760479043, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.134, "step": 18831 }, { "epoch": 28.191616766467067, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1275, "step": 18832 }, { "epoch": 28.19311377245509, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1266, "step": 18833 }, { "epoch": 28.194610778443113, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.13, "step": 18834 }, { "epoch": 28.19610778443114, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1208, "step": 18835 }, { "epoch": 28.197604790419163, "grad_norm": 0.1416015625, "learning_rate": 0.0008, "loss": 1.1306, "step": 18836 }, { "epoch": 28.199101796407184, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1324, "step": 18837 }, { "epoch": 28.20059880239521, "grad_norm": 0.1259765625, "learning_rate": 0.0008, "loss": 1.1302, "step": 18838 }, { "epoch": 28.202095808383234, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1306, "step": 18839 }, { "epoch": 28.20359281437126, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1272, "step": 18840 }, { "epoch": 28.20508982035928, "grad_norm": 0.1689453125, "learning_rate": 0.0008, "loss": 1.1231, "step": 18841 }, { "epoch": 28.206586826347305, "grad_norm": 0.1962890625, "learning_rate": 0.0008, "loss": 1.1236, "step": 18842 }, { "epoch": 28.20808383233533, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1297, "step": 18843 }, { "epoch": 28.209580838323355, "grad_norm": 0.1708984375, "learning_rate": 0.0008, "loss": 1.1337, "step": 18844 }, { "epoch": 28.211077844311376, "grad_norm": 0.107421875, "learning_rate": 0.0008, "loss": 1.1306, "step": 18845 }, { "epoch": 28.2125748502994, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1297, "step": 18846 }, { "epoch": 28.214071856287426, "grad_norm": 0.12890625, "learning_rate": 0.0008, "loss": 1.1344, "step": 18847 }, { "epoch": 28.21556886227545, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1294, "step": 18848 }, { "epoch": 28.21706586826347, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.1266, "step": 18849 }, { "epoch": 28.218562874251496, "grad_norm": 0.1337890625, "learning_rate": 0.0008, "loss": 1.1286, "step": 18850 }, { "epoch": 28.22005988023952, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1338, "step": 18851 }, { "epoch": 28.221556886227546, "grad_norm": 0.11376953125, "learning_rate": 0.0008, "loss": 1.1319, "step": 18852 }, { "epoch": 28.223053892215567, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1276, "step": 18853 }, { "epoch": 28.224550898203592, "grad_norm": 0.10302734375, "learning_rate": 0.0008, "loss": 1.128, "step": 18854 }, { "epoch": 28.226047904191617, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1295, "step": 18855 }, { "epoch": 28.227544910179642, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1309, "step": 18856 }, { "epoch": 28.229041916167663, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1342, "step": 18857 }, { "epoch": 28.230538922155688, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1334, "step": 18858 }, { "epoch": 28.232035928143713, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1261, "step": 18859 }, { "epoch": 28.233532934131738, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1288, "step": 18860 }, { "epoch": 28.23502994011976, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1267, "step": 18861 }, { "epoch": 28.236526946107784, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1333, "step": 18862 }, { "epoch": 28.23802395209581, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1341, "step": 18863 }, { "epoch": 28.239520958083833, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1307, "step": 18864 }, { "epoch": 28.241017964071855, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.129, "step": 18865 }, { "epoch": 28.24251497005988, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1241, "step": 18866 }, { "epoch": 28.244011976047904, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1257, "step": 18867 }, { "epoch": 28.24550898203593, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1245, "step": 18868 }, { "epoch": 28.24700598802395, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1203, "step": 18869 }, { "epoch": 28.248502994011975, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1193, "step": 18870 }, { "epoch": 28.25, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1323, "step": 18871 }, { "epoch": 28.251497005988025, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1299, "step": 18872 }, { "epoch": 28.25299401197605, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1313, "step": 18873 }, { "epoch": 28.25449101796407, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.129, "step": 18874 }, { "epoch": 28.255988023952096, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1266, "step": 18875 }, { "epoch": 28.25748502994012, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1277, "step": 18876 }, { "epoch": 28.258982035928145, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1395, "step": 18877 }, { "epoch": 28.260479041916167, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1279, "step": 18878 }, { "epoch": 28.26197604790419, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1342, "step": 18879 }, { "epoch": 28.263473053892216, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1274, "step": 18880 }, { "epoch": 28.26497005988024, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1296, "step": 18881 }, { "epoch": 28.266467065868262, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1279, "step": 18882 }, { "epoch": 28.267964071856287, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1329, "step": 18883 }, { "epoch": 28.269461077844312, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1328, "step": 18884 }, { "epoch": 28.270958083832337, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1332, "step": 18885 }, { "epoch": 28.272455089820358, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1317, "step": 18886 }, { "epoch": 28.273952095808383, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1336, "step": 18887 }, { "epoch": 28.275449101796408, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.129, "step": 18888 }, { "epoch": 28.276946107784433, "grad_norm": 0.0498046875, "learning_rate": 0.0008, "loss": 1.1249, "step": 18889 }, { "epoch": 28.278443113772454, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1302, "step": 18890 }, { "epoch": 28.27994011976048, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1269, "step": 18891 }, { "epoch": 28.281437125748504, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1341, "step": 18892 }, { "epoch": 28.28293413173653, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1317, "step": 18893 }, { "epoch": 28.28443113772455, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1343, "step": 18894 }, { "epoch": 28.285928143712574, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1299, "step": 18895 }, { "epoch": 28.2874251497006, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1304, "step": 18896 }, { "epoch": 28.288922155688624, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.126, "step": 18897 }, { "epoch": 28.290419161676645, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1291, "step": 18898 }, { "epoch": 28.29191616766467, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1324, "step": 18899 }, { "epoch": 28.293413173652695, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1291, "step": 18900 }, { "epoch": 28.29491017964072, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1231, "step": 18901 }, { "epoch": 28.29640718562874, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1303, "step": 18902 }, { "epoch": 28.297904191616766, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1324, "step": 18903 }, { "epoch": 28.29940119760479, "grad_norm": 0.1015625, "learning_rate": 0.0008, "loss": 1.1242, "step": 18904 }, { "epoch": 28.300898203592816, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1317, "step": 18905 }, { "epoch": 28.302395209580837, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1261, "step": 18906 }, { "epoch": 28.30389221556886, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1316, "step": 18907 }, { "epoch": 28.305389221556887, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1229, "step": 18908 }, { "epoch": 28.30688622754491, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1271, "step": 18909 }, { "epoch": 28.308383233532933, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1209, "step": 18910 }, { "epoch": 28.309880239520957, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.123, "step": 18911 }, { "epoch": 28.311377245508982, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1282, "step": 18912 }, { "epoch": 28.312874251497007, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1269, "step": 18913 }, { "epoch": 28.31437125748503, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1277, "step": 18914 }, { "epoch": 28.315868263473053, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1241, "step": 18915 }, { "epoch": 28.317365269461078, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1247, "step": 18916 }, { "epoch": 28.318862275449103, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.126, "step": 18917 }, { "epoch": 28.320359281437124, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1232, "step": 18918 }, { "epoch": 28.32185628742515, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1311, "step": 18919 }, { "epoch": 28.323353293413174, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1275, "step": 18920 }, { "epoch": 28.3248502994012, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1267, "step": 18921 }, { "epoch": 28.32634730538922, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1341, "step": 18922 }, { "epoch": 28.327844311377245, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1329, "step": 18923 }, { "epoch": 28.32934131736527, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1331, "step": 18924 }, { "epoch": 28.330838323353294, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1241, "step": 18925 }, { "epoch": 28.33233532934132, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1299, "step": 18926 }, { "epoch": 28.33383233532934, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.125, "step": 18927 }, { "epoch": 28.335329341317365, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1263, "step": 18928 }, { "epoch": 28.33682634730539, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.127, "step": 18929 }, { "epoch": 28.338323353293415, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1255, "step": 18930 }, { "epoch": 28.339820359281436, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1279, "step": 18931 }, { "epoch": 28.34131736526946, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1312, "step": 18932 }, { "epoch": 28.342814371257486, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1342, "step": 18933 }, { "epoch": 28.34431137724551, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1254, "step": 18934 }, { "epoch": 28.345808383233532, "grad_norm": 0.054443359375, "learning_rate": 0.0008, "loss": 1.1294, "step": 18935 }, { "epoch": 28.347305389221557, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1344, "step": 18936 }, { "epoch": 28.34880239520958, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1282, "step": 18937 }, { "epoch": 28.350299401197606, "grad_norm": 0.1103515625, "learning_rate": 0.0008, "loss": 1.1302, "step": 18938 }, { "epoch": 28.351796407185628, "grad_norm": 0.11328125, "learning_rate": 0.0008, "loss": 1.1278, "step": 18939 }, { "epoch": 28.353293413173652, "grad_norm": 0.177734375, "learning_rate": 0.0008, "loss": 1.1286, "step": 18940 }, { "epoch": 28.354790419161677, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1297, "step": 18941 }, { "epoch": 28.356287425149702, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1357, "step": 18942 }, { "epoch": 28.357784431137723, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1417, "step": 18943 }, { "epoch": 28.35928143712575, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.134, "step": 18944 }, { "epoch": 28.360778443113773, "grad_norm": 0.10498046875, "learning_rate": 0.0008, "loss": 1.1246, "step": 18945 }, { "epoch": 28.362275449101798, "grad_norm": 0.1552734375, "learning_rate": 0.0008, "loss": 1.1218, "step": 18946 }, { "epoch": 28.36377245508982, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.127, "step": 18947 }, { "epoch": 28.365269461077844, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1272, "step": 18948 }, { "epoch": 28.36676646706587, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1306, "step": 18949 }, { "epoch": 28.368263473053894, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1282, "step": 18950 }, { "epoch": 28.369760479041915, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.13, "step": 18951 }, { "epoch": 28.37125748502994, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1264, "step": 18952 }, { "epoch": 28.372754491017965, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1218, "step": 18953 }, { "epoch": 28.37425149700599, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.136, "step": 18954 }, { "epoch": 28.37574850299401, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1313, "step": 18955 }, { "epoch": 28.377245508982035, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1272, "step": 18956 }, { "epoch": 28.37874251497006, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1316, "step": 18957 }, { "epoch": 28.380239520958085, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1266, "step": 18958 }, { "epoch": 28.381736526946106, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1274, "step": 18959 }, { "epoch": 28.38323353293413, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1272, "step": 18960 }, { "epoch": 28.384730538922156, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1381, "step": 18961 }, { "epoch": 28.38622754491018, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1235, "step": 18962 }, { "epoch": 28.387724550898202, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1255, "step": 18963 }, { "epoch": 28.389221556886227, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1319, "step": 18964 }, { "epoch": 28.39071856287425, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1316, "step": 18965 }, { "epoch": 28.392215568862277, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1299, "step": 18966 }, { "epoch": 28.393712574850298, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1252, "step": 18967 }, { "epoch": 28.395209580838323, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1249, "step": 18968 }, { "epoch": 28.396706586826348, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1311, "step": 18969 }, { "epoch": 28.398203592814372, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1344, "step": 18970 }, { "epoch": 28.399700598802394, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1287, "step": 18971 }, { "epoch": 28.40119760479042, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1228, "step": 18972 }, { "epoch": 28.402694610778443, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1268, "step": 18973 }, { "epoch": 28.404191616766468, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1253, "step": 18974 }, { "epoch": 28.40568862275449, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1242, "step": 18975 }, { "epoch": 28.407185628742514, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1315, "step": 18976 }, { "epoch": 28.40868263473054, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1262, "step": 18977 }, { "epoch": 28.410179640718564, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1278, "step": 18978 }, { "epoch": 28.411676646706585, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1249, "step": 18979 }, { "epoch": 28.41317365269461, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1234, "step": 18980 }, { "epoch": 28.414670658682635, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1297, "step": 18981 }, { "epoch": 28.41616766467066, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1339, "step": 18982 }, { "epoch": 28.41766467065868, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1338, "step": 18983 }, { "epoch": 28.419161676646706, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1291, "step": 18984 }, { "epoch": 28.42065868263473, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1356, "step": 18985 }, { "epoch": 28.422155688622755, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1292, "step": 18986 }, { "epoch": 28.42365269461078, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1241, "step": 18987 }, { "epoch": 28.4251497005988, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1265, "step": 18988 }, { "epoch": 28.426646706586826, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.122, "step": 18989 }, { "epoch": 28.42814371257485, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1293, "step": 18990 }, { "epoch": 28.429640718562876, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1325, "step": 18991 }, { "epoch": 28.431137724550897, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1242, "step": 18992 }, { "epoch": 28.432634730538922, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1293, "step": 18993 }, { "epoch": 28.434131736526947, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.126, "step": 18994 }, { "epoch": 28.43562874251497, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1293, "step": 18995 }, { "epoch": 28.437125748502993, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1227, "step": 18996 }, { "epoch": 28.438622754491018, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1259, "step": 18997 }, { "epoch": 28.440119760479043, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1282, "step": 18998 }, { "epoch": 28.441616766467067, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1255, "step": 18999 }, { "epoch": 28.44311377245509, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1347, "step": 19000 }, { "epoch": 28.444610778443113, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1328, "step": 19001 }, { "epoch": 28.44610778443114, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1285, "step": 19002 }, { "epoch": 28.447604790419163, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1282, "step": 19003 }, { "epoch": 28.449101796407184, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1298, "step": 19004 }, { "epoch": 28.45059880239521, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1318, "step": 19005 }, { "epoch": 28.452095808383234, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1289, "step": 19006 }, { "epoch": 28.45359281437126, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1247, "step": 19007 }, { "epoch": 28.45508982035928, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1253, "step": 19008 }, { "epoch": 28.456586826347305, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1341, "step": 19009 }, { "epoch": 28.45808383233533, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1342, "step": 19010 }, { "epoch": 28.459580838323355, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1289, "step": 19011 }, { "epoch": 28.461077844311376, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.126, "step": 19012 }, { "epoch": 28.4625748502994, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1277, "step": 19013 }, { "epoch": 28.464071856287426, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1224, "step": 19014 }, { "epoch": 28.46556886227545, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1324, "step": 19015 }, { "epoch": 28.46706586826347, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1289, "step": 19016 }, { "epoch": 28.468562874251496, "grad_norm": 0.08935546875, "learning_rate": 0.0008, "loss": 1.1279, "step": 19017 }, { "epoch": 28.47005988023952, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1257, "step": 19018 }, { "epoch": 28.471556886227546, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1242, "step": 19019 }, { "epoch": 28.473053892215567, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1311, "step": 19020 }, { "epoch": 28.474550898203592, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1352, "step": 19021 }, { "epoch": 28.476047904191617, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1259, "step": 19022 }, { "epoch": 28.477544910179642, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1304, "step": 19023 }, { "epoch": 28.479041916167663, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1273, "step": 19024 }, { "epoch": 28.480538922155688, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1282, "step": 19025 }, { "epoch": 28.482035928143713, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1284, "step": 19026 }, { "epoch": 28.483532934131738, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1148, "step": 19027 }, { "epoch": 28.48502994011976, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1329, "step": 19028 }, { "epoch": 28.486526946107784, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1257, "step": 19029 }, { "epoch": 28.48802395209581, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.127, "step": 19030 }, { "epoch": 28.489520958083833, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1315, "step": 19031 }, { "epoch": 28.491017964071855, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1311, "step": 19032 }, { "epoch": 28.49251497005988, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1312, "step": 19033 }, { "epoch": 28.494011976047904, "grad_norm": 0.10400390625, "learning_rate": 0.0008, "loss": 1.1233, "step": 19034 }, { "epoch": 28.49550898203593, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1336, "step": 19035 }, { "epoch": 28.49700598802395, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1381, "step": 19036 }, { "epoch": 28.498502994011975, "grad_norm": 0.0986328125, "learning_rate": 0.0008, "loss": 1.1256, "step": 19037 }, { "epoch": 28.5, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1285, "step": 19038 }, { "epoch": 28.501497005988025, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.1319, "step": 19039 }, { "epoch": 28.50299401197605, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1281, "step": 19040 }, { "epoch": 28.50449101796407, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1269, "step": 19041 }, { "epoch": 28.505988023952096, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1242, "step": 19042 }, { "epoch": 28.50748502994012, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1335, "step": 19043 }, { "epoch": 28.508982035928145, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1319, "step": 19044 }, { "epoch": 28.510479041916167, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1248, "step": 19045 }, { "epoch": 28.51197604790419, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1202, "step": 19046 }, { "epoch": 28.513473053892216, "grad_norm": 0.1064453125, "learning_rate": 0.0008, "loss": 1.1326, "step": 19047 }, { "epoch": 28.51497005988024, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1268, "step": 19048 }, { "epoch": 28.516467065868262, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1287, "step": 19049 }, { "epoch": 28.517964071856287, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1348, "step": 19050 }, { "epoch": 28.519461077844312, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1324, "step": 19051 }, { "epoch": 28.520958083832337, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1311, "step": 19052 }, { "epoch": 28.522455089820358, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1247, "step": 19053 }, { "epoch": 28.523952095808383, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1251, "step": 19054 }, { "epoch": 28.525449101796408, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1273, "step": 19055 }, { "epoch": 28.526946107784433, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1254, "step": 19056 }, { "epoch": 28.528443113772454, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1338, "step": 19057 }, { "epoch": 28.52994011976048, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.128, "step": 19058 }, { "epoch": 28.531437125748504, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1306, "step": 19059 }, { "epoch": 28.53293413173653, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1307, "step": 19060 }, { "epoch": 28.53443113772455, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1303, "step": 19061 }, { "epoch": 28.535928143712574, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1279, "step": 19062 }, { "epoch": 28.5374251497006, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1243, "step": 19063 }, { "epoch": 28.538922155688624, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1287, "step": 19064 }, { "epoch": 28.540419161676645, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.128, "step": 19065 }, { "epoch": 28.54191616766467, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1275, "step": 19066 }, { "epoch": 28.543413173652695, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.1324, "step": 19067 }, { "epoch": 28.54491017964072, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1299, "step": 19068 }, { "epoch": 28.54640718562874, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1315, "step": 19069 }, { "epoch": 28.547904191616766, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1348, "step": 19070 }, { "epoch": 28.54940119760479, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1272, "step": 19071 }, { "epoch": 28.550898203592816, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1322, "step": 19072 }, { "epoch": 28.552395209580837, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1267, "step": 19073 }, { "epoch": 28.55389221556886, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1306, "step": 19074 }, { "epoch": 28.555389221556887, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1312, "step": 19075 }, { "epoch": 28.55688622754491, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1273, "step": 19076 }, { "epoch": 28.558383233532933, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1258, "step": 19077 }, { "epoch": 28.559880239520957, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1293, "step": 19078 }, { "epoch": 28.561377245508982, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1274, "step": 19079 }, { "epoch": 28.562874251497007, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1266, "step": 19080 }, { "epoch": 28.56437125748503, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1271, "step": 19081 }, { "epoch": 28.565868263473053, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1241, "step": 19082 }, { "epoch": 28.567365269461078, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1319, "step": 19083 }, { "epoch": 28.568862275449103, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1265, "step": 19084 }, { "epoch": 28.570359281437124, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1299, "step": 19085 }, { "epoch": 28.57185628742515, "grad_norm": 0.0966796875, "learning_rate": 0.0008, "loss": 1.1309, "step": 19086 }, { "epoch": 28.573353293413174, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1226, "step": 19087 }, { "epoch": 28.5748502994012, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1351, "step": 19088 }, { "epoch": 28.57634730538922, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1292, "step": 19089 }, { "epoch": 28.577844311377245, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1283, "step": 19090 }, { "epoch": 28.57934131736527, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.13, "step": 19091 }, { "epoch": 28.580838323353294, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1181, "step": 19092 }, { "epoch": 28.58233532934132, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1325, "step": 19093 }, { "epoch": 28.58383233532934, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1319, "step": 19094 }, { "epoch": 28.585329341317365, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1324, "step": 19095 }, { "epoch": 28.58682634730539, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1259, "step": 19096 }, { "epoch": 28.58832335329341, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1261, "step": 19097 }, { "epoch": 28.589820359281436, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.1246, "step": 19098 }, { "epoch": 28.59131736526946, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1325, "step": 19099 }, { "epoch": 28.592814371257486, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.129, "step": 19100 }, { "epoch": 28.59431137724551, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.133, "step": 19101 }, { "epoch": 28.595808383233532, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1265, "step": 19102 }, { "epoch": 28.597305389221557, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1261, "step": 19103 }, { "epoch": 28.59880239520958, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1318, "step": 19104 }, { "epoch": 28.600299401197606, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.132, "step": 19105 }, { "epoch": 28.601796407185628, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1326, "step": 19106 }, { "epoch": 28.603293413173652, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1291, "step": 19107 }, { "epoch": 28.604790419161677, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1245, "step": 19108 }, { "epoch": 28.606287425149702, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1215, "step": 19109 }, { "epoch": 28.607784431137723, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1306, "step": 19110 }, { "epoch": 28.60928143712575, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1274, "step": 19111 }, { "epoch": 28.610778443113773, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1276, "step": 19112 }, { "epoch": 28.612275449101798, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1251, "step": 19113 }, { "epoch": 28.61377245508982, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1298, "step": 19114 }, { "epoch": 28.615269461077844, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1325, "step": 19115 }, { "epoch": 28.61676646706587, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1262, "step": 19116 }, { "epoch": 28.618263473053894, "grad_norm": 0.10595703125, "learning_rate": 0.0008, "loss": 1.1294, "step": 19117 }, { "epoch": 28.619760479041915, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1265, "step": 19118 }, { "epoch": 28.62125748502994, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1231, "step": 19119 }, { "epoch": 28.622754491017965, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1271, "step": 19120 }, { "epoch": 28.62425149700599, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1284, "step": 19121 }, { "epoch": 28.62574850299401, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1325, "step": 19122 }, { "epoch": 28.627245508982035, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.134, "step": 19123 }, { "epoch": 28.62874251497006, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1235, "step": 19124 }, { "epoch": 28.630239520958085, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1362, "step": 19125 }, { "epoch": 28.631736526946106, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1268, "step": 19126 }, { "epoch": 28.63323353293413, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1332, "step": 19127 }, { "epoch": 28.634730538922156, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1262, "step": 19128 }, { "epoch": 28.63622754491018, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1274, "step": 19129 }, { "epoch": 28.637724550898202, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.13, "step": 19130 }, { "epoch": 28.639221556886227, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1291, "step": 19131 }, { "epoch": 28.64071856287425, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1234, "step": 19132 }, { "epoch": 28.642215568862277, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1251, "step": 19133 }, { "epoch": 28.643712574850298, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1302, "step": 19134 }, { "epoch": 28.645209580838323, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1266, "step": 19135 }, { "epoch": 28.646706586826348, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1287, "step": 19136 }, { "epoch": 28.648203592814372, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1343, "step": 19137 }, { "epoch": 28.649700598802394, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1244, "step": 19138 }, { "epoch": 28.65119760479042, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1208, "step": 19139 }, { "epoch": 28.652694610778443, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1259, "step": 19140 }, { "epoch": 28.654191616766468, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1247, "step": 19141 }, { "epoch": 28.65568862275449, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.128, "step": 19142 }, { "epoch": 28.657185628742514, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1278, "step": 19143 }, { "epoch": 28.65868263473054, "grad_norm": 0.1025390625, "learning_rate": 0.0008, "loss": 1.1319, "step": 19144 }, { "epoch": 28.660179640718564, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1346, "step": 19145 }, { "epoch": 28.66167664670659, "grad_norm": 0.10107421875, "learning_rate": 0.0008, "loss": 1.1307, "step": 19146 }, { "epoch": 28.66317365269461, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.126, "step": 19147 }, { "epoch": 28.664670658682635, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1254, "step": 19148 }, { "epoch": 28.66616766467066, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1296, "step": 19149 }, { "epoch": 28.66766467065868, "grad_norm": 0.095703125, "learning_rate": 0.0008, "loss": 1.129, "step": 19150 }, { "epoch": 28.669161676646706, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1205, "step": 19151 }, { "epoch": 28.67065868263473, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1235, "step": 19152 }, { "epoch": 28.672155688622755, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1245, "step": 19153 }, { "epoch": 28.67365269461078, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1248, "step": 19154 }, { "epoch": 28.6751497005988, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1242, "step": 19155 }, { "epoch": 28.676646706586826, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1283, "step": 19156 }, { "epoch": 28.67814371257485, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1295, "step": 19157 }, { "epoch": 28.679640718562876, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1248, "step": 19158 }, { "epoch": 28.681137724550897, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1333, "step": 19159 }, { "epoch": 28.682634730538922, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1288, "step": 19160 }, { "epoch": 28.684131736526947, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1317, "step": 19161 }, { "epoch": 28.68562874251497, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1318, "step": 19162 }, { "epoch": 28.687125748502993, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1252, "step": 19163 }, { "epoch": 28.688622754491018, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1235, "step": 19164 }, { "epoch": 28.690119760479043, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1276, "step": 19165 }, { "epoch": 28.691616766467067, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1286, "step": 19166 }, { "epoch": 28.69311377245509, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1293, "step": 19167 }, { "epoch": 28.694610778443113, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1285, "step": 19168 }, { "epoch": 28.69610778443114, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1291, "step": 19169 }, { "epoch": 28.697604790419163, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1376, "step": 19170 }, { "epoch": 28.699101796407184, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1295, "step": 19171 }, { "epoch": 28.70059880239521, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1332, "step": 19172 }, { "epoch": 28.702095808383234, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1171, "step": 19173 }, { "epoch": 28.70359281437126, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1259, "step": 19174 }, { "epoch": 28.70508982035928, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1232, "step": 19175 }, { "epoch": 28.706586826347305, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1276, "step": 19176 }, { "epoch": 28.70808383233533, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.128, "step": 19177 }, { "epoch": 28.709580838323355, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1249, "step": 19178 }, { "epoch": 28.711077844311376, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1331, "step": 19179 }, { "epoch": 28.7125748502994, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1218, "step": 19180 }, { "epoch": 28.714071856287426, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1289, "step": 19181 }, { "epoch": 28.71556886227545, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1285, "step": 19182 }, { "epoch": 28.71706586826347, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1339, "step": 19183 }, { "epoch": 28.718562874251496, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1243, "step": 19184 }, { "epoch": 28.72005988023952, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1258, "step": 19185 }, { "epoch": 28.721556886227546, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1217, "step": 19186 }, { "epoch": 28.723053892215567, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1315, "step": 19187 }, { "epoch": 28.724550898203592, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1229, "step": 19188 }, { "epoch": 28.726047904191617, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1295, "step": 19189 }, { "epoch": 28.727544910179642, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1219, "step": 19190 }, { "epoch": 28.729041916167663, "grad_norm": 0.047607421875, "learning_rate": 0.0008, "loss": 1.1299, "step": 19191 }, { "epoch": 28.730538922155688, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1299, "step": 19192 }, { "epoch": 28.732035928143713, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1309, "step": 19193 }, { "epoch": 28.733532934131738, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.131, "step": 19194 }, { "epoch": 28.73502994011976, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1289, "step": 19195 }, { "epoch": 28.736526946107784, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1352, "step": 19196 }, { "epoch": 28.73802395209581, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1278, "step": 19197 }, { "epoch": 28.739520958083833, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1247, "step": 19198 }, { "epoch": 28.741017964071855, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1344, "step": 19199 }, { "epoch": 28.74251497005988, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1279, "step": 19200 }, { "epoch": 28.744011976047904, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1308, "step": 19201 }, { "epoch": 28.74550898203593, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1318, "step": 19202 }, { "epoch": 28.74700598802395, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1258, "step": 19203 }, { "epoch": 28.748502994011975, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1308, "step": 19204 }, { "epoch": 28.75, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1375, "step": 19205 }, { "epoch": 28.751497005988025, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1289, "step": 19206 }, { "epoch": 28.75299401197605, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1303, "step": 19207 }, { "epoch": 28.75449101796407, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.127, "step": 19208 }, { "epoch": 28.755988023952096, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1186, "step": 19209 }, { "epoch": 28.75748502994012, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1242, "step": 19210 }, { "epoch": 28.758982035928145, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1306, "step": 19211 }, { "epoch": 28.760479041916167, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1362, "step": 19212 }, { "epoch": 28.76197604790419, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1288, "step": 19213 }, { "epoch": 28.763473053892216, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1269, "step": 19214 }, { "epoch": 28.76497005988024, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1367, "step": 19215 }, { "epoch": 28.766467065868262, "grad_norm": 0.057373046875, "learning_rate": 0.0008, "loss": 1.13, "step": 19216 }, { "epoch": 28.767964071856287, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1323, "step": 19217 }, { "epoch": 28.769461077844312, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1276, "step": 19218 }, { "epoch": 28.770958083832337, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1285, "step": 19219 }, { "epoch": 28.772455089820358, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1222, "step": 19220 }, { "epoch": 28.773952095808383, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1301, "step": 19221 }, { "epoch": 28.775449101796408, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.126, "step": 19222 }, { "epoch": 28.776946107784433, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1276, "step": 19223 }, { "epoch": 28.778443113772454, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1343, "step": 19224 }, { "epoch": 28.77994011976048, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.131, "step": 19225 }, { "epoch": 28.781437125748504, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1308, "step": 19226 }, { "epoch": 28.78293413173653, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1349, "step": 19227 }, { "epoch": 28.78443113772455, "grad_norm": 0.09912109375, "learning_rate": 0.0008, "loss": 1.1314, "step": 19228 }, { "epoch": 28.785928143712574, "grad_norm": 0.1142578125, "learning_rate": 0.0008, "loss": 1.127, "step": 19229 }, { "epoch": 28.7874251497006, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1277, "step": 19230 }, { "epoch": 28.788922155688624, "grad_norm": 0.1669921875, "learning_rate": 0.0008, "loss": 1.1313, "step": 19231 }, { "epoch": 28.790419161676645, "grad_norm": 0.1279296875, "learning_rate": 0.0008, "loss": 1.128, "step": 19232 }, { "epoch": 28.79191616766467, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1332, "step": 19233 }, { "epoch": 28.793413173652695, "grad_norm": 0.11474609375, "learning_rate": 0.0008, "loss": 1.1284, "step": 19234 }, { "epoch": 28.79491017964072, "grad_norm": 0.1083984375, "learning_rate": 0.0008, "loss": 1.1241, "step": 19235 }, { "epoch": 28.79640718562874, "grad_norm": 0.1455078125, "learning_rate": 0.0008, "loss": 1.1303, "step": 19236 }, { "epoch": 28.797904191616766, "grad_norm": 0.09765625, "learning_rate": 0.0008, "loss": 1.1305, "step": 19237 }, { "epoch": 28.79940119760479, "grad_norm": 0.1171875, "learning_rate": 0.0008, "loss": 1.1266, "step": 19238 }, { "epoch": 28.800898203592816, "grad_norm": 0.10693359375, "learning_rate": 0.0008, "loss": 1.1246, "step": 19239 }, { "epoch": 28.802395209580837, "grad_norm": 0.10791015625, "learning_rate": 0.0008, "loss": 1.134, "step": 19240 }, { "epoch": 28.80389221556886, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.1313, "step": 19241 }, { "epoch": 28.805389221556887, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1306, "step": 19242 }, { "epoch": 28.80688622754491, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.128, "step": 19243 }, { "epoch": 28.808383233532933, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1303, "step": 19244 }, { "epoch": 28.809880239520957, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1265, "step": 19245 }, { "epoch": 28.811377245508982, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.124, "step": 19246 }, { "epoch": 28.812874251497007, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1285, "step": 19247 }, { "epoch": 28.81437125748503, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1383, "step": 19248 }, { "epoch": 28.815868263473053, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1251, "step": 19249 }, { "epoch": 28.817365269461078, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1222, "step": 19250 }, { "epoch": 28.818862275449103, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1258, "step": 19251 }, { "epoch": 28.820359281437124, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1359, "step": 19252 }, { "epoch": 28.82185628742515, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1335, "step": 19253 }, { "epoch": 28.823353293413174, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1321, "step": 19254 }, { "epoch": 28.8248502994012, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1297, "step": 19255 }, { "epoch": 28.82634730538922, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1293, "step": 19256 }, { "epoch": 28.827844311377245, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1319, "step": 19257 }, { "epoch": 28.82934131736527, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1298, "step": 19258 }, { "epoch": 28.830838323353294, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1267, "step": 19259 }, { "epoch": 28.83233532934132, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1302, "step": 19260 }, { "epoch": 28.83383233532934, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1211, "step": 19261 }, { "epoch": 28.835329341317365, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1238, "step": 19262 }, { "epoch": 28.83682634730539, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1257, "step": 19263 }, { "epoch": 28.83832335329341, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1337, "step": 19264 }, { "epoch": 28.839820359281436, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1293, "step": 19265 }, { "epoch": 28.84131736526946, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1264, "step": 19266 }, { "epoch": 28.842814371257486, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1252, "step": 19267 }, { "epoch": 28.84431137724551, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1379, "step": 19268 }, { "epoch": 28.845808383233532, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1244, "step": 19269 }, { "epoch": 28.847305389221557, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1312, "step": 19270 }, { "epoch": 28.84880239520958, "grad_norm": 0.052734375, "learning_rate": 0.0008, "loss": 1.1221, "step": 19271 }, { "epoch": 28.850299401197606, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1219, "step": 19272 }, { "epoch": 28.851796407185628, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1317, "step": 19273 }, { "epoch": 28.853293413173652, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1281, "step": 19274 }, { "epoch": 28.854790419161677, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1297, "step": 19275 }, { "epoch": 28.856287425149702, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1257, "step": 19276 }, { "epoch": 28.857784431137723, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1279, "step": 19277 }, { "epoch": 28.85928143712575, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1299, "step": 19278 }, { "epoch": 28.860778443113773, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.131, "step": 19279 }, { "epoch": 28.862275449101798, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1292, "step": 19280 }, { "epoch": 28.86377245508982, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1269, "step": 19281 }, { "epoch": 28.865269461077844, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1248, "step": 19282 }, { "epoch": 28.86676646706587, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1263, "step": 19283 }, { "epoch": 28.868263473053894, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1271, "step": 19284 }, { "epoch": 28.869760479041915, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.131, "step": 19285 }, { "epoch": 28.87125748502994, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.13, "step": 19286 }, { "epoch": 28.872754491017965, "grad_norm": 0.053955078125, "learning_rate": 0.0008, "loss": 1.1333, "step": 19287 }, { "epoch": 28.87425149700599, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1314, "step": 19288 }, { "epoch": 28.87574850299401, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1293, "step": 19289 }, { "epoch": 28.877245508982035, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1304, "step": 19290 }, { "epoch": 28.87874251497006, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.131, "step": 19291 }, { "epoch": 28.880239520958085, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1314, "step": 19292 }, { "epoch": 28.881736526946106, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1241, "step": 19293 }, { "epoch": 28.88323353293413, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1274, "step": 19294 }, { "epoch": 28.884730538922156, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1242, "step": 19295 }, { "epoch": 28.88622754491018, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1304, "step": 19296 }, { "epoch": 28.887724550898202, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1337, "step": 19297 }, { "epoch": 28.889221556886227, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1302, "step": 19298 }, { "epoch": 28.89071856287425, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1257, "step": 19299 }, { "epoch": 28.892215568862277, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1249, "step": 19300 }, { "epoch": 28.893712574850298, "grad_norm": 0.109375, "learning_rate": 0.0008, "loss": 1.1315, "step": 19301 }, { "epoch": 28.895209580838323, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1274, "step": 19302 }, { "epoch": 28.896706586826348, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1308, "step": 19303 }, { "epoch": 28.898203592814372, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1321, "step": 19304 }, { "epoch": 28.899700598802394, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.13, "step": 19305 }, { "epoch": 28.90119760479042, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1303, "step": 19306 }, { "epoch": 28.902694610778443, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1186, "step": 19307 }, { "epoch": 28.904191616766468, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1241, "step": 19308 }, { "epoch": 28.90568862275449, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1225, "step": 19309 }, { "epoch": 28.907185628742514, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.131, "step": 19310 }, { "epoch": 28.90868263473054, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1233, "step": 19311 }, { "epoch": 28.910179640718564, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1302, "step": 19312 }, { "epoch": 28.91167664670659, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.127, "step": 19313 }, { "epoch": 28.91317365269461, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1274, "step": 19314 }, { "epoch": 28.914670658682635, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1256, "step": 19315 }, { "epoch": 28.91616766467066, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1216, "step": 19316 }, { "epoch": 28.91766467065868, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.125, "step": 19317 }, { "epoch": 28.919161676646706, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.13, "step": 19318 }, { "epoch": 28.92065868263473, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1335, "step": 19319 }, { "epoch": 28.922155688622755, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1356, "step": 19320 }, { "epoch": 28.92365269461078, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1313, "step": 19321 }, { "epoch": 28.9251497005988, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1309, "step": 19322 }, { "epoch": 28.926646706586826, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1225, "step": 19323 }, { "epoch": 28.92814371257485, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.131, "step": 19324 }, { "epoch": 28.929640718562876, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1279, "step": 19325 }, { "epoch": 28.931137724550897, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1265, "step": 19326 }, { "epoch": 28.932634730538922, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1291, "step": 19327 }, { "epoch": 28.934131736526947, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1318, "step": 19328 }, { "epoch": 28.93562874251497, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1298, "step": 19329 }, { "epoch": 28.937125748502993, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1314, "step": 19330 }, { "epoch": 28.938622754491018, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1305, "step": 19331 }, { "epoch": 28.940119760479043, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1301, "step": 19332 }, { "epoch": 28.941616766467067, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1321, "step": 19333 }, { "epoch": 28.94311377245509, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1269, "step": 19334 }, { "epoch": 28.944610778443113, "grad_norm": 0.08984375, "learning_rate": 0.0008, "loss": 1.1212, "step": 19335 }, { "epoch": 28.94610778443114, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1203, "step": 19336 }, { "epoch": 28.947604790419163, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1339, "step": 19337 }, { "epoch": 28.949101796407184, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1342, "step": 19338 }, { "epoch": 28.95059880239521, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1271, "step": 19339 }, { "epoch": 28.952095808383234, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1269, "step": 19340 }, { "epoch": 28.95359281437126, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1294, "step": 19341 }, { "epoch": 28.95508982035928, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1386, "step": 19342 }, { "epoch": 28.956586826347305, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1293, "step": 19343 }, { "epoch": 28.95808383233533, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.129, "step": 19344 }, { "epoch": 28.959580838323355, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1285, "step": 19345 }, { "epoch": 28.961077844311376, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1272, "step": 19346 }, { "epoch": 28.9625748502994, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1254, "step": 19347 }, { "epoch": 28.964071856287426, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1252, "step": 19348 }, { "epoch": 28.96556886227545, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1295, "step": 19349 }, { "epoch": 28.96706586826347, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1304, "step": 19350 }, { "epoch": 28.968562874251496, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1277, "step": 19351 }, { "epoch": 28.97005988023952, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1236, "step": 19352 }, { "epoch": 28.971556886227546, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1293, "step": 19353 }, { "epoch": 28.973053892215567, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.128, "step": 19354 }, { "epoch": 28.974550898203592, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1285, "step": 19355 }, { "epoch": 28.976047904191617, "grad_norm": 0.09033203125, "learning_rate": 0.0008, "loss": 1.1301, "step": 19356 }, { "epoch": 28.977544910179642, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.127, "step": 19357 }, { "epoch": 28.979041916167663, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1313, "step": 19358 }, { "epoch": 28.980538922155688, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1256, "step": 19359 }, { "epoch": 28.982035928143713, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.135, "step": 19360 }, { "epoch": 28.983532934131738, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1277, "step": 19361 }, { "epoch": 28.98502994011976, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1316, "step": 19362 }, { "epoch": 28.986526946107784, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1234, "step": 19363 }, { "epoch": 28.98802395209581, "grad_norm": 0.10009765625, "learning_rate": 0.0008, "loss": 1.1299, "step": 19364 }, { "epoch": 28.989520958083833, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1323, "step": 19365 }, { "epoch": 28.991017964071855, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.129, "step": 19366 }, { "epoch": 28.99251497005988, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1285, "step": 19367 }, { "epoch": 28.994011976047904, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1327, "step": 19368 }, { "epoch": 28.99550898203593, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1288, "step": 19369 }, { "epoch": 28.99700598802395, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1301, "step": 19370 }, { "epoch": 28.998502994011975, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1278, "step": 19371 }, { "epoch": 29.0, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1191, "step": 19372 }, { "epoch": 29.001497005988025, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1194, "step": 19373 }, { "epoch": 29.00299401197605, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1317, "step": 19374 }, { "epoch": 29.00449101796407, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1301, "step": 19375 }, { "epoch": 29.005988023952096, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1318, "step": 19376 }, { "epoch": 29.00748502994012, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1247, "step": 19377 }, { "epoch": 29.008982035928145, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1278, "step": 19378 }, { "epoch": 29.010479041916167, "grad_norm": 0.11865234375, "learning_rate": 0.0008, "loss": 1.129, "step": 19379 }, { "epoch": 29.01197604790419, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1357, "step": 19380 }, { "epoch": 29.013473053892216, "grad_norm": 0.1123046875, "learning_rate": 0.0008, "loss": 1.1163, "step": 19381 }, { "epoch": 29.01497005988024, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.122, "step": 19382 }, { "epoch": 29.016467065868262, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1323, "step": 19383 }, { "epoch": 29.017964071856287, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1263, "step": 19384 }, { "epoch": 29.019461077844312, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1298, "step": 19385 }, { "epoch": 29.020958083832337, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1254, "step": 19386 }, { "epoch": 29.022455089820358, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1216, "step": 19387 }, { "epoch": 29.023952095808383, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1295, "step": 19388 }, { "epoch": 29.025449101796408, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1231, "step": 19389 }, { "epoch": 29.026946107784433, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1289, "step": 19390 }, { "epoch": 29.028443113772454, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1263, "step": 19391 }, { "epoch": 29.02994011976048, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1252, "step": 19392 }, { "epoch": 29.031437125748504, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1288, "step": 19393 }, { "epoch": 29.03293413173653, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1246, "step": 19394 }, { "epoch": 29.03443113772455, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1271, "step": 19395 }, { "epoch": 29.035928143712574, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1314, "step": 19396 }, { "epoch": 29.0374251497006, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.131, "step": 19397 }, { "epoch": 29.038922155688624, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.1229, "step": 19398 }, { "epoch": 29.040419161676645, "grad_norm": 0.11279296875, "learning_rate": 0.0008, "loss": 1.13, "step": 19399 }, { "epoch": 29.04191616766467, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1244, "step": 19400 }, { "epoch": 29.043413173652695, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1265, "step": 19401 }, { "epoch": 29.04491017964072, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1302, "step": 19402 }, { "epoch": 29.04640718562874, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.124, "step": 19403 }, { "epoch": 29.047904191616766, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1319, "step": 19404 }, { "epoch": 29.04940119760479, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1324, "step": 19405 }, { "epoch": 29.050898203592816, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1269, "step": 19406 }, { "epoch": 29.052395209580837, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1304, "step": 19407 }, { "epoch": 29.05389221556886, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1295, "step": 19408 }, { "epoch": 29.055389221556887, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.132, "step": 19409 }, { "epoch": 29.05688622754491, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1311, "step": 19410 }, { "epoch": 29.058383233532933, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1281, "step": 19411 }, { "epoch": 29.059880239520957, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1237, "step": 19412 }, { "epoch": 29.061377245508982, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1314, "step": 19413 }, { "epoch": 29.062874251497007, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1288, "step": 19414 }, { "epoch": 29.06437125748503, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1308, "step": 19415 }, { "epoch": 29.065868263473053, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1355, "step": 19416 }, { "epoch": 29.067365269461078, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1298, "step": 19417 }, { "epoch": 29.068862275449103, "grad_norm": 0.052490234375, "learning_rate": 0.0008, "loss": 1.1229, "step": 19418 }, { "epoch": 29.070359281437124, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.126, "step": 19419 }, { "epoch": 29.07185628742515, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.129, "step": 19420 }, { "epoch": 29.073353293413174, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1309, "step": 19421 }, { "epoch": 29.0748502994012, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1333, "step": 19422 }, { "epoch": 29.07634730538922, "grad_norm": 0.056396484375, "learning_rate": 0.0008, "loss": 1.1303, "step": 19423 }, { "epoch": 29.077844311377245, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1241, "step": 19424 }, { "epoch": 29.07934131736527, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.126, "step": 19425 }, { "epoch": 29.080838323353294, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1264, "step": 19426 }, { "epoch": 29.082335329341316, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1241, "step": 19427 }, { "epoch": 29.08383233532934, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1221, "step": 19428 }, { "epoch": 29.085329341317365, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1244, "step": 19429 }, { "epoch": 29.08682634730539, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1319, "step": 19430 }, { "epoch": 29.088323353293415, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.122, "step": 19431 }, { "epoch": 29.089820359281436, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1322, "step": 19432 }, { "epoch": 29.09131736526946, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1263, "step": 19433 }, { "epoch": 29.092814371257486, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1258, "step": 19434 }, { "epoch": 29.09431137724551, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1268, "step": 19435 }, { "epoch": 29.095808383233532, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1306, "step": 19436 }, { "epoch": 29.097305389221557, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1283, "step": 19437 }, { "epoch": 29.09880239520958, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.129, "step": 19438 }, { "epoch": 29.100299401197606, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1325, "step": 19439 }, { "epoch": 29.101796407185628, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1284, "step": 19440 }, { "epoch": 29.103293413173652, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1303, "step": 19441 }, { "epoch": 29.104790419161677, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1278, "step": 19442 }, { "epoch": 29.106287425149702, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1325, "step": 19443 }, { "epoch": 29.107784431137723, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.127, "step": 19444 }, { "epoch": 29.10928143712575, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1306, "step": 19445 }, { "epoch": 29.110778443113773, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1172, "step": 19446 }, { "epoch": 29.112275449101798, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1305, "step": 19447 }, { "epoch": 29.11377245508982, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1273, "step": 19448 }, { "epoch": 29.115269461077844, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1376, "step": 19449 }, { "epoch": 29.11676646706587, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1315, "step": 19450 }, { "epoch": 29.118263473053894, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1263, "step": 19451 }, { "epoch": 29.119760479041915, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1259, "step": 19452 }, { "epoch": 29.12125748502994, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1243, "step": 19453 }, { "epoch": 29.122754491017965, "grad_norm": 0.09716796875, "learning_rate": 0.0008, "loss": 1.1323, "step": 19454 }, { "epoch": 29.12425149700599, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1329, "step": 19455 }, { "epoch": 29.12574850299401, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1338, "step": 19456 }, { "epoch": 29.127245508982035, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1307, "step": 19457 }, { "epoch": 29.12874251497006, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1248, "step": 19458 }, { "epoch": 29.130239520958085, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1294, "step": 19459 }, { "epoch": 29.131736526946106, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1315, "step": 19460 }, { "epoch": 29.13323353293413, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1283, "step": 19461 }, { "epoch": 29.134730538922156, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1266, "step": 19462 }, { "epoch": 29.13622754491018, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1319, "step": 19463 }, { "epoch": 29.137724550898202, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1264, "step": 19464 }, { "epoch": 29.139221556886227, "grad_norm": 0.061767578125, "learning_rate": 0.0008, "loss": 1.1205, "step": 19465 }, { "epoch": 29.14071856287425, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1314, "step": 19466 }, { "epoch": 29.142215568862277, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1294, "step": 19467 }, { "epoch": 29.143712574850298, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1251, "step": 19468 }, { "epoch": 29.145209580838323, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1302, "step": 19469 }, { "epoch": 29.146706586826348, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1338, "step": 19470 }, { "epoch": 29.148203592814372, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1271, "step": 19471 }, { "epoch": 29.149700598802394, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1311, "step": 19472 }, { "epoch": 29.15119760479042, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1246, "step": 19473 }, { "epoch": 29.152694610778443, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.1253, "step": 19474 }, { "epoch": 29.154191616766468, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.131, "step": 19475 }, { "epoch": 29.15568862275449, "grad_norm": 0.1044921875, "learning_rate": 0.0008, "loss": 1.1279, "step": 19476 }, { "epoch": 29.157185628742514, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1298, "step": 19477 }, { "epoch": 29.15868263473054, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1256, "step": 19478 }, { "epoch": 29.160179640718564, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1264, "step": 19479 }, { "epoch": 29.161676646706585, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1289, "step": 19480 }, { "epoch": 29.16317365269461, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1241, "step": 19481 }, { "epoch": 29.164670658682635, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1282, "step": 19482 }, { "epoch": 29.16616766467066, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1293, "step": 19483 }, { "epoch": 29.16766467065868, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1299, "step": 19484 }, { "epoch": 29.169161676646706, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1269, "step": 19485 }, { "epoch": 29.17065868263473, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1265, "step": 19486 }, { "epoch": 29.172155688622755, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1298, "step": 19487 }, { "epoch": 29.17365269461078, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1375, "step": 19488 }, { "epoch": 29.1751497005988, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.126, "step": 19489 }, { "epoch": 29.176646706586826, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1331, "step": 19490 }, { "epoch": 29.17814371257485, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1333, "step": 19491 }, { "epoch": 29.179640718562876, "grad_norm": 0.047119140625, "learning_rate": 0.0008, "loss": 1.1303, "step": 19492 }, { "epoch": 29.181137724550897, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1342, "step": 19493 }, { "epoch": 29.182634730538922, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1253, "step": 19494 }, { "epoch": 29.184131736526947, "grad_norm": 0.05126953125, "learning_rate": 0.0008, "loss": 1.1321, "step": 19495 }, { "epoch": 29.18562874251497, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1326, "step": 19496 }, { "epoch": 29.187125748502993, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1247, "step": 19497 }, { "epoch": 29.188622754491018, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1302, "step": 19498 }, { "epoch": 29.190119760479043, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1326, "step": 19499 }, { "epoch": 29.191616766467067, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.123, "step": 19500 }, { "epoch": 29.19311377245509, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1221, "step": 19501 }, { "epoch": 29.194610778443113, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.125, "step": 19502 }, { "epoch": 29.19610778443114, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1234, "step": 19503 }, { "epoch": 29.197604790419163, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1337, "step": 19504 }, { "epoch": 29.199101796407184, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1196, "step": 19505 }, { "epoch": 29.20059880239521, "grad_norm": 0.09814453125, "learning_rate": 0.0008, "loss": 1.124, "step": 19506 }, { "epoch": 29.202095808383234, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1254, "step": 19507 }, { "epoch": 29.20359281437126, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.129, "step": 19508 }, { "epoch": 29.20508982035928, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1288, "step": 19509 }, { "epoch": 29.206586826347305, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1304, "step": 19510 }, { "epoch": 29.20808383233533, "grad_norm": 0.111328125, "learning_rate": 0.0008, "loss": 1.136, "step": 19511 }, { "epoch": 29.209580838323355, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1297, "step": 19512 }, { "epoch": 29.211077844311376, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1245, "step": 19513 }, { "epoch": 29.2125748502994, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1245, "step": 19514 }, { "epoch": 29.214071856287426, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1264, "step": 19515 }, { "epoch": 29.21556886227545, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1268, "step": 19516 }, { "epoch": 29.21706586826347, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1278, "step": 19517 }, { "epoch": 29.218562874251496, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.132, "step": 19518 }, { "epoch": 29.22005988023952, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1294, "step": 19519 }, { "epoch": 29.221556886227546, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1235, "step": 19520 }, { "epoch": 29.223053892215567, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1351, "step": 19521 }, { "epoch": 29.224550898203592, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.13, "step": 19522 }, { "epoch": 29.226047904191617, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1286, "step": 19523 }, { "epoch": 29.227544910179642, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1212, "step": 19524 }, { "epoch": 29.229041916167663, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1282, "step": 19525 }, { "epoch": 29.230538922155688, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1347, "step": 19526 }, { "epoch": 29.232035928143713, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1286, "step": 19527 }, { "epoch": 29.233532934131738, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1316, "step": 19528 }, { "epoch": 29.23502994011976, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1299, "step": 19529 }, { "epoch": 29.236526946107784, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1245, "step": 19530 }, { "epoch": 29.23802395209581, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1276, "step": 19531 }, { "epoch": 29.239520958083833, "grad_norm": 0.0595703125, "learning_rate": 0.0008, "loss": 1.123, "step": 19532 }, { "epoch": 29.241017964071855, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.127, "step": 19533 }, { "epoch": 29.24251497005988, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1283, "step": 19534 }, { "epoch": 29.244011976047904, "grad_norm": 0.052978515625, "learning_rate": 0.0008, "loss": 1.1261, "step": 19535 }, { "epoch": 29.24550898203593, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1335, "step": 19536 }, { "epoch": 29.24700598802395, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1288, "step": 19537 }, { "epoch": 29.248502994011975, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1295, "step": 19538 }, { "epoch": 29.25, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1228, "step": 19539 }, { "epoch": 29.251497005988025, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1256, "step": 19540 }, { "epoch": 29.25299401197605, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1298, "step": 19541 }, { "epoch": 29.25449101796407, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1235, "step": 19542 }, { "epoch": 29.255988023952096, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1285, "step": 19543 }, { "epoch": 29.25748502994012, "grad_norm": 0.05322265625, "learning_rate": 0.0008, "loss": 1.1237, "step": 19544 }, { "epoch": 29.258982035928145, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1284, "step": 19545 }, { "epoch": 29.260479041916167, "grad_norm": 0.061279296875, "learning_rate": 0.0008, "loss": 1.1236, "step": 19546 }, { "epoch": 29.26197604790419, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1262, "step": 19547 }, { "epoch": 29.263473053892216, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1191, "step": 19548 }, { "epoch": 29.26497005988024, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1243, "step": 19549 }, { "epoch": 29.266467065868262, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1195, "step": 19550 }, { "epoch": 29.267964071856287, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1244, "step": 19551 }, { "epoch": 29.269461077844312, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1293, "step": 19552 }, { "epoch": 29.270958083832337, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1233, "step": 19553 }, { "epoch": 29.272455089820358, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1258, "step": 19554 }, { "epoch": 29.273952095808383, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1223, "step": 19555 }, { "epoch": 29.275449101796408, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1198, "step": 19556 }, { "epoch": 29.276946107784433, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1342, "step": 19557 }, { "epoch": 29.278443113772454, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1302, "step": 19558 }, { "epoch": 29.27994011976048, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1292, "step": 19559 }, { "epoch": 29.281437125748504, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1291, "step": 19560 }, { "epoch": 29.28293413173653, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1321, "step": 19561 }, { "epoch": 29.28443113772455, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1296, "step": 19562 }, { "epoch": 29.285928143712574, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1248, "step": 19563 }, { "epoch": 29.2874251497006, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1229, "step": 19564 }, { "epoch": 29.288922155688624, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1272, "step": 19565 }, { "epoch": 29.290419161676645, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.134, "step": 19566 }, { "epoch": 29.29191616766467, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1318, "step": 19567 }, { "epoch": 29.293413173652695, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1318, "step": 19568 }, { "epoch": 29.29491017964072, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1246, "step": 19569 }, { "epoch": 29.29640718562874, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1261, "step": 19570 }, { "epoch": 29.297904191616766, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1316, "step": 19571 }, { "epoch": 29.29940119760479, "grad_norm": 0.057861328125, "learning_rate": 0.0008, "loss": 1.1214, "step": 19572 }, { "epoch": 29.300898203592816, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1271, "step": 19573 }, { "epoch": 29.302395209580837, "grad_norm": 0.0830078125, "learning_rate": 0.0008, "loss": 1.1209, "step": 19574 }, { "epoch": 29.30389221556886, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1259, "step": 19575 }, { "epoch": 29.305389221556887, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1324, "step": 19576 }, { "epoch": 29.30688622754491, "grad_norm": 0.10205078125, "learning_rate": 0.0008, "loss": 1.1273, "step": 19577 }, { "epoch": 29.308383233532933, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1297, "step": 19578 }, { "epoch": 29.309880239520957, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1346, "step": 19579 }, { "epoch": 29.311377245508982, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.125, "step": 19580 }, { "epoch": 29.312874251497007, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.129, "step": 19581 }, { "epoch": 29.31437125748503, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1317, "step": 19582 }, { "epoch": 29.315868263473053, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1351, "step": 19583 }, { "epoch": 29.317365269461078, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1281, "step": 19584 }, { "epoch": 29.318862275449103, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.128, "step": 19585 }, { "epoch": 29.320359281437124, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.125, "step": 19586 }, { "epoch": 29.32185628742515, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1335, "step": 19587 }, { "epoch": 29.323353293413174, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1313, "step": 19588 }, { "epoch": 29.3248502994012, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1302, "step": 19589 }, { "epoch": 29.32634730538922, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.129, "step": 19590 }, { "epoch": 29.327844311377245, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1316, "step": 19591 }, { "epoch": 29.32934131736527, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1327, "step": 19592 }, { "epoch": 29.330838323353294, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1368, "step": 19593 }, { "epoch": 29.33233532934132, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1342, "step": 19594 }, { "epoch": 29.33383233532934, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1276, "step": 19595 }, { "epoch": 29.335329341317365, "grad_norm": 0.059814453125, "learning_rate": 0.0008, "loss": 1.1299, "step": 19596 }, { "epoch": 29.33682634730539, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1316, "step": 19597 }, { "epoch": 29.338323353293415, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.132, "step": 19598 }, { "epoch": 29.339820359281436, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1307, "step": 19599 }, { "epoch": 29.34131736526946, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1251, "step": 19600 }, { "epoch": 29.342814371257486, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1298, "step": 19601 }, { "epoch": 29.34431137724551, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1248, "step": 19602 }, { "epoch": 29.345808383233532, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1232, "step": 19603 }, { "epoch": 29.347305389221557, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1206, "step": 19604 }, { "epoch": 29.34880239520958, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1268, "step": 19605 }, { "epoch": 29.350299401197606, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1228, "step": 19606 }, { "epoch": 29.351796407185628, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1293, "step": 19607 }, { "epoch": 29.353293413173652, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1288, "step": 19608 }, { "epoch": 29.354790419161677, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1332, "step": 19609 }, { "epoch": 29.356287425149702, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1287, "step": 19610 }, { "epoch": 29.357784431137723, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1259, "step": 19611 }, { "epoch": 29.35928143712575, "grad_norm": 0.05859375, "learning_rate": 0.0008, "loss": 1.1266, "step": 19612 }, { "epoch": 29.360778443113773, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1345, "step": 19613 }, { "epoch": 29.362275449101798, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1302, "step": 19614 }, { "epoch": 29.36377245508982, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1276, "step": 19615 }, { "epoch": 29.365269461077844, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1274, "step": 19616 }, { "epoch": 29.36676646706587, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1243, "step": 19617 }, { "epoch": 29.368263473053894, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1347, "step": 19618 }, { "epoch": 29.369760479041915, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1262, "step": 19619 }, { "epoch": 29.37125748502994, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1305, "step": 19620 }, { "epoch": 29.372754491017965, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.128, "step": 19621 }, { "epoch": 29.37425149700599, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1262, "step": 19622 }, { "epoch": 29.37574850299401, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.127, "step": 19623 }, { "epoch": 29.377245508982035, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1261, "step": 19624 }, { "epoch": 29.37874251497006, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1271, "step": 19625 }, { "epoch": 29.380239520958085, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1371, "step": 19626 }, { "epoch": 29.381736526946106, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1264, "step": 19627 }, { "epoch": 29.38323353293413, "grad_norm": 0.13671875, "learning_rate": 0.0008, "loss": 1.1313, "step": 19628 }, { "epoch": 29.384730538922156, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1306, "step": 19629 }, { "epoch": 29.38622754491018, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1288, "step": 19630 }, { "epoch": 29.387724550898202, "grad_norm": 0.1005859375, "learning_rate": 0.0008, "loss": 1.1293, "step": 19631 }, { "epoch": 29.389221556886227, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1249, "step": 19632 }, { "epoch": 29.39071856287425, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1325, "step": 19633 }, { "epoch": 29.392215568862277, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.132, "step": 19634 }, { "epoch": 29.393712574850298, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.133, "step": 19635 }, { "epoch": 29.395209580838323, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1276, "step": 19636 }, { "epoch": 29.396706586826348, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1295, "step": 19637 }, { "epoch": 29.398203592814372, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1254, "step": 19638 }, { "epoch": 29.399700598802394, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1304, "step": 19639 }, { "epoch": 29.40119760479042, "grad_norm": 0.12353515625, "learning_rate": 0.0008, "loss": 1.1301, "step": 19640 }, { "epoch": 29.402694610778443, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1321, "step": 19641 }, { "epoch": 29.404191616766468, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1248, "step": 19642 }, { "epoch": 29.40568862275449, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1228, "step": 19643 }, { "epoch": 29.407185628742514, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1333, "step": 19644 }, { "epoch": 29.40868263473054, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1337, "step": 19645 }, { "epoch": 29.410179640718564, "grad_norm": 0.0927734375, "learning_rate": 0.0008, "loss": 1.1271, "step": 19646 }, { "epoch": 29.411676646706585, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.125, "step": 19647 }, { "epoch": 29.41317365269461, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1317, "step": 19648 }, { "epoch": 29.414670658682635, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1287, "step": 19649 }, { "epoch": 29.41616766467066, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1277, "step": 19650 }, { "epoch": 29.41766467065868, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1316, "step": 19651 }, { "epoch": 29.419161676646706, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1273, "step": 19652 }, { "epoch": 29.42065868263473, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1299, "step": 19653 }, { "epoch": 29.422155688622755, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1199, "step": 19654 }, { "epoch": 29.42365269461078, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1276, "step": 19655 }, { "epoch": 29.4251497005988, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1321, "step": 19656 }, { "epoch": 29.426646706586826, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1281, "step": 19657 }, { "epoch": 29.42814371257485, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1284, "step": 19658 }, { "epoch": 29.429640718562876, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.124, "step": 19659 }, { "epoch": 29.431137724550897, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1279, "step": 19660 }, { "epoch": 29.432634730538922, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1303, "step": 19661 }, { "epoch": 29.434131736526947, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1258, "step": 19662 }, { "epoch": 29.43562874251497, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1285, "step": 19663 }, { "epoch": 29.437125748502993, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1305, "step": 19664 }, { "epoch": 29.438622754491018, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1261, "step": 19665 }, { "epoch": 29.440119760479043, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1303, "step": 19666 }, { "epoch": 29.441616766467067, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1349, "step": 19667 }, { "epoch": 29.44311377245509, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1265, "step": 19668 }, { "epoch": 29.444610778443113, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1342, "step": 19669 }, { "epoch": 29.44610778443114, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1223, "step": 19670 }, { "epoch": 29.447604790419163, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1333, "step": 19671 }, { "epoch": 29.449101796407184, "grad_norm": 0.08349609375, "learning_rate": 0.0008, "loss": 1.1337, "step": 19672 }, { "epoch": 29.45059880239521, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1291, "step": 19673 }, { "epoch": 29.452095808383234, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1224, "step": 19674 }, { "epoch": 29.45359281437126, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1286, "step": 19675 }, { "epoch": 29.45508982035928, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.122, "step": 19676 }, { "epoch": 29.456586826347305, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1254, "step": 19677 }, { "epoch": 29.45808383233533, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.131, "step": 19678 }, { "epoch": 29.459580838323355, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1289, "step": 19679 }, { "epoch": 29.461077844311376, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1269, "step": 19680 }, { "epoch": 29.4625748502994, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1292, "step": 19681 }, { "epoch": 29.464071856287426, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1255, "step": 19682 }, { "epoch": 29.46556886227545, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1247, "step": 19683 }, { "epoch": 29.46706586826347, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.124, "step": 19684 }, { "epoch": 29.468562874251496, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1259, "step": 19685 }, { "epoch": 29.47005988023952, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1269, "step": 19686 }, { "epoch": 29.471556886227546, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1292, "step": 19687 }, { "epoch": 29.473053892215567, "grad_norm": 0.050537109375, "learning_rate": 0.0008, "loss": 1.1275, "step": 19688 }, { "epoch": 29.474550898203592, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.13, "step": 19689 }, { "epoch": 29.476047904191617, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1279, "step": 19690 }, { "epoch": 29.477544910179642, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1331, "step": 19691 }, { "epoch": 29.479041916167663, "grad_norm": 0.1591796875, "learning_rate": 0.0008, "loss": 1.1321, "step": 19692 }, { "epoch": 29.480538922155688, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1327, "step": 19693 }, { "epoch": 29.482035928143713, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1332, "step": 19694 }, { "epoch": 29.483532934131738, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1274, "step": 19695 }, { "epoch": 29.48502994011976, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1277, "step": 19696 }, { "epoch": 29.486526946107784, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1262, "step": 19697 }, { "epoch": 29.48802395209581, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1212, "step": 19698 }, { "epoch": 29.489520958083833, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1218, "step": 19699 }, { "epoch": 29.491017964071855, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1297, "step": 19700 }, { "epoch": 29.49251497005988, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1264, "step": 19701 }, { "epoch": 29.494011976047904, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1282, "step": 19702 }, { "epoch": 29.49550898203593, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1273, "step": 19703 }, { "epoch": 29.49700598802395, "grad_norm": 0.08740234375, "learning_rate": 0.0008, "loss": 1.1344, "step": 19704 }, { "epoch": 29.498502994011975, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.13, "step": 19705 }, { "epoch": 29.5, "grad_norm": 0.055419921875, "learning_rate": 0.0008, "loss": 1.1282, "step": 19706 }, { "epoch": 29.501497005988025, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1251, "step": 19707 }, { "epoch": 29.50299401197605, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1284, "step": 19708 }, { "epoch": 29.50449101796407, "grad_norm": 0.056884765625, "learning_rate": 0.0008, "loss": 1.1346, "step": 19709 }, { "epoch": 29.505988023952096, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1274, "step": 19710 }, { "epoch": 29.50748502994012, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1218, "step": 19711 }, { "epoch": 29.508982035928145, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1309, "step": 19712 }, { "epoch": 29.510479041916167, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1274, "step": 19713 }, { "epoch": 29.51197604790419, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1263, "step": 19714 }, { "epoch": 29.513473053892216, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1291, "step": 19715 }, { "epoch": 29.51497005988024, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1258, "step": 19716 }, { "epoch": 29.516467065868262, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1365, "step": 19717 }, { "epoch": 29.517964071856287, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1212, "step": 19718 }, { "epoch": 29.519461077844312, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1323, "step": 19719 }, { "epoch": 29.520958083832337, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1299, "step": 19720 }, { "epoch": 29.522455089820358, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1261, "step": 19721 }, { "epoch": 29.523952095808383, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1268, "step": 19722 }, { "epoch": 29.525449101796408, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1248, "step": 19723 }, { "epoch": 29.526946107784433, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1261, "step": 19724 }, { "epoch": 29.528443113772454, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1277, "step": 19725 }, { "epoch": 29.52994011976048, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1308, "step": 19726 }, { "epoch": 29.531437125748504, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1209, "step": 19727 }, { "epoch": 29.53293413173653, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1266, "step": 19728 }, { "epoch": 29.53443113772455, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1344, "step": 19729 }, { "epoch": 29.535928143712574, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1327, "step": 19730 }, { "epoch": 29.5374251497006, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1336, "step": 19731 }, { "epoch": 29.538922155688624, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1294, "step": 19732 }, { "epoch": 29.540419161676645, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1304, "step": 19733 }, { "epoch": 29.54191616766467, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1299, "step": 19734 }, { "epoch": 29.543413173652695, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1289, "step": 19735 }, { "epoch": 29.54491017964072, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1237, "step": 19736 }, { "epoch": 29.54640718562874, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1304, "step": 19737 }, { "epoch": 29.547904191616766, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1252, "step": 19738 }, { "epoch": 29.54940119760479, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1291, "step": 19739 }, { "epoch": 29.550898203592816, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1274, "step": 19740 }, { "epoch": 29.552395209580837, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1267, "step": 19741 }, { "epoch": 29.55389221556886, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1344, "step": 19742 }, { "epoch": 29.555389221556887, "grad_norm": 0.056640625, "learning_rate": 0.0008, "loss": 1.1281, "step": 19743 }, { "epoch": 29.55688622754491, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1217, "step": 19744 }, { "epoch": 29.558383233532933, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1237, "step": 19745 }, { "epoch": 29.559880239520957, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1258, "step": 19746 }, { "epoch": 29.561377245508982, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1242, "step": 19747 }, { "epoch": 29.562874251497007, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1276, "step": 19748 }, { "epoch": 29.56437125748503, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1271, "step": 19749 }, { "epoch": 29.565868263473053, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1299, "step": 19750 }, { "epoch": 29.567365269461078, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1231, "step": 19751 }, { "epoch": 29.568862275449103, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1242, "step": 19752 }, { "epoch": 29.570359281437124, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1285, "step": 19753 }, { "epoch": 29.57185628742515, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1287, "step": 19754 }, { "epoch": 29.573353293413174, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1278, "step": 19755 }, { "epoch": 29.5748502994012, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1287, "step": 19756 }, { "epoch": 29.57634730538922, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1234, "step": 19757 }, { "epoch": 29.577844311377245, "grad_norm": 0.05908203125, "learning_rate": 0.0008, "loss": 1.1331, "step": 19758 }, { "epoch": 29.57934131736527, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1281, "step": 19759 }, { "epoch": 29.580838323353294, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1332, "step": 19760 }, { "epoch": 29.58233532934132, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1252, "step": 19761 }, { "epoch": 29.58383233532934, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1289, "step": 19762 }, { "epoch": 29.585329341317365, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.125, "step": 19763 }, { "epoch": 29.58682634730539, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1285, "step": 19764 }, { "epoch": 29.58832335329341, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1306, "step": 19765 }, { "epoch": 29.589820359281436, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1314, "step": 19766 }, { "epoch": 29.59131736526946, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1267, "step": 19767 }, { "epoch": 29.592814371257486, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.129, "step": 19768 }, { "epoch": 29.59431137724551, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1217, "step": 19769 }, { "epoch": 29.595808383233532, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1223, "step": 19770 }, { "epoch": 29.597305389221557, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1214, "step": 19771 }, { "epoch": 29.59880239520958, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1267, "step": 19772 }, { "epoch": 29.600299401197606, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1265, "step": 19773 }, { "epoch": 29.601796407185628, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1308, "step": 19774 }, { "epoch": 29.603293413173652, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1287, "step": 19775 }, { "epoch": 29.604790419161677, "grad_norm": 0.09326171875, "learning_rate": 0.0008, "loss": 1.1311, "step": 19776 }, { "epoch": 29.606287425149702, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1251, "step": 19777 }, { "epoch": 29.607784431137723, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1244, "step": 19778 }, { "epoch": 29.60928143712575, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1262, "step": 19779 }, { "epoch": 29.610778443113773, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1254, "step": 19780 }, { "epoch": 29.612275449101798, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1255, "step": 19781 }, { "epoch": 29.61377245508982, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1279, "step": 19782 }, { "epoch": 29.615269461077844, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1247, "step": 19783 }, { "epoch": 29.61676646706587, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1266, "step": 19784 }, { "epoch": 29.618263473053894, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1319, "step": 19785 }, { "epoch": 29.619760479041915, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1252, "step": 19786 }, { "epoch": 29.62125748502994, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1245, "step": 19787 }, { "epoch": 29.622754491017965, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.1294, "step": 19788 }, { "epoch": 29.62425149700599, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1213, "step": 19789 }, { "epoch": 29.62574850299401, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1308, "step": 19790 }, { "epoch": 29.627245508982035, "grad_norm": 0.09130859375, "learning_rate": 0.0008, "loss": 1.1243, "step": 19791 }, { "epoch": 29.62874251497006, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1343, "step": 19792 }, { "epoch": 29.630239520958085, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.135, "step": 19793 }, { "epoch": 29.631736526946106, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1283, "step": 19794 }, { "epoch": 29.63323353293413, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.127, "step": 19795 }, { "epoch": 29.634730538922156, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1299, "step": 19796 }, { "epoch": 29.63622754491018, "grad_norm": 0.09423828125, "learning_rate": 0.0008, "loss": 1.1314, "step": 19797 }, { "epoch": 29.637724550898202, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1345, "step": 19798 }, { "epoch": 29.639221556886227, "grad_norm": 0.091796875, "learning_rate": 0.0008, "loss": 1.1344, "step": 19799 }, { "epoch": 29.64071856287425, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1271, "step": 19800 }, { "epoch": 29.642215568862277, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1247, "step": 19801 }, { "epoch": 29.643712574850298, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1242, "step": 19802 }, { "epoch": 29.645209580838323, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1237, "step": 19803 }, { "epoch": 29.646706586826348, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1271, "step": 19804 }, { "epoch": 29.648203592814372, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1257, "step": 19805 }, { "epoch": 29.649700598802394, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1311, "step": 19806 }, { "epoch": 29.65119760479042, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.128, "step": 19807 }, { "epoch": 29.652694610778443, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1211, "step": 19808 }, { "epoch": 29.654191616766468, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1301, "step": 19809 }, { "epoch": 29.65568862275449, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1296, "step": 19810 }, { "epoch": 29.657185628742514, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1267, "step": 19811 }, { "epoch": 29.65868263473054, "grad_norm": 0.05517578125, "learning_rate": 0.0008, "loss": 1.1249, "step": 19812 }, { "epoch": 29.660179640718564, "grad_norm": 0.087890625, "learning_rate": 0.0008, "loss": 1.1208, "step": 19813 }, { "epoch": 29.66167664670659, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.128, "step": 19814 }, { "epoch": 29.66317365269461, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1195, "step": 19815 }, { "epoch": 29.664670658682635, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1284, "step": 19816 }, { "epoch": 29.66616766467066, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1239, "step": 19817 }, { "epoch": 29.66766467065868, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1248, "step": 19818 }, { "epoch": 29.669161676646706, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1177, "step": 19819 }, { "epoch": 29.67065868263473, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.1227, "step": 19820 }, { "epoch": 29.672155688622755, "grad_norm": 0.1318359375, "learning_rate": 0.0008, "loss": 1.1241, "step": 19821 }, { "epoch": 29.67365269461078, "grad_norm": 0.0888671875, "learning_rate": 0.0008, "loss": 1.1235, "step": 19822 }, { "epoch": 29.6751497005988, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1257, "step": 19823 }, { "epoch": 29.676646706586826, "grad_norm": 0.09619140625, "learning_rate": 0.0008, "loss": 1.1306, "step": 19824 }, { "epoch": 29.67814371257485, "grad_norm": 0.08642578125, "learning_rate": 0.0008, "loss": 1.1319, "step": 19825 }, { "epoch": 29.679640718562876, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.125, "step": 19826 }, { "epoch": 29.681137724550897, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1272, "step": 19827 }, { "epoch": 29.682634730538922, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1286, "step": 19828 }, { "epoch": 29.684131736526947, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1292, "step": 19829 }, { "epoch": 29.68562874251497, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1326, "step": 19830 }, { "epoch": 29.687125748502993, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.123, "step": 19831 }, { "epoch": 29.688622754491018, "grad_norm": 0.058349609375, "learning_rate": 0.0008, "loss": 1.1313, "step": 19832 }, { "epoch": 29.690119760479043, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1336, "step": 19833 }, { "epoch": 29.691616766467067, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1245, "step": 19834 }, { "epoch": 29.69311377245509, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1254, "step": 19835 }, { "epoch": 29.694610778443113, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1207, "step": 19836 }, { "epoch": 29.69610778443114, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1281, "step": 19837 }, { "epoch": 29.697604790419163, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1262, "step": 19838 }, { "epoch": 29.699101796407184, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1271, "step": 19839 }, { "epoch": 29.70059880239521, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1305, "step": 19840 }, { "epoch": 29.702095808383234, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1273, "step": 19841 }, { "epoch": 29.70359281437126, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1272, "step": 19842 }, { "epoch": 29.70508982035928, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1273, "step": 19843 }, { "epoch": 29.706586826347305, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1274, "step": 19844 }, { "epoch": 29.70808383233533, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1281, "step": 19845 }, { "epoch": 29.709580838323355, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1229, "step": 19846 }, { "epoch": 29.711077844311376, "grad_norm": 0.0693359375, "learning_rate": 0.0008, "loss": 1.125, "step": 19847 }, { "epoch": 29.7125748502994, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1315, "step": 19848 }, { "epoch": 29.714071856287426, "grad_norm": 0.0556640625, "learning_rate": 0.0008, "loss": 1.1275, "step": 19849 }, { "epoch": 29.71556886227545, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1303, "step": 19850 }, { "epoch": 29.71706586826347, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.124, "step": 19851 }, { "epoch": 29.718562874251496, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1255, "step": 19852 }, { "epoch": 29.72005988023952, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1246, "step": 19853 }, { "epoch": 29.721556886227546, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1287, "step": 19854 }, { "epoch": 29.723053892215567, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.126, "step": 19855 }, { "epoch": 29.724550898203592, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1256, "step": 19856 }, { "epoch": 29.726047904191617, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1262, "step": 19857 }, { "epoch": 29.727544910179642, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1292, "step": 19858 }, { "epoch": 29.729041916167663, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.1288, "step": 19859 }, { "epoch": 29.730538922155688, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1319, "step": 19860 }, { "epoch": 29.732035928143713, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1259, "step": 19861 }, { "epoch": 29.733532934131738, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1282, "step": 19862 }, { "epoch": 29.73502994011976, "grad_norm": 0.07373046875, "learning_rate": 0.0008, "loss": 1.1213, "step": 19863 }, { "epoch": 29.736526946107784, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1284, "step": 19864 }, { "epoch": 29.73802395209581, "grad_norm": 0.07568359375, "learning_rate": 0.0008, "loss": 1.1314, "step": 19865 }, { "epoch": 29.739520958083833, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1268, "step": 19866 }, { "epoch": 29.741017964071855, "grad_norm": 0.08056640625, "learning_rate": 0.0008, "loss": 1.1306, "step": 19867 }, { "epoch": 29.74251497005988, "grad_norm": 0.055908203125, "learning_rate": 0.0008, "loss": 1.1317, "step": 19868 }, { "epoch": 29.744011976047904, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1292, "step": 19869 }, { "epoch": 29.74550898203593, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1266, "step": 19870 }, { "epoch": 29.74700598802395, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1228, "step": 19871 }, { "epoch": 29.748502994011975, "grad_norm": 0.05615234375, "learning_rate": 0.0008, "loss": 1.1259, "step": 19872 }, { "epoch": 29.75, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1354, "step": 19873 }, { "epoch": 29.751497005988025, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1254, "step": 19874 }, { "epoch": 29.75299401197605, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.125, "step": 19875 }, { "epoch": 29.75449101796407, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1222, "step": 19876 }, { "epoch": 29.755988023952096, "grad_norm": 0.0751953125, "learning_rate": 0.0008, "loss": 1.1317, "step": 19877 }, { "epoch": 29.75748502994012, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1282, "step": 19878 }, { "epoch": 29.758982035928145, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1359, "step": 19879 }, { "epoch": 29.760479041916167, "grad_norm": 0.08837890625, "learning_rate": 0.0008, "loss": 1.1285, "step": 19880 }, { "epoch": 29.76197604790419, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1186, "step": 19881 }, { "epoch": 29.763473053892216, "grad_norm": 0.05810546875, "learning_rate": 0.0008, "loss": 1.1325, "step": 19882 }, { "epoch": 29.76497005988024, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1233, "step": 19883 }, { "epoch": 29.766467065868262, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1227, "step": 19884 }, { "epoch": 29.767964071856287, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1275, "step": 19885 }, { "epoch": 29.769461077844312, "grad_norm": 0.099609375, "learning_rate": 0.0008, "loss": 1.1286, "step": 19886 }, { "epoch": 29.770958083832337, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1271, "step": 19887 }, { "epoch": 29.772455089820358, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1267, "step": 19888 }, { "epoch": 29.773952095808383, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1281, "step": 19889 }, { "epoch": 29.775449101796408, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1334, "step": 19890 }, { "epoch": 29.776946107784433, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1288, "step": 19891 }, { "epoch": 29.778443113772454, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1283, "step": 19892 }, { "epoch": 29.77994011976048, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1239, "step": 19893 }, { "epoch": 29.781437125748504, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1318, "step": 19894 }, { "epoch": 29.78293413173653, "grad_norm": 0.0673828125, "learning_rate": 0.0008, "loss": 1.1235, "step": 19895 }, { "epoch": 29.78443113772455, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1319, "step": 19896 }, { "epoch": 29.785928143712574, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.129, "step": 19897 }, { "epoch": 29.7874251497006, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1289, "step": 19898 }, { "epoch": 29.788922155688624, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1238, "step": 19899 }, { "epoch": 29.790419161676645, "grad_norm": 0.09521484375, "learning_rate": 0.0008, "loss": 1.1175, "step": 19900 }, { "epoch": 29.79191616766467, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1254, "step": 19901 }, { "epoch": 29.793413173652695, "grad_norm": 0.06201171875, "learning_rate": 0.0008, "loss": 1.1228, "step": 19902 }, { "epoch": 29.79491017964072, "grad_norm": 0.07470703125, "learning_rate": 0.0008, "loss": 1.1266, "step": 19903 }, { "epoch": 29.79640718562874, "grad_norm": 0.0859375, "learning_rate": 0.0008, "loss": 1.1321, "step": 19904 }, { "epoch": 29.797904191616766, "grad_norm": 0.058837890625, "learning_rate": 0.0008, "loss": 1.1355, "step": 19905 }, { "epoch": 29.79940119760479, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1287, "step": 19906 }, { "epoch": 29.800898203592816, "grad_norm": 0.09228515625, "learning_rate": 0.0008, "loss": 1.1321, "step": 19907 }, { "epoch": 29.802395209580837, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1269, "step": 19908 }, { "epoch": 29.80389221556886, "grad_norm": 0.0791015625, "learning_rate": 0.0008, "loss": 1.1214, "step": 19909 }, { "epoch": 29.805389221556887, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.134, "step": 19910 }, { "epoch": 29.80688622754491, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1258, "step": 19911 }, { "epoch": 29.808383233532933, "grad_norm": 0.07080078125, "learning_rate": 0.0008, "loss": 1.1297, "step": 19912 }, { "epoch": 29.809880239520957, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1243, "step": 19913 }, { "epoch": 29.811377245508982, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1256, "step": 19914 }, { "epoch": 29.812874251497007, "grad_norm": 0.08154296875, "learning_rate": 0.0008, "loss": 1.1259, "step": 19915 }, { "epoch": 29.81437125748503, "grad_norm": 0.060791015625, "learning_rate": 0.0008, "loss": 1.1325, "step": 19916 }, { "epoch": 29.815868263473053, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1353, "step": 19917 }, { "epoch": 29.817365269461078, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1346, "step": 19918 }, { "epoch": 29.818862275449103, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1297, "step": 19919 }, { "epoch": 29.820359281437124, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1301, "step": 19920 }, { "epoch": 29.82185628742515, "grad_norm": 0.0908203125, "learning_rate": 0.0008, "loss": 1.1349, "step": 19921 }, { "epoch": 29.823353293413174, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.12, "step": 19922 }, { "epoch": 29.8248502994012, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1273, "step": 19923 }, { "epoch": 29.82634730538922, "grad_norm": 0.06640625, "learning_rate": 0.0008, "loss": 1.1337, "step": 19924 }, { "epoch": 29.827844311377245, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1231, "step": 19925 }, { "epoch": 29.82934131736527, "grad_norm": 0.060302734375, "learning_rate": 0.0008, "loss": 1.1231, "step": 19926 }, { "epoch": 29.830838323353294, "grad_norm": 0.059326171875, "learning_rate": 0.0008, "loss": 1.1181, "step": 19927 }, { "epoch": 29.83233532934132, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1276, "step": 19928 }, { "epoch": 29.83383233532934, "grad_norm": 0.0634765625, "learning_rate": 0.0008, "loss": 1.1297, "step": 19929 }, { "epoch": 29.835329341317365, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1329, "step": 19930 }, { "epoch": 29.83682634730539, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1298, "step": 19931 }, { "epoch": 29.83832335329341, "grad_norm": 0.06982421875, "learning_rate": 0.0008, "loss": 1.1238, "step": 19932 }, { "epoch": 29.839820359281436, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1259, "step": 19933 }, { "epoch": 29.84131736526946, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1225, "step": 19934 }, { "epoch": 29.842814371257486, "grad_norm": 0.05078125, "learning_rate": 0.0008, "loss": 1.1273, "step": 19935 }, { "epoch": 29.84431137724551, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1321, "step": 19936 }, { "epoch": 29.845808383233532, "grad_norm": 0.08544921875, "learning_rate": 0.0008, "loss": 1.127, "step": 19937 }, { "epoch": 29.847305389221557, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1269, "step": 19938 }, { "epoch": 29.84880239520958, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1295, "step": 19939 }, { "epoch": 29.850299401197606, "grad_norm": 0.08251953125, "learning_rate": 0.0008, "loss": 1.1336, "step": 19940 }, { "epoch": 29.851796407185628, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1236, "step": 19941 }, { "epoch": 29.853293413173652, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1287, "step": 19942 }, { "epoch": 29.854790419161677, "grad_norm": 0.07421875, "learning_rate": 0.0008, "loss": 1.1252, "step": 19943 }, { "epoch": 29.856287425149702, "grad_norm": 0.06005859375, "learning_rate": 0.0008, "loss": 1.1219, "step": 19944 }, { "epoch": 29.857784431137723, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1339, "step": 19945 }, { "epoch": 29.85928143712575, "grad_norm": 0.06591796875, "learning_rate": 0.0008, "loss": 1.1268, "step": 19946 }, { "epoch": 29.860778443113773, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.123, "step": 19947 }, { "epoch": 29.862275449101798, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1252, "step": 19948 }, { "epoch": 29.86377245508982, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1319, "step": 19949 }, { "epoch": 29.865269461077844, "grad_norm": 0.068359375, "learning_rate": 0.0008, "loss": 1.1243, "step": 19950 }, { "epoch": 29.86676646706587, "grad_norm": 0.07763671875, "learning_rate": 0.0008, "loss": 1.1233, "step": 19951 }, { "epoch": 29.868263473053894, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1281, "step": 19952 }, { "epoch": 29.869760479041915, "grad_norm": 0.06494140625, "learning_rate": 0.0008, "loss": 1.1258, "step": 19953 }, { "epoch": 29.87125748502994, "grad_norm": 0.0810546875, "learning_rate": 0.0008, "loss": 1.1252, "step": 19954 }, { "epoch": 29.872754491017965, "grad_norm": 0.0576171875, "learning_rate": 0.0008, "loss": 1.1247, "step": 19955 }, { "epoch": 29.87425149700599, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1231, "step": 19956 }, { "epoch": 29.87574850299401, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1257, "step": 19957 }, { "epoch": 29.877245508982035, "grad_norm": 0.062255859375, "learning_rate": 0.0008, "loss": 1.1262, "step": 19958 }, { "epoch": 29.87874251497006, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.125, "step": 19959 }, { "epoch": 29.880239520958085, "grad_norm": 0.06396484375, "learning_rate": 0.0008, "loss": 1.1238, "step": 19960 }, { "epoch": 29.881736526946106, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.1246, "step": 19961 }, { "epoch": 29.88323353293413, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1268, "step": 19962 }, { "epoch": 29.884730538922156, "grad_norm": 0.06787109375, "learning_rate": 0.0008, "loss": 1.1262, "step": 19963 }, { "epoch": 29.88622754491018, "grad_norm": 0.083984375, "learning_rate": 0.0008, "loss": 1.1219, "step": 19964 }, { "epoch": 29.887724550898202, "grad_norm": 0.0703125, "learning_rate": 0.0008, "loss": 1.1267, "step": 19965 }, { "epoch": 29.889221556886227, "grad_norm": 0.0947265625, "learning_rate": 0.0008, "loss": 1.1279, "step": 19966 }, { "epoch": 29.89071856287425, "grad_norm": 0.076171875, "learning_rate": 0.0008, "loss": 1.1217, "step": 19967 }, { "epoch": 29.892215568862277, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1284, "step": 19968 }, { "epoch": 29.893712574850298, "grad_norm": 0.0771484375, "learning_rate": 0.0008, "loss": 1.1299, "step": 19969 }, { "epoch": 29.895209580838323, "grad_norm": 0.06689453125, "learning_rate": 0.0008, "loss": 1.1277, "step": 19970 }, { "epoch": 29.896706586826348, "grad_norm": 0.0615234375, "learning_rate": 0.0008, "loss": 1.1261, "step": 19971 }, { "epoch": 29.898203592814372, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1258, "step": 19972 }, { "epoch": 29.899700598802394, "grad_norm": 0.05712890625, "learning_rate": 0.0008, "loss": 1.1279, "step": 19973 }, { "epoch": 29.90119760479042, "grad_norm": 0.0625, "learning_rate": 0.0008, "loss": 1.1249, "step": 19974 }, { "epoch": 29.902694610778443, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1298, "step": 19975 }, { "epoch": 29.904191616766468, "grad_norm": 0.0849609375, "learning_rate": 0.0008, "loss": 1.1255, "step": 19976 }, { "epoch": 29.90568862275449, "grad_norm": 0.060546875, "learning_rate": 0.0008, "loss": 1.1306, "step": 19977 }, { "epoch": 29.907185628742514, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1271, "step": 19978 }, { "epoch": 29.90868263473054, "grad_norm": 0.0732421875, "learning_rate": 0.0008, "loss": 1.1347, "step": 19979 }, { "epoch": 29.910179640718564, "grad_norm": 0.07177734375, "learning_rate": 0.0008, "loss": 1.132, "step": 19980 }, { "epoch": 29.91167664670659, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1325, "step": 19981 }, { "epoch": 29.91317365269461, "grad_norm": 0.0712890625, "learning_rate": 0.0008, "loss": 1.1234, "step": 19982 }, { "epoch": 29.914670658682635, "grad_norm": 0.0654296875, "learning_rate": 0.0008, "loss": 1.1306, "step": 19983 }, { "epoch": 29.91616766467066, "grad_norm": 0.0869140625, "learning_rate": 0.0008, "loss": 1.1237, "step": 19984 }, { "epoch": 29.91766467065868, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1259, "step": 19985 }, { "epoch": 29.919161676646706, "grad_norm": 0.078125, "learning_rate": 0.0008, "loss": 1.1277, "step": 19986 }, { "epoch": 29.92065868263473, "grad_norm": 0.09375, "learning_rate": 0.0008, "loss": 1.1237, "step": 19987 }, { "epoch": 29.922155688622755, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1366, "step": 19988 }, { "epoch": 29.92365269461078, "grad_norm": 0.072265625, "learning_rate": 0.0008, "loss": 1.1272, "step": 19989 }, { "epoch": 29.9251497005988, "grad_norm": 0.06298828125, "learning_rate": 0.0008, "loss": 1.127, "step": 19990 }, { "epoch": 29.926646706586826, "grad_norm": 0.08203125, "learning_rate": 0.0008, "loss": 1.1278, "step": 19991 }, { "epoch": 29.92814371257485, "grad_norm": 0.07666015625, "learning_rate": 0.0008, "loss": 1.1304, "step": 19992 }, { "epoch": 29.929640718562876, "grad_norm": 0.07861328125, "learning_rate": 0.0008, "loss": 1.1305, "step": 19993 }, { "epoch": 29.931137724550897, "grad_norm": 0.07275390625, "learning_rate": 0.0008, "loss": 1.1208, "step": 19994 }, { "epoch": 29.932634730538922, "grad_norm": 0.06884765625, "learning_rate": 0.0008, "loss": 1.1281, "step": 19995 }, { "epoch": 29.934131736526947, "grad_norm": 0.064453125, "learning_rate": 0.0008, "loss": 1.1308, "step": 19996 }, { "epoch": 29.93562874251497, "grad_norm": 0.080078125, "learning_rate": 0.0008, "loss": 1.1283, "step": 19997 }, { "epoch": 29.937125748502993, "grad_norm": 0.06103515625, "learning_rate": 0.0008, "loss": 1.1325, "step": 19998 }, { "epoch": 29.938622754491018, "grad_norm": 0.08447265625, "learning_rate": 0.0008, "loss": 1.1304, "step": 19999 }, { "epoch": 29.940119760479043, "grad_norm": 0.07958984375, "learning_rate": 0.0008, "loss": 1.1267, "step": 20000 } ], "logging_steps": 1.0, "max_steps": 1002000, "num_input_tokens_seen": 0, "num_train_epochs": 1500, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": false }, "attributes": {} } }, "total_flos": 1.1928940848575676e+19, "train_batch_size": 120, "trial_name": null, "trial_params": null }