{ "best_metric": null, "best_model_checkpoint": null, "epoch": 3.9548387096774196, "eval_steps": 500, "global_step": 308, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.012903225806451613, "grad_norm": 3.865084409713745, "learning_rate": 1.6129032258064516e-06, "loss": 0.2085, "step": 1 }, { "epoch": 0.025806451612903226, "grad_norm": 4.324934482574463, "learning_rate": 3.225806451612903e-06, "loss": 0.2026, "step": 2 }, { "epoch": 0.03870967741935484, "grad_norm": 3.4087750911712646, "learning_rate": 4.838709677419355e-06, "loss": 0.2162, "step": 3 }, { "epoch": 0.05161290322580645, "grad_norm": 1.6109209060668945, "learning_rate": 6.451612903225806e-06, "loss": 0.1839, "step": 4 }, { "epoch": 0.06451612903225806, "grad_norm": 1.3727400302886963, "learning_rate": 8.064516129032258e-06, "loss": 0.1952, "step": 5 }, { "epoch": 0.07741935483870968, "grad_norm": 1.283871054649353, "learning_rate": 9.67741935483871e-06, "loss": 0.1879, "step": 6 }, { "epoch": 0.09032258064516129, "grad_norm": 1.0045589208602905, "learning_rate": 1.129032258064516e-05, "loss": 0.1501, "step": 7 }, { "epoch": 0.1032258064516129, "grad_norm": 1.3846447467803955, "learning_rate": 1.2903225806451613e-05, "loss": 0.1489, "step": 8 }, { "epoch": 0.11612903225806452, "grad_norm": 1.364700198173523, "learning_rate": 1.4516129032258066e-05, "loss": 0.1484, "step": 9 }, { "epoch": 0.12903225806451613, "grad_norm": 0.9269119501113892, "learning_rate": 1.6129032258064517e-05, "loss": 0.1388, "step": 10 }, { "epoch": 0.14193548387096774, "grad_norm": 2.6850790977478027, "learning_rate": 1.774193548387097e-05, "loss": 0.1319, "step": 11 }, { "epoch": 0.15483870967741936, "grad_norm": 1.2811543941497803, "learning_rate": 1.935483870967742e-05, "loss": 0.1246, "step": 12 }, { "epoch": 0.16774193548387098, "grad_norm": 1.0256060361862183, "learning_rate": 2.0967741935483873e-05, "loss": 0.1319, "step": 13 }, { "epoch": 0.18064516129032257, "grad_norm": 0.9394522905349731, "learning_rate": 2.258064516129032e-05, "loss": 0.1191, "step": 14 }, { "epoch": 0.1935483870967742, "grad_norm": 0.8834514021873474, "learning_rate": 2.4193548387096777e-05, "loss": 0.1248, "step": 15 }, { "epoch": 0.2064516129032258, "grad_norm": 1.0034328699111938, "learning_rate": 2.5806451612903226e-05, "loss": 0.124, "step": 16 }, { "epoch": 0.21935483870967742, "grad_norm": 0.8772470951080322, "learning_rate": 2.7419354838709678e-05, "loss": 0.1068, "step": 17 }, { "epoch": 0.23225806451612904, "grad_norm": 0.9788472652435303, "learning_rate": 2.9032258064516133e-05, "loss": 0.1245, "step": 18 }, { "epoch": 0.24516129032258063, "grad_norm": 0.8098297715187073, "learning_rate": 3.0645161290322585e-05, "loss": 0.1099, "step": 19 }, { "epoch": 0.25806451612903225, "grad_norm": 0.8270663619041443, "learning_rate": 3.2258064516129034e-05, "loss": 0.0959, "step": 20 }, { "epoch": 0.2709677419354839, "grad_norm": 0.6456682682037354, "learning_rate": 3.387096774193548e-05, "loss": 0.0923, "step": 21 }, { "epoch": 0.2838709677419355, "grad_norm": 0.8719914555549622, "learning_rate": 3.548387096774194e-05, "loss": 0.1088, "step": 22 }, { "epoch": 0.2967741935483871, "grad_norm": 0.6625027060508728, "learning_rate": 3.7096774193548386e-05, "loss": 0.0882, "step": 23 }, { "epoch": 0.3096774193548387, "grad_norm": 0.6829620003700256, "learning_rate": 3.870967741935484e-05, "loss": 0.0992, "step": 24 }, { "epoch": 0.3225806451612903, "grad_norm": 0.8227680921554565, "learning_rate": 4.032258064516129e-05, "loss": 0.1224, "step": 25 }, { "epoch": 0.33548387096774196, "grad_norm": 0.5915015339851379, "learning_rate": 4.1935483870967746e-05, "loss": 0.0846, "step": 26 }, { "epoch": 0.34838709677419355, "grad_norm": 0.6318536400794983, "learning_rate": 4.3548387096774194e-05, "loss": 0.0864, "step": 27 }, { "epoch": 0.36129032258064514, "grad_norm": 0.6643650531768799, "learning_rate": 4.516129032258064e-05, "loss": 0.103, "step": 28 }, { "epoch": 0.3741935483870968, "grad_norm": 0.6708557605743408, "learning_rate": 4.67741935483871e-05, "loss": 0.1148, "step": 29 }, { "epoch": 0.3870967741935484, "grad_norm": 0.5628843307495117, "learning_rate": 4.8387096774193554e-05, "loss": 0.0789, "step": 30 }, { "epoch": 0.4, "grad_norm": 0.5701764225959778, "learning_rate": 5e-05, "loss": 0.0844, "step": 31 }, { "epoch": 0.4129032258064516, "grad_norm": 0.565487265586853, "learning_rate": 4.999839214987031e-05, "loss": 0.0813, "step": 32 }, { "epoch": 0.4258064516129032, "grad_norm": 0.5022628903388977, "learning_rate": 4.9993568806295786e-05, "loss": 0.0818, "step": 33 }, { "epoch": 0.43870967741935485, "grad_norm": 0.5096875429153442, "learning_rate": 4.9985530589693516e-05, "loss": 0.0844, "step": 34 }, { "epoch": 0.45161290322580644, "grad_norm": 0.6019471883773804, "learning_rate": 4.997427853400333e-05, "loss": 0.0903, "step": 35 }, { "epoch": 0.4645161290322581, "grad_norm": 0.4620855152606964, "learning_rate": 4.995981408655473e-05, "loss": 0.0772, "step": 36 }, { "epoch": 0.4774193548387097, "grad_norm": 0.46718630194664, "learning_rate": 4.9942139107880844e-05, "loss": 0.0786, "step": 37 }, { "epoch": 0.49032258064516127, "grad_norm": 0.6058530211448669, "learning_rate": 4.9921255871479e-05, "loss": 0.0948, "step": 38 }, { "epoch": 0.5032258064516129, "grad_norm": 0.5089570879936218, "learning_rate": 4.989716706351835e-05, "loss": 0.0805, "step": 39 }, { "epoch": 0.5161290322580645, "grad_norm": 0.4403476417064667, "learning_rate": 4.986987578249432e-05, "loss": 0.0899, "step": 40 }, { "epoch": 0.5290322580645161, "grad_norm": 0.640173077583313, "learning_rate": 4.9839385538830104e-05, "loss": 0.0849, "step": 41 }, { "epoch": 0.5419354838709678, "grad_norm": 0.5224989056587219, "learning_rate": 4.980570025442507e-05, "loss": 0.0845, "step": 42 }, { "epoch": 0.5548387096774193, "grad_norm": 0.5968887805938721, "learning_rate": 4.9768824262150344e-05, "loss": 0.0912, "step": 43 }, { "epoch": 0.567741935483871, "grad_norm": 0.5556889772415161, "learning_rate": 4.972876230529142e-05, "loss": 0.0907, "step": 44 }, { "epoch": 0.5806451612903226, "grad_norm": 0.4258240759372711, "learning_rate": 4.968551953693813e-05, "loss": 0.063, "step": 45 }, { "epoch": 0.5935483870967742, "grad_norm": 0.5726487636566162, "learning_rate": 4.963910151932171e-05, "loss": 0.0816, "step": 46 }, { "epoch": 0.6064516129032258, "grad_norm": 0.5164178609848022, "learning_rate": 4.958951422309942e-05, "loss": 0.0815, "step": 47 }, { "epoch": 0.6193548387096774, "grad_norm": 0.5187662839889526, "learning_rate": 4.95367640265865e-05, "loss": 0.0834, "step": 48 }, { "epoch": 0.632258064516129, "grad_norm": 0.4768002927303314, "learning_rate": 4.948085771493579e-05, "loss": 0.0818, "step": 49 }, { "epoch": 0.6451612903225806, "grad_norm": 1.3129404783248901, "learning_rate": 4.942180247926492e-05, "loss": 0.077, "step": 50 }, { "epoch": 0.6580645161290323, "grad_norm": 0.5702396631240845, "learning_rate": 4.935960591573135e-05, "loss": 0.0869, "step": 51 }, { "epoch": 0.6709677419354839, "grad_norm": 0.5151292085647583, "learning_rate": 4.929427602455532e-05, "loss": 0.091, "step": 52 }, { "epoch": 0.6838709677419355, "grad_norm": 0.4023990035057068, "learning_rate": 4.922582120899072e-05, "loss": 0.0772, "step": 53 }, { "epoch": 0.6967741935483871, "grad_norm": 0.5167486667633057, "learning_rate": 4.91542502742443e-05, "loss": 0.0815, "step": 54 }, { "epoch": 0.7096774193548387, "grad_norm": 0.4807862341403961, "learning_rate": 4.907957242634299e-05, "loss": 0.0819, "step": 55 }, { "epoch": 0.7225806451612903, "grad_norm": 0.5813907384872437, "learning_rate": 4.900179727094978e-05, "loss": 0.0852, "step": 56 }, { "epoch": 0.7354838709677419, "grad_norm": 0.44726839661598206, "learning_rate": 4.8920934812128164e-05, "loss": 0.0846, "step": 57 }, { "epoch": 0.7483870967741936, "grad_norm": 0.4237375557422638, "learning_rate": 4.8836995451055344e-05, "loss": 0.0918, "step": 58 }, { "epoch": 0.7612903225806451, "grad_norm": 0.3706034719944, "learning_rate": 4.874998998468433e-05, "loss": 0.0713, "step": 59 }, { "epoch": 0.7741935483870968, "grad_norm": 0.5643853545188904, "learning_rate": 4.865992960435514e-05, "loss": 0.0838, "step": 60 }, { "epoch": 0.7870967741935484, "grad_norm": 0.4395618140697479, "learning_rate": 4.856682589435531e-05, "loss": 0.0948, "step": 61 }, { "epoch": 0.8, "grad_norm": 0.4633603096008301, "learning_rate": 4.8470690830429823e-05, "loss": 0.076, "step": 62 }, { "epoch": 0.8129032258064516, "grad_norm": 0.44225528836250305, "learning_rate": 4.837153677824067e-05, "loss": 0.0887, "step": 63 }, { "epoch": 0.8258064516129032, "grad_norm": 0.3647634983062744, "learning_rate": 4.826937649177632e-05, "loss": 0.0752, "step": 64 }, { "epoch": 0.8387096774193549, "grad_norm": 0.4087846577167511, "learning_rate": 4.816422311171115e-05, "loss": 0.0701, "step": 65 }, { "epoch": 0.8516129032258064, "grad_norm": 0.3962288796901703, "learning_rate": 4.805609016371522e-05, "loss": 0.072, "step": 66 }, { "epoch": 0.864516129032258, "grad_norm": 0.35116565227508545, "learning_rate": 4.7944991556714496e-05, "loss": 0.0639, "step": 67 }, { "epoch": 0.8774193548387097, "grad_norm": 0.37532997131347656, "learning_rate": 4.783094158110174e-05, "loss": 0.0804, "step": 68 }, { "epoch": 0.8903225806451613, "grad_norm": 0.4580182433128357, "learning_rate": 4.7713954906898415e-05, "loss": 0.0883, "step": 69 }, { "epoch": 0.9032258064516129, "grad_norm": 0.46939095854759216, "learning_rate": 4.759404658186765e-05, "loss": 0.086, "step": 70 }, { "epoch": 0.9161290322580645, "grad_norm": 0.38567599654197693, "learning_rate": 4.747123202957872e-05, "loss": 0.0627, "step": 71 }, { "epoch": 0.9290322580645162, "grad_norm": 0.5291863083839417, "learning_rate": 4.7345527047423135e-05, "loss": 0.087, "step": 72 }, { "epoch": 0.9419354838709677, "grad_norm": 0.5343989133834839, "learning_rate": 4.721694780458266e-05, "loss": 0.0817, "step": 73 }, { "epoch": 0.9548387096774194, "grad_norm": 0.5286762118339539, "learning_rate": 4.7085510839949444e-05, "loss": 0.0824, "step": 74 }, { "epoch": 0.967741935483871, "grad_norm": 0.6843681931495667, "learning_rate": 4.695123305999877e-05, "loss": 0.0757, "step": 75 }, { "epoch": 0.9806451612903225, "grad_norm": 0.48566552996635437, "learning_rate": 4.681413173661429e-05, "loss": 0.0793, "step": 76 }, { "epoch": 0.9935483870967742, "grad_norm": 0.5594713091850281, "learning_rate": 4.667422450486646e-05, "loss": 0.0919, "step": 77 }, { "epoch": 1.0, "grad_norm": 0.5594713091850281, "learning_rate": 4.653152936074413e-05, "loss": 0.0812, "step": 78 }, { "epoch": 1.0129032258064516, "grad_norm": 0.6600771546363831, "learning_rate": 4.638606465883979e-05, "loss": 0.0528, "step": 79 }, { "epoch": 1.0258064516129033, "grad_norm": 0.4466298520565033, "learning_rate": 4.623784910998862e-05, "loss": 0.0562, "step": 80 }, { "epoch": 1.038709677419355, "grad_norm": 0.46932584047317505, "learning_rate": 4.608690177886177e-05, "loss": 0.052, "step": 81 }, { "epoch": 1.0516129032258064, "grad_norm": 0.3475061357021332, "learning_rate": 4.593324208151412e-05, "loss": 0.0477, "step": 82 }, { "epoch": 1.064516129032258, "grad_norm": 0.3392031788825989, "learning_rate": 4.577688978288681e-05, "loss": 0.0552, "step": 83 }, { "epoch": 1.0774193548387097, "grad_norm": 0.36944520473480225, "learning_rate": 4.5617864994264936e-05, "loss": 0.0623, "step": 84 }, { "epoch": 1.0903225806451613, "grad_norm": 0.3976334035396576, "learning_rate": 4.545618817069066e-05, "loss": 0.055, "step": 85 }, { "epoch": 1.103225806451613, "grad_norm": 0.3573772609233856, "learning_rate": 4.529188010833212e-05, "loss": 0.0652, "step": 86 }, { "epoch": 1.1161290322580646, "grad_norm": 0.37986183166503906, "learning_rate": 4.512496194180846e-05, "loss": 0.0506, "step": 87 }, { "epoch": 1.129032258064516, "grad_norm": 0.45936280488967896, "learning_rate": 4.495545514147134e-05, "loss": 0.0671, "step": 88 }, { "epoch": 1.1419354838709677, "grad_norm": 0.3391832113265991, "learning_rate": 4.478338151064323e-05, "loss": 0.0537, "step": 89 }, { "epoch": 1.1548387096774193, "grad_norm": 0.30911317467689514, "learning_rate": 4.460876318281291e-05, "loss": 0.0531, "step": 90 }, { "epoch": 1.167741935483871, "grad_norm": 0.4350668489933014, "learning_rate": 4.443162261878846e-05, "loss": 0.051, "step": 91 }, { "epoch": 1.1806451612903226, "grad_norm": 0.34030795097351074, "learning_rate": 4.425198260380818e-05, "loss": 0.0509, "step": 92 }, { "epoch": 1.1935483870967742, "grad_norm": 0.31266239285469055, "learning_rate": 4.406986624460979e-05, "loss": 0.0493, "step": 93 }, { "epoch": 1.206451612903226, "grad_norm": 0.3814278244972229, "learning_rate": 4.3885296966458224e-05, "loss": 0.0575, "step": 94 }, { "epoch": 1.2193548387096773, "grad_norm": 0.2861946225166321, "learning_rate": 4.369829851013251e-05, "loss": 0.0518, "step": 95 }, { "epoch": 1.232258064516129, "grad_norm": 0.35761791467666626, "learning_rate": 4.350889492887203e-05, "loss": 0.0509, "step": 96 }, { "epoch": 1.2451612903225806, "grad_norm": 0.321609765291214, "learning_rate": 4.3317110585282595e-05, "loss": 0.0468, "step": 97 }, { "epoch": 1.2580645161290323, "grad_norm": 0.33591514825820923, "learning_rate": 4.3122970148202744e-05, "loss": 0.055, "step": 98 }, { "epoch": 1.270967741935484, "grad_norm": 0.38554760813713074, "learning_rate": 4.292649858953063e-05, "loss": 0.07, "step": 99 }, { "epoch": 1.2838709677419355, "grad_norm": 0.5272489786148071, "learning_rate": 4.272772118101195e-05, "loss": 0.0564, "step": 100 }, { "epoch": 1.2967741935483872, "grad_norm": 0.4388972818851471, "learning_rate": 4.252666349098926e-05, "loss": 0.0536, "step": 101 }, { "epoch": 1.3096774193548386, "grad_norm": 0.33931440114974976, "learning_rate": 4.2323351381113206e-05, "loss": 0.0472, "step": 102 }, { "epoch": 1.3225806451612903, "grad_norm": 0.2954395115375519, "learning_rate": 4.211781100301596e-05, "loss": 0.0645, "step": 103 }, { "epoch": 1.335483870967742, "grad_norm": 0.5045889019966125, "learning_rate": 4.1910068794947404e-05, "loss": 0.056, "step": 104 }, { "epoch": 1.3483870967741935, "grad_norm": 0.3853508532047272, "learning_rate": 4.170015147837444e-05, "loss": 0.0518, "step": 105 }, { "epoch": 1.3612903225806452, "grad_norm": 0.27302005887031555, "learning_rate": 4.148808605454385e-05, "loss": 0.0487, "step": 106 }, { "epoch": 1.3741935483870968, "grad_norm": 0.41000276803970337, "learning_rate": 4.127389980100915e-05, "loss": 0.0528, "step": 107 }, { "epoch": 1.3870967741935485, "grad_norm": 0.3967251479625702, "learning_rate": 4.1057620268122004e-05, "loss": 0.0485, "step": 108 }, { "epoch": 1.4, "grad_norm": 0.30169478058815, "learning_rate": 4.08392752754884e-05, "loss": 0.0485, "step": 109 }, { "epoch": 1.4129032258064516, "grad_norm": 0.3874256908893585, "learning_rate": 4.0618892908390316e-05, "loss": 0.0514, "step": 110 }, { "epoch": 1.4258064516129032, "grad_norm": 0.29030513763427734, "learning_rate": 4.039650151417316e-05, "loss": 0.0442, "step": 111 }, { "epoch": 1.4387096774193548, "grad_norm": 0.3206923007965088, "learning_rate": 4.0172129698599484e-05, "loss": 0.052, "step": 112 }, { "epoch": 1.4516129032258065, "grad_norm": 0.3883202075958252, "learning_rate": 3.994580632216952e-05, "loss": 0.057, "step": 113 }, { "epoch": 1.4645161290322581, "grad_norm": 0.38505852222442627, "learning_rate": 3.971756049640888e-05, "loss": 0.0515, "step": 114 }, { "epoch": 1.4774193548387098, "grad_norm": 0.27141115069389343, "learning_rate": 3.9487421580124e-05, "loss": 0.0524, "step": 115 }, { "epoch": 1.4903225806451612, "grad_norm": 0.4098750054836273, "learning_rate": 3.92554191756258e-05, "loss": 0.0544, "step": 116 }, { "epoch": 1.5032258064516129, "grad_norm": 0.3381025791168213, "learning_rate": 3.902158312492196e-05, "loss": 0.0468, "step": 117 }, { "epoch": 1.5161290322580645, "grad_norm": 0.3622021973133087, "learning_rate": 3.878594350587844e-05, "loss": 0.0451, "step": 118 }, { "epoch": 1.5290322580645161, "grad_norm": 0.31938228011131287, "learning_rate": 3.8548530628350623e-05, "loss": 0.0509, "step": 119 }, { "epoch": 1.5419354838709678, "grad_norm": 0.3087027370929718, "learning_rate": 3.8309375030284566e-05, "loss": 0.0616, "step": 120 }, { "epoch": 1.5548387096774192, "grad_norm": 0.36196205019950867, "learning_rate": 3.806850747378902e-05, "loss": 0.0514, "step": 121 }, { "epoch": 1.567741935483871, "grad_norm": 0.31175723671913147, "learning_rate": 3.782595894117854e-05, "loss": 0.0478, "step": 122 }, { "epoch": 1.5806451612903225, "grad_norm": 0.3437715768814087, "learning_rate": 3.758176063098829e-05, "loss": 0.0489, "step": 123 }, { "epoch": 1.5935483870967742, "grad_norm": 0.3346206247806549, "learning_rate": 3.733594395396106e-05, "loss": 0.0516, "step": 124 }, { "epoch": 1.6064516129032258, "grad_norm": 0.38583481311798096, "learning_rate": 3.7088540529006924e-05, "loss": 0.0461, "step": 125 }, { "epoch": 1.6193548387096774, "grad_norm": 0.25379714369773865, "learning_rate": 3.68395821791362e-05, "loss": 0.0414, "step": 126 }, { "epoch": 1.632258064516129, "grad_norm": 0.42217546701431274, "learning_rate": 3.65891009273661e-05, "loss": 0.0556, "step": 127 }, { "epoch": 1.6451612903225805, "grad_norm": 0.36978331208229065, "learning_rate": 3.633712899260166e-05, "loss": 0.0575, "step": 128 }, { "epoch": 1.6580645161290324, "grad_norm": 0.3139955401420593, "learning_rate": 3.608369878549154e-05, "loss": 0.0451, "step": 129 }, { "epoch": 1.6709677419354838, "grad_norm": 0.3436100482940674, "learning_rate": 3.582884290425903e-05, "loss": 0.0522, "step": 130 }, { "epoch": 1.6838709677419355, "grad_norm": 0.2942555248737335, "learning_rate": 3.557259413050907e-05, "loss": 0.049, "step": 131 }, { "epoch": 1.696774193548387, "grad_norm": 0.32971078157424927, "learning_rate": 3.531498542501161e-05, "loss": 0.0495, "step": 132 }, { "epoch": 1.7096774193548387, "grad_norm": 0.2842581272125244, "learning_rate": 3.505604992346187e-05, "loss": 0.0494, "step": 133 }, { "epoch": 1.7225806451612904, "grad_norm": 0.37220752239227295, "learning_rate": 3.4795820932218235e-05, "loss": 0.0489, "step": 134 }, { "epoch": 1.7354838709677418, "grad_norm": 0.3396855592727661, "learning_rate": 3.4534331924018095e-05, "loss": 0.058, "step": 135 }, { "epoch": 1.7483870967741937, "grad_norm": 0.30913567543029785, "learning_rate": 3.427161653367231e-05, "loss": 0.0505, "step": 136 }, { "epoch": 1.761290322580645, "grad_norm": 0.3557908535003662, "learning_rate": 3.4007708553738836e-05, "loss": 0.0454, "step": 137 }, { "epoch": 1.7741935483870968, "grad_norm": 0.32114386558532715, "learning_rate": 3.3742641930176045e-05, "loss": 0.0462, "step": 138 }, { "epoch": 1.7870967741935484, "grad_norm": 0.264426589012146, "learning_rate": 3.347645075797635e-05, "loss": 0.0436, "step": 139 }, { "epoch": 1.8, "grad_norm": 0.3021955192089081, "learning_rate": 3.320916927678061e-05, "loss": 0.0386, "step": 140 }, { "epoch": 1.8129032258064517, "grad_norm": 0.35488826036453247, "learning_rate": 3.2940831866473966e-05, "loss": 0.0462, "step": 141 }, { "epoch": 1.8258064516129031, "grad_norm": 0.2876163125038147, "learning_rate": 3.2671473042763604e-05, "loss": 0.0545, "step": 142 }, { "epoch": 1.838709677419355, "grad_norm": 0.3992978036403656, "learning_rate": 3.240112745273911e-05, "loss": 0.0434, "step": 143 }, { "epoch": 1.8516129032258064, "grad_norm": 0.3140002489089966, "learning_rate": 3.212982987041582e-05, "loss": 0.0459, "step": 144 }, { "epoch": 1.864516129032258, "grad_norm": 0.357721209526062, "learning_rate": 3.1857615192261984e-05, "loss": 0.0548, "step": 145 }, { "epoch": 1.8774193548387097, "grad_norm": 0.32279306650161743, "learning_rate": 3.158451843271005e-05, "loss": 0.0593, "step": 146 }, { "epoch": 1.8903225806451613, "grad_norm": 0.3754781484603882, "learning_rate": 3.131057471965283e-05, "loss": 0.0508, "step": 147 }, { "epoch": 1.903225806451613, "grad_norm": 0.42713454365730286, "learning_rate": 3.10358192899251e-05, "loss": 0.0423, "step": 148 }, { "epoch": 1.9161290322580644, "grad_norm": 0.28894898295402527, "learning_rate": 3.076028748477112e-05, "loss": 0.0442, "step": 149 }, { "epoch": 1.9290322580645163, "grad_norm": 0.32115331292152405, "learning_rate": 3.0484014745298784e-05, "loss": 0.0571, "step": 150 }, { "epoch": 1.9419354838709677, "grad_norm": 0.36180686950683594, "learning_rate": 3.0207036607920892e-05, "loss": 0.0498, "step": 151 }, { "epoch": 1.9548387096774194, "grad_norm": 0.36315590143203735, "learning_rate": 2.9929388699784163e-05, "loss": 0.0548, "step": 152 }, { "epoch": 1.967741935483871, "grad_norm": 0.24509944021701813, "learning_rate": 2.9651106734186618e-05, "loss": 0.0473, "step": 153 }, { "epoch": 1.9806451612903224, "grad_norm": 0.30309274792671204, "learning_rate": 2.9372226505983802e-05, "loss": 0.0515, "step": 154 }, { "epoch": 1.9935483870967743, "grad_norm": 0.31110507249832153, "learning_rate": 2.9092783886984615e-05, "loss": 0.049, "step": 155 }, { "epoch": 2.0, "grad_norm": 0.5940459966659546, "learning_rate": 2.8812814821337147e-05, "loss": 0.0358, "step": 156 }, { "epoch": 2.0129032258064514, "grad_norm": 0.23582255840301514, "learning_rate": 2.8532355320905274e-05, "loss": 0.0252, "step": 157 }, { "epoch": 2.0258064516129033, "grad_norm": 0.2490931898355484, "learning_rate": 2.825144146063654e-05, "loss": 0.0242, "step": 158 }, { "epoch": 2.0387096774193547, "grad_norm": 0.300824373960495, "learning_rate": 2.7970109373921878e-05, "loss": 0.0278, "step": 159 }, { "epoch": 2.0516129032258066, "grad_norm": 0.2460525780916214, "learning_rate": 2.7688395247947842e-05, "loss": 0.0292, "step": 160 }, { "epoch": 2.064516129032258, "grad_norm": 0.2941141724586487, "learning_rate": 2.740633531904196e-05, "loss": 0.0257, "step": 161 }, { "epoch": 2.07741935483871, "grad_norm": 0.27139392495155334, "learning_rate": 2.712396586801169e-05, "loss": 0.0246, "step": 162 }, { "epoch": 2.0903225806451613, "grad_norm": 0.3047543168067932, "learning_rate": 2.6841323215477714e-05, "loss": 0.0313, "step": 163 }, { "epoch": 2.1032258064516127, "grad_norm": 0.3697696924209595, "learning_rate": 2.6558443717202076e-05, "loss": 0.0311, "step": 164 }, { "epoch": 2.1161290322580646, "grad_norm": 0.4939499497413635, "learning_rate": 2.6275363759411815e-05, "loss": 0.036, "step": 165 }, { "epoch": 2.129032258064516, "grad_norm": 0.2502232789993286, "learning_rate": 2.599211975411867e-05, "loss": 0.0194, "step": 166 }, { "epoch": 2.141935483870968, "grad_norm": 0.2647090554237366, "learning_rate": 2.57087481344355e-05, "loss": 0.0259, "step": 167 }, { "epoch": 2.1548387096774193, "grad_norm": 0.2818278968334198, "learning_rate": 2.5425285349889938e-05, "loss": 0.0239, "step": 168 }, { "epoch": 2.167741935483871, "grad_norm": 0.28670820593833923, "learning_rate": 2.5141767861735976e-05, "loss": 0.0248, "step": 169 }, { "epoch": 2.1806451612903226, "grad_norm": 0.2521524727344513, "learning_rate": 2.485823213826403e-05, "loss": 0.0259, "step": 170 }, { "epoch": 2.193548387096774, "grad_norm": 0.2689652740955353, "learning_rate": 2.457471465011007e-05, "loss": 0.026, "step": 171 }, { "epoch": 2.206451612903226, "grad_norm": 0.24064841866493225, "learning_rate": 2.4291251865564505e-05, "loss": 0.0252, "step": 172 }, { "epoch": 2.2193548387096773, "grad_norm": 0.26753664016723633, "learning_rate": 2.400788024588134e-05, "loss": 0.0236, "step": 173 }, { "epoch": 2.232258064516129, "grad_norm": 0.30342286825180054, "learning_rate": 2.3724636240588194e-05, "loss": 0.0346, "step": 174 }, { "epoch": 2.2451612903225806, "grad_norm": 0.36332225799560547, "learning_rate": 2.3441556282797934e-05, "loss": 0.0351, "step": 175 }, { "epoch": 2.258064516129032, "grad_norm": 0.2872660458087921, "learning_rate": 2.315867678452229e-05, "loss": 0.0257, "step": 176 }, { "epoch": 2.270967741935484, "grad_norm": 0.33576029539108276, "learning_rate": 2.287603413198831e-05, "loss": 0.0307, "step": 177 }, { "epoch": 2.2838709677419353, "grad_norm": 0.42973652482032776, "learning_rate": 2.2593664680958045e-05, "loss": 0.0325, "step": 178 }, { "epoch": 2.296774193548387, "grad_norm": 0.28366559743881226, "learning_rate": 2.231160475205216e-05, "loss": 0.0279, "step": 179 }, { "epoch": 2.3096774193548386, "grad_norm": 0.23520596325397491, "learning_rate": 2.202989062607813e-05, "loss": 0.0242, "step": 180 }, { "epoch": 2.3225806451612905, "grad_norm": 0.23934811353683472, "learning_rate": 2.1748558539363458e-05, "loss": 0.0213, "step": 181 }, { "epoch": 2.335483870967742, "grad_norm": 0.2189173400402069, "learning_rate": 2.1467644679094728e-05, "loss": 0.0202, "step": 182 }, { "epoch": 2.3483870967741938, "grad_norm": 0.42822569608688354, "learning_rate": 2.118718517866286e-05, "loss": 0.0361, "step": 183 }, { "epoch": 2.361290322580645, "grad_norm": 0.2864365875720978, "learning_rate": 2.0907216113015397e-05, "loss": 0.0303, "step": 184 }, { "epoch": 2.3741935483870966, "grad_norm": 0.24325411021709442, "learning_rate": 2.0627773494016207e-05, "loss": 0.0265, "step": 185 }, { "epoch": 2.3870967741935485, "grad_norm": 0.23324590921401978, "learning_rate": 2.0348893265813394e-05, "loss": 0.0233, "step": 186 }, { "epoch": 2.4, "grad_norm": 0.2864069640636444, "learning_rate": 2.0070611300215843e-05, "loss": 0.0259, "step": 187 }, { "epoch": 2.412903225806452, "grad_norm": 0.2672640085220337, "learning_rate": 1.979296339207912e-05, "loss": 0.0267, "step": 188 }, { "epoch": 2.425806451612903, "grad_norm": 0.22086627781391144, "learning_rate": 1.951598525470122e-05, "loss": 0.02, "step": 189 }, { "epoch": 2.4387096774193546, "grad_norm": 0.24434693157672882, "learning_rate": 1.923971251522888e-05, "loss": 0.0235, "step": 190 }, { "epoch": 2.4516129032258065, "grad_norm": 0.2901481091976166, "learning_rate": 1.8964180710074905e-05, "loss": 0.0242, "step": 191 }, { "epoch": 2.464516129032258, "grad_norm": 0.2797795236110687, "learning_rate": 1.8689425280347166e-05, "loss": 0.0246, "step": 192 }, { "epoch": 2.47741935483871, "grad_norm": 0.22296951711177826, "learning_rate": 1.8415481567289954e-05, "loss": 0.0223, "step": 193 }, { "epoch": 2.490322580645161, "grad_norm": 0.2699308395385742, "learning_rate": 1.8142384807738022e-05, "loss": 0.027, "step": 194 }, { "epoch": 2.5032258064516126, "grad_norm": 0.23768910765647888, "learning_rate": 1.787017012958419e-05, "loss": 0.0241, "step": 195 }, { "epoch": 2.5161290322580645, "grad_norm": 0.2236274927854538, "learning_rate": 1.75988725472609e-05, "loss": 0.0241, "step": 196 }, { "epoch": 2.5290322580645164, "grad_norm": 0.31210729479789734, "learning_rate": 1.7328526957236406e-05, "loss": 0.0302, "step": 197 }, { "epoch": 2.541935483870968, "grad_norm": 0.22504620254039764, "learning_rate": 1.7059168133526043e-05, "loss": 0.0236, "step": 198 }, { "epoch": 2.554838709677419, "grad_norm": 0.24498535692691803, "learning_rate": 1.6790830723219398e-05, "loss": 0.0228, "step": 199 }, { "epoch": 2.567741935483871, "grad_norm": 0.2544174790382385, "learning_rate": 1.6523549242023655e-05, "loss": 0.027, "step": 200 }, { "epoch": 2.5806451612903225, "grad_norm": 0.2597277760505676, "learning_rate": 1.6257358069823965e-05, "loss": 0.0304, "step": 201 }, { "epoch": 2.5935483870967744, "grad_norm": 0.2677987515926361, "learning_rate": 1.5992291446261166e-05, "loss": 0.0254, "step": 202 }, { "epoch": 2.606451612903226, "grad_norm": 0.27171266078948975, "learning_rate": 1.5728383466327684e-05, "loss": 0.0222, "step": 203 }, { "epoch": 2.6193548387096772, "grad_norm": 0.2346351146697998, "learning_rate": 1.5465668075981904e-05, "loss": 0.0219, "step": 204 }, { "epoch": 2.632258064516129, "grad_norm": 0.25438180565834045, "learning_rate": 1.520417906778176e-05, "loss": 0.0264, "step": 205 }, { "epoch": 2.6451612903225805, "grad_norm": 0.2690618336200714, "learning_rate": 1.4943950076538135e-05, "loss": 0.0237, "step": 206 }, { "epoch": 2.6580645161290324, "grad_norm": 0.29505690932273865, "learning_rate": 1.4685014574988393e-05, "loss": 0.0313, "step": 207 }, { "epoch": 2.670967741935484, "grad_norm": 0.26153627038002014, "learning_rate": 1.4427405869490923e-05, "loss": 0.0236, "step": 208 }, { "epoch": 2.6838709677419352, "grad_norm": 0.2462136447429657, "learning_rate": 1.4171157095740976e-05, "loss": 0.027, "step": 209 }, { "epoch": 2.696774193548387, "grad_norm": 0.33909931778907776, "learning_rate": 1.391630121450847e-05, "loss": 0.0257, "step": 210 }, { "epoch": 2.709677419354839, "grad_norm": 0.2736709415912628, "learning_rate": 1.3662871007398348e-05, "loss": 0.0274, "step": 211 }, { "epoch": 2.7225806451612904, "grad_norm": 0.3460533022880554, "learning_rate": 1.3410899072633915e-05, "loss": 0.0292, "step": 212 }, { "epoch": 2.735483870967742, "grad_norm": 0.28637826442718506, "learning_rate": 1.3160417820863807e-05, "loss": 0.0249, "step": 213 }, { "epoch": 2.7483870967741937, "grad_norm": 0.21176108717918396, "learning_rate": 1.2911459470993085e-05, "loss": 0.019, "step": 214 }, { "epoch": 2.761290322580645, "grad_norm": 0.22695419192314148, "learning_rate": 1.2664056046038942e-05, "loss": 0.0248, "step": 215 }, { "epoch": 2.774193548387097, "grad_norm": 0.3826678693294525, "learning_rate": 1.2418239369011712e-05, "loss": 0.0206, "step": 216 }, { "epoch": 2.7870967741935484, "grad_norm": 0.21120679378509521, "learning_rate": 1.217404105882147e-05, "loss": 0.02, "step": 217 }, { "epoch": 2.8, "grad_norm": 0.266303151845932, "learning_rate": 1.1931492526210988e-05, "loss": 0.0217, "step": 218 }, { "epoch": 2.8129032258064517, "grad_norm": 0.2666541635990143, "learning_rate": 1.1690624969715441e-05, "loss": 0.0248, "step": 219 }, { "epoch": 2.825806451612903, "grad_norm": 0.2325655072927475, "learning_rate": 1.145146937164938e-05, "loss": 0.0226, "step": 220 }, { "epoch": 2.838709677419355, "grad_norm": 0.277413934469223, "learning_rate": 1.121405649412156e-05, "loss": 0.0272, "step": 221 }, { "epoch": 2.8516129032258064, "grad_norm": 0.23814360797405243, "learning_rate": 1.0978416875078042e-05, "loss": 0.0195, "step": 222 }, { "epoch": 2.864516129032258, "grad_norm": 0.3257719576358795, "learning_rate": 1.0744580824374217e-05, "loss": 0.0336, "step": 223 }, { "epoch": 2.8774193548387097, "grad_norm": 0.24910488724708557, "learning_rate": 1.0512578419876004e-05, "loss": 0.0236, "step": 224 }, { "epoch": 2.8903225806451616, "grad_norm": 0.27725309133529663, "learning_rate": 1.0282439503591135e-05, "loss": 0.0272, "step": 225 }, { "epoch": 2.903225806451613, "grad_norm": 0.23216702044010162, "learning_rate": 1.0054193677830481e-05, "loss": 0.0244, "step": 226 }, { "epoch": 2.9161290322580644, "grad_norm": 0.27499204874038696, "learning_rate": 9.827870301400527e-06, "loss": 0.0214, "step": 227 }, { "epoch": 2.9290322580645163, "grad_norm": 0.2503328323364258, "learning_rate": 9.603498485826848e-06, "loss": 0.0223, "step": 228 }, { "epoch": 2.9419354838709677, "grad_norm": 0.23868584632873535, "learning_rate": 9.38110709160969e-06, "loss": 0.0227, "step": 229 }, { "epoch": 2.9548387096774196, "grad_norm": 0.3298948407173157, "learning_rate": 9.160724724511608e-06, "loss": 0.0288, "step": 230 }, { "epoch": 2.967741935483871, "grad_norm": 0.23431891202926636, "learning_rate": 8.942379731877992e-06, "loss": 0.023, "step": 231 }, { "epoch": 2.9806451612903224, "grad_norm": 0.2671021819114685, "learning_rate": 8.726100198990853e-06, "loss": 0.0213, "step": 232 }, { "epoch": 2.9935483870967743, "grad_norm": 0.29010385274887085, "learning_rate": 8.51191394545615e-06, "loss": 0.0314, "step": 233 }, { "epoch": 3.0, "grad_norm": 0.29010385274887085, "learning_rate": 8.299848521625563e-06, "loss": 0.0159, "step": 234 }, { "epoch": 3.0129032258064514, "grad_norm": 0.28189539909362793, "learning_rate": 8.089931205052597e-06, "loss": 0.0106, "step": 235 }, { "epoch": 3.0258064516129033, "grad_norm": 0.1651533842086792, "learning_rate": 7.882188996984046e-06, "loss": 0.0122, "step": 236 }, { "epoch": 3.0387096774193547, "grad_norm": 0.1792869120836258, "learning_rate": 7.676648618886798e-06, "loss": 0.0129, "step": 237 }, { "epoch": 3.0516129032258066, "grad_norm": 0.16915558278560638, "learning_rate": 7.473336509010742e-06, "loss": 0.0099, "step": 238 }, { "epoch": 3.064516129032258, "grad_norm": 0.15584175288677216, "learning_rate": 7.2722788189880545e-06, "loss": 0.0123, "step": 239 }, { "epoch": 3.07741935483871, "grad_norm": 0.1726021021604538, "learning_rate": 7.073501410469371e-06, "loss": 0.0114, "step": 240 }, { "epoch": 3.0903225806451613, "grad_norm": 0.15609051287174225, "learning_rate": 6.877029851797265e-06, "loss": 0.0091, "step": 241 }, { "epoch": 3.1032258064516127, "grad_norm": 0.14097200334072113, "learning_rate": 6.6828894147174165e-06, "loss": 0.0084, "step": 242 }, { "epoch": 3.1161290322580646, "grad_norm": 0.17083188891410828, "learning_rate": 6.491105071127984e-06, "loss": 0.0118, "step": 243 }, { "epoch": 3.129032258064516, "grad_norm": 0.18476150929927826, "learning_rate": 6.3017014898674955e-06, "loss": 0.0079, "step": 244 }, { "epoch": 3.141935483870968, "grad_norm": 0.22733564674854279, "learning_rate": 6.114703033541783e-06, "loss": 0.013, "step": 245 }, { "epoch": 3.1548387096774193, "grad_norm": 0.15913838148117065, "learning_rate": 5.930133755390216e-06, "loss": 0.0068, "step": 246 }, { "epoch": 3.167741935483871, "grad_norm": 0.1369076371192932, "learning_rate": 5.7480173961918266e-06, "loss": 0.0077, "step": 247 }, { "epoch": 3.1806451612903226, "grad_norm": 0.2318694293498993, "learning_rate": 5.568377381211548e-06, "loss": 0.0095, "step": 248 }, { "epoch": 3.193548387096774, "grad_norm": 0.21019090712070465, "learning_rate": 5.391236817187095e-06, "loss": 0.0095, "step": 249 }, { "epoch": 3.206451612903226, "grad_norm": 0.1967868059873581, "learning_rate": 5.216618489356773e-06, "loss": 0.0111, "step": 250 }, { "epoch": 3.2193548387096773, "grad_norm": 0.24731586873531342, "learning_rate": 5.0445448585286684e-06, "loss": 0.0072, "step": 251 }, { "epoch": 3.232258064516129, "grad_norm": 0.19581767916679382, "learning_rate": 4.87503805819155e-06, "loss": 0.0055, "step": 252 }, { "epoch": 3.2451612903225806, "grad_norm": 0.1791073977947235, "learning_rate": 4.708119891667892e-06, "loss": 0.0088, "step": 253 }, { "epoch": 3.258064516129032, "grad_norm": 0.25945040583610535, "learning_rate": 4.5438118293093416e-06, "loss": 0.0071, "step": 254 }, { "epoch": 3.270967741935484, "grad_norm": 0.21385855972766876, "learning_rate": 4.382135005735072e-06, "loss": 0.0173, "step": 255 }, { "epoch": 3.2838709677419353, "grad_norm": 0.5789936780929565, "learning_rate": 4.223110217113191e-06, "loss": 0.0079, "step": 256 }, { "epoch": 3.296774193548387, "grad_norm": 0.23074735701084137, "learning_rate": 4.066757918485886e-06, "loss": 0.0069, "step": 257 }, { "epoch": 3.3096774193548386, "grad_norm": 0.20934291183948517, "learning_rate": 3.913098221138237e-06, "loss": 0.0088, "step": 258 }, { "epoch": 3.3225806451612905, "grad_norm": 0.2551226019859314, "learning_rate": 3.762150890011387e-06, "loss": 0.0104, "step": 259 }, { "epoch": 3.335483870967742, "grad_norm": 0.22543351352214813, "learning_rate": 3.6139353411602154e-06, "loss": 0.009, "step": 260 }, { "epoch": 3.3483870967741938, "grad_norm": 0.22513934969902039, "learning_rate": 3.468470639255872e-06, "loss": 0.0103, "step": 261 }, { "epoch": 3.361290322580645, "grad_norm": 0.23426640033721924, "learning_rate": 3.325775495133546e-06, "loss": 0.0086, "step": 262 }, { "epoch": 3.3741935483870966, "grad_norm": 0.17604398727416992, "learning_rate": 3.1858682633857105e-06, "loss": 0.0118, "step": 263 }, { "epoch": 3.3870967741935485, "grad_norm": 0.22662527859210968, "learning_rate": 3.0487669400012382e-06, "loss": 0.0089, "step": 264 }, { "epoch": 3.4, "grad_norm": 0.1959221363067627, "learning_rate": 2.914489160050554e-06, "loss": 0.0101, "step": 265 }, { "epoch": 3.412903225806452, "grad_norm": 0.20400796830654144, "learning_rate": 2.7830521954173543e-06, "loss": 0.0088, "step": 266 }, { "epoch": 3.425806451612903, "grad_norm": 0.21109241247177124, "learning_rate": 2.654472952576864e-06, "loss": 0.0083, "step": 267 }, { "epoch": 3.4387096774193546, "grad_norm": 0.20984111726284027, "learning_rate": 2.5287679704212836e-06, "loss": 0.0091, "step": 268 }, { "epoch": 3.4516129032258065, "grad_norm": 0.22513136267662048, "learning_rate": 2.4059534181323588e-06, "loss": 0.0102, "step": 269 }, { "epoch": 3.464516129032258, "grad_norm": 0.23501525819301605, "learning_rate": 2.2860450931015853e-06, "loss": 0.0095, "step": 270 }, { "epoch": 3.47741935483871, "grad_norm": 0.19732145965099335, "learning_rate": 2.1690584188982592e-06, "loss": 0.0087, "step": 271 }, { "epoch": 3.490322580645161, "grad_norm": 0.19582946598529816, "learning_rate": 2.0550084432855056e-06, "loss": 0.0074, "step": 272 }, { "epoch": 3.5032258064516126, "grad_norm": 0.19379809498786926, "learning_rate": 1.9439098362847825e-06, "loss": 0.008, "step": 273 }, { "epoch": 3.5161290322580645, "grad_norm": 0.20334380865097046, "learning_rate": 1.835776888288851e-06, "loss": 0.0104, "step": 274 }, { "epoch": 3.5290322580645164, "grad_norm": 0.1963813155889511, "learning_rate": 1.730623508223686e-06, "loss": 0.0114, "step": 275 }, { "epoch": 3.541935483870968, "grad_norm": 0.2684271037578583, "learning_rate": 1.62846322175933e-06, "loss": 0.0106, "step": 276 }, { "epoch": 3.554838709677419, "grad_norm": 0.19259046018123627, "learning_rate": 1.5293091695701839e-06, "loss": 0.0083, "step": 277 }, { "epoch": 3.567741935483871, "grad_norm": 0.19738364219665527, "learning_rate": 1.4331741056446968e-06, "loss": 0.0166, "step": 278 }, { "epoch": 3.5806451612903225, "grad_norm": 0.22224943339824677, "learning_rate": 1.3400703956448684e-06, "loss": 0.0074, "step": 279 }, { "epoch": 3.5935483870967744, "grad_norm": 0.23735593259334564, "learning_rate": 1.2500100153156779e-06, "loss": 0.0091, "step": 280 }, { "epoch": 3.606451612903226, "grad_norm": 0.1716444343328476, "learning_rate": 1.1630045489446572e-06, "loss": 0.0078, "step": 281 }, { "epoch": 3.6193548387096772, "grad_norm": 0.1820307821035385, "learning_rate": 1.079065187871839e-06, "loss": 0.009, "step": 282 }, { "epoch": 3.632258064516129, "grad_norm": 0.1972561627626419, "learning_rate": 9.982027290502238e-07, "loss": 0.0122, "step": 283 }, { "epoch": 3.6451612903225805, "grad_norm": 0.21676687896251678, "learning_rate": 9.204275736570151e-07, "loss": 0.0082, "step": 284 }, { "epoch": 3.6580645161290324, "grad_norm": 0.1852135956287384, "learning_rate": 8.45749725755704e-07, "loss": 0.0078, "step": 285 }, { "epoch": 3.670967741935484, "grad_norm": 0.15526923537254333, "learning_rate": 7.741787910092807e-07, "loss": 0.0058, "step": 286 }, { "epoch": 3.6838709677419352, "grad_norm": 0.25308313965797424, "learning_rate": 7.05723975444686e-07, "loss": 0.0072, "step": 287 }, { "epoch": 3.696774193548387, "grad_norm": 0.16028578579425812, "learning_rate": 6.403940842686474e-07, "loss": 0.0086, "step": 288 }, { "epoch": 3.709677419354839, "grad_norm": 0.164954274892807, "learning_rate": 5.781975207350826e-07, "loss": 0.0074, "step": 289 }, { "epoch": 3.7225806451612904, "grad_norm": 0.21616578102111816, "learning_rate": 5.191422850642113e-07, "loss": 0.0111, "step": 290 }, { "epoch": 3.735483870967742, "grad_norm": 0.23164188861846924, "learning_rate": 4.6323597341350054e-07, "loss": 0.0093, "step": 291 }, { "epoch": 3.7483870967741937, "grad_norm": 0.2091163843870163, "learning_rate": 4.1048577690058766e-07, "loss": 0.008, "step": 292 }, { "epoch": 3.761290322580645, "grad_norm": 0.17184507846832275, "learning_rate": 3.608984806782928e-07, "loss": 0.0089, "step": 293 }, { "epoch": 3.774193548387097, "grad_norm": 0.2208150029182434, "learning_rate": 3.144804630618725e-07, "loss": 0.0098, "step": 294 }, { "epoch": 3.7870967741935484, "grad_norm": 0.18591627478599548, "learning_rate": 2.71237694708576e-07, "loss": 0.0091, "step": 295 }, { "epoch": 3.8, "grad_norm": 0.22179189324378967, "learning_rate": 2.3117573784966206e-07, "loss": 0.0096, "step": 296 }, { "epoch": 3.8129032258064517, "grad_norm": 0.247935950756073, "learning_rate": 1.9429974557493014e-07, "loss": 0.0075, "step": 297 }, { "epoch": 3.825806451612903, "grad_norm": 0.19308875501155853, "learning_rate": 1.6061446116990008e-07, "loss": 0.0097, "step": 298 }, { "epoch": 3.838709677419355, "grad_norm": 0.28245624899864197, "learning_rate": 1.3012421750568105e-07, "loss": 0.0105, "step": 299 }, { "epoch": 3.8516129032258064, "grad_norm": 0.15781083703041077, "learning_rate": 1.0283293648165605e-07, "loss": 0.0087, "step": 300 }, { "epoch": 3.864516129032258, "grad_norm": 0.20132358372211456, "learning_rate": 7.874412852099944e-08, "loss": 0.0081, "step": 301 }, { "epoch": 3.8774193548387097, "grad_norm": 0.19697970151901245, "learning_rate": 5.786089211915202e-08, "loss": 0.0116, "step": 302 }, { "epoch": 3.8903225806451616, "grad_norm": 0.189448282122612, "learning_rate": 4.018591344526479e-08, "loss": 0.0079, "step": 303 }, { "epoch": 3.903225806451613, "grad_norm": 0.1814439296722412, "learning_rate": 2.5721465996675355e-08, "loss": 0.0061, "step": 304 }, { "epoch": 3.9161290322580644, "grad_norm": 0.17018909752368927, "learning_rate": 1.4469410306480746e-08, "loss": 0.0074, "step": 305 }, { "epoch": 3.9290322580645163, "grad_norm": 0.2005801647901535, "learning_rate": 6.431193704217741e-09, "loss": 0.0084, "step": 306 }, { "epoch": 3.9419354838709677, "grad_norm": 0.2100239396095276, "learning_rate": 1.6078501296951099e-09, "loss": 0.0118, "step": 307 }, { "epoch": 3.9548387096774196, "grad_norm": 0.17311210930347443, "learning_rate": 0.0, "loss": 0.0059, "step": 308 }, { "epoch": 3.9548387096774196, "step": 308, "total_flos": 5.413829559357276e+17, "train_loss": 0.04737277891541311, "train_runtime": 1301.7465, "train_samples_per_second": 15.216, "train_steps_per_second": 0.237 } ], "logging_steps": 1, "max_steps": 308, "num_input_tokens_seen": 0, "num_train_epochs": 4, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.413829559357276e+17, "train_batch_size": 4, "trial_name": null, "trial_params": null }