diff --git "a/trainer_state.json" "b/trainer_state.json" new file mode 100644--- /dev/null +++ "b/trainer_state.json" @@ -0,0 +1,4354 @@ +{ + "best_metric": null, + "best_model_checkpoint": null, + "epoch": 3.975767366720517, + "eval_steps": 500, + "global_step": 616, + "is_hyper_param_search": false, + "is_local_process_zero": true, + "is_world_process_zero": true, + "log_history": [ + { + "epoch": 0.006462035541195477, + "grad_norm": 15.905838012695312, + "learning_rate": 1.6129032258064518e-07, + "loss": 0.1599, + "step": 1 + }, + { + "epoch": 0.012924071082390954, + "grad_norm": 27.293901443481445, + "learning_rate": 3.2258064516129035e-07, + "loss": 0.2199, + "step": 2 + }, + { + "epoch": 0.01938610662358643, + "grad_norm": 18.97646141052246, + "learning_rate": 4.838709677419355e-07, + "loss": 0.2004, + "step": 3 + }, + { + "epoch": 0.025848142164781908, + "grad_norm": 26.008861541748047, + "learning_rate": 6.451612903225807e-07, + "loss": 0.2176, + "step": 4 + }, + { + "epoch": 0.03231017770597738, + "grad_norm": 50.078155517578125, + "learning_rate": 8.064516129032258e-07, + "loss": 0.1862, + "step": 5 + }, + { + "epoch": 0.03877221324717286, + "grad_norm": 3807.063720703125, + "learning_rate": 9.67741935483871e-07, + "loss": 0.2138, + "step": 6 + }, + { + "epoch": 0.045234248788368334, + "grad_norm": 14.361485481262207, + "learning_rate": 1.1290322580645162e-06, + "loss": 0.1844, + "step": 7 + }, + { + "epoch": 0.051696284329563816, + "grad_norm": 8.04658031463623, + "learning_rate": 1.2903225806451614e-06, + "loss": 0.1894, + "step": 8 + }, + { + "epoch": 0.05815831987075929, + "grad_norm": 7.152590274810791, + "learning_rate": 1.4516129032258066e-06, + "loss": 0.1658, + "step": 9 + }, + { + "epoch": 0.06462035541195477, + "grad_norm": 4.961307525634766, + "learning_rate": 1.6129032258064516e-06, + "loss": 0.1739, + "step": 10 + }, + { + "epoch": 0.07108239095315025, + "grad_norm": 6.286808013916016, + "learning_rate": 1.774193548387097e-06, + "loss": 0.1784, + "step": 11 + }, + { + "epoch": 0.07754442649434572, + "grad_norm": 3.9918179512023926, + "learning_rate": 1.935483870967742e-06, + "loss": 0.1839, + "step": 12 + }, + { + "epoch": 0.0840064620355412, + "grad_norm": 4.566269874572754, + "learning_rate": 2.096774193548387e-06, + "loss": 0.1794, + "step": 13 + }, + { + "epoch": 0.09046849757673667, + "grad_norm": 2.4546995162963867, + "learning_rate": 2.2580645161290324e-06, + "loss": 0.155, + "step": 14 + }, + { + "epoch": 0.09693053311793215, + "grad_norm": 19.017515182495117, + "learning_rate": 2.4193548387096776e-06, + "loss": 0.1428, + "step": 15 + }, + { + "epoch": 0.10339256865912763, + "grad_norm": 1.8924509286880493, + "learning_rate": 2.580645161290323e-06, + "loss": 0.1424, + "step": 16 + }, + { + "epoch": 0.1098546042003231, + "grad_norm": 1.193750023841858, + "learning_rate": 2.7419354838709676e-06, + "loss": 0.1369, + "step": 17 + }, + { + "epoch": 0.11631663974151858, + "grad_norm": 0.9094203114509583, + "learning_rate": 2.903225806451613e-06, + "loss": 0.1269, + "step": 18 + }, + { + "epoch": 0.12277867528271405, + "grad_norm": 1.336350440979004, + "learning_rate": 3.0645161290322584e-06, + "loss": 0.1243, + "step": 19 + }, + { + "epoch": 0.12924071082390953, + "grad_norm": 26.0351505279541, + "learning_rate": 3.225806451612903e-06, + "loss": 0.1165, + "step": 20 + }, + { + "epoch": 0.13570274636510501, + "grad_norm": 1.1493116617202759, + "learning_rate": 3.3870967741935484e-06, + "loss": 0.1543, + "step": 21 + }, + { + "epoch": 0.1421647819063005, + "grad_norm": 0.8133248686790466, + "learning_rate": 3.548387096774194e-06, + "loss": 0.1064, + "step": 22 + }, + { + "epoch": 0.14862681744749595, + "grad_norm": 0.8336551785469055, + "learning_rate": 3.7096774193548392e-06, + "loss": 0.1192, + "step": 23 + }, + { + "epoch": 0.15508885298869143, + "grad_norm": 0.7685548067092896, + "learning_rate": 3.870967741935484e-06, + "loss": 0.1051, + "step": 24 + }, + { + "epoch": 0.16155088852988692, + "grad_norm": 1.217137336730957, + "learning_rate": 4.032258064516129e-06, + "loss": 0.1377, + "step": 25 + }, + { + "epoch": 0.1680129240710824, + "grad_norm": 1.1432764530181885, + "learning_rate": 4.193548387096774e-06, + "loss": 0.1151, + "step": 26 + }, + { + "epoch": 0.17447495961227788, + "grad_norm": 2.054145097732544, + "learning_rate": 4.35483870967742e-06, + "loss": 0.1178, + "step": 27 + }, + { + "epoch": 0.18093699515347333, + "grad_norm": 0.9552474021911621, + "learning_rate": 4.516129032258065e-06, + "loss": 0.1363, + "step": 28 + }, + { + "epoch": 0.18739903069466882, + "grad_norm": 1.1134074926376343, + "learning_rate": 4.67741935483871e-06, + "loss": 0.0971, + "step": 29 + }, + { + "epoch": 0.1938610662358643, + "grad_norm": 1.2744982242584229, + "learning_rate": 4.838709677419355e-06, + "loss": 0.1177, + "step": 30 + }, + { + "epoch": 0.20032310177705978, + "grad_norm": 0.8065481185913086, + "learning_rate": 5e-06, + "loss": 0.1089, + "step": 31 + }, + { + "epoch": 0.20678513731825526, + "grad_norm": 0.8988919854164124, + "learning_rate": 5.161290322580646e-06, + "loss": 0.1324, + "step": 32 + }, + { + "epoch": 0.21324717285945072, + "grad_norm": 0.7514846324920654, + "learning_rate": 5.322580645161291e-06, + "loss": 0.1045, + "step": 33 + }, + { + "epoch": 0.2197092084006462, + "grad_norm": 0.8767650723457336, + "learning_rate": 5.483870967741935e-06, + "loss": 0.1188, + "step": 34 + }, + { + "epoch": 0.22617124394184168, + "grad_norm": 0.9772207736968994, + "learning_rate": 5.645161290322582e-06, + "loss": 0.0941, + "step": 35 + }, + { + "epoch": 0.23263327948303716, + "grad_norm": 1.004211664199829, + "learning_rate": 5.806451612903226e-06, + "loss": 0.1063, + "step": 36 + }, + { + "epoch": 0.23909531502423265, + "grad_norm": 0.753398597240448, + "learning_rate": 5.967741935483872e-06, + "loss": 0.0908, + "step": 37 + }, + { + "epoch": 0.2455573505654281, + "grad_norm": 0.8244940042495728, + "learning_rate": 6.129032258064517e-06, + "loss": 0.1037, + "step": 38 + }, + { + "epoch": 0.2520193861066236, + "grad_norm": 0.9269906282424927, + "learning_rate": 6.290322580645162e-06, + "loss": 0.121, + "step": 39 + }, + { + "epoch": 0.25848142164781907, + "grad_norm": 0.8812072277069092, + "learning_rate": 6.451612903225806e-06, + "loss": 0.1229, + "step": 40 + }, + { + "epoch": 0.2649434571890145, + "grad_norm": 1.0441511869430542, + "learning_rate": 6.612903225806452e-06, + "loss": 0.1201, + "step": 41 + }, + { + "epoch": 0.27140549273021003, + "grad_norm": 0.819930374622345, + "learning_rate": 6.774193548387097e-06, + "loss": 0.1175, + "step": 42 + }, + { + "epoch": 0.2778675282714055, + "grad_norm": 0.7328965067863464, + "learning_rate": 6.935483870967743e-06, + "loss": 0.0865, + "step": 43 + }, + { + "epoch": 0.284329563812601, + "grad_norm": 0.7510896325111389, + "learning_rate": 7.096774193548388e-06, + "loss": 0.0874, + "step": 44 + }, + { + "epoch": 0.29079159935379645, + "grad_norm": 0.8940995335578918, + "learning_rate": 7.258064516129033e-06, + "loss": 0.1069, + "step": 45 + }, + { + "epoch": 0.2972536348949919, + "grad_norm": 1.0601608753204346, + "learning_rate": 7.4193548387096784e-06, + "loss": 0.1034, + "step": 46 + }, + { + "epoch": 0.3037156704361874, + "grad_norm": 0.6812165379524231, + "learning_rate": 7.580645161290323e-06, + "loss": 0.083, + "step": 47 + }, + { + "epoch": 0.31017770597738287, + "grad_norm": 0.7888991236686707, + "learning_rate": 7.741935483870968e-06, + "loss": 0.1021, + "step": 48 + }, + { + "epoch": 0.3166397415185784, + "grad_norm": 0.7519361972808838, + "learning_rate": 7.903225806451613e-06, + "loss": 0.095, + "step": 49 + }, + { + "epoch": 0.32310177705977383, + "grad_norm": 0.8417134881019592, + "learning_rate": 8.064516129032258e-06, + "loss": 0.0904, + "step": 50 + }, + { + "epoch": 0.3295638126009693, + "grad_norm": 0.7418419122695923, + "learning_rate": 8.225806451612904e-06, + "loss": 0.1081, + "step": 51 + }, + { + "epoch": 0.3360258481421648, + "grad_norm": 0.7588947415351868, + "learning_rate": 8.387096774193549e-06, + "loss": 0.106, + "step": 52 + }, + { + "epoch": 0.34248788368336025, + "grad_norm": 0.5819141864776611, + "learning_rate": 8.548387096774194e-06, + "loss": 0.0767, + "step": 53 + }, + { + "epoch": 0.34894991922455576, + "grad_norm": 0.7023757100105286, + "learning_rate": 8.70967741935484e-06, + "loss": 0.1041, + "step": 54 + }, + { + "epoch": 0.3554119547657512, + "grad_norm": 0.6532196402549744, + "learning_rate": 8.870967741935484e-06, + "loss": 0.0815, + "step": 55 + }, + { + "epoch": 0.36187399030694667, + "grad_norm": 0.7267298102378845, + "learning_rate": 9.03225806451613e-06, + "loss": 0.0994, + "step": 56 + }, + { + "epoch": 0.3683360258481422, + "grad_norm": 0.6799736618995667, + "learning_rate": 9.193548387096775e-06, + "loss": 0.0867, + "step": 57 + }, + { + "epoch": 0.37479806138933763, + "grad_norm": 0.6375831961631775, + "learning_rate": 9.35483870967742e-06, + "loss": 0.082, + "step": 58 + }, + { + "epoch": 0.38126009693053314, + "grad_norm": 0.5718384385108948, + "learning_rate": 9.516129032258065e-06, + "loss": 0.0769, + "step": 59 + }, + { + "epoch": 0.3877221324717286, + "grad_norm": 0.682659924030304, + "learning_rate": 9.67741935483871e-06, + "loss": 0.0744, + "step": 60 + }, + { + "epoch": 0.39418416801292405, + "grad_norm": 0.7422767877578735, + "learning_rate": 9.838709677419356e-06, + "loss": 0.0838, + "step": 61 + }, + { + "epoch": 0.40064620355411956, + "grad_norm": 0.6545633673667908, + "learning_rate": 1e-05, + "loss": 0.083, + "step": 62 + }, + { + "epoch": 0.407108239095315, + "grad_norm": 0.6176275014877319, + "learning_rate": 9.99991960684721e-06, + "loss": 0.0819, + "step": 63 + }, + { + "epoch": 0.4135702746365105, + "grad_norm": 0.6768634915351868, + "learning_rate": 9.999678429974063e-06, + "loss": 0.0742, + "step": 64 + }, + { + "epoch": 0.420032310177706, + "grad_norm": 0.727836549282074, + "learning_rate": 9.999276477136145e-06, + "loss": 0.0915, + "step": 65 + }, + { + "epoch": 0.42649434571890144, + "grad_norm": 0.6889824867248535, + "learning_rate": 9.998713761259157e-06, + "loss": 0.0805, + "step": 66 + }, + { + "epoch": 0.43295638126009695, + "grad_norm": 0.7547912001609802, + "learning_rate": 9.997990300438505e-06, + "loss": 0.0852, + "step": 67 + }, + { + "epoch": 0.4394184168012924, + "grad_norm": 0.5975009202957153, + "learning_rate": 9.997106117938704e-06, + "loss": 0.0818, + "step": 68 + }, + { + "epoch": 0.4458804523424879, + "grad_norm": 1.1194217205047607, + "learning_rate": 9.996061242192645e-06, + "loss": 0.0961, + "step": 69 + }, + { + "epoch": 0.45234248788368336, + "grad_norm": 0.6122073531150818, + "learning_rate": 9.994855706800666e-06, + "loss": 0.0714, + "step": 70 + }, + { + "epoch": 0.4588045234248788, + "grad_norm": 19.22555923461914, + "learning_rate": 9.993489550529486e-06, + "loss": 0.0691, + "step": 71 + }, + { + "epoch": 0.46526655896607433, + "grad_norm": 0.468181848526001, + "learning_rate": 9.991962817310947e-06, + "loss": 0.0661, + "step": 72 + }, + { + "epoch": 0.4717285945072698, + "grad_norm": 13.211414337158203, + "learning_rate": 9.990275556240612e-06, + "loss": 0.0849, + "step": 73 + }, + { + "epoch": 0.4781906300484653, + "grad_norm": 0.6043105125427246, + "learning_rate": 9.98842782157617e-06, + "loss": 0.0803, + "step": 74 + }, + { + "epoch": 0.48465266558966075, + "grad_norm": 0.6836904883384705, + "learning_rate": 9.986419672735712e-06, + "loss": 0.0984, + "step": 75 + }, + { + "epoch": 0.4911147011308562, + "grad_norm": 0.7549227476119995, + "learning_rate": 9.9842511742958e-06, + "loss": 0.0682, + "step": 76 + }, + { + "epoch": 0.4975767366720517, + "grad_norm": 0.671994149684906, + "learning_rate": 9.981922395989409e-06, + "loss": 0.07, + "step": 77 + }, + { + "epoch": 0.5040387722132472, + "grad_norm": 0.8292804956436157, + "learning_rate": 9.97943341270367e-06, + "loss": 0.0918, + "step": 78 + }, + { + "epoch": 0.5105008077544426, + "grad_norm": 0.5530771017074585, + "learning_rate": 9.976784304477467e-06, + "loss": 0.0768, + "step": 79 + }, + { + "epoch": 0.5169628432956381, + "grad_norm": 0.5503768920898438, + "learning_rate": 9.973975156498866e-06, + "loss": 0.0923, + "step": 80 + }, + { + "epoch": 0.5234248788368336, + "grad_norm": 0.46011969447135925, + "learning_rate": 9.971006059102369e-06, + "loss": 0.063, + "step": 81 + }, + { + "epoch": 0.529886914378029, + "grad_norm": 0.5449568033218384, + "learning_rate": 9.96787710776602e-06, + "loss": 0.0667, + "step": 82 + }, + { + "epoch": 0.5363489499192245, + "grad_norm": 0.5621317028999329, + "learning_rate": 9.964588403108324e-06, + "loss": 0.0568, + "step": 83 + }, + { + "epoch": 0.5428109854604201, + "grad_norm": 0.4867885410785675, + "learning_rate": 9.961140050885014e-06, + "loss": 0.0681, + "step": 84 + }, + { + "epoch": 0.5492730210016155, + "grad_norm": 0.5386512279510498, + "learning_rate": 9.957532161985654e-06, + "loss": 0.0711, + "step": 85 + }, + { + "epoch": 0.555735056542811, + "grad_norm": 0.4690256416797638, + "learning_rate": 9.95376485243007e-06, + "loss": 0.0667, + "step": 86 + }, + { + "epoch": 0.5621970920840065, + "grad_norm": 0.5720282793045044, + "learning_rate": 9.949838243364614e-06, + "loss": 0.0939, + "step": 87 + }, + { + "epoch": 0.568659127625202, + "grad_norm": 0.6215516924858093, + "learning_rate": 9.945752461058286e-06, + "loss": 0.0691, + "step": 88 + }, + { + "epoch": 0.5751211631663974, + "grad_norm": 0.4669437110424042, + "learning_rate": 9.941507636898651e-06, + "loss": 0.0685, + "step": 89 + }, + { + "epoch": 0.5815831987075929, + "grad_norm": 0.4649712145328522, + "learning_rate": 9.937103907387626e-06, + "loss": 0.074, + "step": 90 + }, + { + "epoch": 0.5880452342487884, + "grad_norm": 0.7192949056625366, + "learning_rate": 9.932541414137096e-06, + "loss": 0.0965, + "step": 91 + }, + { + "epoch": 0.5945072697899838, + "grad_norm": 13.985039710998535, + "learning_rate": 9.927820303864342e-06, + "loss": 0.0651, + "step": 92 + }, + { + "epoch": 0.6009693053311793, + "grad_norm": 0.5609593987464905, + "learning_rate": 9.922940728387345e-06, + "loss": 0.0796, + "step": 93 + }, + { + "epoch": 0.6074313408723748, + "grad_norm": 1.4020167589187622, + "learning_rate": 9.917902844619885e-06, + "loss": 0.0657, + "step": 94 + }, + { + "epoch": 0.6138933764135702, + "grad_norm": 0.5554618239402771, + "learning_rate": 9.912706814566504e-06, + "loss": 0.0676, + "step": 95 + }, + { + "epoch": 0.6203554119547657, + "grad_norm": 0.6547884345054626, + "learning_rate": 9.907352805317301e-06, + "loss": 0.109, + "step": 96 + }, + { + "epoch": 0.6268174474959612, + "grad_norm": 0.6062126755714417, + "learning_rate": 9.901840989042547e-06, + "loss": 0.0864, + "step": 97 + }, + { + "epoch": 0.6332794830371568, + "grad_norm": 0.4889056980609894, + "learning_rate": 9.896171542987158e-06, + "loss": 0.0632, + "step": 98 + }, + { + "epoch": 0.6397415185783522, + "grad_norm": 0.6889820098876953, + "learning_rate": 9.890344649464992e-06, + "loss": 0.0782, + "step": 99 + }, + { + "epoch": 0.6462035541195477, + "grad_norm": 1.3493472337722778, + "learning_rate": 9.884360495852984e-06, + "loss": 0.0779, + "step": 100 + }, + { + "epoch": 0.6526655896607432, + "grad_norm": 0.5208078026771545, + "learning_rate": 9.878219274585125e-06, + "loss": 0.072, + "step": 101 + }, + { + "epoch": 0.6591276252019386, + "grad_norm": 0.6799489259719849, + "learning_rate": 9.871921183146272e-06, + "loss": 0.0718, + "step": 102 + }, + { + "epoch": 0.6655896607431341, + "grad_norm": 0.5299189686775208, + "learning_rate": 9.865466424065792e-06, + "loss": 0.0687, + "step": 103 + }, + { + "epoch": 0.6720516962843296, + "grad_norm": 0.6351456046104431, + "learning_rate": 9.858855204911065e-06, + "loss": 0.0892, + "step": 104 + }, + { + "epoch": 0.678513731825525, + "grad_norm": 0.567482054233551, + "learning_rate": 9.852087738280784e-06, + "loss": 0.0778, + "step": 105 + }, + { + "epoch": 0.6849757673667205, + "grad_norm": 0.5811858773231506, + "learning_rate": 9.845164241798143e-06, + "loss": 0.0709, + "step": 106 + }, + { + "epoch": 0.691437802907916, + "grad_norm": 0.600745677947998, + "learning_rate": 9.838084938103832e-06, + "loss": 0.0721, + "step": 107 + }, + { + "epoch": 0.6978998384491115, + "grad_norm": 0.5404493808746338, + "learning_rate": 9.830850054848859e-06, + "loss": 0.0651, + "step": 108 + }, + { + "epoch": 0.7043618739903069, + "grad_norm": 0.6496622562408447, + "learning_rate": 9.823459824687262e-06, + "loss": 0.0794, + "step": 109 + }, + { + "epoch": 0.7108239095315024, + "grad_norm": 0.5033782124519348, + "learning_rate": 9.815914485268598e-06, + "loss": 0.0866, + "step": 110 + }, + { + "epoch": 0.7172859450726979, + "grad_norm": 0.48423922061920166, + "learning_rate": 9.808214279230317e-06, + "loss": 0.0681, + "step": 111 + }, + { + "epoch": 0.7237479806138933, + "grad_norm": 0.5603338479995728, + "learning_rate": 9.800359454189955e-06, + "loss": 0.0894, + "step": 112 + }, + { + "epoch": 0.7302100161550888, + "grad_norm": 0.554610550403595, + "learning_rate": 9.792350262737173e-06, + "loss": 0.0735, + "step": 113 + }, + { + "epoch": 0.7366720516962844, + "grad_norm": 0.4603743553161621, + "learning_rate": 9.784186962425633e-06, + "loss": 0.0546, + "step": 114 + }, + { + "epoch": 0.7431340872374798, + "grad_norm": 0.4438991844654083, + "learning_rate": 9.775869815764713e-06, + "loss": 0.0571, + "step": 115 + }, + { + "epoch": 0.7495961227786753, + "grad_norm": 0.6239739060401917, + "learning_rate": 9.76739909021107e-06, + "loss": 0.0997, + "step": 116 + }, + { + "epoch": 0.7560581583198708, + "grad_norm": 0.47505640983581543, + "learning_rate": 9.758775058160037e-06, + "loss": 0.0755, + "step": 117 + }, + { + "epoch": 0.7625201938610663, + "grad_norm": 0.5345163345336914, + "learning_rate": 9.749997996936866e-06, + "loss": 0.0754, + "step": 118 + }, + { + "epoch": 0.7689822294022617, + "grad_norm": 0.4881122410297394, + "learning_rate": 9.741068188787806e-06, + "loss": 0.0787, + "step": 119 + }, + { + "epoch": 0.7754442649434572, + "grad_norm": 0.6172494888305664, + "learning_rate": 9.731985920871028e-06, + "loss": 0.0737, + "step": 120 + }, + { + "epoch": 0.7819063004846527, + "grad_norm": 0.4486066401004791, + "learning_rate": 9.722751485247393e-06, + "loss": 0.0702, + "step": 121 + }, + { + "epoch": 0.7883683360258481, + "grad_norm": 1.1322523355484009, + "learning_rate": 9.713365178871061e-06, + "loss": 0.0766, + "step": 122 + }, + { + "epoch": 0.7948303715670436, + "grad_norm": 0.44906798005104065, + "learning_rate": 9.703827303579936e-06, + "loss": 0.0635, + "step": 123 + }, + { + "epoch": 0.8012924071082391, + "grad_norm": 0.5776321887969971, + "learning_rate": 9.694138166085964e-06, + "loss": 0.0733, + "step": 124 + }, + { + "epoch": 0.8077544426494345, + "grad_norm": 0.45236676931381226, + "learning_rate": 9.684298077965269e-06, + "loss": 0.0653, + "step": 125 + }, + { + "epoch": 0.81421647819063, + "grad_norm": 0.4732246696949005, + "learning_rate": 9.674307355648136e-06, + "loss": 0.0647, + "step": 126 + }, + { + "epoch": 0.8206785137318255, + "grad_norm": 0.5663241744041443, + "learning_rate": 9.664166320408828e-06, + "loss": 0.0634, + "step": 127 + }, + { + "epoch": 0.827140549273021, + "grad_norm": 0.6197142004966736, + "learning_rate": 9.653875298355264e-06, + "loss": 0.0854, + "step": 128 + }, + { + "epoch": 0.8336025848142165, + "grad_norm": 0.6459212899208069, + "learning_rate": 9.64343462041853e-06, + "loss": 0.0744, + "step": 129 + }, + { + "epoch": 0.840064620355412, + "grad_norm": 0.6052826046943665, + "learning_rate": 9.63284462234223e-06, + "loss": 0.0882, + "step": 130 + }, + { + "epoch": 0.8465266558966075, + "grad_norm": 0.6480473875999451, + "learning_rate": 9.622105644671698e-06, + "loss": 0.0721, + "step": 131 + }, + { + "epoch": 0.8529886914378029, + "grad_norm": 0.5728505849838257, + "learning_rate": 9.611218032743044e-06, + "loss": 0.0929, + "step": 132 + }, + { + "epoch": 0.8594507269789984, + "grad_norm": 0.5130248665809631, + "learning_rate": 9.600182136672048e-06, + "loss": 0.0738, + "step": 133 + }, + { + "epoch": 0.8659127625201939, + "grad_norm": 0.4673842787742615, + "learning_rate": 9.5889983113429e-06, + "loss": 0.0631, + "step": 134 + }, + { + "epoch": 0.8723747980613893, + "grad_norm": 0.4138599932193756, + "learning_rate": 9.57766691639679e-06, + "loss": 0.0639, + "step": 135 + }, + { + "epoch": 0.8788368336025848, + "grad_norm": 0.520550549030304, + "learning_rate": 9.56618831622035e-06, + "loss": 0.0615, + "step": 136 + }, + { + "epoch": 0.8852988691437803, + "grad_norm": 0.5906829833984375, + "learning_rate": 9.554562879933917e-06, + "loss": 0.0767, + "step": 137 + }, + { + "epoch": 0.8917609046849758, + "grad_norm": 0.528221070766449, + "learning_rate": 9.542790981379683e-06, + "loss": 0.0668, + "step": 138 + }, + { + "epoch": 0.8982229402261712, + "grad_norm": 0.5368471741676331, + "learning_rate": 9.530872999109665e-06, + "loss": 0.068, + "step": 139 + }, + { + "epoch": 0.9046849757673667, + "grad_norm": 0.5090873837471008, + "learning_rate": 9.51880931637353e-06, + "loss": 0.0615, + "step": 140 + }, + { + "epoch": 0.9111470113085622, + "grad_norm": 0.4326649308204651, + "learning_rate": 9.506600321106273e-06, + "loss": 0.0612, + "step": 141 + }, + { + "epoch": 0.9176090468497576, + "grad_norm": 0.5169522762298584, + "learning_rate": 9.494246405915743e-06, + "loss": 0.0707, + "step": 142 + }, + { + "epoch": 0.9240710823909531, + "grad_norm": 0.5026911497116089, + "learning_rate": 9.481747968070018e-06, + "loss": 0.064, + "step": 143 + }, + { + "epoch": 0.9305331179321487, + "grad_norm": 0.45077645778656006, + "learning_rate": 9.469105409484628e-06, + "loss": 0.053, + "step": 144 + }, + { + "epoch": 0.9369951534733441, + "grad_norm": 0.46447330713272095, + "learning_rate": 9.456319136709628e-06, + "loss": 0.0679, + "step": 145 + }, + { + "epoch": 0.9434571890145396, + "grad_norm": 0.43717750906944275, + "learning_rate": 9.443389560916532e-06, + "loss": 0.0734, + "step": 146 + }, + { + "epoch": 0.9499192245557351, + "grad_norm": 0.4999764561653137, + "learning_rate": 9.430317097885082e-06, + "loss": 0.0764, + "step": 147 + }, + { + "epoch": 0.9563812600969306, + "grad_norm": 0.5041486024856567, + "learning_rate": 9.417102167989888e-06, + "loss": 0.0776, + "step": 148 + }, + { + "epoch": 0.962843295638126, + "grad_norm": 0.5824252367019653, + "learning_rate": 9.403745196186904e-06, + "loss": 0.0871, + "step": 149 + }, + { + "epoch": 0.9693053311793215, + "grad_norm": 0.44613441824913025, + "learning_rate": 9.390246611999754e-06, + "loss": 0.0753, + "step": 150 + }, + { + "epoch": 0.975767366720517, + "grad_norm": 0.5553467273712158, + "learning_rate": 9.376606849505939e-06, + "loss": 0.0788, + "step": 151 + }, + { + "epoch": 0.9822294022617124, + "grad_norm": 0.47406890988349915, + "learning_rate": 9.362826347322857e-06, + "loss": 0.0612, + "step": 152 + }, + { + "epoch": 0.9886914378029079, + "grad_norm": 0.36964505910873413, + "learning_rate": 9.348905548593722e-06, + "loss": 0.0683, + "step": 153 + }, + { + "epoch": 0.9951534733441034, + "grad_norm": 0.4598817229270935, + "learning_rate": 9.334844900973292e-06, + "loss": 0.0718, + "step": 154 + }, + { + "epoch": 1.0, + "grad_norm": 0.4598817229270935, + "learning_rate": 9.320644856613482e-06, + "loss": 0.0629, + "step": 155 + }, + { + "epoch": 1.0064620355411955, + "grad_norm": 0.4689626395702362, + "learning_rate": 9.306305872148826e-06, + "loss": 0.0471, + "step": 156 + }, + { + "epoch": 1.012924071082391, + "grad_norm": 0.37304380536079407, + "learning_rate": 9.291828408681796e-06, + "loss": 0.0471, + "step": 157 + }, + { + "epoch": 1.0193861066235865, + "grad_norm": 0.45355430245399475, + "learning_rate": 9.277212931767958e-06, + "loss": 0.0497, + "step": 158 + }, + { + "epoch": 1.0258481421647818, + "grad_norm": 0.4865805506706238, + "learning_rate": 9.262459911401025e-06, + "loss": 0.0662, + "step": 159 + }, + { + "epoch": 1.0323101777059773, + "grad_norm": 0.35557398200035095, + "learning_rate": 9.247569821997724e-06, + "loss": 0.0397, + "step": 160 + }, + { + "epoch": 1.0387722132471728, + "grad_norm": 0.34070008993148804, + "learning_rate": 9.232543142382546e-06, + "loss": 0.0614, + "step": 161 + }, + { + "epoch": 1.0452342487883683, + "grad_norm": 0.4690254330635071, + "learning_rate": 9.217380355772353e-06, + "loss": 0.0484, + "step": 162 + }, + { + "epoch": 1.0516962843295639, + "grad_norm": 0.3816601634025574, + "learning_rate": 9.202081949760833e-06, + "loss": 0.045, + "step": 163 + }, + { + "epoch": 1.0581583198707594, + "grad_norm": 0.7278222441673279, + "learning_rate": 9.186648416302823e-06, + "loss": 0.0497, + "step": 164 + }, + { + "epoch": 1.0646203554119547, + "grad_norm": 0.3829418420791626, + "learning_rate": 9.171080251698488e-06, + "loss": 0.0371, + "step": 165 + }, + { + "epoch": 1.0710823909531502, + "grad_norm": 0.33008721470832825, + "learning_rate": 9.155377956577363e-06, + "loss": 0.0561, + "step": 166 + }, + { + "epoch": 1.0775444264943457, + "grad_norm": 0.47766488790512085, + "learning_rate": 9.13954203588225e-06, + "loss": 0.0528, + "step": 167 + }, + { + "epoch": 1.0840064620355412, + "grad_norm": 0.3429339528083801, + "learning_rate": 9.123572998852988e-06, + "loss": 0.0524, + "step": 168 + }, + { + "epoch": 1.0904684975767367, + "grad_norm": 0.4901416003704071, + "learning_rate": 9.107471359010069e-06, + "loss": 0.0512, + "step": 169 + }, + { + "epoch": 1.0969305331179322, + "grad_norm": 0.4330100119113922, + "learning_rate": 9.091237634138133e-06, + "loss": 0.054, + "step": 170 + }, + { + "epoch": 1.1033925686591277, + "grad_norm": 0.3829297721385956, + "learning_rate": 9.074872346269305e-06, + "loss": 0.0414, + "step": 171 + }, + { + "epoch": 1.109854604200323, + "grad_norm": 0.3282840847969055, + "learning_rate": 9.058376021666424e-06, + "loss": 0.0476, + "step": 172 + }, + { + "epoch": 1.1163166397415185, + "grad_norm": 0.39873006939888, + "learning_rate": 9.041749190806105e-06, + "loss": 0.0514, + "step": 173 + }, + { + "epoch": 1.122778675282714, + "grad_norm": 0.44557517766952515, + "learning_rate": 9.024992388361691e-06, + "loss": 0.0435, + "step": 174 + }, + { + "epoch": 1.1292407108239095, + "grad_norm": 3.866434097290039, + "learning_rate": 9.008106153186055e-06, + "loss": 0.0497, + "step": 175 + }, + { + "epoch": 1.135702746365105, + "grad_norm": 0.3858250677585602, + "learning_rate": 8.991091028294268e-06, + "loss": 0.0464, + "step": 176 + }, + { + "epoch": 1.1421647819063006, + "grad_norm": 0.47651711106300354, + "learning_rate": 8.973947560846146e-06, + "loss": 0.0376, + "step": 177 + }, + { + "epoch": 1.148626817447496, + "grad_norm": 0.7432534694671631, + "learning_rate": 8.956676302128646e-06, + "loss": 0.0436, + "step": 178 + }, + { + "epoch": 1.1550888529886914, + "grad_norm": 0.523420512676239, + "learning_rate": 8.939277807538147e-06, + "loss": 0.0619, + "step": 179 + }, + { + "epoch": 1.1615508885298869, + "grad_norm": 0.463724821805954, + "learning_rate": 8.921752636562582e-06, + "loss": 0.0455, + "step": 180 + }, + { + "epoch": 1.1680129240710824, + "grad_norm": 0.45723769068717957, + "learning_rate": 8.90410135276345e-06, + "loss": 0.0421, + "step": 181 + }, + { + "epoch": 1.1744749596122779, + "grad_norm": 0.38577011227607727, + "learning_rate": 8.886324523757692e-06, + "loss": 0.0531, + "step": 182 + }, + { + "epoch": 1.1809369951534734, + "grad_norm": 0.5292965173721313, + "learning_rate": 8.868422721199442e-06, + "loss": 0.0635, + "step": 183 + }, + { + "epoch": 1.187399030694669, + "grad_norm": 0.3577275574207306, + "learning_rate": 8.850396520761636e-06, + "loss": 0.0458, + "step": 184 + }, + { + "epoch": 1.1938610662358644, + "grad_norm": 0.4251251518726349, + "learning_rate": 8.832246502117512e-06, + "loss": 0.0363, + "step": 185 + }, + { + "epoch": 1.2003231017770597, + "grad_norm": 3.846379041671753, + "learning_rate": 8.813973248921958e-06, + "loss": 0.0512, + "step": 186 + }, + { + "epoch": 1.2067851373182552, + "grad_norm": 0.41917675733566284, + "learning_rate": 8.795577348792748e-06, + "loss": 0.0421, + "step": 187 + }, + { + "epoch": 1.2132471728594507, + "grad_norm": 0.5150820016860962, + "learning_rate": 8.777059393291645e-06, + "loss": 0.0413, + "step": 188 + }, + { + "epoch": 1.2197092084006462, + "grad_norm": 0.3357498049736023, + "learning_rate": 8.75841997790538e-06, + "loss": 0.0422, + "step": 189 + }, + { + "epoch": 1.2261712439418417, + "grad_norm": 0.37925323843955994, + "learning_rate": 8.739659702026502e-06, + "loss": 0.0448, + "step": 190 + }, + { + "epoch": 1.2326332794830372, + "grad_norm": 0.8641751408576965, + "learning_rate": 8.7207791689341e-06, + "loss": 0.0599, + "step": 191 + }, + { + "epoch": 1.2390953150242328, + "grad_norm": 0.4052288234233856, + "learning_rate": 8.701778985774405e-06, + "loss": 0.0503, + "step": 192 + }, + { + "epoch": 1.245557350565428, + "grad_norm": 0.45619818568229675, + "learning_rate": 8.68265976354127e-06, + "loss": 0.0483, + "step": 193 + }, + { + "epoch": 1.2520193861066236, + "grad_norm": 0.35979220271110535, + "learning_rate": 8.663422117056519e-06, + "loss": 0.0413, + "step": 194 + }, + { + "epoch": 1.258481421647819, + "grad_norm": 0.3667192757129669, + "learning_rate": 8.644066664950169e-06, + "loss": 0.0383, + "step": 195 + }, + { + "epoch": 1.2649434571890146, + "grad_norm": 0.3911844491958618, + "learning_rate": 8.62459402964055e-06, + "loss": 0.0461, + "step": 196 + }, + { + "epoch": 1.27140549273021, + "grad_norm": 0.5011119842529297, + "learning_rate": 8.605004837314277e-06, + "loss": 0.0492, + "step": 197 + }, + { + "epoch": 1.2778675282714054, + "grad_norm": 0.562304675579071, + "learning_rate": 8.585299717906127e-06, + "loss": 0.0668, + "step": 198 + }, + { + "epoch": 1.284329563812601, + "grad_norm": 0.5580583214759827, + "learning_rate": 8.565479305078767e-06, + "loss": 0.0352, + "step": 199 + }, + { + "epoch": 1.2907915993537964, + "grad_norm": 0.5646708011627197, + "learning_rate": 8.54554423620239e-06, + "loss": 0.0489, + "step": 200 + }, + { + "epoch": 1.297253634894992, + "grad_norm": 0.5223414897918701, + "learning_rate": 8.525495152334211e-06, + "loss": 0.0447, + "step": 201 + }, + { + "epoch": 1.3037156704361874, + "grad_norm": 0.4284669756889343, + "learning_rate": 8.505332698197853e-06, + "loss": 0.0518, + "step": 202 + }, + { + "epoch": 1.310177705977383, + "grad_norm": 0.4862108826637268, + "learning_rate": 8.48505752216262e-06, + "loss": 0.0389, + "step": 203 + }, + { + "epoch": 1.3166397415185784, + "grad_norm": 0.4296407103538513, + "learning_rate": 8.464670276222642e-06, + "loss": 0.0509, + "step": 204 + }, + { + "epoch": 1.3231017770597737, + "grad_norm": 0.40503671765327454, + "learning_rate": 8.444171615975909e-06, + "loss": 0.0493, + "step": 205 + }, + { + "epoch": 1.3295638126009692, + "grad_norm": 0.39369431138038635, + "learning_rate": 8.423562200603192e-06, + "loss": 0.0594, + "step": 206 + }, + { + "epoch": 1.3360258481421647, + "grad_norm": 0.5622990727424622, + "learning_rate": 8.402842692846842e-06, + "loss": 0.039, + "step": 207 + }, + { + "epoch": 1.3424878836833603, + "grad_norm": 0.4047924280166626, + "learning_rate": 8.38201375898948e-06, + "loss": 0.0363, + "step": 208 + }, + { + "epoch": 1.3489499192245558, + "grad_norm": 0.4414857029914856, + "learning_rate": 8.361076068832574e-06, + "loss": 0.0582, + "step": 209 + }, + { + "epoch": 1.3554119547657513, + "grad_norm": 0.43151915073394775, + "learning_rate": 8.340030295674887e-06, + "loss": 0.0433, + "step": 210 + }, + { + "epoch": 1.3618739903069468, + "grad_norm": 0.49384137988090515, + "learning_rate": 8.31887711629085e-06, + "loss": 0.0514, + "step": 211 + }, + { + "epoch": 1.368336025848142, + "grad_norm": 0.4050130844116211, + "learning_rate": 8.29761721090877e-06, + "loss": 0.0448, + "step": 212 + }, + { + "epoch": 1.3747980613893376, + "grad_norm": 0.351788192987442, + "learning_rate": 8.276251263188976e-06, + "loss": 0.0415, + "step": 213 + }, + { + "epoch": 1.381260096930533, + "grad_norm": 0.3712752163410187, + "learning_rate": 8.254779960201831e-06, + "loss": 0.0428, + "step": 214 + }, + { + "epoch": 1.3877221324717286, + "grad_norm": 0.408925861120224, + "learning_rate": 8.23320399240563e-06, + "loss": 0.0527, + "step": 215 + }, + { + "epoch": 1.394184168012924, + "grad_norm": 0.3616960942745209, + "learning_rate": 8.2115240536244e-06, + "loss": 0.0323, + "step": 216 + }, + { + "epoch": 1.4006462035541196, + "grad_norm": 0.3473566472530365, + "learning_rate": 8.1897408410256e-06, + "loss": 0.0463, + "step": 217 + }, + { + "epoch": 1.4071082390953151, + "grad_norm": 0.477464497089386, + "learning_rate": 8.16785505509768e-06, + "loss": 0.0521, + "step": 218 + }, + { + "epoch": 1.4135702746365104, + "grad_norm": 0.6549597978591919, + "learning_rate": 8.145867399627575e-06, + "loss": 0.0467, + "step": 219 + }, + { + "epoch": 1.420032310177706, + "grad_norm": 0.46223175525665283, + "learning_rate": 8.123778581678064e-06, + "loss": 0.0444, + "step": 220 + }, + { + "epoch": 1.4264943457189014, + "grad_norm": 0.5096275210380554, + "learning_rate": 8.10158931156503e-06, + "loss": 0.0451, + "step": 221 + }, + { + "epoch": 1.432956381260097, + "grad_norm": 0.37000322341918945, + "learning_rate": 8.079300302834632e-06, + "loss": 0.051, + "step": 222 + }, + { + "epoch": 1.4394184168012925, + "grad_norm": 0.671576738357544, + "learning_rate": 8.056912272240338e-06, + "loss": 0.0466, + "step": 223 + }, + { + "epoch": 1.445880452342488, + "grad_norm": 0.35391512513160706, + "learning_rate": 8.034425939719896e-06, + "loss": 0.0528, + "step": 224 + }, + { + "epoch": 1.4523424878836835, + "grad_norm": 0.35577550530433655, + "learning_rate": 8.011842028372175e-06, + "loss": 0.047, + "step": 225 + }, + { + "epoch": 1.4588045234248788, + "grad_norm": 0.4128228425979614, + "learning_rate": 7.989161264433904e-06, + "loss": 0.04, + "step": 226 + }, + { + "epoch": 1.4652665589660743, + "grad_norm": 3.7340619564056396, + "learning_rate": 7.966384377256335e-06, + "loss": 0.0509, + "step": 227 + }, + { + "epoch": 1.4717285945072698, + "grad_norm": 0.4408370852470398, + "learning_rate": 7.943512099281776e-06, + "loss": 0.0533, + "step": 228 + }, + { + "epoch": 1.4781906300484653, + "grad_norm": 0.3726482093334198, + "learning_rate": 7.92054516602004e-06, + "loss": 0.0395, + "step": 229 + }, + { + "epoch": 1.4846526655896608, + "grad_norm": 0.40403544902801514, + "learning_rate": 7.897484316024799e-06, + "loss": 0.0582, + "step": 230 + }, + { + "epoch": 1.491114701130856, + "grad_norm": 0.4779585301876068, + "learning_rate": 7.874330290869829e-06, + "loss": 0.0422, + "step": 231 + }, + { + "epoch": 1.4975767366720518, + "grad_norm": 0.46821513772010803, + "learning_rate": 7.85108383512516e-06, + "loss": 0.0551, + "step": 232 + }, + { + "epoch": 1.504038772213247, + "grad_norm": 0.3851401209831238, + "learning_rate": 7.827745696333139e-06, + "loss": 0.0542, + "step": 233 + }, + { + "epoch": 1.5105008077544426, + "grad_norm": 0.3976247310638428, + "learning_rate": 7.804316624984391e-06, + "loss": 0.0444, + "step": 234 + }, + { + "epoch": 1.5169628432956381, + "grad_norm": 0.3913721740245819, + "learning_rate": 7.780797374493683e-06, + "loss": 0.0421, + "step": 235 + }, + { + "epoch": 1.5234248788368336, + "grad_norm": 0.38424035906791687, + "learning_rate": 7.757188701175688e-06, + "loss": 0.057, + "step": 236 + }, + { + "epoch": 1.5298869143780292, + "grad_norm": 0.36402031779289246, + "learning_rate": 7.733491364220686e-06, + "loss": 0.0599, + "step": 237 + }, + { + "epoch": 1.5363489499192244, + "grad_norm": 0.5822838544845581, + "learning_rate": 7.709706125670124e-06, + "loss": 0.0418, + "step": 238 + }, + { + "epoch": 1.5428109854604202, + "grad_norm": 0.38915562629699707, + "learning_rate": 7.685833750392131e-06, + "loss": 0.0462, + "step": 239 + }, + { + "epoch": 1.5492730210016155, + "grad_norm": 0.41611573100090027, + "learning_rate": 7.661875006056914e-06, + "loss": 0.0363, + "step": 240 + }, + { + "epoch": 1.555735056542811, + "grad_norm": 0.3718394637107849, + "learning_rate": 7.637830663112064e-06, + "loss": 0.0562, + "step": 241 + }, + { + "epoch": 1.5621970920840065, + "grad_norm": 0.5133814811706543, + "learning_rate": 7.613701494757803e-06, + "loss": 0.0566, + "step": 242 + }, + { + "epoch": 1.568659127625202, + "grad_norm": 0.496139794588089, + "learning_rate": 7.589488276922095e-06, + "loss": 0.053, + "step": 243 + }, + { + "epoch": 1.5751211631663975, + "grad_norm": 0.43612685799598694, + "learning_rate": 7.5651917882357075e-06, + "loss": 0.0371, + "step": 244 + }, + { + "epoch": 1.5815831987075928, + "grad_norm": 0.47427716851234436, + "learning_rate": 7.540812810007172e-06, + "loss": 0.0652, + "step": 245 + }, + { + "epoch": 1.5880452342487885, + "grad_norm": 0.48456844687461853, + "learning_rate": 7.516352126197658e-06, + "loss": 0.05, + "step": 246 + }, + { + "epoch": 1.5945072697899838, + "grad_norm": 0.4757959246635437, + "learning_rate": 7.491810523395762e-06, + "loss": 0.051, + "step": 247 + }, + { + "epoch": 1.6009693053311793, + "grad_norm": 0.44984814524650574, + "learning_rate": 7.467188790792213e-06, + "loss": 0.0469, + "step": 248 + }, + { + "epoch": 1.6074313408723748, + "grad_norm": 0.3804328739643097, + "learning_rate": 7.442487720154494e-06, + "loss": 0.0438, + "step": 249 + }, + { + "epoch": 1.6138933764135701, + "grad_norm": 0.3791749179363251, + "learning_rate": 7.417708105801386e-06, + "loss": 0.0488, + "step": 250 + }, + { + "epoch": 1.6203554119547658, + "grad_norm": 0.3565092980861664, + "learning_rate": 7.392850744577416e-06, + "loss": 0.0369, + "step": 251 + }, + { + "epoch": 1.6268174474959611, + "grad_norm": 0.3572937250137329, + "learning_rate": 7.36791643582724e-06, + "loss": 0.0368, + "step": 252 + }, + { + "epoch": 1.6332794830371569, + "grad_norm": 3.2710509300231934, + "learning_rate": 7.342905981369937e-06, + "loss": 0.061, + "step": 253 + }, + { + "epoch": 1.6397415185783522, + "grad_norm": 0.4295857548713684, + "learning_rate": 7.31782018547322e-06, + "loss": 0.0428, + "step": 254 + }, + { + "epoch": 1.6462035541195477, + "grad_norm": 0.40736424922943115, + "learning_rate": 7.2926598548275765e-06, + "loss": 0.0419, + "step": 255 + }, + { + "epoch": 1.6526655896607432, + "grad_norm": 0.3768783211708069, + "learning_rate": 7.267425798520333e-06, + "loss": 0.0417, + "step": 256 + }, + { + "epoch": 1.6591276252019385, + "grad_norm": 0.39212289452552795, + "learning_rate": 7.242118828009622e-06, + "loss": 0.0538, + "step": 257 + }, + { + "epoch": 1.6655896607431342, + "grad_norm": 0.4783058166503906, + "learning_rate": 7.2167397570983075e-06, + "loss": 0.0402, + "step": 258 + }, + { + "epoch": 1.6720516962843295, + "grad_norm": 0.4783688485622406, + "learning_rate": 7.191289401907796e-06, + "loss": 0.0435, + "step": 259 + }, + { + "epoch": 1.678513731825525, + "grad_norm": 0.44153332710266113, + "learning_rate": 7.165768580851806e-06, + "loss": 0.0429, + "step": 260 + }, + { + "epoch": 1.6849757673667205, + "grad_norm": 0.37209975719451904, + "learning_rate": 7.140178114610045e-06, + "loss": 0.0613, + "step": 261 + }, + { + "epoch": 1.691437802907916, + "grad_norm": 0.4744909703731537, + "learning_rate": 7.114518826101815e-06, + "loss": 0.048, + "step": 262 + }, + { + "epoch": 1.6978998384491115, + "grad_norm": 0.5942760705947876, + "learning_rate": 7.088791540459562e-06, + "loss": 0.043, + "step": 263 + }, + { + "epoch": 1.7043618739903068, + "grad_norm": 0.43296632170677185, + "learning_rate": 7.062997085002322e-06, + "loss": 0.0486, + "step": 264 + }, + { + "epoch": 1.7108239095315025, + "grad_norm": 0.4319445788860321, + "learning_rate": 7.03713628920914e-06, + "loss": 0.0417, + "step": 265 + }, + { + "epoch": 1.7172859450726978, + "grad_norm": 2.305532932281494, + "learning_rate": 7.011209984692375e-06, + "loss": 0.04, + "step": 266 + }, + { + "epoch": 1.7237479806138933, + "grad_norm": 0.3852980434894562, + "learning_rate": 6.985219005170973e-06, + "loss": 0.0432, + "step": 267 + }, + { + "epoch": 1.7302100161550888, + "grad_norm": 0.43154263496398926, + "learning_rate": 6.959164186443648e-06, + "loss": 0.0457, + "step": 268 + }, + { + "epoch": 1.7366720516962844, + "grad_norm": 0.4496629238128662, + "learning_rate": 6.933046366362011e-06, + "loss": 0.0557, + "step": 269 + }, + { + "epoch": 1.7431340872374799, + "grad_norm": 0.5219531059265137, + "learning_rate": 6.90686638480362e-06, + "loss": 0.0364, + "step": 270 + }, + { + "epoch": 1.7495961227786752, + "grad_norm": 0.525452196598053, + "learning_rate": 6.88062508364498e-06, + "loss": 0.0567, + "step": 271 + }, + { + "epoch": 1.7560581583198709, + "grad_norm": 0.3721238672733307, + "learning_rate": 6.8543233067344625e-06, + "loss": 0.0569, + "step": 272 + }, + { + "epoch": 1.7625201938610662, + "grad_norm": 0.49607154726982117, + "learning_rate": 6.827961899865178e-06, + "loss": 0.0419, + "step": 273 + }, + { + "epoch": 1.7689822294022617, + "grad_norm": 0.3797336220741272, + "learning_rate": 6.801541710747767e-06, + "loss": 0.0427, + "step": 274 + }, + { + "epoch": 1.7754442649434572, + "grad_norm": 0.3607645332813263, + "learning_rate": 6.775063588983153e-06, + "loss": 0.0411, + "step": 275 + }, + { + "epoch": 1.7819063004846527, + "grad_norm": 0.3796041011810303, + "learning_rate": 6.748528386035209e-06, + "loss": 0.0448, + "step": 276 + }, + { + "epoch": 1.7883683360258482, + "grad_norm": 0.39435023069381714, + "learning_rate": 6.7219369552033865e-06, + "loss": 0.0429, + "step": 277 + }, + { + "epoch": 1.7948303715670435, + "grad_norm": 0.6904699802398682, + "learning_rate": 6.695290151595271e-06, + "loss": 0.0507, + "step": 278 + }, + { + "epoch": 1.8012924071082392, + "grad_norm": 0.45726367831230164, + "learning_rate": 6.668588832099081e-06, + "loss": 0.0378, + "step": 279 + }, + { + "epoch": 1.8077544426494345, + "grad_norm": 0.31050002574920654, + "learning_rate": 6.6418338553561225e-06, + "loss": 0.0496, + "step": 280 + }, + { + "epoch": 1.81421647819063, + "grad_norm": 0.42013099789619446, + "learning_rate": 6.615026081733168e-06, + "loss": 0.0467, + "step": 281 + }, + { + "epoch": 1.8206785137318255, + "grad_norm": 0.3072008490562439, + "learning_rate": 6.5881663732947935e-06, + "loss": 0.0458, + "step": 282 + }, + { + "epoch": 1.827140549273021, + "grad_norm": 3.501995801925659, + "learning_rate": 6.561255593775656e-06, + "loss": 0.0385, + "step": 283 + }, + { + "epoch": 1.8336025848142166, + "grad_norm": 0.3589681386947632, + "learning_rate": 6.5342946085527205e-06, + "loss": 0.0368, + "step": 284 + }, + { + "epoch": 1.8400646203554119, + "grad_norm": 0.405106782913208, + "learning_rate": 6.507284284617427e-06, + "loss": 0.0499, + "step": 285 + }, + { + "epoch": 1.8465266558966076, + "grad_norm": 0.41938844323158264, + "learning_rate": 6.480225490547821e-06, + "loss": 0.056, + "step": 286 + }, + { + "epoch": 1.8529886914378029, + "grad_norm": 0.43309280276298523, + "learning_rate": 6.4531190964806005e-06, + "loss": 0.0434, + "step": 287 + }, + { + "epoch": 1.8594507269789984, + "grad_norm": 0.3927607238292694, + "learning_rate": 6.425965974083164e-06, + "loss": 0.0461, + "step": 288 + }, + { + "epoch": 1.865912762520194, + "grad_norm": 0.38023263216018677, + "learning_rate": 6.398766996525554e-06, + "loss": 0.0373, + "step": 289 + }, + { + "epoch": 1.8723747980613892, + "grad_norm": 0.3614664673805237, + "learning_rate": 6.371523038452398e-06, + "loss": 0.0421, + "step": 290 + }, + { + "epoch": 1.878836833602585, + "grad_norm": 0.40119874477386475, + "learning_rate": 6.344234975954765e-06, + "loss": 0.0433, + "step": 291 + }, + { + "epoch": 1.8852988691437802, + "grad_norm": 0.5142848491668701, + "learning_rate": 6.316903686542011e-06, + "loss": 0.0411, + "step": 292 + }, + { + "epoch": 1.891760904684976, + "grad_norm": 0.40000125765800476, + "learning_rate": 6.289530049113543e-06, + "loss": 0.0544, + "step": 293 + }, + { + "epoch": 1.8982229402261712, + "grad_norm": 3.5423994064331055, + "learning_rate": 6.262114943930566e-06, + "loss": 0.0466, + "step": 294 + }, + { + "epoch": 1.9046849757673667, + "grad_norm": 0.42096999287605286, + "learning_rate": 6.234659252587782e-06, + "loss": 0.0488, + "step": 295 + }, + { + "epoch": 1.9111470113085622, + "grad_norm": 0.45695486664772034, + "learning_rate": 6.20716385798502e-06, + "loss": 0.053, + "step": 296 + }, + { + "epoch": 1.9176090468497575, + "grad_norm": 0.42890581488609314, + "learning_rate": 6.17962964429887e-06, + "loss": 0.0476, + "step": 297 + }, + { + "epoch": 1.9240710823909533, + "grad_norm": 0.4009229242801666, + "learning_rate": 6.152057496954225e-06, + "loss": 0.0429, + "step": 298 + }, + { + "epoch": 1.9305331179321485, + "grad_norm": 0.7172302007675171, + "learning_rate": 6.12444830259583e-06, + "loss": 0.0504, + "step": 299 + }, + { + "epoch": 1.936995153473344, + "grad_norm": 0.4833920896053314, + "learning_rate": 6.096802949059757e-06, + "loss": 0.0538, + "step": 300 + }, + { + "epoch": 1.9434571890145396, + "grad_norm": 3.44974684715271, + "learning_rate": 6.069122325344857e-06, + "loss": 0.0408, + "step": 301 + }, + { + "epoch": 1.949919224555735, + "grad_norm": 0.7173548340797424, + "learning_rate": 6.041407321584178e-06, + "loss": 0.0485, + "step": 302 + }, + { + "epoch": 1.9563812600969306, + "grad_norm": 0.4045400023460388, + "learning_rate": 6.013658829016328e-06, + "loss": 0.0433, + "step": 303 + }, + { + "epoch": 1.9628432956381259, + "grad_norm": 0.38726726174354553, + "learning_rate": 5.9858777399568325e-06, + "loss": 0.044, + "step": 304 + }, + { + "epoch": 1.9693053311793216, + "grad_norm": 0.4238860011100769, + "learning_rate": 5.958064947769423e-06, + "loss": 0.0488, + "step": 305 + }, + { + "epoch": 1.975767366720517, + "grad_norm": 0.41494399309158325, + "learning_rate": 5.930221346837324e-06, + "loss": 0.0444, + "step": 306 + }, + { + "epoch": 1.9822294022617124, + "grad_norm": 1.2054787874221802, + "learning_rate": 5.902347832534475e-06, + "loss": 0.0639, + "step": 307 + }, + { + "epoch": 1.988691437802908, + "grad_norm": 0.4015323519706726, + "learning_rate": 5.874445301196761e-06, + "loss": 0.0392, + "step": 308 + }, + { + "epoch": 1.9951534733441034, + "grad_norm": 0.4661884307861328, + "learning_rate": 5.846514650093162e-06, + "loss": 0.0487, + "step": 309 + }, + { + "epoch": 2.0, + "grad_norm": 0.43507838249206543, + "learning_rate": 5.818556777396923e-06, + "loss": 0.0566, + "step": 310 + }, + { + "epoch": 2.0064620355411953, + "grad_norm": 0.52005934715271, + "learning_rate": 5.790572582156654e-06, + "loss": 0.0266, + "step": 311 + }, + { + "epoch": 2.012924071082391, + "grad_norm": 0.3060389757156372, + "learning_rate": 5.76256296426743e-06, + "loss": 0.0255, + "step": 312 + }, + { + "epoch": 2.0193861066235863, + "grad_norm": 0.26340675354003906, + "learning_rate": 5.734528824441845e-06, + "loss": 0.0226, + "step": 313 + }, + { + "epoch": 2.025848142164782, + "grad_norm": 0.29653921723365784, + "learning_rate": 5.706471064181055e-06, + "loss": 0.0246, + "step": 314 + }, + { + "epoch": 2.0323101777059773, + "grad_norm": 0.2679958939552307, + "learning_rate": 5.678390585745784e-06, + "loss": 0.019, + "step": 315 + }, + { + "epoch": 2.038772213247173, + "grad_norm": 0.40780436992645264, + "learning_rate": 5.6502882921273084e-06, + "loss": 0.0223, + "step": 316 + }, + { + "epoch": 2.0452342487883683, + "grad_norm": 4.025135517120361, + "learning_rate": 5.6221650870184215e-06, + "loss": 0.0289, + "step": 317 + }, + { + "epoch": 2.0516962843295636, + "grad_norm": 0.3368605971336365, + "learning_rate": 5.594021874784376e-06, + "loss": 0.0216, + "step": 318 + }, + { + "epoch": 2.0581583198707594, + "grad_norm": 0.37941357493400574, + "learning_rate": 5.565859560433792e-06, + "loss": 0.028, + "step": 319 + }, + { + "epoch": 2.0646203554119547, + "grad_norm": 0.39574962854385376, + "learning_rate": 5.537679049589568e-06, + "loss": 0.0359, + "step": 320 + }, + { + "epoch": 2.0710823909531504, + "grad_norm": 0.3605610728263855, + "learning_rate": 5.50948124845975e-06, + "loss": 0.0228, + "step": 321 + }, + { + "epoch": 2.0775444264943457, + "grad_norm": 0.3201454281806946, + "learning_rate": 5.481267063808392e-06, + "loss": 0.0241, + "step": 322 + }, + { + "epoch": 2.0840064620355414, + "grad_norm": 0.32988935708999634, + "learning_rate": 5.453037402926397e-06, + "loss": 0.0232, + "step": 323 + }, + { + "epoch": 2.0904684975767367, + "grad_norm": 0.43368834257125854, + "learning_rate": 5.4247931736023385e-06, + "loss": 0.0219, + "step": 324 + }, + { + "epoch": 2.096930533117932, + "grad_norm": 0.37438321113586426, + "learning_rate": 5.396535284093278e-06, + "loss": 0.0216, + "step": 325 + }, + { + "epoch": 2.1033925686591277, + "grad_norm": 0.41498661041259766, + "learning_rate": 5.368264643095543e-06, + "loss": 0.0214, + "step": 326 + }, + { + "epoch": 2.109854604200323, + "grad_norm": 0.43471306562423706, + "learning_rate": 5.3399821597155225e-06, + "loss": 0.0194, + "step": 327 + }, + { + "epoch": 2.1163166397415187, + "grad_norm": 0.550999104976654, + "learning_rate": 5.3116887434404155e-06, + "loss": 0.0264, + "step": 328 + }, + { + "epoch": 2.122778675282714, + "grad_norm": 0.4273984730243683, + "learning_rate": 5.283385304109e-06, + "loss": 0.0238, + "step": 329 + }, + { + "epoch": 2.1292407108239093, + "grad_norm": 0.34016546607017517, + "learning_rate": 5.255072751882363e-06, + "loss": 0.0221, + "step": 330 + }, + { + "epoch": 2.135702746365105, + "grad_norm": 0.43326249718666077, + "learning_rate": 5.22675199721464e-06, + "loss": 0.0273, + "step": 331 + }, + { + "epoch": 2.1421647819063003, + "grad_norm": 0.3180527985095978, + "learning_rate": 5.198423950823734e-06, + "loss": 0.0243, + "step": 332 + }, + { + "epoch": 2.148626817447496, + "grad_norm": 0.28859448432922363, + "learning_rate": 5.170089523662028e-06, + "loss": 0.0272, + "step": 333 + }, + { + "epoch": 2.1550888529886914, + "grad_norm": 11.164939880371094, + "learning_rate": 5.141749626887101e-06, + "loss": 0.0306, + "step": 334 + }, + { + "epoch": 2.161550888529887, + "grad_norm": 0.3972926139831543, + "learning_rate": 5.113405171832404e-06, + "loss": 0.0209, + "step": 335 + }, + { + "epoch": 2.1680129240710824, + "grad_norm": 0.32792770862579346, + "learning_rate": 5.0850570699779875e-06, + "loss": 0.029, + "step": 336 + }, + { + "epoch": 2.1744749596122777, + "grad_norm": 0.5258669853210449, + "learning_rate": 5.05670623292116e-06, + "loss": 0.0469, + "step": 337 + }, + { + "epoch": 2.1809369951534734, + "grad_norm": 4.376863479614258, + "learning_rate": 5.028353572347195e-06, + "loss": 0.0329, + "step": 338 + }, + { + "epoch": 2.1873990306946687, + "grad_norm": 0.38787195086479187, + "learning_rate": 5e-06, + "loss": 0.025, + "step": 339 + }, + { + "epoch": 2.1938610662358644, + "grad_norm": 0.3971186578273773, + "learning_rate": 4.971646427652806e-06, + "loss": 0.022, + "step": 340 + }, + { + "epoch": 2.2003231017770597, + "grad_norm": 0.4147299528121948, + "learning_rate": 4.94329376707884e-06, + "loss": 0.0353, + "step": 341 + }, + { + "epoch": 2.2067851373182554, + "grad_norm": 0.3218820095062256, + "learning_rate": 4.914942930022014e-06, + "loss": 0.0223, + "step": 342 + }, + { + "epoch": 2.2132471728594507, + "grad_norm": 0.3497336208820343, + "learning_rate": 4.8865948281675976e-06, + "loss": 0.0245, + "step": 343 + }, + { + "epoch": 2.219709208400646, + "grad_norm": 0.5387941002845764, + "learning_rate": 4.858250373112901e-06, + "loss": 0.0375, + "step": 344 + }, + { + "epoch": 2.2261712439418417, + "grad_norm": 0.3683622181415558, + "learning_rate": 4.829910476337972e-06, + "loss": 0.0174, + "step": 345 + }, + { + "epoch": 2.232633279483037, + "grad_norm": 0.40947481989860535, + "learning_rate": 4.801576049176269e-06, + "loss": 0.0211, + "step": 346 + }, + { + "epoch": 2.2390953150242328, + "grad_norm": 0.38499361276626587, + "learning_rate": 4.773248002785362e-06, + "loss": 0.0229, + "step": 347 + }, + { + "epoch": 2.245557350565428, + "grad_norm": 0.45877766609191895, + "learning_rate": 4.744927248117639e-06, + "loss": 0.0271, + "step": 348 + }, + { + "epoch": 2.2520193861066238, + "grad_norm": 0.6704347133636475, + "learning_rate": 4.716614695891002e-06, + "loss": 0.0223, + "step": 349 + }, + { + "epoch": 2.258481421647819, + "grad_norm": 21.48920440673828, + "learning_rate": 4.688311256559587e-06, + "loss": 0.0235, + "step": 350 + }, + { + "epoch": 2.2649434571890144, + "grad_norm": 0.33492201566696167, + "learning_rate": 4.66001784028448e-06, + "loss": 0.0269, + "step": 351 + }, + { + "epoch": 2.27140549273021, + "grad_norm": 0.39123740792274475, + "learning_rate": 4.631735356904458e-06, + "loss": 0.0222, + "step": 352 + }, + { + "epoch": 2.2778675282714054, + "grad_norm": 0.3031889796257019, + "learning_rate": 4.6034647159067234e-06, + "loss": 0.0225, + "step": 353 + }, + { + "epoch": 2.284329563812601, + "grad_norm": 0.3864370584487915, + "learning_rate": 4.575206826397662e-06, + "loss": 0.0222, + "step": 354 + }, + { + "epoch": 2.2907915993537964, + "grad_norm": 0.6137961149215698, + "learning_rate": 4.546962597073607e-06, + "loss": 0.0276, + "step": 355 + }, + { + "epoch": 2.297253634894992, + "grad_norm": 0.4349140226840973, + "learning_rate": 4.5187329361916095e-06, + "loss": 0.0248, + "step": 356 + }, + { + "epoch": 2.3037156704361874, + "grad_norm": 0.3713863492012024, + "learning_rate": 4.490518751540251e-06, + "loss": 0.0268, + "step": 357 + }, + { + "epoch": 2.3101777059773827, + "grad_norm": 0.39676621556282043, + "learning_rate": 4.462320950410432e-06, + "loss": 0.0262, + "step": 358 + }, + { + "epoch": 2.3166397415185784, + "grad_norm": 0.37068402767181396, + "learning_rate": 4.4341404395662105e-06, + "loss": 0.0211, + "step": 359 + }, + { + "epoch": 2.3231017770597737, + "grad_norm": 0.29026004672050476, + "learning_rate": 4.405978125215627e-06, + "loss": 0.0208, + "step": 360 + }, + { + "epoch": 2.3295638126009695, + "grad_norm": 0.46004900336265564, + "learning_rate": 4.377834912981579e-06, + "loss": 0.0267, + "step": 361 + }, + { + "epoch": 2.3360258481421647, + "grad_norm": 0.38983336091041565, + "learning_rate": 4.3497117078726915e-06, + "loss": 0.026, + "step": 362 + }, + { + "epoch": 2.3424878836833605, + "grad_norm": 0.485213965177536, + "learning_rate": 4.321609414254217e-06, + "loss": 0.0224, + "step": 363 + }, + { + "epoch": 2.3489499192245558, + "grad_norm": 0.3934685289859772, + "learning_rate": 4.2935289358189454e-06, + "loss": 0.022, + "step": 364 + }, + { + "epoch": 2.355411954765751, + "grad_norm": 0.4222012460231781, + "learning_rate": 4.265471175558156e-06, + "loss": 0.0198, + "step": 365 + }, + { + "epoch": 2.361873990306947, + "grad_norm": 0.3215023875236511, + "learning_rate": 4.237437035732572e-06, + "loss": 0.0252, + "step": 366 + }, + { + "epoch": 2.368336025848142, + "grad_norm": 0.38228940963745117, + "learning_rate": 4.2094274178433455e-06, + "loss": 0.026, + "step": 367 + }, + { + "epoch": 2.374798061389338, + "grad_norm": 0.44622451066970825, + "learning_rate": 4.18144322260308e-06, + "loss": 0.0323, + "step": 368 + }, + { + "epoch": 2.381260096930533, + "grad_norm": 0.42110568284988403, + "learning_rate": 4.153485349906839e-06, + "loss": 0.0175, + "step": 369 + }, + { + "epoch": 2.387722132471729, + "grad_norm": 0.4341752827167511, + "learning_rate": 4.125554698803241e-06, + "loss": 0.0224, + "step": 370 + }, + { + "epoch": 2.394184168012924, + "grad_norm": 0.6297674775123596, + "learning_rate": 4.0976521674655255e-06, + "loss": 0.0223, + "step": 371 + }, + { + "epoch": 2.4006462035541194, + "grad_norm": 0.34877192974090576, + "learning_rate": 4.069778653162679e-06, + "loss": 0.0223, + "step": 372 + }, + { + "epoch": 2.407108239095315, + "grad_norm": 0.3711933493614197, + "learning_rate": 4.041935052230579e-06, + "loss": 0.018, + "step": 373 + }, + { + "epoch": 2.4135702746365104, + "grad_norm": 0.2846461236476898, + "learning_rate": 4.014122260043169e-06, + "loss": 0.0196, + "step": 374 + }, + { + "epoch": 2.420032310177706, + "grad_norm": 0.3798210024833679, + "learning_rate": 3.986341170983672e-06, + "loss": 0.0252, + "step": 375 + }, + { + "epoch": 2.4264943457189014, + "grad_norm": 0.45135289430618286, + "learning_rate": 3.958592678415825e-06, + "loss": 0.0272, + "step": 376 + }, + { + "epoch": 2.432956381260097, + "grad_norm": 0.3350200355052948, + "learning_rate": 3.9308776746551444e-06, + "loss": 0.0208, + "step": 377 + }, + { + "epoch": 2.4394184168012925, + "grad_norm": 0.3789970576763153, + "learning_rate": 3.903197050940244e-06, + "loss": 0.0206, + "step": 378 + }, + { + "epoch": 2.4458804523424877, + "grad_norm": 0.40493834018707275, + "learning_rate": 3.875551697404172e-06, + "loss": 0.0225, + "step": 379 + }, + { + "epoch": 2.4523424878836835, + "grad_norm": 0.45063942670822144, + "learning_rate": 3.847942503045776e-06, + "loss": 0.0246, + "step": 380 + }, + { + "epoch": 2.4588045234248788, + "grad_norm": 0.36469295620918274, + "learning_rate": 3.820370355701133e-06, + "loss": 0.0288, + "step": 381 + }, + { + "epoch": 2.4652665589660745, + "grad_norm": 0.3883582055568695, + "learning_rate": 3.792836142014981e-06, + "loss": 0.0305, + "step": 382 + }, + { + "epoch": 2.47172859450727, + "grad_norm": 0.41473427414894104, + "learning_rate": 3.7653407474122195e-06, + "loss": 0.0216, + "step": 383 + }, + { + "epoch": 2.4781906300484655, + "grad_norm": 0.44368648529052734, + "learning_rate": 3.7378850560694337e-06, + "loss": 0.0297, + "step": 384 + }, + { + "epoch": 2.484652665589661, + "grad_norm": 0.5202524662017822, + "learning_rate": 3.7104699508864606e-06, + "loss": 0.0258, + "step": 385 + }, + { + "epoch": 2.491114701130856, + "grad_norm": 0.41755107045173645, + "learning_rate": 3.683096313457991e-06, + "loss": 0.0319, + "step": 386 + }, + { + "epoch": 2.497576736672052, + "grad_norm": 0.4544621706008911, + "learning_rate": 3.6557650240452358e-06, + "loss": 0.027, + "step": 387 + }, + { + "epoch": 2.504038772213247, + "grad_norm": 0.3956614136695862, + "learning_rate": 3.6284769615476045e-06, + "loss": 0.018, + "step": 388 + }, + { + "epoch": 2.5105008077544424, + "grad_norm": 0.3755435049533844, + "learning_rate": 3.601233003474448e-06, + "loss": 0.0223, + "step": 389 + }, + { + "epoch": 2.516962843295638, + "grad_norm": 0.36021357774734497, + "learning_rate": 3.5740340259168383e-06, + "loss": 0.0203, + "step": 390 + }, + { + "epoch": 2.523424878836834, + "grad_norm": 0.35905206203460693, + "learning_rate": 3.5468809035194008e-06, + "loss": 0.0213, + "step": 391 + }, + { + "epoch": 2.529886914378029, + "grad_norm": 0.35023632645606995, + "learning_rate": 3.519774509452181e-06, + "loss": 0.0194, + "step": 392 + }, + { + "epoch": 2.5363489499192244, + "grad_norm": 0.3874582052230835, + "learning_rate": 3.4927157153825717e-06, + "loss": 0.0255, + "step": 393 + }, + { + "epoch": 2.54281098546042, + "grad_norm": 0.3666871190071106, + "learning_rate": 3.4657053914472816e-06, + "loss": 0.0205, + "step": 394 + }, + { + "epoch": 2.5492730210016155, + "grad_norm": 0.47673851251602173, + "learning_rate": 3.4387444062243453e-06, + "loss": 0.0271, + "step": 395 + }, + { + "epoch": 2.5557350565428107, + "grad_norm": 0.3491911292076111, + "learning_rate": 3.4118336267052086e-06, + "loss": 0.0191, + "step": 396 + }, + { + "epoch": 2.5621970920840065, + "grad_norm": 0.3430013656616211, + "learning_rate": 3.384973918266834e-06, + "loss": 0.019, + "step": 397 + }, + { + "epoch": 2.568659127625202, + "grad_norm": 0.3667357265949249, + "learning_rate": 3.3581661446438796e-06, + "loss": 0.0241, + "step": 398 + }, + { + "epoch": 2.5751211631663975, + "grad_norm": 0.395435094833374, + "learning_rate": 3.3314111679009203e-06, + "loss": 0.0295, + "step": 399 + }, + { + "epoch": 2.581583198707593, + "grad_norm": 0.4434990882873535, + "learning_rate": 3.3047098484047314e-06, + "loss": 0.023, + "step": 400 + }, + { + "epoch": 2.5880452342487885, + "grad_norm": 0.40480467677116394, + "learning_rate": 3.2780630447966135e-06, + "loss": 0.0245, + "step": 401 + }, + { + "epoch": 2.594507269789984, + "grad_norm": 0.39888110756874084, + "learning_rate": 3.251471613964793e-06, + "loss": 0.0356, + "step": 402 + }, + { + "epoch": 2.600969305331179, + "grad_norm": 0.3975083529949188, + "learning_rate": 3.224936411016849e-06, + "loss": 0.0212, + "step": 403 + }, + { + "epoch": 2.607431340872375, + "grad_norm": 0.31860825419425964, + "learning_rate": 3.198458289252234e-06, + "loss": 0.0148, + "step": 404 + }, + { + "epoch": 2.61389337641357, + "grad_norm": 0.5908383727073669, + "learning_rate": 3.172038100134823e-06, + "loss": 0.0215, + "step": 405 + }, + { + "epoch": 2.620355411954766, + "grad_norm": 0.32025986909866333, + "learning_rate": 3.145676693265537e-06, + "loss": 0.0189, + "step": 406 + }, + { + "epoch": 2.626817447495961, + "grad_norm": 0.3670276403427124, + "learning_rate": 3.1193749163550226e-06, + "loss": 0.0229, + "step": 407 + }, + { + "epoch": 2.633279483037157, + "grad_norm": 0.3176894187927246, + "learning_rate": 3.093133615196381e-06, + "loss": 0.0212, + "step": 408 + }, + { + "epoch": 2.639741518578352, + "grad_norm": 0.41281652450561523, + "learning_rate": 3.0669536336379906e-06, + "loss": 0.0191, + "step": 409 + }, + { + "epoch": 2.6462035541195474, + "grad_norm": 0.38180580735206604, + "learning_rate": 3.040835813556352e-06, + "loss": 0.02, + "step": 410 + }, + { + "epoch": 2.652665589660743, + "grad_norm": 0.2951391041278839, + "learning_rate": 3.014780994829029e-06, + "loss": 0.0162, + "step": 411 + }, + { + "epoch": 2.6591276252019385, + "grad_norm": 0.45817264914512634, + "learning_rate": 2.988790015307627e-06, + "loss": 0.0215, + "step": 412 + }, + { + "epoch": 2.665589660743134, + "grad_norm": 0.4247450828552246, + "learning_rate": 2.9628637107908614e-06, + "loss": 0.0271, + "step": 413 + }, + { + "epoch": 2.6720516962843295, + "grad_norm": 0.3811360001564026, + "learning_rate": 2.937002914997679e-06, + "loss": 0.0276, + "step": 414 + }, + { + "epoch": 2.678513731825525, + "grad_norm": 0.3762235641479492, + "learning_rate": 2.911208459540442e-06, + "loss": 0.021, + "step": 415 + }, + { + "epoch": 2.6849757673667205, + "grad_norm": 0.4333343207836151, + "learning_rate": 2.8854811738981848e-06, + "loss": 0.0247, + "step": 416 + }, + { + "epoch": 2.691437802907916, + "grad_norm": 0.36235716938972473, + "learning_rate": 2.859821885389957e-06, + "loss": 0.024, + "step": 417 + }, + { + "epoch": 2.6978998384491115, + "grad_norm": 0.3801264762878418, + "learning_rate": 2.8342314191481952e-06, + "loss": 0.0243, + "step": 418 + }, + { + "epoch": 2.704361873990307, + "grad_norm": 0.37116238474845886, + "learning_rate": 2.808710598092206e-06, + "loss": 0.0202, + "step": 419 + }, + { + "epoch": 2.7108239095315025, + "grad_norm": 0.36050671339035034, + "learning_rate": 2.783260242901694e-06, + "loss": 0.023, + "step": 420 + }, + { + "epoch": 2.717285945072698, + "grad_norm": 0.4219389855861664, + "learning_rate": 2.7578811719903788e-06, + "loss": 0.0201, + "step": 421 + }, + { + "epoch": 2.7237479806138936, + "grad_norm": 0.40774789452552795, + "learning_rate": 2.7325742014796695e-06, + "loss": 0.0272, + "step": 422 + }, + { + "epoch": 2.730210016155089, + "grad_norm": 0.3500642478466034, + "learning_rate": 2.707340145172423e-06, + "loss": 0.0209, + "step": 423 + }, + { + "epoch": 2.736672051696284, + "grad_norm": 0.40537238121032715, + "learning_rate": 2.682179814526783e-06, + "loss": 0.0192, + "step": 424 + }, + { + "epoch": 2.74313408723748, + "grad_norm": 0.346743643283844, + "learning_rate": 2.6570940186300655e-06, + "loss": 0.0208, + "step": 425 + }, + { + "epoch": 2.749596122778675, + "grad_norm": 0.504565954208374, + "learning_rate": 2.6320835641727615e-06, + "loss": 0.026, + "step": 426 + }, + { + "epoch": 2.756058158319871, + "grad_norm": 0.4314068853855133, + "learning_rate": 2.607149255422584e-06, + "loss": 0.0193, + "step": 427 + }, + { + "epoch": 2.762520193861066, + "grad_norm": 0.32916414737701416, + "learning_rate": 2.582291894198617e-06, + "loss": 0.0196, + "step": 428 + }, + { + "epoch": 2.768982229402262, + "grad_norm": 1.6414940357208252, + "learning_rate": 2.557512279845509e-06, + "loss": 0.0211, + "step": 429 + }, + { + "epoch": 2.775444264943457, + "grad_norm": 0.4827597141265869, + "learning_rate": 2.5328112092077882e-06, + "loss": 0.0234, + "step": 430 + }, + { + "epoch": 2.7819063004846525, + "grad_norm": 0.45988279581069946, + "learning_rate": 2.5081894766042393e-06, + "loss": 0.0282, + "step": 431 + }, + { + "epoch": 2.788368336025848, + "grad_norm": 0.5003206133842468, + "learning_rate": 2.4836478738023424e-06, + "loss": 0.0236, + "step": 432 + }, + { + "epoch": 2.7948303715670435, + "grad_norm": 0.44246071577072144, + "learning_rate": 2.4591871899928286e-06, + "loss": 0.0374, + "step": 433 + }, + { + "epoch": 2.8012924071082392, + "grad_norm": 0.4109882712364197, + "learning_rate": 2.434808211764294e-06, + "loss": 0.0218, + "step": 434 + }, + { + "epoch": 2.8077544426494345, + "grad_norm": 0.46064233779907227, + "learning_rate": 2.410511723077907e-06, + "loss": 0.0178, + "step": 435 + }, + { + "epoch": 2.8142164781906303, + "grad_norm": 0.33899685740470886, + "learning_rate": 2.386298505242198e-06, + "loss": 0.0296, + "step": 436 + }, + { + "epoch": 2.8206785137318255, + "grad_norm": 0.37225639820098877, + "learning_rate": 2.3621693368879363e-06, + "loss": 0.0268, + "step": 437 + }, + { + "epoch": 2.827140549273021, + "grad_norm": 0.49789291620254517, + "learning_rate": 2.3381249939430882e-06, + "loss": 0.0288, + "step": 438 + }, + { + "epoch": 2.8336025848142166, + "grad_norm": 0.4957059919834137, + "learning_rate": 2.3141662496078695e-06, + "loss": 0.0194, + "step": 439 + }, + { + "epoch": 2.840064620355412, + "grad_norm": 0.36811667680740356, + "learning_rate": 2.2902938743298765e-06, + "loss": 0.0219, + "step": 440 + }, + { + "epoch": 2.8465266558966076, + "grad_norm": 0.3737165927886963, + "learning_rate": 2.2665086357793155e-06, + "loss": 0.0217, + "step": 441 + }, + { + "epoch": 2.852988691437803, + "grad_norm": 0.46944284439086914, + "learning_rate": 2.242811298824312e-06, + "loss": 0.0218, + "step": 442 + }, + { + "epoch": 2.8594507269789986, + "grad_norm": 2.569209098815918, + "learning_rate": 2.21920262550632e-06, + "loss": 0.0226, + "step": 443 + }, + { + "epoch": 2.865912762520194, + "grad_norm": 0.46464183926582336, + "learning_rate": 2.1956833750156086e-06, + "loss": 0.0273, + "step": 444 + }, + { + "epoch": 2.872374798061389, + "grad_norm": 0.36627525091171265, + "learning_rate": 2.1722543036668613e-06, + "loss": 0.0237, + "step": 445 + }, + { + "epoch": 2.878836833602585, + "grad_norm": 0.42264869809150696, + "learning_rate": 2.1489161648748436e-06, + "loss": 0.0207, + "step": 446 + }, + { + "epoch": 2.88529886914378, + "grad_norm": 0.32069268822669983, + "learning_rate": 2.125669709130174e-06, + "loss": 0.0187, + "step": 447 + }, + { + "epoch": 2.891760904684976, + "grad_norm": 0.3201678693294525, + "learning_rate": 2.102515683975201e-06, + "loss": 0.0184, + "step": 448 + }, + { + "epoch": 2.898222940226171, + "grad_norm": 0.3414503037929535, + "learning_rate": 2.0794548339799605e-06, + "loss": 0.0191, + "step": 449 + }, + { + "epoch": 2.904684975767367, + "grad_norm": 0.33603090047836304, + "learning_rate": 2.056487900718227e-06, + "loss": 0.0186, + "step": 450 + }, + { + "epoch": 2.9111470113085622, + "grad_norm": 0.40074023604393005, + "learning_rate": 2.0336156227436653e-06, + "loss": 0.0184, + "step": 451 + }, + { + "epoch": 2.9176090468497575, + "grad_norm": 0.37722983956336975, + "learning_rate": 2.010838735566096e-06, + "loss": 0.0257, + "step": 452 + }, + { + "epoch": 2.9240710823909533, + "grad_norm": 0.3689229488372803, + "learning_rate": 1.9881579716278267e-06, + "loss": 0.0182, + "step": 453 + }, + { + "epoch": 2.9305331179321485, + "grad_norm": 0.4667937755584717, + "learning_rate": 1.9655740602801055e-06, + "loss": 0.0373, + "step": 454 + }, + { + "epoch": 2.936995153473344, + "grad_norm": 0.3728683888912201, + "learning_rate": 1.943087727759663e-06, + "loss": 0.0221, + "step": 455 + }, + { + "epoch": 2.9434571890145396, + "grad_norm": 0.38799959421157837, + "learning_rate": 1.92069969716537e-06, + "loss": 0.0238, + "step": 456 + }, + { + "epoch": 2.9499192245557353, + "grad_norm": 0.5497745275497437, + "learning_rate": 1.8984106884349702e-06, + "loss": 0.0291, + "step": 457 + }, + { + "epoch": 2.9563812600969306, + "grad_norm": 0.3938653767108917, + "learning_rate": 1.8762214183219379e-06, + "loss": 0.0215, + "step": 458 + }, + { + "epoch": 2.962843295638126, + "grad_norm": 0.35690999031066895, + "learning_rate": 1.8541326003724258e-06, + "loss": 0.0191, + "step": 459 + }, + { + "epoch": 2.9693053311793216, + "grad_norm": 0.3347409665584564, + "learning_rate": 1.8321449449023215e-06, + "loss": 0.0205, + "step": 460 + }, + { + "epoch": 2.975767366720517, + "grad_norm": 0.3588216006755829, + "learning_rate": 1.8102591589744016e-06, + "loss": 0.0246, + "step": 461 + }, + { + "epoch": 2.982229402261712, + "grad_norm": 0.4010322391986847, + "learning_rate": 1.7884759463755984e-06, + "loss": 0.0196, + "step": 462 + }, + { + "epoch": 2.988691437802908, + "grad_norm": 0.3672601878643036, + "learning_rate": 1.7667960075943723e-06, + "loss": 0.023, + "step": 463 + }, + { + "epoch": 2.9951534733441036, + "grad_norm": 0.5553978085517883, + "learning_rate": 1.7452200397981706e-06, + "loss": 0.0167, + "step": 464 + }, + { + "epoch": 3.0, + "grad_norm": 0.5560592412948608, + "learning_rate": 1.723748736811025e-06, + "loss": 0.0229, + "step": 465 + }, + { + "epoch": 3.0064620355411953, + "grad_norm": 0.32263022661209106, + "learning_rate": 1.7023827890912302e-06, + "loss": 0.016, + "step": 466 + }, + { + "epoch": 3.012924071082391, + "grad_norm": 0.3601260185241699, + "learning_rate": 1.681122883709152e-06, + "loss": 0.013, + "step": 467 + }, + { + "epoch": 3.0193861066235863, + "grad_norm": 2.2087621688842773, + "learning_rate": 1.6599697043251128e-06, + "loss": 0.0129, + "step": 468 + }, + { + "epoch": 3.025848142164782, + "grad_norm": 0.4122757315635681, + "learning_rate": 1.638923931167427e-06, + "loss": 0.0119, + "step": 469 + }, + { + "epoch": 3.0323101777059773, + "grad_norm": 0.24214933812618256, + "learning_rate": 1.6179862410105197e-06, + "loss": 0.0079, + "step": 470 + }, + { + "epoch": 3.038772213247173, + "grad_norm": 0.2189481109380722, + "learning_rate": 1.5971573071531588e-06, + "loss": 0.0124, + "step": 471 + }, + { + "epoch": 3.0452342487883683, + "grad_norm": 1.217540979385376, + "learning_rate": 1.5764377993968094e-06, + "loss": 0.011, + "step": 472 + }, + { + "epoch": 3.0516962843295636, + "grad_norm": 0.3144000768661499, + "learning_rate": 1.5558283840240924e-06, + "loss": 0.0115, + "step": 473 + }, + { + "epoch": 3.0581583198707594, + "grad_norm": 0.23900973796844482, + "learning_rate": 1.5353297237773595e-06, + "loss": 0.0088, + "step": 474 + }, + { + "epoch": 3.0646203554119547, + "grad_norm": 0.2415008693933487, + "learning_rate": 1.5149424778373811e-06, + "loss": 0.0097, + "step": 475 + }, + { + "epoch": 3.0710823909531504, + "grad_norm": 0.24851197004318237, + "learning_rate": 1.4946673018021484e-06, + "loss": 0.0114, + "step": 476 + }, + { + "epoch": 3.0775444264943457, + "grad_norm": 0.32280173897743225, + "learning_rate": 1.474504847665791e-06, + "loss": 0.0113, + "step": 477 + }, + { + "epoch": 3.0840064620355414, + "grad_norm": 0.2650688588619232, + "learning_rate": 1.4544557637976108e-06, + "loss": 0.0092, + "step": 478 + }, + { + "epoch": 3.0904684975767367, + "grad_norm": 0.39706680178642273, + "learning_rate": 1.4345206949212338e-06, + "loss": 0.0107, + "step": 479 + }, + { + "epoch": 3.096930533117932, + "grad_norm": 0.27723875641822815, + "learning_rate": 1.4147002820938743e-06, + "loss": 0.0124, + "step": 480 + }, + { + "epoch": 3.1033925686591277, + "grad_norm": 0.35606086254119873, + "learning_rate": 1.3949951626857244e-06, + "loss": 0.0092, + "step": 481 + }, + { + "epoch": 3.109854604200323, + "grad_norm": 0.3037882447242737, + "learning_rate": 1.375405970359453e-06, + "loss": 0.0111, + "step": 482 + }, + { + "epoch": 3.1163166397415187, + "grad_norm": 0.3035484254360199, + "learning_rate": 1.3559333350498332e-06, + "loss": 0.0118, + "step": 483 + }, + { + "epoch": 3.122778675282714, + "grad_norm": 0.32289376854896545, + "learning_rate": 1.3365778829434834e-06, + "loss": 0.0123, + "step": 484 + }, + { + "epoch": 3.1292407108239093, + "grad_norm": 0.3762624263763428, + "learning_rate": 1.3173402364587307e-06, + "loss": 0.0087, + "step": 485 + }, + { + "epoch": 3.135702746365105, + "grad_norm": 0.21013927459716797, + "learning_rate": 1.298221014225597e-06, + "loss": 0.0065, + "step": 486 + }, + { + "epoch": 3.1421647819063003, + "grad_norm": 0.352093368768692, + "learning_rate": 1.2792208310659015e-06, + "loss": 0.0135, + "step": 487 + }, + { + "epoch": 3.148626817447496, + "grad_norm": 0.25764092803001404, + "learning_rate": 1.2603402979734992e-06, + "loss": 0.0092, + "step": 488 + }, + { + "epoch": 3.1550888529886914, + "grad_norm": 0.26913541555404663, + "learning_rate": 1.2415800220946223e-06, + "loss": 0.0057, + "step": 489 + }, + { + "epoch": 3.161550888529887, + "grad_norm": 0.3840892016887665, + "learning_rate": 1.2229406067083566e-06, + "loss": 0.0126, + "step": 490 + }, + { + "epoch": 3.1680129240710824, + "grad_norm": 0.6032068133354187, + "learning_rate": 1.2044226512072537e-06, + "loss": 0.0129, + "step": 491 + }, + { + "epoch": 3.1744749596122777, + "grad_norm": 0.2559524178504944, + "learning_rate": 1.1860267510780432e-06, + "loss": 0.0078, + "step": 492 + }, + { + "epoch": 3.1809369951534734, + "grad_norm": 1.6579912900924683, + "learning_rate": 1.1677534978824906e-06, + "loss": 0.0126, + "step": 493 + }, + { + "epoch": 3.1873990306946687, + "grad_norm": 0.2948991060256958, + "learning_rate": 1.1496034792383654e-06, + "loss": 0.0087, + "step": 494 + }, + { + "epoch": 3.1938610662358644, + "grad_norm": 0.32418084144592285, + "learning_rate": 1.1315772788005603e-06, + "loss": 0.0093, + "step": 495 + }, + { + "epoch": 3.2003231017770597, + "grad_norm": 0.42377397418022156, + "learning_rate": 1.1136754762423097e-06, + "loss": 0.0102, + "step": 496 + }, + { + "epoch": 3.2067851373182554, + "grad_norm": 0.34062591195106506, + "learning_rate": 1.0958986472365518e-06, + "loss": 0.0176, + "step": 497 + }, + { + "epoch": 3.2132471728594507, + "grad_norm": 0.43844881653785706, + "learning_rate": 1.0782473634374191e-06, + "loss": 0.0132, + "step": 498 + }, + { + "epoch": 3.219709208400646, + "grad_norm": 0.4194018244743347, + "learning_rate": 1.0607221924618533e-06, + "loss": 0.0103, + "step": 499 + }, + { + "epoch": 3.2261712439418417, + "grad_norm": 0.27430325746536255, + "learning_rate": 1.0433236978713546e-06, + "loss": 0.0085, + "step": 500 + }, + { + "epoch": 3.232633279483037, + "grad_norm": 0.1858188509941101, + "learning_rate": 1.0260524391538546e-06, + "loss": 0.0092, + "step": 501 + }, + { + "epoch": 3.2390953150242328, + "grad_norm": 0.43904629349708557, + "learning_rate": 1.0089089717057337e-06, + "loss": 0.0107, + "step": 502 + }, + { + "epoch": 3.245557350565428, + "grad_norm": 0.3002469539642334, + "learning_rate": 9.91893846813947e-07, + "loss": 0.0086, + "step": 503 + }, + { + "epoch": 3.2520193861066238, + "grad_norm": 0.2562623918056488, + "learning_rate": 9.7500761163831e-07, + "loss": 0.007, + "step": 504 + }, + { + "epoch": 3.258481421647819, + "grad_norm": 0.29067540168762207, + "learning_rate": 9.582508091938953e-07, + "loss": 0.0081, + "step": 505 + }, + { + "epoch": 3.2649434571890144, + "grad_norm": 0.2679020166397095, + "learning_rate": 9.416239783335785e-07, + "loss": 0.0098, + "step": 506 + }, + { + "epoch": 3.27140549273021, + "grad_norm": 0.32440176606178284, + "learning_rate": 9.251276537306969e-07, + "loss": 0.0074, + "step": 507 + }, + { + "epoch": 3.2778675282714054, + "grad_norm": 0.2440427541732788, + "learning_rate": 9.087623658618682e-07, + "loss": 0.009, + "step": 508 + }, + { + "epoch": 3.284329563812601, + "grad_norm": 0.3681286871433258, + "learning_rate": 8.925286409899308e-07, + "loss": 0.0071, + "step": 509 + }, + { + "epoch": 3.2907915993537964, + "grad_norm": 0.2806476950645447, + "learning_rate": 8.764270011470144e-07, + "loss": 0.0099, + "step": 510 + }, + { + "epoch": 3.297253634894992, + "grad_norm": 0.4697900414466858, + "learning_rate": 8.604579641177524e-07, + "loss": 0.0152, + "step": 511 + }, + { + "epoch": 3.3037156704361874, + "grad_norm": 0.3158174753189087, + "learning_rate": 8.446220434226382e-07, + "loss": 0.0085, + "step": 512 + }, + { + "epoch": 3.3101777059773827, + "grad_norm": 0.42739158868789673, + "learning_rate": 8.289197483015127e-07, + "loss": 0.0132, + "step": 513 + }, + { + "epoch": 3.3166397415185784, + "grad_norm": 0.2981080114841461, + "learning_rate": 8.133515836971773e-07, + "loss": 0.0064, + "step": 514 + }, + { + "epoch": 3.3231017770597737, + "grad_norm": 0.41693729162216187, + "learning_rate": 7.97918050239167e-07, + "loss": 0.0114, + "step": 515 + }, + { + "epoch": 3.3295638126009695, + "grad_norm": 0.44874924421310425, + "learning_rate": 7.826196442276473e-07, + "loss": 0.0099, + "step": 516 + }, + { + "epoch": 3.3360258481421647, + "grad_norm": 0.2718696892261505, + "learning_rate": 7.674568576174546e-07, + "loss": 0.0118, + "step": 517 + }, + { + "epoch": 3.3424878836833605, + "grad_norm": 0.39589688181877136, + "learning_rate": 7.524301780022774e-07, + "loss": 0.0143, + "step": 518 + }, + { + "epoch": 3.3489499192245558, + "grad_norm": 0.31575649976730347, + "learning_rate": 7.375400885989758e-07, + "loss": 0.0131, + "step": 519 + }, + { + "epoch": 3.355411954765751, + "grad_norm": 0.4461822807788849, + "learning_rate": 7.227870682320432e-07, + "loss": 0.0086, + "step": 520 + }, + { + "epoch": 3.361873990306947, + "grad_norm": 0.34174418449401855, + "learning_rate": 7.081715913182069e-07, + "loss": 0.0069, + "step": 521 + }, + { + "epoch": 3.368336025848142, + "grad_norm": 0.2855561375617981, + "learning_rate": 6.936941278511744e-07, + "loss": 0.008, + "step": 522 + }, + { + "epoch": 3.374798061389338, + "grad_norm": 0.3220924139022827, + "learning_rate": 6.793551433865198e-07, + "loss": 0.009, + "step": 523 + }, + { + "epoch": 3.381260096930533, + "grad_norm": 0.28071683645248413, + "learning_rate": 6.651550990267091e-07, + "loss": 0.0069, + "step": 524 + }, + { + "epoch": 3.387722132471729, + "grad_norm": 0.26843127608299255, + "learning_rate": 6.510944514062784e-07, + "loss": 0.0074, + "step": 525 + }, + { + "epoch": 3.394184168012924, + "grad_norm": 0.3113328218460083, + "learning_rate": 6.371736526771421e-07, + "loss": 0.0091, + "step": 526 + }, + { + "epoch": 3.4006462035541194, + "grad_norm": 0.22835861146450043, + "learning_rate": 6.233931504940633e-07, + "loss": 0.0072, + "step": 527 + }, + { + "epoch": 3.407108239095315, + "grad_norm": 8068.888671875, + "learning_rate": 6.097533880002476e-07, + "loss": 0.0093, + "step": 528 + }, + { + "epoch": 3.4135702746365104, + "grad_norm": 0.39374470710754395, + "learning_rate": 5.962548038130972e-07, + "loss": 0.012, + "step": 529 + }, + { + "epoch": 3.420032310177706, + "grad_norm": 0.4547136723995209, + "learning_rate": 5.828978320101109e-07, + "loss": 0.0098, + "step": 530 + }, + { + "epoch": 3.4264943457189014, + "grad_norm": 0.4039766490459442, + "learning_rate": 5.696829021149181e-07, + "loss": 0.0077, + "step": 531 + }, + { + "epoch": 3.432956381260097, + "grad_norm": 0.32658788561820984, + "learning_rate": 5.566104390834709e-07, + "loss": 0.008, + "step": 532 + }, + { + "epoch": 3.4394184168012925, + "grad_norm": 0.3402605950832367, + "learning_rate": 5.436808632903729e-07, + "loss": 0.0082, + "step": 533 + }, + { + "epoch": 3.4458804523424877, + "grad_norm": 0.31616881489753723, + "learning_rate": 5.308945905153729e-07, + "loss": 0.0129, + "step": 534 + }, + { + "epoch": 3.4523424878836835, + "grad_norm": 0.43558764457702637, + "learning_rate": 5.182520319299816e-07, + "loss": 0.0117, + "step": 535 + }, + { + "epoch": 3.4588045234248788, + "grad_norm": 0.337622731924057, + "learning_rate": 5.057535940842567e-07, + "loss": 0.007, + "step": 536 + }, + { + "epoch": 3.4652665589660745, + "grad_norm": 0.3466811180114746, + "learning_rate": 4.933996788937279e-07, + "loss": 0.011, + "step": 537 + }, + { + "epoch": 3.47172859450727, + "grad_norm": 0.35212457180023193, + "learning_rate": 4.811906836264718e-07, + "loss": 0.0107, + "step": 538 + }, + { + "epoch": 3.4781906300484655, + "grad_norm": 0.3255941867828369, + "learning_rate": 4.691270008903365e-07, + "loss": 0.0113, + "step": 539 + }, + { + "epoch": 3.484652665589661, + "grad_norm": 0.3242311477661133, + "learning_rate": 4.572090186203171e-07, + "loss": 0.0116, + "step": 540 + }, + { + "epoch": 3.491114701130856, + "grad_norm": 0.35152673721313477, + "learning_rate": 4.4543712006608507e-07, + "loss": 0.0101, + "step": 541 + }, + { + "epoch": 3.497576736672052, + "grad_norm": 0.31033825874328613, + "learning_rate": 4.338116837796519e-07, + "loss": 0.0101, + "step": 542 + }, + { + "epoch": 3.504038772213247, + "grad_norm": 0.3278428316116333, + "learning_rate": 4.2233308360321024e-07, + "loss": 0.0077, + "step": 543 + }, + { + "epoch": 3.5105008077544424, + "grad_norm": 1.6630736589431763, + "learning_rate": 4.110016886571011e-07, + "loss": 0.0087, + "step": 544 + }, + { + "epoch": 3.516962843295638, + "grad_norm": 0.49623093008995056, + "learning_rate": 3.998178633279537e-07, + "loss": 0.0075, + "step": 545 + }, + { + "epoch": 3.523424878836834, + "grad_norm": 0.24495165050029755, + "learning_rate": 3.887819672569565e-07, + "loss": 0.0076, + "step": 546 + }, + { + "epoch": 3.529886914378029, + "grad_norm": 1.0713497400283813, + "learning_rate": 3.778943553283015e-07, + "loss": 0.0087, + "step": 547 + }, + { + "epoch": 3.5363489499192244, + "grad_norm": 0.35764917731285095, + "learning_rate": 3.671553776577702e-07, + "loss": 0.0145, + "step": 548 + }, + { + "epoch": 3.54281098546042, + "grad_norm": 0.4270109534263611, + "learning_rate": 3.5656537958147164e-07, + "loss": 0.0196, + "step": 549 + }, + { + "epoch": 3.5492730210016155, + "grad_norm": 0.36601418256759644, + "learning_rate": 3.461247016447372e-07, + "loss": 0.0128, + "step": 550 + }, + { + "epoch": 3.5557350565428107, + "grad_norm": 0.2740163505077362, + "learning_rate": 3.3583367959117374e-07, + "loss": 0.0073, + "step": 551 + }, + { + "epoch": 3.5621970920840065, + "grad_norm": 0.30819642543792725, + "learning_rate": 3.2569264435186597e-07, + "loss": 0.011, + "step": 552 + }, + { + "epoch": 3.568659127625202, + "grad_norm": 0.33172082901000977, + "learning_rate": 3.1570192203473183e-07, + "loss": 0.0092, + "step": 553 + }, + { + "epoch": 3.5751211631663975, + "grad_norm": 0.23010392487049103, + "learning_rate": 3.058618339140368e-07, + "loss": 0.0035, + "step": 554 + }, + { + "epoch": 3.581583198707593, + "grad_norm": 0.22640886902809143, + "learning_rate": 2.961726964200645e-07, + "loss": 0.0056, + "step": 555 + }, + { + "epoch": 3.5880452342487885, + "grad_norm": 0.21879248321056366, + "learning_rate": 2.8663482112893936e-07, + "loss": 0.0066, + "step": 556 + }, + { + "epoch": 3.594507269789984, + "grad_norm": 0.2790137827396393, + "learning_rate": 2.772485147526077e-07, + "loss": 0.0082, + "step": 557 + }, + { + "epoch": 3.600969305331179, + "grad_norm": 2.688298225402832, + "learning_rate": 2.680140791289737e-07, + "loss": 0.015, + "step": 558 + }, + { + "epoch": 3.607431340872375, + "grad_norm": 0.29820549488067627, + "learning_rate": 2.5893181121219637e-07, + "loss": 0.0097, + "step": 559 + }, + { + "epoch": 3.61389337641357, + "grad_norm": 0.33826717734336853, + "learning_rate": 2.500020030631356e-07, + "loss": 0.0087, + "step": 560 + }, + { + "epoch": 3.620355411954766, + "grad_norm": 0.43633323907852173, + "learning_rate": 2.4122494183996426e-07, + "loss": 0.0127, + "step": 561 + }, + { + "epoch": 3.626817447495961, + "grad_norm": 0.2826526165008545, + "learning_rate": 2.3260090978893146e-07, + "loss": 0.0089, + "step": 562 + }, + { + "epoch": 3.633279483037157, + "grad_norm": 0.27118533849716187, + "learning_rate": 2.2413018423528832e-07, + "loss": 0.0059, + "step": 563 + }, + { + "epoch": 3.639741518578352, + "grad_norm": 2.0000076293945312, + "learning_rate": 2.1581303757436778e-07, + "loss": 0.0088, + "step": 564 + }, + { + "epoch": 3.6462035541195474, + "grad_norm": 0.2869947850704193, + "learning_rate": 2.076497372628261e-07, + "loss": 0.0084, + "step": 565 + }, + { + "epoch": 3.652665589660743, + "grad_norm": 0.35098886489868164, + "learning_rate": 1.9964054581004476e-07, + "loss": 0.0133, + "step": 566 + }, + { + "epoch": 3.6591276252019385, + "grad_norm": 0.32275232672691345, + "learning_rate": 1.9178572076968437e-07, + "loss": 0.0109, + "step": 567 + }, + { + "epoch": 3.665589660743134, + "grad_norm": 0.44486019015312195, + "learning_rate": 1.84085514731403e-07, + "loss": 0.0093, + "step": 568 + }, + { + "epoch": 3.6720516962843295, + "grad_norm": 0.30504077672958374, + "learning_rate": 1.7654017531273882e-07, + "loss": 0.0097, + "step": 569 + }, + { + "epoch": 3.678513731825525, + "grad_norm": 0.4125590920448303, + "learning_rate": 1.6914994515114082e-07, + "loss": 0.01, + "step": 570 + }, + { + "epoch": 3.6849757673667205, + "grad_norm": 0.44932204484939575, + "learning_rate": 1.619150618961701e-07, + "loss": 0.01, + "step": 571 + }, + { + "epoch": 3.691437802907916, + "grad_norm": 0.39718618988990784, + "learning_rate": 1.5483575820185615e-07, + "loss": 0.0096, + "step": 572 + }, + { + "epoch": 3.6978998384491115, + "grad_norm": 0.2970835864543915, + "learning_rate": 1.4791226171921748e-07, + "loss": 0.0076, + "step": 573 + }, + { + "epoch": 3.704361873990307, + "grad_norm": 0.33279949426651, + "learning_rate": 1.411447950889372e-07, + "loss": 0.0108, + "step": 574 + }, + { + "epoch": 3.7108239095315025, + "grad_norm": 0.33126136660575867, + "learning_rate": 1.3453357593420757e-07, + "loss": 0.0084, + "step": 575 + }, + { + "epoch": 3.717285945072698, + "grad_norm": 0.41318628191947937, + "learning_rate": 1.2807881685372947e-07, + "loss": 0.007, + "step": 576 + }, + { + "epoch": 3.7237479806138936, + "grad_norm": 0.3176984488964081, + "learning_rate": 1.2178072541487508e-07, + "loss": 0.0103, + "step": 577 + }, + { + "epoch": 3.730210016155089, + "grad_norm": 0.35345304012298584, + "learning_rate": 1.1563950414701653e-07, + "loss": 0.0082, + "step": 578 + }, + { + "epoch": 3.736672051696284, + "grad_norm": 0.37485456466674805, + "learning_rate": 1.0965535053500843e-07, + "loss": 0.0072, + "step": 579 + }, + { + "epoch": 3.74313408723748, + "grad_norm": 0.426248699426651, + "learning_rate": 1.0382845701284228e-07, + "loss": 0.0107, + "step": 580 + }, + { + "epoch": 3.749596122778675, + "grad_norm": 0.7327535152435303, + "learning_rate": 9.815901095745373e-08, + "loss": 0.0108, + "step": 581 + }, + { + "epoch": 3.756058158319871, + "grad_norm": 0.39146602153778076, + "learning_rate": 9.264719468270011e-08, + "loss": 0.0165, + "step": 582 + }, + { + "epoch": 3.762520193861066, + "grad_norm": 0.35452187061309814, + "learning_rate": 8.729318543349685e-08, + "loss": 0.0068, + "step": 583 + }, + { + "epoch": 3.768982229402262, + "grad_norm": 0.3429497480392456, + "learning_rate": 8.209715538011753e-08, + "loss": 0.0101, + "step": 584 + }, + { + "epoch": 3.775444264943457, + "grad_norm": 1.697080373764038, + "learning_rate": 7.70592716126567e-08, + "loss": 0.0105, + "step": 585 + }, + { + "epoch": 3.7819063004846525, + "grad_norm": 0.27320510149002075, + "learning_rate": 7.217969613565856e-08, + "loss": 0.0105, + "step": 586 + }, + { + "epoch": 3.788368336025848, + "grad_norm": 0.288004070520401, + "learning_rate": 6.745858586290566e-08, + "loss": 0.0083, + "step": 587 + }, + { + "epoch": 3.7948303715670435, + "grad_norm": 0.42086732387542725, + "learning_rate": 6.28960926123745e-08, + "loss": 0.0143, + "step": 588 + }, + { + "epoch": 3.8012924071082392, + "grad_norm": 0.30742815136909485, + "learning_rate": 5.84923631013512e-08, + "loss": 0.0148, + "step": 589 + }, + { + "epoch": 3.8077544426494345, + "grad_norm": 0.4420062005519867, + "learning_rate": 5.424753894171519e-08, + "loss": 0.0116, + "step": 590 + }, + { + "epoch": 3.8142164781906303, + "grad_norm": 0.2861470580101013, + "learning_rate": 5.016175663538625e-08, + "loss": 0.0043, + "step": 591 + }, + { + "epoch": 3.8206785137318255, + "grad_norm": 0.26148203015327454, + "learning_rate": 4.623514756993241e-08, + "loss": 0.0062, + "step": 592 + }, + { + "epoch": 3.827140549273021, + "grad_norm": 0.2247193306684494, + "learning_rate": 4.246783801434617e-08, + "loss": 0.0068, + "step": 593 + }, + { + "epoch": 3.8336025848142166, + "grad_norm": 0.2534724175930023, + "learning_rate": 3.885994911498603e-08, + "loss": 0.0065, + "step": 594 + }, + { + "epoch": 3.840064620355412, + "grad_norm": 0.31165772676467896, + "learning_rate": 3.541159689167628e-08, + "loss": 0.0097, + "step": 595 + }, + { + "epoch": 3.8465266558966076, + "grad_norm": 0.47996920347213745, + "learning_rate": 3.212289223398002e-08, + "loss": 0.0057, + "step": 596 + }, + { + "epoch": 3.852988691437803, + "grad_norm": 0.2576350271701813, + "learning_rate": 2.8993940897631412e-08, + "loss": 0.0073, + "step": 597 + }, + { + "epoch": 3.8594507269789986, + "grad_norm": 0.3661513030529022, + "learning_rate": 2.602484350113621e-08, + "loss": 0.0085, + "step": 598 + }, + { + "epoch": 3.865912762520194, + "grad_norm": 0.4161849915981293, + "learning_rate": 2.321569552253433e-08, + "loss": 0.0116, + "step": 599 + }, + { + "epoch": 3.872374798061389, + "grad_norm": 0.25652387738227844, + "learning_rate": 2.056658729633121e-08, + "loss": 0.0078, + "step": 600 + }, + { + "epoch": 3.878836833602585, + "grad_norm": 0.31865638494491577, + "learning_rate": 1.807760401059122e-08, + "loss": 0.0111, + "step": 601 + }, + { + "epoch": 3.88529886914378, + "grad_norm": 0.4232560396194458, + "learning_rate": 1.5748825704199887e-08, + "loss": 0.0146, + "step": 602 + }, + { + "epoch": 3.891760904684976, + "grad_norm": 0.3084033727645874, + "learning_rate": 1.3580327264289261e-08, + "loss": 0.0066, + "step": 603 + }, + { + "epoch": 3.898222940226171, + "grad_norm": 0.2703307271003723, + "learning_rate": 1.1572178423830405e-08, + "loss": 0.0091, + "step": 604 + }, + { + "epoch": 3.904684975767367, + "grad_norm": 0.34035852551460266, + "learning_rate": 9.724443759389635e-09, + "loss": 0.0059, + "step": 605 + }, + { + "epoch": 3.9111470113085622, + "grad_norm": 0.5055537819862366, + "learning_rate": 8.037182689052958e-09, + "loss": 0.0087, + "step": 606 + }, + { + "epoch": 3.9176090468497575, + "grad_norm": 0.4783738851547241, + "learning_rate": 6.510449470514824e-09, + "loss": 0.0115, + "step": 607 + }, + { + "epoch": 3.9240710823909533, + "grad_norm": 0.24573703110218048, + "learning_rate": 5.1442931993350705e-09, + "loss": 0.007, + "step": 608 + }, + { + "epoch": 3.9305331179321485, + "grad_norm": 0.2814759314060211, + "learning_rate": 3.9387578073563086e-09, + "loss": 0.0105, + "step": 609 + }, + { + "epoch": 3.936995153473344, + "grad_norm": 0.30954548716545105, + "learning_rate": 2.8938820612961494e-09, + "loss": 0.0057, + "step": 610 + }, + { + "epoch": 3.9434571890145396, + "grad_norm": 0.25791114568710327, + "learning_rate": 2.0096995614959924e-09, + "loss": 0.0094, + "step": 611 + }, + { + "epoch": 3.9499192245557353, + "grad_norm": 0.2954118549823761, + "learning_rate": 1.2862387408435483e-09, + "loss": 0.0053, + "step": 612 + }, + { + "epoch": 3.9563812600969306, + "grad_norm": 0.3213749825954437, + "learning_rate": 7.235228638574621e-10, + "loss": 0.0145, + "step": 613 + }, + { + "epoch": 3.962843295638126, + "grad_norm": 0.2650403380393982, + "learning_rate": 3.2157002593902196e-10, + "loss": 0.0072, + "step": 614 + }, + { + "epoch": 3.9693053311793216, + "grad_norm": 0.32353588938713074, + "learning_rate": 8.039315279040338e-11, + "loss": 0.0067, + "step": 615 + }, + { + "epoch": 3.975767366720517, + "grad_norm": 0.2813875377178192, + "learning_rate": 0.0, + "loss": 0.007, + "step": 616 + }, + { + "epoch": 3.975767366720517, + "step": 616, + "total_flos": 8.773760444520202e+17, + "train_loss": 0.04414510081163849, + "train_runtime": 2220.3071, + "train_samples_per_second": 8.921, + "train_steps_per_second": 0.277 + } + ], + "logging_steps": 1, + "max_steps": 616, + "num_input_tokens_seen": 0, + "num_train_epochs": 4, + "save_steps": 500, + "stateful_callbacks": { + "TrainerControl": { + "args": { + "should_epoch_stop": false, + "should_evaluate": false, + "should_log": false, + "should_save": true, + "should_training_stop": true + }, + "attributes": {} + } + }, + "total_flos": 8.773760444520202e+17, + "train_batch_size": 1, + "trial_name": null, + "trial_params": null +}