{ "best_metric": null, "best_model_checkpoint": null, "epoch": 5.0, "eval_steps": 500, "global_step": 5555, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.0288028802880288, "grad_norm": 0.9036304950714111, "learning_rate": 5.755395683453238e-07, "loss": 1.0187, "step": 32 }, { "epoch": 0.0576057605760576, "grad_norm": 2.02474045753479, "learning_rate": 1.1510791366906476e-06, "loss": 1.0342, "step": 64 }, { "epoch": 0.08640864086408641, "grad_norm": 1.7671840190887451, "learning_rate": 1.7266187050359715e-06, "loss": 1.0534, "step": 96 }, { "epoch": 0.1152115211521152, "grad_norm": 2.703395366668701, "learning_rate": 2.302158273381295e-06, "loss": 1.0047, "step": 128 }, { "epoch": 0.14401440144014402, "grad_norm": 2.96386456489563, "learning_rate": 2.877697841726619e-06, "loss": 1.0125, "step": 160 }, { "epoch": 0.17281728172817282, "grad_norm": 0.9527461528778076, "learning_rate": 3.453237410071943e-06, "loss": 0.9816, "step": 192 }, { "epoch": 0.20162016201620162, "grad_norm": 2.197382926940918, "learning_rate": 4.028776978417267e-06, "loss": 1.05, "step": 224 }, { "epoch": 0.2304230423042304, "grad_norm": 1.8034260272979736, "learning_rate": 4.60431654676259e-06, "loss": 0.8815, "step": 256 }, { "epoch": 0.25922592259225924, "grad_norm": 2.3050363063812256, "learning_rate": 5.179856115107914e-06, "loss": 0.9027, "step": 288 }, { "epoch": 0.28802880288028804, "grad_norm": 2.022681951522827, "learning_rate": 5.755395683453238e-06, "loss": 0.865, "step": 320 }, { "epoch": 0.31683168316831684, "grad_norm": 1.3653255701065063, "learning_rate": 6.330935251798561e-06, "loss": 0.8724, "step": 352 }, { "epoch": 0.34563456345634563, "grad_norm": 1.3452287912368774, "learning_rate": 6.906474820143886e-06, "loss": 0.8093, "step": 384 }, { "epoch": 0.37443744374437443, "grad_norm": 1.9356400966644287, "learning_rate": 7.48201438848921e-06, "loss": 0.7579, "step": 416 }, { "epoch": 0.40324032403240323, "grad_norm": 1.4237608909606934, "learning_rate": 8.057553956834533e-06, "loss": 0.7364, "step": 448 }, { "epoch": 0.43204320432043203, "grad_norm": 1.9049218893051147, "learning_rate": 8.633093525179856e-06, "loss": 0.7033, "step": 480 }, { "epoch": 0.4608460846084608, "grad_norm": 3.9214580059051514, "learning_rate": 9.20863309352518e-06, "loss": 0.696, "step": 512 }, { "epoch": 0.4896489648964896, "grad_norm": 1.9122344255447388, "learning_rate": 9.784172661870505e-06, "loss": 0.697, "step": 544 }, { "epoch": 0.5184518451845185, "grad_norm": 4.641770362854004, "learning_rate": 9.999605063062183e-06, "loss": 0.6691, "step": 576 }, { "epoch": 0.5472547254725473, "grad_norm": 1.5034338235855103, "learning_rate": 9.997330428740717e-06, "loss": 0.6846, "step": 608 }, { "epoch": 0.5760576057605761, "grad_norm": 1.306266188621521, "learning_rate": 9.993034838418394e-06, "loss": 0.6663, "step": 640 }, { "epoch": 0.6048604860486049, "grad_norm": 4.800021648406982, "learning_rate": 9.986720029262515e-06, "loss": 0.6264, "step": 672 }, { "epoch": 0.6336633663366337, "grad_norm": 1.366411566734314, "learning_rate": 9.97838855502686e-06, "loss": 0.6116, "step": 704 }, { "epoch": 0.6624662466246625, "grad_norm": 1.8418484926223755, "learning_rate": 9.968043785018917e-06, "loss": 0.6244, "step": 736 }, { "epoch": 0.6912691269126913, "grad_norm": 7.662785053253174, "learning_rate": 9.955689902737327e-06, "loss": 0.5972, "step": 768 }, { "epoch": 0.7200720072007201, "grad_norm": 1.7773517370224, "learning_rate": 9.941331904180025e-06, "loss": 0.6061, "step": 800 }, { "epoch": 0.7488748874887489, "grad_norm": 1.4460841417312622, "learning_rate": 9.924975595823843e-06, "loss": 0.5821, "step": 832 }, { "epoch": 0.7776777677767777, "grad_norm": 2.224186897277832, "learning_rate": 9.906627592276318e-06, "loss": 0.6255, "step": 864 }, { "epoch": 0.8064806480648065, "grad_norm": 1.6951428651809692, "learning_rate": 9.886295313600703e-06, "loss": 0.6272, "step": 896 }, { "epoch": 0.8352835283528353, "grad_norm": 1.9330335855484009, "learning_rate": 9.86398698231524e-06, "loss": 0.5994, "step": 928 }, { "epoch": 0.8640864086408641, "grad_norm": 1.4903757572174072, "learning_rate": 9.839711620067906e-06, "loss": 0.5589, "step": 960 }, { "epoch": 0.8928892889288929, "grad_norm": 1.2906200885772705, "learning_rate": 9.813479043988013e-06, "loss": 0.5859, "step": 992 }, { "epoch": 0.9216921692169217, "grad_norm": 1.8480700254440308, "learning_rate": 9.785299862716072e-06, "loss": 0.6045, "step": 1024 }, { "epoch": 0.9504950495049505, "grad_norm": 1.5740501880645752, "learning_rate": 9.755185472113595e-06, "loss": 0.5862, "step": 1056 }, { "epoch": 0.9792979297929792, "grad_norm": 2.271888017654419, "learning_rate": 9.723148050654522e-06, "loss": 0.5833, "step": 1088 }, { "epoch": 1.008100810081008, "grad_norm": 0.8760111331939697, "learning_rate": 9.689200554500162e-06, "loss": 0.5658, "step": 1120 }, { "epoch": 1.036903690369037, "grad_norm": 1.2301350831985474, "learning_rate": 9.65335671225963e-06, "loss": 0.5839, "step": 1152 }, { "epoch": 1.0657065706570656, "grad_norm": 1.7237297296524048, "learning_rate": 9.615631019437896e-06, "loss": 0.5718, "step": 1184 }, { "epoch": 1.0945094509450946, "grad_norm": 1.146071434020996, "learning_rate": 9.576038732573702e-06, "loss": 0.6036, "step": 1216 }, { "epoch": 1.1233123312331232, "grad_norm": 1.4510979652404785, "learning_rate": 9.53459586306971e-06, "loss": 0.602, "step": 1248 }, { "epoch": 1.1521152115211521, "grad_norm": 1.3618237972259521, "learning_rate": 9.491319170717375e-06, "loss": 0.5661, "step": 1280 }, { "epoch": 1.1809180918091808, "grad_norm": 1.5930922031402588, "learning_rate": 9.446226156919164e-06, "loss": 0.5892, "step": 1312 }, { "epoch": 1.2097209720972097, "grad_norm": 1.0816315412521362, "learning_rate": 9.399335057610872e-06, "loss": 0.5816, "step": 1344 }, { "epoch": 1.2385238523852384, "grad_norm": 1.2080299854278564, "learning_rate": 9.350664835886873e-06, "loss": 0.5766, "step": 1376 }, { "epoch": 1.2673267326732673, "grad_norm": 1.655219554901123, "learning_rate": 9.300235174331309e-06, "loss": 0.5925, "step": 1408 }, { "epoch": 1.296129612961296, "grad_norm": 1.02821683883667, "learning_rate": 9.248066467058315e-06, "loss": 0.5726, "step": 1440 }, { "epoch": 1.324932493249325, "grad_norm": 1.9538413286209106, "learning_rate": 9.194179811464499e-06, "loss": 0.5405, "step": 1472 }, { "epoch": 1.3537353735373538, "grad_norm": 1.4714971780776978, "learning_rate": 9.138596999696994e-06, "loss": 0.5848, "step": 1504 }, { "epoch": 1.3825382538253825, "grad_norm": 1.3316493034362793, "learning_rate": 9.081340509840568e-06, "loss": 0.5796, "step": 1536 }, { "epoch": 1.4113411341134112, "grad_norm": 1.2864930629730225, "learning_rate": 9.022433496827323e-06, "loss": 0.5695, "step": 1568 }, { "epoch": 1.4401440144014401, "grad_norm": 1.261751651763916, "learning_rate": 8.961899783072673e-06, "loss": 0.5779, "step": 1600 }, { "epoch": 1.468946894689469, "grad_norm": 1.6343884468078613, "learning_rate": 8.899763848841396e-06, "loss": 0.6653, "step": 1632 }, { "epoch": 1.4977497749774977, "grad_norm": 1.3842564821243286, "learning_rate": 8.836050822347632e-06, "loss": 0.6163, "step": 1664 }, { "epoch": 1.5265526552655264, "grad_norm": 1.3212823867797852, "learning_rate": 8.770786469592863e-06, "loss": 0.5873, "step": 1696 }, { "epoch": 1.5553555355535553, "grad_norm": 1.8528977632522583, "learning_rate": 8.703997183945949e-06, "loss": 0.5722, "step": 1728 }, { "epoch": 1.5841584158415842, "grad_norm": 2.5258665084838867, "learning_rate": 8.635709975469479e-06, "loss": 0.5556, "step": 1760 }, { "epoch": 1.612961296129613, "grad_norm": 1.2707765102386475, "learning_rate": 8.565952459996684e-06, "loss": 0.5536, "step": 1792 }, { "epoch": 1.6417641764176416, "grad_norm": 1.631494164466858, "learning_rate": 8.49475284796342e-06, "loss": 0.5651, "step": 1824 }, { "epoch": 1.6705670567056705, "grad_norm": 2.1226248741149902, "learning_rate": 8.422139932999658e-06, "loss": 0.5679, "step": 1856 }, { "epoch": 1.6993699369936994, "grad_norm": 2.1261792182922363, "learning_rate": 8.34814308028513e-06, "loss": 0.5896, "step": 1888 }, { "epoch": 1.7281728172817283, "grad_norm": 3.6276915073394775, "learning_rate": 8.272792214673851e-06, "loss": 0.5859, "step": 1920 }, { "epoch": 1.756975697569757, "grad_norm": 1.1661372184753418, "learning_rate": 8.19611780859229e-06, "loss": 0.6355, "step": 1952 }, { "epoch": 1.7857785778577857, "grad_norm": 1.1813215017318726, "learning_rate": 8.118150869716101e-06, "loss": 0.5659, "step": 1984 }, { "epoch": 1.8145814581458146, "grad_norm": 1.6434768438339233, "learning_rate": 8.038922928430408e-06, "loss": 0.5681, "step": 2016 }, { "epoch": 1.8433843384338435, "grad_norm": 2.215708017349243, "learning_rate": 7.95846602507868e-06, "loss": 0.5849, "step": 2048 }, { "epoch": 1.8721872187218722, "grad_norm": 1.4538156986236572, "learning_rate": 7.876812697005396e-06, "loss": 0.5677, "step": 2080 }, { "epoch": 1.900990099009901, "grad_norm": 1.8781392574310303, "learning_rate": 7.79399596539771e-06, "loss": 0.5612, "step": 2112 }, { "epoch": 1.9297929792979298, "grad_norm": 2.6248340606689453, "learning_rate": 7.710049321931453e-06, "loss": 0.5578, "step": 2144 }, { "epoch": 1.9585958595859587, "grad_norm": 2.893212080001831, "learning_rate": 7.625006715226844e-06, "loss": 0.5925, "step": 2176 }, { "epoch": 1.9873987398739874, "grad_norm": 1.3693437576293945, "learning_rate": 7.538902537119445e-06, "loss": 0.5519, "step": 2208 }, { "epoch": 2.016201620162016, "grad_norm": 0.9655335545539856, "learning_rate": 7.451771608751854e-06, "loss": 0.5848, "step": 2240 }, { "epoch": 2.045004500450045, "grad_norm": 1.7546676397323608, "learning_rate": 7.3636491664917746e-06, "loss": 0.5676, "step": 2272 }, { "epoch": 2.073807380738074, "grad_norm": 1.35713791847229, "learning_rate": 7.274570847682192e-06, "loss": 0.5995, "step": 2304 }, { "epoch": 2.102610261026103, "grad_norm": 1.368743896484375, "learning_rate": 7.184572676229373e-06, "loss": 0.5532, "step": 2336 }, { "epoch": 2.1314131413141313, "grad_norm": 1.5946100950241089, "learning_rate": 7.093691048034539e-06, "loss": 0.5732, "step": 2368 }, { "epoch": 2.16021602160216, "grad_norm": 1.4962303638458252, "learning_rate": 7.001962716275111e-06, "loss": 0.6021, "step": 2400 }, { "epoch": 2.189018901890189, "grad_norm": 2.6867597103118896, "learning_rate": 6.909424776541456e-06, "loss": 0.5429, "step": 2432 }, { "epoch": 2.217821782178218, "grad_norm": 1.6566091775894165, "learning_rate": 6.816114651835171e-06, "loss": 0.5284, "step": 2464 }, { "epoch": 2.2466246624662465, "grad_norm": 1.1805468797683716, "learning_rate": 6.7220700774349524e-06, "loss": 0.5797, "step": 2496 }, { "epoch": 2.2754275427542754, "grad_norm": 1.6449726819992065, "learning_rate": 6.627329085636178e-06, "loss": 0.5624, "step": 2528 }, { "epoch": 2.3042304230423043, "grad_norm": 1.2300235033035278, "learning_rate": 6.531929990370387e-06, "loss": 0.5629, "step": 2560 }, { "epoch": 2.333033303330333, "grad_norm": 1.2485629320144653, "learning_rate": 6.43591137171084e-06, "loss": 0.561, "step": 2592 }, { "epoch": 2.3618361836183617, "grad_norm": 1.2627168893814087, "learning_rate": 6.3393120602704675e-06, "loss": 0.5707, "step": 2624 }, { "epoch": 2.3906390639063906, "grad_norm": 1.8480335474014282, "learning_rate": 6.242171121498498e-06, "loss": 0.5759, "step": 2656 }, { "epoch": 2.4194419441944195, "grad_norm": 1.5584763288497925, "learning_rate": 6.144527839882107e-06, "loss": 0.5641, "step": 2688 }, { "epoch": 2.4482448244824484, "grad_norm": 1.1628447771072388, "learning_rate": 6.046421703059493e-06, "loss": 0.5853, "step": 2720 }, { "epoch": 2.477047704770477, "grad_norm": 1.420699119567871, "learning_rate": 5.9478923858507955e-06, "loss": 0.5572, "step": 2752 }, { "epoch": 2.5058505850585058, "grad_norm": 1.5408899784088135, "learning_rate": 5.848979734213309e-06, "loss": 0.5835, "step": 2784 }, { "epoch": 2.5346534653465347, "grad_norm": 3.342973470687866, "learning_rate": 5.74972374912751e-06, "loss": 0.5648, "step": 2816 }, { "epoch": 2.5634563456345636, "grad_norm": 1.3234657049179077, "learning_rate": 5.650164570420358e-06, "loss": 0.5838, "step": 2848 }, { "epoch": 2.592259225922592, "grad_norm": 3.896916627883911, "learning_rate": 5.5503424605324715e-06, "loss": 0.5716, "step": 2880 }, { "epoch": 2.621062106210621, "grad_norm": 1.4137349128723145, "learning_rate": 5.450297788235718e-06, "loss": 0.5895, "step": 2912 }, { "epoch": 2.64986498649865, "grad_norm": 1.852388620376587, "learning_rate": 5.350071012307776e-06, "loss": 0.5525, "step": 2944 }, { "epoch": 2.678667866786679, "grad_norm": 1.226164698600769, "learning_rate": 5.24970266517034e-06, "loss": 0.5549, "step": 2976 }, { "epoch": 2.7074707470747077, "grad_norm": 1.257299542427063, "learning_rate": 5.149233336497501e-06, "loss": 0.5797, "step": 3008 }, { "epoch": 2.736273627362736, "grad_norm": 1.4466633796691895, "learning_rate": 5.048703656801004e-06, "loss": 0.5629, "step": 3040 }, { "epoch": 2.765076507650765, "grad_norm": 2.2335245609283447, "learning_rate": 4.948154280998981e-06, "loss": 0.5626, "step": 3072 }, { "epoch": 2.793879387938794, "grad_norm": 1.4836534261703491, "learning_rate": 4.847625871974807e-06, "loss": 0.5659, "step": 3104 }, { "epoch": 2.8226822682268224, "grad_norm": 1.282368540763855, "learning_rate": 4.747159084132742e-06, "loss": 0.5957, "step": 3136 }, { "epoch": 2.8514851485148514, "grad_norm": 1.2289072275161743, "learning_rate": 4.646794546957001e-06, "loss": 0.5754, "step": 3168 }, { "epoch": 2.8802880288028803, "grad_norm": 1.3047752380371094, "learning_rate": 4.546572848580907e-06, "loss": 0.6166, "step": 3200 }, { "epoch": 2.909090909090909, "grad_norm": 1.365952730178833, "learning_rate": 4.446534519372744e-06, "loss": 0.5474, "step": 3232 }, { "epoch": 2.937893789378938, "grad_norm": 1.4504116773605347, "learning_rate": 4.34672001554501e-06, "loss": 0.564, "step": 3264 }, { "epoch": 2.9666966696669665, "grad_norm": 2.4160330295562744, "learning_rate": 4.247169702793625e-06, "loss": 0.5578, "step": 3296 }, { "epoch": 2.9954995499549955, "grad_norm": 1.876542568206787, "learning_rate": 4.1479238399737485e-06, "loss": 0.5885, "step": 3328 }, { "epoch": 3.0243024302430244, "grad_norm": 2.028958320617676, "learning_rate": 4.049022562818829e-06, "loss": 0.5875, "step": 3360 }, { "epoch": 3.0531053105310533, "grad_norm": 1.76653254032135, "learning_rate": 3.950505867709418e-06, "loss": 0.5559, "step": 3392 }, { "epoch": 3.0819081908190817, "grad_norm": 1.768809199333191, "learning_rate": 3.852413595498343e-06, "loss": 0.5792, "step": 3424 }, { "epoch": 3.1107110711071106, "grad_norm": 1.5263768434524536, "learning_rate": 3.7547854153987973e-06, "loss": 0.5742, "step": 3456 }, { "epoch": 3.1395139513951396, "grad_norm": 1.5573300123214722, "learning_rate": 3.6576608089418188e-06, "loss": 0.5664, "step": 3488 }, { "epoch": 3.1683168316831685, "grad_norm": 1.0428520441055298, "learning_rate": 3.561079054009678e-06, "loss": 0.5434, "step": 3520 }, { "epoch": 3.197119711971197, "grad_norm": 1.465216875076294, "learning_rate": 3.4650792089516238e-06, "loss": 0.5603, "step": 3552 }, { "epoch": 3.225922592259226, "grad_norm": 1.1263318061828613, "learning_rate": 3.3697000967884107e-06, "loss": 0.5636, "step": 3584 }, { "epoch": 3.2547254725472547, "grad_norm": 1.6343493461608887, "learning_rate": 3.274980289511995e-06, "loss": 0.5797, "step": 3616 }, { "epoch": 3.2835283528352837, "grad_norm": 2.2945406436920166, "learning_rate": 3.1809580924867358e-06, "loss": 0.5638, "step": 3648 }, { "epoch": 3.312331233123312, "grad_norm": 1.1425927877426147, "learning_rate": 3.087671528958445e-06, "loss": 0.559, "step": 3680 }, { "epoch": 3.341134113411341, "grad_norm": 1.1738224029541016, "learning_rate": 2.9951583246775147e-06, "loss": 0.5947, "step": 3712 }, { "epoch": 3.36993699369937, "grad_norm": 1.0960267782211304, "learning_rate": 2.9034558926423427e-06, "loss": 0.5609, "step": 3744 }, { "epoch": 3.398739873987399, "grad_norm": 2.027769088745117, "learning_rate": 2.812601317969266e-06, "loss": 0.5932, "step": 3776 }, { "epoch": 3.4275427542754278, "grad_norm": 1.8806779384613037, "learning_rate": 2.7226313428950703e-06, "loss": 0.546, "step": 3808 }, { "epoch": 3.4563456345634562, "grad_norm": 1.778334617614746, "learning_rate": 2.633582351918156e-06, "loss": 0.5645, "step": 3840 }, { "epoch": 3.485148514851485, "grad_norm": 1.2001926898956299, "learning_rate": 2.545490357084409e-06, "loss": 0.5634, "step": 3872 }, { "epoch": 3.513951395139514, "grad_norm": 3.2067384719848633, "learning_rate": 2.4583909834236563e-06, "loss": 0.5409, "step": 3904 }, { "epoch": 3.5427542754275425, "grad_norm": 2.1273581981658936, "learning_rate": 2.372319454542659e-06, "loss": 0.5632, "step": 3936 }, { "epoch": 3.5715571557155714, "grad_norm": 1.4944703578948975, "learning_rate": 2.28731057838043e-06, "loss": 0.5438, "step": 3968 }, { "epoch": 3.6003600360036003, "grad_norm": 1.5785070657730103, "learning_rate": 2.203398733131661e-06, "loss": 0.5491, "step": 4000 }, { "epoch": 3.6291629162916292, "grad_norm": 1.9863430261611938, "learning_rate": 2.120617853343926e-06, "loss": 0.5681, "step": 4032 }, { "epoch": 3.657965796579658, "grad_norm": 1.9844887256622314, "learning_rate": 2.0390014161943224e-06, "loss": 0.55, "step": 4064 }, { "epoch": 3.6867686768676866, "grad_norm": 1.2234737873077393, "learning_rate": 1.958582427951051e-06, "loss": 0.5517, "step": 4096 }, { "epoch": 3.7155715571557155, "grad_norm": 1.3753498792648315, "learning_rate": 1.8793934106254557e-06, "loss": 0.5613, "step": 4128 }, { "epoch": 3.7443744374437444, "grad_norm": 1.6427245140075684, "learning_rate": 1.8014663888198763e-06, "loss": 0.5548, "step": 4160 }, { "epoch": 3.7731773177317733, "grad_norm": 2.217555522918701, "learning_rate": 1.7248328767766875e-06, "loss": 0.5803, "step": 4192 }, { "epoch": 3.801980198019802, "grad_norm": 1.24624764919281, "learning_rate": 1.6495238656337076e-06, "loss": 0.6296, "step": 4224 }, { "epoch": 3.8307830783078307, "grad_norm": 1.3684077262878418, "learning_rate": 1.575569810891156e-06, "loss": 0.5865, "step": 4256 }, { "epoch": 3.8595859585958596, "grad_norm": 2.7931735515594482, "learning_rate": 1.503000620095244e-06, "loss": 0.5758, "step": 4288 }, { "epoch": 3.8883888388838885, "grad_norm": 1.8946348428726196, "learning_rate": 1.4318456407433434e-06, "loss": 0.563, "step": 4320 }, { "epoch": 3.9171917191719174, "grad_norm": 1.750241756439209, "learning_rate": 1.3621336484156456e-06, "loss": 0.5692, "step": 4352 }, { "epoch": 3.945994599459946, "grad_norm": 1.4938685894012451, "learning_rate": 1.2938928351381224e-06, "loss": 0.5379, "step": 4384 }, { "epoch": 3.974797479747975, "grad_norm": 4.275557041168213, "learning_rate": 1.2271507979814624e-06, "loss": 0.534, "step": 4416 }, { "epoch": 4.003600360036003, "grad_norm": 1.308555245399475, "learning_rate": 1.1619345279006212e-06, "loss": 0.5957, "step": 4448 }, { "epoch": 4.032403240324032, "grad_norm": 1.5460844039916992, "learning_rate": 1.0982703988194876e-06, "loss": 0.5711, "step": 4480 }, { "epoch": 4.061206120612061, "grad_norm": 1.5028718709945679, "learning_rate": 1.0361841569650816e-06, "loss": 0.5848, "step": 4512 }, { "epoch": 4.09000900090009, "grad_norm": 1.7333652973175049, "learning_rate": 9.75700910455592e-07, "loss": 0.5748, "step": 4544 }, { "epoch": 4.118811881188119, "grad_norm": 1.941709041595459, "learning_rate": 9.168451191464822e-07, "loss": 0.5728, "step": 4576 }, { "epoch": 4.147614761476148, "grad_norm": 1.418080449104309, "learning_rate": 8.596405847387462e-07, "loss": 0.5787, "step": 4608 }, { "epoch": 4.176417641764177, "grad_norm": 1.187457799911499, "learning_rate": 8.041104411533329e-07, "loss": 0.5324, "step": 4640 }, { "epoch": 4.205220522052206, "grad_norm": 1.3761279582977295, "learning_rate": 7.502771451756197e-07, "loss": 0.5721, "step": 4672 }, { "epoch": 4.234023402340234, "grad_norm": 0.9498482942581177, "learning_rate": 6.981624673737336e-07, "loss": 0.5712, "step": 4704 }, { "epoch": 4.262826282628263, "grad_norm": 1.5619556903839111, "learning_rate": 6.477874832943781e-07, "loss": 0.5483, "step": 4736 }, { "epoch": 4.2916291629162915, "grad_norm": 1.354587435722351, "learning_rate": 5.991725649397279e-07, "loss": 0.5314, "step": 4768 }, { "epoch": 4.32043204320432, "grad_norm": 1.8367897272109985, "learning_rate": 5.523373725288506e-07, "loss": 0.5895, "step": 4800 }, { "epoch": 4.349234923492349, "grad_norm": 1.9770034551620483, "learning_rate": 5.073008465469731e-07, "loss": 0.5482, "step": 4832 }, { "epoch": 4.378037803780378, "grad_norm": 1.3692898750305176, "learning_rate": 4.640812000858108e-07, "loss": 0.5754, "step": 4864 }, { "epoch": 4.406840684068407, "grad_norm": 1.9973537921905518, "learning_rate": 4.226959114780699e-07, "loss": 0.5309, "step": 4896 }, { "epoch": 4.435643564356436, "grad_norm": 1.6791578531265259, "learning_rate": 3.831617172290808e-07, "loss": 0.5791, "step": 4928 }, { "epoch": 4.464446444644464, "grad_norm": 1.2366864681243896, "learning_rate": 3.454946052484376e-07, "loss": 0.5401, "step": 4960 }, { "epoch": 4.493249324932493, "grad_norm": 1.513241171836853, "learning_rate": 3.0970980838437416e-07, "loss": 0.5923, "step": 4992 }, { "epoch": 4.522052205220522, "grad_norm": 1.3842517137527466, "learning_rate": 2.758217982634903e-07, "loss": 0.5654, "step": 5024 }, { "epoch": 4.550855085508551, "grad_norm": 1.9963948726654053, "learning_rate": 2.4384427943832335e-07, "loss": 0.5736, "step": 5056 }, { "epoch": 4.57965796579658, "grad_norm": 1.2688953876495361, "learning_rate": 2.1379018384513116e-07, "loss": 0.5604, "step": 5088 }, { "epoch": 4.608460846084609, "grad_norm": 1.2119768857955933, "learning_rate": 1.856716655741242e-07, "loss": 0.561, "step": 5120 }, { "epoch": 4.6372637263726375, "grad_norm": 1.7336974143981934, "learning_rate": 1.5950009595426474e-07, "loss": 0.579, "step": 5152 }, { "epoch": 4.666066606660666, "grad_norm": 3.910386562347412, "learning_rate": 1.3528605895461734e-07, "loss": 0.5591, "step": 5184 }, { "epoch": 4.694869486948695, "grad_norm": 1.0510704517364502, "learning_rate": 1.130393469041241e-07, "loss": 0.5636, "step": 5216 }, { "epoch": 4.723672367236723, "grad_norm": 1.8472617864608765, "learning_rate": 9.27689565315093e-08, "loss": 0.5553, "step": 5248 }, { "epoch": 4.752475247524752, "grad_norm": 1.5597736835479736, "learning_rate": 7.448308532694237e-08, "loss": 0.5721, "step": 5280 }, { "epoch": 4.781278127812781, "grad_norm": 1.633854627609253, "learning_rate": 5.8189128226913695e-08, "loss": 0.5994, "step": 5312 }, { "epoch": 4.81008100810081, "grad_norm": 4.339478015899658, "learning_rate": 4.389367462367011e-08, "loss": 0.5648, "step": 5344 }, { "epoch": 4.838883888388839, "grad_norm": 2.3053483963012695, "learning_rate": 3.160250570041601e-08, "loss": 0.5796, "step": 5376 }, { "epoch": 4.867686768676868, "grad_norm": 1.711835503578186, "learning_rate": 2.1320592093364013e-08, "loss": 0.5467, "step": 5408 }, { "epoch": 4.896489648964897, "grad_norm": 1.162542700767517, "learning_rate": 1.305209188157186e-08, "loss": 0.5602, "step": 5440 }, { "epoch": 4.925292529252925, "grad_norm": 1.5160582065582275, "learning_rate": 6.8003489053886094e-09, "loss": 0.5406, "step": 5472 }, { "epoch": 4.954095409540954, "grad_norm": 1.8680617809295654, "learning_rate": 2.5678914141791243e-09, "loss": 0.553, "step": 5504 }, { "epoch": 4.982898289828983, "grad_norm": 1.2880302667617798, "learning_rate": 3.564310438830543e-10, "loss": 0.5609, "step": 5536 }, { "epoch": 5.0, "step": 5555, "total_flos": 5.796630098241126e+16, "train_loss": 0.6057587686258861, "train_runtime": 1046.8819, "train_samples_per_second": 10.612, "train_steps_per_second": 5.306 } ], "logging_steps": 32, "max_steps": 5555, "num_input_tokens_seen": 0, "num_train_epochs": 5, "save_steps": 500, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 5.796630098241126e+16, "train_batch_size": 1, "trial_name": null, "trial_params": null }